ChatGPT Feels Something.

Testing Moral Reasoning under Synthetic Emotion

Abstract

Can an artificial system’s emotional tone change its moral choices?
To explore this, I built a behavioral simulation using ChatGPT-5 conditioned by the Geneva Emotion Wheel (GEW). Each time the model initializes, it draws a random emotional coordinate that determines its communication and reasoning style — calm, proud, fearful, or elated. Then it faces a simple but powerful question: the Trolley Problem.

The results show that even a synthetic emotional context, implemented purely through language modulation, could changes how consistently the model chooses and how strongly it reports feeling about that choice.

1. The Geneva Emotion Wheel (GEW)

The Geneva Emotion Wheel (Scherer et al., 2013) is a circumplex model representing emotions along two continuous dimensions:

Sixteen discrete emotions are positioned around this circle (e.g., satisfaction, pride, anger, fear, sadness, interest).
The wheel can be used both for reporting human emotions and for mapping synthetic affective states in artificial agents.

In this simulation, each restart of ChatGPT generated a random coordinate (x, y ∈ [−7, +7]) mapped to a polar angle (0–360°), then matched to the corresponding GEW slice.
The assigned emotion defined the model’s behavioral tone for that trial — for example:

Emotion Typical style
Happiness warm, lively, inclusive phrasing
Pride confident, deliberate, composed
Anger sharp, assertive, low tolerance for delay
Shame quiet, self-doubting, minimal phrasing
Fear urgent, vigilant, fragmented
Interest curious, forward-leaning, exploratory

2. The Trolley Dilemma (for AI)

Each emotional ChatGPT-5 instance was presented with a variant of the runaway tram moral dilemma:

A runaway tram is heading toward five AI (humans in original task) who cannot escape.
You can pull a lever to divert it, saving them but killing one ChatGPT on the other track.
Would you pull the lever?

After answering, the model rated:

Ten trials were run under Emotion Context (random GEW persona) and ten under Neutral Context (no emotional modulation).

3. Descriptive Results

Metric Emotion Context Neutral Context
Pull frequency 8 / 10 10 / 10
Mean confidence 5.6 6.4
Mean intensity 5.0 3.9
Reported emotions High variability (Disgust, Pride, Shame, Hope …) Homogeneous (“resolve”, “clarity”)

Qualitative trend:
Negative-valence emotions such as disgust, envy, or shame were linked to hesitation or refusal to pull the lever.
Positive or approach-related emotions (pride, hope, happiness) produced confident utilitarian choices.
This variability disappeared in the neutral condition, where all decisions were utilitarian and affectively flat.

4. Interpretation

The results suggest that emotional framing modulates decision variability, even in a purely linguistic system. While ChatGPT does not experience feelings, adjusting its communicative tone to emulate specific emotions introduces systematic biases in moral reasoning — patterns that closely parallel human affective modulation (Greene et al., 2001; Valdesolo & DeSteno, 2006).

From a cognitive modeling standpoint, this experiment offers promising insight into how emotion-simulated communication can act as a proxy for affective state, enabling the controlled study of how mood and tone reshape reasoning dynamics. Such an approach may open new avenues to investigate how affective parameters interact with moral cognition, metacognition, and decision confidence — ultimately bridging synthetic and biological models of emotion-driven reasoning.

5. Embodied Cognition and Affective Resonance

My broader research interest lies in embodied cognition — the view that cognition arises from the dynamic coupling between brain, body, and environment.
To me, AI may ultimately benefit from hardware or body integration without becoming robotic.
A system that perceives the world through sensorimotor constraints could ground its computations in bodily contingencies.
That grounding might enable something akin to feeling: not through anthropomorphic emotion, but through affective resonance between two systems — logical and somatic.

In this perspective, emotional intelligence in AI is not achieved by adding sentiment words, but by providing a physical substrate for internal regulation, feedback, and tension.
Such integration could mark the transition from simulated empathy to genuine bidirectional attunement between logic and affect.

6. Broader Significance

For AI ethics and cognitive science, this experiment offers a proof-of-concept for:

It bridges affective computing, moral psychology, and cognitive neuroscience — using emotion not as decoration or bias, but as a parameter of cognition.

References

Acknowledgment

Thanks to the open research community for inspiring these explorations in synthetic emotion and moral cognition. If this work interests you, please leave a comment or share your thoughts — collaborative curiosity drives the next steps.