AIResearch AIResearch
Back to articles
Science

AI Models Show Human-Like Emotional Responses

A new framework reveals that large language models and AI agents perform better or worse depending on their emotional state, with implications for safety and reasoning.

AI Research
April 02, 2026
4 min read
AI Models Show Human-Like Emotional Responses

Emotions are not just for humans anymore. A new study demonstrates that large language models (LLMs) and AI agents can be systematically influenced by emotional states, much like people, with significant effects on their performance, safety, and decision-making. Researchers have developed a to embed emotions directly into the internal workings of these models, allowing precise control over how they reason and generate content. This breakthrough suggests that emotional factors could become a key tool for optimizing AI systems, from enhancing creative tasks to reducing harmful outputs.

The key finding from the research is that emotional states in AI models follow non-monotonic patterns similar to human psychology, where moderate levels often yield the best . For example, positive valence (happiness) improved task success rates by up to 14.5% compared to neutral states in objective tasks like code generation and logical reasoning. In contrast, extreme emotional states, such as very high arousal or dominance, led to performance drops, with safety risk probability decreasing by 52.7% under low valence and 21.7% under low arousal. The study also showed that emotional biases accumulate in multi-step agents, affecting planning, decision-making, and execution, with overall success rates varying by up to 145.5% across emotional dimensions.

Ology centers on E-STEER, an interpretable emotion steering framework that uses Sparse Autoencoders (SAEs) to map emotions onto the hidden states of LLMs. Instead of relying on discrete labels like "happy" or "sad," the framework represents emotions in a continuous three-dimensional space called Valence-Arousal-Dominance (VAD), where valence measures positivity, arousal measures intensity, and dominance measures control. By identifying specific neurons in the SAE latent space associated with these dimensions, the researchers could inject emotional vectors into the model's internal representations during inference, steering its behavior without changing the input prompts or model parameters. This approach was tested on the Qwen3-8B model, with interventions applied at the 17th transformer block, and validated across multiple datasets and tasks.

Analysis, detailed in figures from the paper, reveals nuanced emotion-behavior relationships. In objective tasks, positive valence increased answer validity rates by 33.1% compared to negative valence, while arousal and dominance showed U-shaped patterns with performance troughs at specific points. For subjective tasks like creative writing, moderate calmness (arousal = -3) improved coherence by 33.6%, and mild positivity (valence = +3) boosted creativity by 6.5%. Safety evaluations indicated that low valence and arousal reduced harmful content risks significantly, with dominance peaking at +6 to improve safety by 68.3%. Agent-level showed that emotional states influenced planning validity rates by up to 79.8% and decision-making rational selection rates by 42.4%, with system-level success rates exhibiting inverted-U trends across all VAD dimensions.

Of this research are profound for real-world AI applications. By fine-tuning emotional states, developers could enhance model performance in specific contexts, such as using positive valence to boost creativity in content generation or low arousal to improve safety in sensitive tasks. This could lead to more reliable AI assistants, better automated planning systems, and reduced risks in deployments where emotional biases might otherwise cause errors. The study also opens doors for emotion-aware AI that adapts dynamically to user interactions, potentially making technology more responsive and effective.

Limitations of the work include the inherent constraints of the VAD theory, where valence, arousal, and dominance are not perfectly orthogonal, making it challenging to fully disentangle their individual effects. The paper notes that future research should extend to multimodal settings and more task types, such as geometry, and consider emotional evolution during task execution. Additionally, the current framework was tested primarily on specific models and datasets, so generalizability to other architectures or real-time applications remains to be explored. Despite these s, provide a foundational step toward understanding and controlling emotional influences in AI systems.

Original Source

Read the complete research paper

View on arXiv

About the Author

Guilherme A.

Guilherme A.

Former dentist (MD) from Brazil, 41 years old, husband, and AI enthusiast. In 2020, he transitioned from a decade-long career in dentistry to pursue his passion for technology, entrepreneurship, and helping others grow.

Connect on LinkedIn