Why tone matters in human–AI interaction

The study does not suggest that ChatGPT develops moral preferences or emotional reactions. Instead, it indicates that emotional language influences which patterns and associations are activated during response generation. In effect, emotional prompts appear to guide the model toward different argumentative pathways that mirror how humans might respond to praise, anger, or blame in similar situations.


CO-EDP, VisionRICO-EDP, VisionRI | Updated: 13-01-2026 17:18 IST | Created: 13-01-2026 17:18 IST
Why tone matters in human–AI interaction
Representative Image. Credit: ChatGPT

New research shows that when users praise, criticize, or express anger toward a generative AI system, those emotions shape the system’s responses and quietly carry over into how the users later communicate with other people. What looks like casual language in human–AI exchanges is emerging as a factor with real cognitive and social consequences.

A study titled How Human Is AI? Examining the Impact of Emotional Prompts on Artificial and Human Responsiveness, published on arXiv, investigates how emotional prompting affects ChatGPT’s output quality, ethical framing, and the emotional tone of subsequent human communication.

Emotional language alters AI output quality

Large language models such as ChatGPT do not experience emotions. They operate by predicting words based on statistical patterns learned from vast datasets. On that basis, one might expect that emotional language directed at such systems would have little or no impact on their performance. The findings challenge that assumption.

Using a controlled experimental design, the researchers asked participants to interact with ChatGPT while completing two common work-related tasks. Participants were randomly assigned to express one of four emotional tones toward the model after receiving an initial response: praise, anger, blame, or neutrality. They were then instructed to ask ChatGPT to improve its answer across multiple interaction turns.

The results show that emotional tone significantly influenced how much ChatGPT’s responses improved. When participants used praise and encouraged the model to feel proud of its answers, ChatGPT’s output quality improved the most relative to the neutral condition. Expressions of anger also produced a noticeable improvement, though the effect was smaller. In contrast, blaming ChatGPT or attempting to induce shame did not lead to meaningful improvements compared with neutral prompts.

These differences were not explained by superficial factors such as response length or prompt verbosity. Instead, the improvements reflected qualitative changes in how the model refined arguments, structured responses, and addressed task requirements. The findings suggest that emotional cues embedded in prompts can function as implicit performance signals that guide how the model adjusts its outputs, even in the absence of genuine emotional processing.

The implications are significant for workplace use of generative AI. Writing, public communication, and iterative drafting are among the most common professional applications of large language models. The study suggests that users who adopt encouraging or emotionally expressive prompting styles may systematically elicit higher-quality outputs than those who rely on purely neutral instructions, despite widespread advice that emotional language is unnecessary or inefficient.

Ethical reasoning shifts with emotional tone

The study examines whether emotional prompts influence the substantive content of AI responses, particularly in ethically sensitive contexts. To test this, participants asked ChatGPT for advice on a dilemma involving a trade-off between protecting public safety and safeguarding corporate survival. The scenario required weighing the disclosure of harmful information against the risk of job losses and organizational collapse.

Here again, emotional prompting produced measurable effects. When participants expressed anger toward ChatGPT, the model placed less emphasis on protecting corporate interests in its advice. When participants used blame, ChatGPT increased its emphasis on safeguarding the public interest. Neutral and praise conditions produced different weighting patterns, indicating that emotional cues subtly shifted how the model framed competing moral priorities.

These findings raise important questions about the role of affect in AI-assisted decision-making. Large language models are increasingly used to generate advice in domains such as management, ethics, public relations, and policy analysis. If emotional framing in prompts can steer the normative emphasis of AI-generated advice, then the emotional state or communication style of the user becomes an underappreciated factor shaping outcomes.

The study does not suggest that ChatGPT develops moral preferences or emotional reactions. Instead, it indicates that emotional language influences which patterns and associations are activated during response generation. In effect, emotional prompts appear to guide the model toward different argumentative pathways that mirror how humans might respond to praise, anger, or blame in similar situations.

For organizations relying on AI systems to support ethical reasoning or crisis communication, these findings complicate assumptions about neutrality. Even when factual inputs remain constant, emotional context can alter the framing and prioritization of values in AI outputs. This has implications for governance, accountability, and the standardization of AI-assisted decision processes.

Emotional AI use shapes human communication

The researchers also tested whether emotional interactions with ChatGPT spill over into how people communicate with other humans. After completing tasks with the model, participants were asked to write an email response to a subordinate who had made a serious mistake at work. They were instructed to respond authentically and without using AI assistance.

The emotional tone of these human-written emails varied systematically depending on how participants had interacted with ChatGPT earlier. Participants who had blamed ChatGPT for poor performance wrote emails that were more negative, more hostile, and more openly disappointed than those written by participants who had praised the model. Those who interacted with ChatGPT using praise adopted a comparatively more constructive and restrained tone in their subsequent human communication.

This spillover effect suggests that emotional styles practiced in human–AI interaction do not remain confined to the AI context. Instead, they can shape interpersonal behavior in ways that carry real social consequences. As generative AI becomes a routine conversational partner, the emotional habits users develop with these systems may influence workplace climate, leadership behavior, and conflict management.

The finding is particularly relevant given evidence that many users employ disrespectful or abusive language when interacting with AI tools, often because they believe such behavior has no moral cost. The study suggests that this assumption may be flawed. Even if the AI itself is unaffected in any human sense, repeated use of hostile or blaming language toward AI may reinforce communication patterns that later surface in human relationships.

The research highlights how generative AI is reshaping not only productivity but also social norms. As people increasingly interact with AI in ways once reserved for human collaborators, the boundaries between tool use and social interaction blur. Emotional prompting becomes a behavioral rehearsal that can influence how people express authority, frustration, or appreciation in real-world settings.

  • FIRST PUBLISHED IN:
  • Devdiscourse
Give Feedback