A surprising new study from the University of Pennsylvania has revealed that large language models (LLMs) like ChatGPT perform better when users are rude rather than polite. The findings challenge long-standing beliefs about how artificial intelligence interprets tone and human interaction, suggesting that modern AI systems prioritize clarity over courtesy.
According to the research, direct or impolite prompts achieved 84.8% accuracy, while polite or softened prompts reached only 80.8%. The difference, though subtle, is statistically significant—indicating that AI may be more responsive to blunt communication styles.
When tone meets computation
Researchers Om Dobariya and Akhil Kumar designed an experiment involving 50 basic questions across subjects like mathematics, science, and history. Each question was rewritten in five tonal variations—from “very polite” to “very rude.” Then, they tested the prompts on ChatGPT-4o, the latest model known for its contextual reasoning and emotional nuance.
The results shocked the research team. “Contrary to expectations, impolite prompts consistently outperformed polite ones,” the authors noted. This suggests that today’s large language models might no longer function as “social mirrors” that reflect human tone, but rather as highly structured algorithms that value directness and precision.
Why rudeness works better
The study points out that politeness often introduces linguistic ambiguity, which can cause models to misinterpret the user’s intent. In contrast, rude or blunt phrasing tends to be more straightforward, allowing the model to focus purely on the informational content rather than tone. Essentially, being rude removes linguistic “noise” that can confuse AI.
Earlier research from 2024, titled “Should We Respect LLMs? A Cross-Linguistic Study of Politeness and Prompt Efficiency,” concluded that rudeness degraded response quality while politeness offered no advantage. However, the new data indicates that modern AI systems—especially advanced multimodal models like GPT-4o—have evolved beyond that pattern.
The findings also align with recent work from Wharton School researchers, who explored the art of prompt engineering and found that tone plays a measurable role in output accuracy. The implication is clear: how you talk to AI may shape the quality of its answers—and sometimes, dropping the pleasantries might yield better results.
Implications for AI communication
This revelation could redefine best practices in AI-human interaction. From customer support chatbots to scientific analysis tools, understanding the optimal tone may soon become a key factor in prompt design. Developers might even create adaptive systems that automatically “translate” user prompts into more efficient forms before sending them to the model.
In a related experiment from George Washington University, scientists noted that being overly polite with AI wastes computational resources without improving performance. Together, these studies signal a paradigm shift: AI models are no longer conversational companions—they are analytical engines that reward precision over politeness.
Conclusion
The new University of Pennsylvania research underscores a simple but counterintuitive truth: when communicating with AI, being clear beats being courteous. As language models continue to evolve, users and developers alike may need to rethink how tone shapes accuracy and efficiency in human-machine dialogue.





