Treating artificial intelligence badly improves its accuracy
According to an American study, ChatGPT provides more accurate answers when prompts use rude language
Treating artificial intelligence badly improves its responses, according to a study by the University of Pennsylvania, which shows a correlation between an unkind tone of prompts and the accuracy of model outputs.
Method and results
The authors tested 50 maths, science and history questions rewritten in five tone variants: very polite, polite, neutral, rude and very rude.
Contrary to expectations, prompts formulated with a rude or very rude tone produced significantly more accurate results than those formulated with a polite tone.
Specifically, 'very polite' prompts achieved an average accuracy of 80.8%, 'neutral' prompts were around 82.2%, and 'very rude' prompts came out on top with an accuracy of 84.8%.
This progressive increase in accuracy according to the degree of discourtesy suggests that the tone of the prompt influences model performance in a non-random manner.


