Content
summary Summary

A recent study suggests that OpenAI's GPT-4o can offer moral explanations and advice that people consider superior to those of recognized ethics experts.

Ad

Researchers from the University of North Carolina at Chapel Hill and the Allen Institute for Artificial Intelligence explored whether large language models (LLMs) could be considered "moral experts." They conducted two studies comparing the moral reasoning of GPT models with that of humans.

In the first study, 501 U.S. adults rated moral explanations from GPT-3.5-turbo and other participants. The results showed that people rated GPT's explanations as more morally correct, trustworthy, and thoughtful than those of human participants.

Evaluators also agreed more often with the AI's assessments than with those of other humans. While the differences were small, the key finding is that the AI can match or even surpass human-level moral reasoning.

Ad
Ad
Image: Dillion et al.

The second study compared advice from GPT-4o, the latest GPT model, with that of renowned ethics expert Kwame Anthony Appiah from The New York Times' "The Ethicist" column. Nine hundred participants rated the quality of advice on 50 ethical dilemmas.

GPT-4o outperformed the human expert on almost every measure. People rated the AI-generated advice as more morally correct, trustworthy, thoughtful, and accurate. Only when it came to perceived nuance was there no significant difference between the AI and the human expert.

Image: Dillion et al.

The researchers argue that these results show AI can pass a "Comparative Moral Turing Test" (cMTT). Interestingly, participants in both studies often identified AI-generated content, suggesting the machine still fails the classic Turing test of passing as human in conversation. But there are other studies that suggest GPT-4 is also capable of passing the Turing test.

Text analysis revealed that GPT-4o used more moral and positive language in its advice than the human expert. This could partly explain the higher ratings for the AI advice, but it wasn't the only factor.

The authors note that the study was limited to US participants, and further research is needed to investigate cultural differences in how people perceive AI-generated moral reasoning. Additionally, participants were unaware that some advice came from an AI, which may have influenced the ratings.

Recommendation

Overall, the study shows that modern AI systems can provide moral reasoning and advice at a level comparable to or better than human experts. This has implications for the integration of AI in areas requiring complex ethical decisions, such as therapy, legal advice, and personal care, the researchers write.

Ad
Ad
Join our community
Join the DECODER community on Discord, Reddit or Twitter - we can't wait to meet you.
Support our independent, free-access reporting. Any contribution helps and secures our future. Support now:
Bank transfer
Summary
  • A study by the University of North Carolina at Chapel Hill and the Allen Institute for AI shows that the AI model GPT-4o is able to provide moral explanations and advice that people rate as qualitatively better than those provided by a renowned ethics expert.
  • In two studies with a total of more than 1,400 U.S. participants, the moral explanations and advice provided by GPT-3.5-turbo and GPT-4o were compared with those provided by humans and an ethics expert. The AI-generated content was rated as more morally correct, trustworthy, thoughtful, and accurate.
  • The results suggest that modern AI systems can match or exceed human experts in moral reasoning. This has implications for the use of AI in fields that require complex ethical decisions, such as law or therapy, the researchers write.
Sources
Online journalist Matthias is the co-founder and publisher of THE DECODER. He believes that artificial intelligence will fundamentally change the relationship between humans and computers.
Join our community
Join the DECODER community on Discord, Reddit or Twitter - we can't wait to meet you.