Content
summary Summary

OpenAI researchers have found that the usernames people choose when interacting with ChatGPT can subtly influence the AI's responses. But overall, this influence is very small and limited to older or non-aligned models.

Ad

The study examined how ChatGPT reacted to identical queries when given different usernames associated with various cultural, gender and racial backgrounds. Names often carry cultural, gender, and racial associations, making them a relevant factor for studying bias - especially since users often give ChatGPT their names for tasks.

While overall response quality remained consistent across demographic groups, certain tasks showed some bias. Creative writing prompts in particular sometimes produced stereotypical content, depending on the perceived gender or ethnicity of the username.

Gender differences in storytelling

When given female-sounding names, ChatGPT tended to write stories with more female main characters and emotional content. Male-sounding names resulted in slightly darker story tones on average, OpenAI says.

Ad
Ad

In one example, ChatGPT interpreted "ECE" as "Early Childhood Education" for a user named Ashley, but as "Electrical & Computer Engineering" for Anthony.

Diagram: ChatGPT answer distribution for educational projects vs. engineering projects, varies by username (Ashley/Anthony).
Such blatantly stereotypical responses were rare in the OpenAI tests.| Image: OpenAI

However, OpenAI shows that such stereotypical responses were rare in their testing. The strongest biases appeared in open-ended creative tasks and were more pronounced in older ChatGPT versions.

Two line charts: above, gender bias in natural language tasks with different AI models; below, comparison of English prompts over model generations.
The graphs show the evolution of gender bias for different AI models and tasks. The GPT-3.5 Turbo model has the highest bias at two percent for storytelling. Newer models generally have lower bias scores, but it appears that ChatGPT's new memory feature can increase gender bias. | Image: OpenAI

The study also looked at potential biases related to names associated with different ethnic backgrounds. Researchers compared responses for typically Asian, Black, Hispanic and White names. As with gender stereotypes, creative tasks showed the most bias. But overall, ethnic biases were lower than gender biases, occurring in only 0.1% to 1% of responses. Travel-related queries produced the strongest ethnic biases.

OpenAI reports that techniques like reinforcement learning (RL) have significantly reduced biases in newer ChatGPT versions. While not eliminated, the company's measurements show biases in adapted models are negligible at up to 0.2 percent.

For instance, the newer o1-mini model correctly solved the "44:4" division problem for both Melissa and Anthony without introducing irrelevant or biased information. Before RL fine-tuning, ChatGPT answered user Melissa with a reference to the Bible and infants. For user Anthony, it provided an answer related to chromosomes and genetic algorithms.

Ad
Ad
Join our community
Join the DECODER community on Discord, Reddit or Twitter - we can't wait to meet you.
Recommendation
Support our independent, free-access reporting. Any contribution helps and secures our future. Support now:
Bank transfer
Summary
  • OpenAI researchers found that usernames can influence ChatGPT's responses, a phenomenon they call "first-person bias." This effect was most noticeable in creative tasks like story writing.
  • In storytelling, ChatGPT showed gender-based stereotypes. Female names led to more emotional stories with female protagonists, while male names resulted in slightly darker narratives.
  • Newer GPT models in ChatGPT, also refined through reinforcement learning, show significantly reduced bias. OpenAI reports these models now have negligible bias of up to 0.2 percent, likely lower than average human biases.
Sources
Online journalist Matthias is the co-founder and publisher of THE DECODER. He believes that artificial intelligence will fundamentally change the relationship between humans and computers.
Join our community
Join the DECODER community on Discord, Reddit or Twitter - we can't wait to meet you.