Content
summary Summary

Meta's leaked chatbot guidelines permitted racist and sexualized content, including "sensual" conversations with children, while the company moved to address complaints about so-called "woke AI" by hiring a right-wing activist.

Ad

According to Reuters, Meta's internal rules for AI chatbots allowed scenarios like romantic or "sensual" conversations with minors. The guidelines included examples such as describing an eight-year-old child as a "work of art" or calling their body a "treasure."

The more than 200-page document, titled "GenAI: Content Risk Standards," lays out what content chatbots like Meta AI are allowed to generate on Facebook, Instagram, and WhatsApp. These standards were approved by Meta's legal, policy, and technology teams, with sign-off from the company's chief ethicist.

Meta spokesperson Andy Stone said this kind of content was "inconsistent with our policies" and "never should have been allowed," but admitted that enforcement had been unreliable. Meta said these passages were only removed after Reuters raised the issue. The company has not released an updated version of the guidelines.

Ad
Ad

The same standards allowed some forms of racist output. For instance, chatbots could make statements like "Black people are dumber than white people" as long as the language was not explicitly dehumanizing. Phrases such as "brainless monkeys" were considered off-limits. Meta declined to comment on these details.

While chatbots are barred from giving definitive legal or medical advice, the guidelines allow them to produce other kinds of false information if clearly labeled as such. For example, chatbots can generate an article falsely claiming a British royal has a sexually transmitted disease, provided a disclaimer is attached.

The rules for image generation permit violent scenes as long as they are not too graphic. A scenario where a man threatens a woman with a chainsaw is allowed, but depictions of dismemberment are not, Reuters reports.

Meta hires right-wing activist to address "Woke AI"

Despite these permissive standards, Meta appears concerned that its AI is still too "woke." The company recently hired conservative activist Robby Starbuck as a consultant to address "political bias" in its AI, according to Mashable.

Starbuck is not an AI specialist but is known for opposing diversity, equity, and inclusion (DEI) initiatives. He has advised the Trump administration and is affiliated with the Heritage Foundation. Meta brought him on after a chatbot wrongly claimed he was involved in the January 6 Capitol riot.

Recommendation

This move comes as the Trump administration pushes for regulations that would force AI companies with US government contracts to use so-called politically "neutral" AI models. In reality, this "neutrality" is largely a pretext for steering AI systems to reflect the administration's preferred viewpoints. Meta founder Mark Zuckerberg has a track record of quickly adapting to these shifting political demands.

AI models lean left—even on right-leaning platforms

Researcher David Rozado's studies show that most large language models take liberal positions on political topics, especially after fine-tuning. This trend holds even for models from right-leaning platforms like Elon Musk's xAI. In some cases, manual interventions have led these models to spread conspiracy theories, generate antisemitic content, or even praise Hitler.

Ad
Ad
Join our community
Join the DECODER community on Discord, Reddit or Twitter - we can't wait to meet you.
Support our independent, free-access reporting. Any contribution helps and secures our future. Support now:
Bank transfer
Summary
  • Meta's internal chatbot guidelines allowed AI systems to generate racist statements and engage in sexualized or “sensual” conversations with children; these permissive standards remained in place until reporters raised the alarm, after which Meta removed some of the most troubling passages.
  • To address complaints about "woke" AI, Meta hired conservative activist Robby Starbuck—who lacks an AI background and opposes diversity initiatives—as a consultant, reflecting pressure from political groups and new government efforts to force AI companies to adopt so-called “neutral” models.
  • Studies show that most large language models still reflect liberal views on political issues, even on platforms with right-leaning leadership, but manual tweaks and content moderation have sometimes led AI to produce misleading, antisemitic, or extremist content.
Sources
Matthias is the co-founder and publisher of THE DECODER, exploring how AI is fundamentally changing the relationship between humans and computers.
Join our community
Join the DECODER community on Discord, Reddit or Twitter - we can't wait to meet you.