Ad
Skip to content

AI expert Andrej Karpathy envisions a web where 99.9% of content is optimized for AI, not humans

Image description
Midjourney prompted by THE DECODER

Key Points

  • AI researcher Andrej Karpathy predicts that in the future, content will primarily be created for AI systems to consume, rather than for human readers.
  • Karpathy suggests that current content formats like websites, PDFs, and videos are not well-suited for AI systems. New standards and formats may need to be developed to make content more AI-friendly in the future.
  • This shift presents new challenges for the digital economy, as existing business models rely on direct human consumption of content. As AI increasingly mediates content, new compensation and licensing systems will need to be developed.

Former OpenAI researcher Andrej Karpathy envisions a future where large language models (LLMs) become the primary interface for content.

In a recent post on X, he suggests that while most content will still be written for humans, 99.9 percent of optimization efforts will focus on making content digestible for LLMs rather than human readers. This shift could fundamentally change how documentation and other content is structured.

Karpathy points to current documentation practices as an example: "99% of libraries still have docs that basically render to some pretty .html static pages assuming a human will click through them." By 2025, he argues, documentation should exist as a single project file optimized for an LLM's context window.

Image: Screenshot via X

Moving toward AI-friendly web standards

Karpathy notes that while combining code bases into single files is technically straightforward, the real challenge lies with content stored in human-centric formats - websites, PDFs, images, videos, and audio files.

Ad
DEC_D_Incontent-1

These "pre-LLM era" formats make AI optimization difficult. According to Karpathy, the industry needs new standards that work equally well for both human and machine consumption.

A new proposed web standard called "llms.txt" aligns with Karpathy's vision for AI-optimized content structure. Developed by Jeremy Howard, the specification works similarly to index.html but for AI systems. While index.html directs users to a page's HTML version, llms.txt would guide AI systems to a machine-readable Markdown (.md) version.

This dual approach allows websites to maintain both human-readable and AI-optimized versions of their content. Companies like Anthropic have already implemented this standard.

LLM companies as the new gatekeepers

The implications go far beyond just technical changes. Today's digital content economy runs on human attention - through advertising and subscriptions. The industry now faces the challenge of completely reimagining its value chains and revenue models as content shifts toward AI consumption.

Ad
DEC_D_Incontent-2

AI companies have started licensing live news feeds, and this also raises some serious questions. When companies like OpenAI get to decide what content their AI systems see, they're essentially becoming powerful gatekeepers of information.

This shift threatens to reshape how everyone discovers and consumes online content, raising serious questions about who controls our access to information. The stakes are even higher considering that LLMs still frequently make mistakes when processing and reproducing information.

AI News Without the Hype – Curated by Humans

As a THE DECODER subscriber, you get ad-free reading, our weekly AI newsletter, the exclusive "AI Radar" Frontier Report 6× per year, access to comments, and our complete archive.

Source: Karpathy via X