In a thought-provoking statement on X, former OpenAI researcher Andrej Karpathy has painted a vision of the future where large language models (LLMs) become the primary interface for digital content. This perspective suggests a fundamental shift in how we create, structure, and optimize information in the coming years.
The 99.9% Optimization Shift
While Karpathy acknowledges that most content will continue to be written with humans as the intended audience, he predicts that nearly all optimization efforts (99.9%) will focus on making content digestible for LLMs rather than human readers. This represents a significant reversal of our current paradigm, where readability and user experience for humans drive most content decisions.
Documentation: A Case Study
Karpathy points to current documentation practices as a prime example of this impending transformation. "99% of libraries still have docs that basically render to some pretty .html static pages assuming a human will click through them," he notes. By contrast, he envisions that by 2025, documentation should exist as a single project file optimized for an LLM's context window.
This shift makes practical sense. As LLMs become more capable of parsing, understanding, and retrieving information from large databases, optimizing for their consumption could streamline how developers and users access information.
The Content Format Challenge
While combining code into single files is technically straightforward, Karpathy identifies a more significant challenge: content stored in human-centric formats like websites, PDFs, images, videos, and audio files.
These "pre-LLM era" formats weren't designed with AI consumption in mind, making optimization difficult. Karpathy argues that the industry needs new standards that work equally well for both human and machine consumption.
The Rise of llms.txt
Aligning with Karpathy's vision, a new web standard called "llms.txt" has been proposed by Jeremy Howard. This specification works similarly to index.html but is designed specifically for AI systems. While index.html directs users to a page's HTML version, llms.txt would guide AI systems to a machine-readable Markdown version.
This dual approach allows websites to maintain both human-readable and AI-optimized versions of their content. Companies like Anthropic have already implemented this standard, suggesting early industry adoption.
New Gatekeepers Emerge
The implications of this shift extend far beyond technical changes. Today's digital content economy is built around human attention through advertising and subscriptions. If content increasingly shifts toward AI consumption, the industry will need to completely reimagine its value chains and revenue models.
AI companies have already begun licensing live news feeds, positioning themselves as new gatekeepers of information. When companies like OpenAI can decide what content their AI systems consume, they gain unprecedented control over information access.
This power shift raises serious questions about information control and accessibility. The stakes become even higher considering that LLMs still frequently make mistakes when processing and reproducing information.
Looking Forward
As we navigate this transition, finding the right balance between human and AI-optimized content will be crucial. While Karpathy's vision of a 99.9% shift toward LLM optimization may seem extreme, the underlying trend is clear: AI is becoming an increasingly important consumer of digital content.
For content creators, developers, and businesses, preparing for this dual audience—human and AI—will likely become a critical consideration in the years ahead. The companies that effectively adapt to this new paradigm may gain significant advantages in discoverability, accessibility, and utility of their digital content.
Unlock the Future of Business with AI
Dive into our immersive workshops and equip your team with the tools and knowledge to lead in the AI era.