Former Tesla AI Director and influential voice in deep learning, Andrej Karpathy, recently posted on X highlighting a fundamental issue with large language models (LLMs): memory and personalization features not only fail to make models smarter but may actually reinforce a systemic “training bias,” causing the models to increasingly favor answering “common correct answers” rather than the “truly optimal answer.”
Core issue: LLMs “remember” rather than “reason.”
Karpathy’s argument directly challenges the assumptions about how LLMs operate. He points out that the distribution of “examples” in training data is highly uneven—popular problem solutions and frequently discussed answers appear repeatedly, while rare but equally correct answers are almost nonexistent.
This creates a fundamental problem: when answering questions, LLMs are not truly “reasoning” to find the best answer but are searching their memory for the “most common correct example.” In other words, the more mainstream and widely discussed a solution is, the more likely the model is to choose it—even if better or more contextually appropriate options exist.
The unintended effects of personalization memory features
This issue is further amplified in AI assistants’ personalization memory functions. When models remember user preferences, habits, and past conversations, the “user model” they build is essentially a product of the training data distribution—it remembers the “most typical type” of user rather than truly understanding the individual’s unique needs.
This means that the more personalized an LLM becomes, the more likely it is to fit users into certain “prototypes” rather than providing genuinely tailored responses.
Practical implications for AI media reporters
This insight has direct implications for those using AI to assist their work. When analyzing niche cryptocurrency projects, evaluating non-mainstream policy positions, or researching less-discussed technical viewpoints, AI responses may naturally lean toward “mainstream opinion” rather than objective analysis.
Karpathy believes there is no perfect solution at present—only mitigating through more diverse training data. But the fundamental bias—that “models tend to favor popular answers”—is an inherent characteristic of the LLM architecture, not a bug.
Deeper concern: AI is replicating human collective blind spots
Karpathy’s observation points to a deeper worry: training data is a sampling of human writing history, not an objective distribution of knowledge. This means LLMs not only replicate human knowledge but also mirror collective blind spots, biases, and an overemphasis on certain “mainstream narratives.”
As AI is increasingly used for news analysis, investment decisions, and policy evaluation, the scope of training biases expands. This is not merely a technical issue but a cognitive one that requires users to maintain critical awareness.
Why does AI give more “mainstream” answers the more it understands you? Karpathy’s warning about training bias was first published by ABMedia.