01:07:09
In a revealing conversation with Dwarkesh Patel, Richard Sutton—a pioneering figure in reinforcement learning and recent Turing Award recipient—shared his critical perspective on large language models (LLMs). Sutton argues that LLMs, despite their current popularity, represent a misguided direction in artificial intelligence because they lack fundamental elements like true goals and world models. Instead, he champions reinforcement learning as the foundational approach for building intelligent systems that learn from experience.
Sutton emphasizes that reinforcement learning (RL) and LLMs stem from fundamentally different philosophies. RL focuses on learning from direct experience—taking actions, observing outcomes, and adapting to maximize rewards—which Sutton views as the essence of intelligence. In contrast, LLMs are trained on vast datasets of human-generated text, learning to mimic patterns without inherent goals or the ability to interact with the world. This makes LLMs more about imitation than genuine understanding, limiting their potential for true AI advancement.
He challenges the notion that LLMs possess robust world models, pointing out that they predict text rather than real-world consequences. For instance, while an LLM might generate a plausible response, it doesn't experience surprise or adjust its model based on unforeseen events, a key aspect of learning in biological systems.
A central theme in Sutton's critique is the absence of goals in LLMs. He defines intelligence as the computational ability to achieve goals, something RL systems embody through reward mechanisms. LLMs, however, operate on next-token prediction, which Sutton describes as a superficial objective that doesn't translate to meaningful world interaction. Without a goal-oriented framework, LLMs cannot distinguish between right and wrong actions in a dynamic environment, hindering their ability to learn continuously or adapt to new situations.
Sutton also disputes the idea that LLMs can serve as a prior for experiential learning. He argues that without a grounding in real-world truth, any prior knowledge derived from LLMs is inherently flawed. In RL, prior knowledge can be tested and refined against experiential data, but LLMs lack this feedback loop, making them unreliable as a foundation for advanced AI.
The conversation delves into human learning analogies, where Sutton diverges from the view that imitation is a primary driver. He asserts that most learning in animals and humans stems from trial-and-error and prediction-based mechanisms, not imitation. While cultural transmission in humans involves some imitation, Sutton sees this as a minor layer atop more fundamental RL processes. This perspective reinforces his belief that AI should prioritize experiential learning over mimicking human data, as the latter cannot capture the richness of real-world interactions.
Sutton envisions a future where AI systems learn continually from their experiences, much like mammals do. He outlines key components for such systems: a policy for decision-making, a value function for predicting long-term rewards, and a world model for understanding consequences. These elements enable agents to handle sparse rewards and complex tasks, such as long-term projects like startups, by breaking them down into manageable steps using temporal difference learning.
However, he acknowledges current limitations in generalization and transfer learning within RL systems. Unlike LLMs, which show broad generalization across language tasks, RL methods often struggle with catastrophic interference when faced with new data. Sutton attributes this to a lack of automated techniques for promoting good generalization, noting that human intervention is often required to achieve robust performance.
Beyond technical debates, Sutton reflects on the societal impact of AI advancement. He presents a four-part argument for the inevitability of AI succession: the lack of global consensus, the eventual understanding of intelligence, the achievement of superintelligence, and the natural tendency for smarter entities to gain power. Sutton encourages a positive outlook, framing this transition as a monumental step in the universe's evolution—from replication-based life to design-based intelligence.
He suggests that humanity should focus on instilling robust, prosocial values in AI systems, similar to educating children, rather than trying to control every aspect of the future. This approach, he believes, can lead to voluntary and beneficial evolution, mitigating risks while embracing the potential of digital intelligences.
Key Insight: Richard Sutton's views highlight a pivotal divide in AI research: while LLMs excel at pattern recognition, they fall short in embodying true intelligence. Reinforcement learning, with its focus on goals and experience, offers a more scalable path toward artificial general intelligence, urging the field to prioritize fundamental principles over short-term trends.