What Kind of Computation Is Cognition?
TLDRIn a talk at Yale, Professor Josh Tenenbaum explores the nature of cognition as computation. He discusses the reverse-engineering approach to understanding the human mind by building models akin to AI systems, focusing on intuitive physics and psychology. Tenenbaum emphasizes the importance of integrating neural network pattern recognition with symbolic languages and probabilistic inference to capture human common sense. He also highlights the challenges in cognitive science and AI, including the current limitations in self-driving cars and the promise of probabilistic programming in advancing our understanding of the mind.
Takeaways
- 🌟 Professor Josh Tenenbaum, a computational cognitive scientist at MIT, discusses the nature of cognition and its computational aspects in a seminar at Yale.
- 🤔 The central question of the talk is 'What kind of computation is cognition?', aiming to explore the computational processes underlying human thinking.
- 🏆 Tenenbaum's work and genius were recognized by the MacArthur Foundation in 2019, highlighting his contributions to the field of cognitive science.
- 📊 Tenenbaum presents the 'reverse engineering approach' to understanding cognition, which involves characterizing the human mind using terms from intelligent machine engineering.
- 🤖 He discusses the limitations of current AI technologies, emphasizing the gap between artificial intelligence and the flexible, general-purpose intelligence humans possess.
- 🚗 The discussion includes the challenges in developing self-driving cars, illustrating the long tail of problems that arise even after significant investment and progress.
- 🧠 Tenenbaum touches on the importance of understanding the brain's computations and how they might be represented in computational models.
- 🎯 The goal is to build models that not only match human behavior qualitatively but also solve real-world problems, providing a quantitative account of human cognition.
- 📈 He introduces the concept of 'probabilistic programs' which combine neural network pattern recognition, symbolic languages for knowledge representation, and probabilistic inference.
- 🎮 Tenenbaum also brings up the idea of a 'game engine in your head', suggesting that our minds use similar principles to game engines to simulate and predict outcomes.
- 📚 The talk concludes with a call for further research into how these computational models can be instantiated in the brain and the implications for understanding human learning and intelligence.
Q & A
What is the main topic of Josh Tenenbaum's talk?
-The main topic of Josh Tenenbaum's talk is the nature of cognition as a form of computation, specifically exploring the question, 'What kind of computation is cognition?'
How does Tenenbaum approach the study of cognition?
-Tenenbaum approaches the study of cognition using the reverse-engineering approach, which involves characterizing how the human mind works in terms that would be used to engineer an intelligent machine, focusing on natural intelligence rather than artificial intelligence.
What does Tenenbaum mean by 'reverse engineering' in the context of cognitive science?
-In the context of cognitive science, 'reverse engineering' refers to the process of trying to characterize how the human mind works using the same terms and principles that would be applied to engineer an intelligent machine, with the goal of building models that resemble AI systems but are focused on natural intelligence.
What are the key components of the 'game engine in your head' concept?
-The 'game engine in your head' concept involves using tools from game engines, such as physics engines, to create a rich, immersive, interactive experience within the mind. This includes models of physical objects, agents with goals and beliefs, and the ability to simulate and predict outcomes based on these models.
How does Tenenbaum's research on intuitive physics and intuitive psychology relate to the study of metaphysics?
-Tenenbaum's research on intuitive physics and intuitive psychology is deeply connected to the study of metaphysics because it involves understanding the basic concepts of physical objects, causal interactions, and goal-driven intentional agents, which are fundamental to our common sense and philosophical inquiry.
What are the limitations of current AI technologies like self-driving cars in terms of cognitive abilities?
-Current AI technologies like self-driving cars are limited in their cognitive abilities because they lack the flexible, general-purpose kind of common sense intelligence that humans possess. They are good at specific tasks but struggle with the unpredicted cases and the need for robust, real-world problem-solving.
What is the significance of the Heider and Simmel movie in cognitive science?
-The Heider and Simmel movie is significant in cognitive science because it demonstrates the human ability to attribute goals, emotions, and social interactions to simple moving shapes, highlighting our innate capacity for understanding the world in terms of physical objects and goal-driven agents.
How does Tenenbaum's work on probabilistic programming languages contribute to the understanding of cognition?
-Tenenbaum's work on probabilistic programming languages contributes to the understanding of cognition by providing a computational paradigm that combines neural network pattern recognition, symbolic language representation, and probabilistic inference to model the common sense intuitive physics and psychology that underlie human thought and behavior.
What is the role of symbolic languages in cognitive science?
-Symbolic languages play a crucial role in cognitive science as they are essential for abstract knowledge representation and reasoning. They form the basis of natural language, mathematics, and all of computing, including programming languages used for AI and machine learning.
How does Tenenbaum's research on the 'common sense core' relate to developmental psychology?
-Tenenbaum's research on the 'common sense core' relates to developmental psychology by exploring the basic concepts that humans are built to understand from the earliest stages of development, such as intuitive physics and psychology, which are also central topics in the study of human brain development and cognitive growth.
Outlines
🎤 Welcoming Professor Josh Tenenbaum
The speaker warmly welcomes Professor Josh Tenenbaum, a computational cognitive scientist at MIT, to speak about the nature of cognition. His achievements and recognition by the MacArthur Foundation are highlighted, and the anticipation for his talk on the computation of cognition is expressed.
🤔 The Central Question of Cognition
Josh Tenenbaum introduces the central question of his talk: what kind of computation is cognition? He discusses the importance of understanding cognition in terms of computation and the challenges of finding meaningful answers. The talk aims to raise foundational questions about modeling the mind and the relationship between cognitive science and metaphysics.
🧠 The Gap Between AI and True Intelligence
Tenenbaum discusses the current state of artificial intelligence, noting its increasing utility but lack of true general intelligence. He uses the example of self-driving cars to illustrate the challenges and investments in AI, emphasizing the need for fundamental advances in AI to bridge the gap between current technologies and human-like intelligence.
🚗 The Progress and Limitations of Autonomous Driving
The speaker delves into the specifics of autonomous driving, highlighting the progress made and the underlying AI technologies like deep learning and neural networks. Despite advancements, he points out the limitations and the 'long tail' of problems that remain unsolved, indicating that AI still has a long way to go to achieve true autonomy.
🧠🤖 The Computational Aspect of Intelligence
Tenenbaum explores the computational aspects of intelligence, focusing on the 'common sense core' of human cognition. He discusses the intuitive understanding of physics and psychology that humans possess from a young age and the challenge of capturing these aspects computationally.
👶 Intuitive Physics and Psychology in Infants
The speaker uses examples of infants engaging with objects to illustrate intuitive physics and psychology. He discusses how young children demonstrate an understanding of physical objects and goal-driven agents through their actions and reactions, highlighting the innate human capacity for modeling the world.
🎥 Heider and Simmel: A Cognitive Science Classic
Tenenbaum presents the famous Heider and Simmel animation as a key example of how humans naturally interpret interactions between agents, even from simple shapes. He discusses the rich understanding of physical dynamics, goals, and emotions that humans exhibit when viewing such animations, emphasizing the computational challenge of replicating this understanding in AI.
💡 Probabilistic Programs and the Future of AI
The speaker introduces the concept of probabilistic programming as a way to integrate neural networks, symbolic languages, and probabilistic inference to model common sense. He discusses the potential of this computational paradigm to advance AI and our understanding of cognition, including the role of game engines in simulating physical and social interactions.
🧠 Intuitive Physics and Causal Responsibility
Tenenbaum discusses research on intuitive physics and how people judge causal responsibility. He describes experiments that use eye-tracking to understand how people's gaze patterns change based on whether they are making predictions or causal judgments, revealing the computational processes underlying human intuition about physical interactions.
🛠️ Problem-Solving with Probabilistic Intuitive Physics
The speaker talks about recent work on using probabilistic intuitive physics for problem-solving, particularly in creative tool use. He describes a virtual tools game that demonstrates how people use trial and error to solve problems and how a probabilistic simulation-based model can capture this process, including learning dynamics.
🤖 AI Models of Multi-Agent Interactions
Tenenbaum presents a model called 'Flatland' that simulates multi-agent interactions with physical objects, capturing complex behaviors like collaboration and competition. He emphasizes the potential of these models to understand and predict social interactions and causal relationships, highlighting their success in mimicking human-like interactions.
🧠 The Brain and Learning in Computational Terms
In the final paragraph, Tenenbaum touches on the relationship between these computational models and the brain, and the potential for understanding how humans learn. He suggests that future learning algorithms may need to be 'program learning programs,' capable of writing their own algorithms, much like how humans construct knowledge.
🎓 Yale and the Future of Cognitive Science
Tenenbaum concludes his talk by reflecting on his educational experience at Yale and the future of cognitive science. He discusses the potential of algorithms that write algorithms, and the importance of understanding the brain's computational processes. He ends with a call to explore the synthesis of symbolic representations, probabilistic inference, and learning in understanding cognition.
Mindmap
Keywords
💡Computational Cognitive Science
💡Cognition
💡Metaphysics
💡Epistemology
💡Reverse Engineering
💡Artificial Intelligence (AI)
💡Deep Learning
💡Intuitive Physics
💡Intuitive Psychology
💡Probabilistic Programming
Highlights
Professor Josh Tenenbaum discusses the computation of cognition and the reverse-engineering approach to understanding the human mind.
Tenenbaum's research focuses on building models that mimic AI systems to understand natural intelligence.
The importance of comparing models to data, both qualitatively and quantitatively, to ensure accuracy in cognitive modeling.
The gap between current AI technologies and the flexible, general-purpose intelligence of humans.
Self-driving cars exemplify the challenges and limitations of current AI in real-world scenarios.
The potential of probabilistic programming to integrate neural network pattern recognition with symbolic languages and probabilistic inference.
The concept of 'the game engine in your head,' which models the intuitive physics and psychology of everyday interactions.
Intuitive physics in young children and the computational models that attempt to replicate their understanding of the physical world.
The use of simulation programs and probabilistic inference to create computational models of the human mind's theory of mind.
The significance of counterfactual analysis in understanding causality and causal responsibility in cognitive science.
The application of probabilistic intuitive physics in predicting human behavior and creative problem-solving.
The development of a probabilistic generative model that captures multi-agent interactions in a virtual environment.
The future of human learning and machine learning involves algorithms that write algorithms, moving beyond traditional neural network training.
The challenge of creating physical circuits or chips that implement the probabilistic computations of cognitive models.
The coherence of scientific models and the quest for a satisfying understanding of the mind through computational approaches.
The integration of symbolic representations, probabilistic inference, and learning as essential components of understanding cognition.