NTT R&D Forum2023 Special session2:Natured Inspired Intelligence and a New Paradigm for LLM

NTT official channel
7 Dec 202335:28

TLDRIn this special session of the NTT R&D Forum 2023, David and Lion Jones from Sakana AI discuss their nature-inspired approach to AI research. David shares his unconventional journey from studying neural networks at the University of Toronto to working on generative AI at Google. Lion Jones, co-inventor of the Transformer architecture, advocates for character-level language modeling, highlighting its advantages, especially for languages like Japanese. They emphasize the need for a new paradigm in AI development, focusing on collective intelligence and complex adaptive systems, and outline their vision for advancing AI research in Japan.

Takeaways

  • 🐟 Sakana AI is an AI R&D company founded by the speaker and his friend Lion Jones, inspired by the concept of collective intelligence and evolution.
  • 🤖 The speaker's background includes working on Wall Street as a derivatives trader and later as a research scientist at Google Brain, focusing on generative AI.
  • 🎓 Lion Jones is known as the co-inventor of the Transformer architecture, which is foundational to many modern AI applications.
  • 🔍 The speaker believes that current large-scale AI models are not the path to strong AI, suggesting that a new approach inspired by complex adaptive systems is needed.
  • 🌐 The speaker proposes that intelligence is not just about the parameters of a neural network, but also about how systems are organized and adapt to their environment.
  • 📚 Sakana AI aims to challenge conventional AI development methods and explore nature-inspired collective intelligence systems.
  • 🌏 The decision to start the company in Japan is motivated by a desire to develop a strong AI R&D ecosystem in the country and to diversify the global AI landscape.
  • 💡 The speaker emphasizes the importance of character-level language modeling, suggesting it offers advantages for languages like Japanese and could be beneficial for AI development.
  • 🔠 Character-level language models can potentially improve the accuracy of spelling and understanding within AI systems, including for languages with complex scripts.
  • 🌐 The speaker calls for a more diverse and inclusive approach to AI development, including considering non-Western-centric perspectives and languages.
  • 🚀 Sakana AI's long-term vision includes creating new types of foundation models that are more adaptive, resilient, and possibly based on swarm intelligence principles.

Q & A

  • What is the meaning behind the name 'Sakana AI' and how does it reflect the company's philosophy?

    -The name 'Sakana AI' is derived from the Japanese word 'Sakana,' which means fish. The company's logo represents a swarm of fish forming a coherent entity from independent rules, symbolizing the company's research ideas that are inspired by nature, such as collective intelligence and evolution. The red fish in the logo signifies the company's desire to not just follow the crowd but to pursue innovative ideas in AI.

  • Who is Lion Jones and what is his significant contribution to the field of AI?

    -Lion Jones is the co-founder of Sakana AI and is known as the co-inventor of the Transformer architecture, which is now a fundamental component powering various AI applications, including chat GPT and stable diffusion.

  • What is the background of Sakana AI's other co-founder, and how did his previous work lead him to Google Brain?

    -The other co-founder comes from a non-conventional background, having studied engineering science with a focus on neural networks at the University of Toronto during the first neural network winter. After working on Wall Street as a derivatives trader, he was recruited by Google to join the Google Brain team due to his continued research in generative AI and nature-inspired approaches.

  • What is the speaker's view on the current approach to training large machine learning models, and what does he propose instead?

    -The speaker believes that the current approach of training very large language models is energy inefficient, rigid, and prone to security attacks. He proposes a new type of foundation model inspired by the principles of complex adaptive systems, which would be more adaptive and part of the environment, rather than large engineering projects.

  • What is the concept of 'collective intelligence' as discussed in the paper by the speaker, and how does it relate to AI?

    -Collective intelligence in the context of the paper refers to the idea of building AI systems that mimic the way natural systems adapt and become part of their environment. It suggests that intelligence is not just about the parameters or weights in a neural network but also about how systems are organized and interact with each other.

  • What is the 'Evo Jax' project mentioned in the script, and how does it demonstrate the potential of evolution and collective intelligence in AI?

    -Evo Jax is a framework built on top of Jax that allows for the use of evolution and collective intelligence algorithms to train artificial life creatures on TPUs and GPUs. It shows that these nature-inspired approaches can scale to Google's level and have the potential to be part of future foundation models.

  • Why did the speaker choose to start Sakana AI in Japan, and what is his vision for the company's role in the global AI ecosystem?

    -The speaker chose to start Sakana AI in Japan because he sees the country's potential to develop its own strong AI R&D ecosystem. He believes that Japan, with its supportive culture for AI, can become a leader in the AI space, especially in Asia, and provide an alternative to the Western or Chinese-centric AI development.

  • What is Lion Jones' perspective on character-level language modeling, and why does he advocate for its use?

    -Lion Jones believes that character-level language modeling is advantageous because it can handle out-of-vocabulary words better and requires a smaller number of parameters. He also suggests that character-level modeling is a better fit for languages like Japanese, where the computational difference between word-level and character-level is not as significant as in English.

  • What are the challenges and potential solutions discussed in the script regarding the current state of AI and the development of large language models?

    -The script discusses the challenges of energy inefficiency, rigidity, and security vulnerabilities in current large language models. The potential solution proposed is the development of new foundation models inspired by complex adaptive systems, which would be more adaptive and resilient, possibly using a swarm of agents rather than a single large model.

  • How does the speaker's background in studying neural networks and working on Wall Street influence his approach to AI research?

    -The speaker's background in neural networks during the first neural network winter and his experience as a derivatives trader on Wall Street during the financial crisis have given him a unique perspective on the complexities of the world and the limitations of models. This has influenced his approach to AI research, focusing on developing new ideas and approaches that can make machine learning and AI do new things.

  • What is the significance of the 'weight agnostic neuron networks' research mentioned in the script, and what does it imply for the future of AI?

    -The 'weight agnostic neuron networks' research demonstrated that it is possible to evolve neural network architectures that can perform tasks even if the weights are randomized, suggesting that intelligence is not solely about the parameters or weights in the network. This implies that the future of AI could involve more adaptive and organized systems that go beyond the current paradigm of large neural networks.

Outlines

00:00

🐟 Introduction to Sakana AI and Its Nature-Inspired Philosophy

The speaker expresses gratitude for being invited to the NTT R&D Forum and introduces Sakana AI, an AI R&D company co-founded with Lion Jones. The company's name, meaning 'fish' in Japanese, symbolizes the collective behavior of fish and the founders' desire to pursue independent and innovative ideas in AI, inspired by nature and evolution. The speaker outlines the talk's agenda, which includes discussing the company's team, vision, technical bets, and the rationale for starting the company in Japan. The speaker's background is highlighted, including his unconventional journey from studying neural networks during a period of limited interest to working in investment banking during the financial crisis, and eventually joining Google Brain to work on generative AI. His research has been nature-inspired, focusing on collective intelligence, evolution, and training AI systems to perform tasks akin to learning and adaptation.

05:01

🧠 Challenging Conventional AI and the Vision for Sakana AI

The speaker discusses his skepticism about the current approach to training machine learning models and the pursuit of strong AI through ever-larger models. He advocates for a new foundation model inspired by complex adaptive systems and collective intelligence. The speaker's research background includes work on hyper networks, language models for generating sketches, and video generation models. He emphasizes the need for a more adaptive and energy-efficient approach to AI development, drawing parallels to natural systems that adapt and integrate into their environments. The speaker also touches on his belief that intelligence is about organization and collective behavior, not just the parameters within a neural network.

10:03

🌐 The Importance of Developing a Japanese AI Ecosystem

The speaker shares his perspective on the importance of establishing a strong AI R&D ecosystem in Japan, independent of Western or Chinese influence. He believes that Japan, with its supportive culture for AI and strong economy, is well-positioned to develop a global AI ecosystem. The speaker expresses his intention for Sakana AI to act as a catalyst in this development. He also addresses the challenges of attracting top AI talent in Japan but argues that the country's appeal makes it easier to draw talent, both domestically and internationally. The speaker reflects on his personal journey and the significance of collective intelligence in the pursuit of strong AI.

15:04

📚 The Evolution of Language Modeling and the Case for Character-Level Models

The speaker, known for his work on the Transformer model, discusses his earlier and subsequent work centered around character-level language modeling. He explains the concept of character-level modeling, where inputs are fed character by character rather than word by word, and its advantages, such as handling out-of-vocabulary words. The speaker details the computational and data challenges of character-level models and the development of techniques to make them effective. He also presents the benefits of character-level models for languages like Japanese, which have rich morphological variations, and encourages the adoption of character-level modeling in the AI community.

20:06

🔠 Advocating for Character-Level Language Models in Japanese and Other Languages

The speaker reiterates the benefits of character-level language models, especially for languages like Japanese, and urges the industry to adopt this approach. He contrasts the handling of English and Japanese by word-level and character-level models, highlighting the efficiency and accuracy of the latter for Japanese. The speaker also addresses the potential of character-level models for other character-based languages like Korean and Chinese, suggesting that understanding the deeper structures of language could facilitate learning and translation across various languages.

25:07

🤖 The Power and Limitations of Large Language Models

The speaker explores the capabilities and limitations of large language models, using examples of image generation and spelling accuracy. He points out that while large models can achieve impressive results, they still struggle with tasks that character-level models handle more naturally, such as spelling and reversing words. The speaker emphasizes the need to simplify AI approaches by adopting character-level modeling, which can solve many of the issues associated with word-level models. He concludes by encouraging the use of character-level models, particularly for the Japanese language, and hints at the broader implications for AI development.

30:07

💬 Q&A: Discussing the Impact of Language on AI and the Universality of Language Structures

In the concluding section, the speaker engages in a Q&A session, addressing questions about the influence of studying Japanese on one's perspective and the potential benefits of character-level modeling for developing multilingual AI systems. He discusses the cross-cultural understanding that multilingual training can provide and the unique challenges and opportunities presented by the Japanese-English language pair. The speaker also contemplates the idea of a universal language structure that AI models can learn, suggesting that training on multiple languages can lead to a deeper understanding of language that transcends specific linguistic pairs.

Mindmap

Keywords

💡NTT R&D Forum

The NTT R&D Forum is a conference where the latest research and development activities in the field of information and communication technology are presented and discussed. In the script, it is the event where the speaker has been invited to present Sakana AI's work, indicating its significance and relevance in the tech community.

💡Sakana AI

Sakana AI is an artificial intelligence research and development company founded by the speaker and his friend Lion Jones. The name 'Sakana' is derived from the Japanese word for fish, symbolizing the concept of collective intelligence and the independent, non-conformist approach of the company to AI innovation.

💡Nature-Inspired Intelligence

Nature-Inspired Intelligence refers to the approach of drawing inspiration from natural phenomena and biological processes to develop intelligent systems. The speaker mentions this concept as a core principle guiding Sakana AI's research, emphasizing the importance of collective intelligence, evolution, and adaptation from nature in creating advanced AI.

💡Transformer Architecture

The Transformer Architecture is a type of deep learning model introduced by Lion Jones, co-founder of Sakana AI. It has become a foundational component in various AI applications, including language processing models like chat GPT and image generation models like stable diffusion. The script highlights its significance in the current landscape of AI technologies.

💡Generative AI

Generative AI refers to artificial intelligence systems that can create new content, such as text, images, or music, that is similar to the content they have been trained on. The speaker discusses his work on generative AI, including training AI systems to generate new types of kanji and high-resolution images, showcasing the creative potential of this technology.

💡Evolution and Self-Play

Evolution and self-play are techniques used in training AI agents where agents learn through repeated play against copies of themselves or through evolutionary processes. The speaker used these methods to train agents for tasks like playing volleyball, which demonstrates the ability of AI to learn complex behaviors through iterative self-improvement.

💡Hypernetworks

Hypernetworks are a type of neural network that generates the weights for another neural network rather than generating data. The speaker mentions this concept as part of his research, indicating its role in fine-tuning modern large language models like those used in chat GPT.

💡Language Model

A language model is an AI system that is trained to predict the likelihood of a sequence of words appearing in a given context. The speaker discusses his work on language models, not just for generating language but also for creating sketches and other abstract representations, which illustrates the versatility of language models in AI.

💡World Models

World Models are AI systems that generate a simplified representation of the environment to train reinforcement learning agents more efficiently. The speaker collaborated on one of the first video generation models of this type, demonstrating the use of generative AI for training purposes in machine learning.

💡Collective Intelligence

Collective Intelligence is the idea that intelligence can emerge from the collaboration and competition of many individuals, rather than being centralized in a single entity. The speaker argues for the development of AI systems based on this principle, suggesting that it could lead to more adaptive and resilient AI models.

💡EvoJax

EvoJax is a framework that combines evolutionary algorithms with neural network training to create artificial life creatures. The speaker's work on EvoJax at Google demonstrates the potential of using nature-inspired techniques to train AI at a large scale.

💡Character-Level Language Modeling

Character-Level Language Modeling is an approach where language models process text at the character level rather than the word level. Lion Jones, in the script, advocates for this method due to its advantages in handling out-of-vocabulary words and its potential benefits for languages like Japanese that have a large character set.

Highlights

Sakana AI is founded by the speaker and his friend Lion Jones, inspired by the concept of collective intelligence from nature.

The company's name 'Sakana' is derived from the Japanese word for fish, symbolizing a group forming a coherent entity.

Lion Jones is the co-inventor of the Transformer architecture, which powers AI models like chat GPT and stable diffusion.

The speaker has a non-conventional background, with experience in investment banking and a focus on generative AI.

Sakana AI aims to challenge conventional AI approaches and develop new models inspired by complex adaptive systems.

The speaker believes that current large AI models are energy-inefficient and not adaptive enough.

A paper by the speaker, 'Collective Intelligence for Deep Learning,' explores nature-inspired AI systems.

The speaker's research includes training AI systems using evolution and self-play for tasks like volleyball agents.

Sakana AI's vision is to create AI systems that are more like a swarm of agents working together, rather than a single model.

The company is based in Japan, aiming to develop a strong AI R&D ecosystem in the country.

The speaker discusses the importance of Japan developing its own AI ecosystem, independent of Western or Chinese influence.

Sakana AI has received significant interest from AI talents globally, attracted by the unique opportunity in Japan.

Lion Jones advocates for character-level language modeling, which has advantages for languages like Japanese.

Character-level models can improve the accuracy of spelling in generated content, as demonstrated in image generation.

Lion Jones believes the industry will eventually move towards character or even bit-level modeling.

The speaker emphasizes the importance of diverse perspectives in AI development, including non-Western languages.

Lion Jones discusses the potential for character-level modeling to improve understanding across different languages.