Sakana.ai Introduces Transformer2, a Self-Adaptive AI

The researchers explain that Transformer2 can adapt like a living brain.
Sakana.ai Introduces Transformer2, a Self-Adaptive AI

Sakana.ai, a Tokyo-based AI and R&D startup, just released a self-adaptive AI system called Transformer2. The company proposed this as an ML system that dynamically adjusts its weights for various tasks.

Sharing a video on X, the company announced on Wednesday, saying, “Adaptation is a remarkable natural phenomenon, like how the octopus can blend in with its environment or how the brain rewires itself after injury.”

This system, introduced in a research paper ‘Transformer2: Self-adaptive LLMs’, offers a dynamic approach to task handling, setting it apart from traditional, static AI models.

Transformer² utilises a two-step process to adapt its weight matrices in real time, tailoring its operations for specific tasks such as mathematics, coding, reasoning, and visual understanding. 

“This model analyses incoming tasks, adjusts its weights, and delivers optimal results through task-specific adaptations,” the researchers said

Check out the GitHub repository here: https://github.com/SakanaAI/self-adaptive-llms

The Brain of LLMs

The system uses a mathematical technique called Singular Value Decomposition (SVD) to understand which parts of the AI are important for different tasks. 

Left: LLM’s “brain” (i.e., weight matrices) into several independent components using SVD. | Right: Using RL to train the combination of these components for various tasks. 

It then uses a method combining SVD fine-tuning with reinforcement learning (RL) to create instructions for adjusting the model’s behaviour, represented as compact ‘z-vectors’.

During inference, the tool employs three strategies—prompt-based, classifier-based, and few-shot adaptation—to detect task types and adjust accordingly. 

The researchers noted, “This approach ensures robust and efficient adaptation, outperforming static systems like LoRA across a range of scenarios.”

Superior Performance

Tests on tasks across both the Llama and Mistral LLMs, including GSM8K (math), HumanEval (coding), and TextVQA (visual understanding), revealed superior performance, with significant gains in adaptability and efficiency.

One surprising discovery was that when solving complex math problems, Transformer2 combines different types of reasoning—not just mathematical but also programming and logical thinking—similar to how humans approach complex problems.

In an unexpected breakthrough, the researchers found that knowledge gained by one AI model could be transferred to another. 

When they moved the learning patterns from one model (Llama) to another (Mistral), the second model showed improved performance on most tasks. However, the researchers note that this worked because both systems had similar underlying structures.

Left: Self-adaptation on unseen tasks. | Right: Learned z-vectors interpolation weights.

“This marks a significant step toward creating ‘living intelligence’ in AI systems,” the research team explained. They envision future AI systems that can continuously learn and adapt like living beings rather than remaining fixed after their initial training.

They concluded, “This marks a shift from static AI to dynamic models capable of lifelong learning and adaptation, redefining how we interact with intelligent systems.”

📣 Want to advertise in AIM? Book here

Picture of Sanjana Gupta
Sanjana Gupta
An information designer by training, Sanjana likes to delve into deep tech and enjoys learning about quantum, space, robotics and chips that build up our world. Outside of work, she likes to spend her time with books, especially those that explore the absurd.
Related Posts
AIM Print and TV
Don’t Miss the Next Big Shift in AI.
Get one year subscription for ₹5999
Download the easiest way to
stay informed