Sakana AI’s TreeQuest Boosts Performance 30% With Multi-Model Teams

▼ Summary
– Sakana AI developed Multi-LLM AB-MCTS, a technique enabling multiple LLMs to collaborate on complex tasks by combining their unique strengths.
– The approach allows enterprises to dynamically leverage different AI models for specific task components, improving overall system robustness and performance.
– The method uses Adaptive Branching Monte Carlo Tree Search (AB-MCTS) to balance refining existing solutions and generating new ones, optimizing decision-making.
– Testing on the ARC-AGI-2 benchmark showed the system solved 30% of problems, outperforming individual models and correcting errors through collective intelligence.
– Sakana AI released TreeQuest, an open-source framework, to help businesses apply Multi-LLM AB-MCTS for tasks like coding, optimization, and reducing AI hallucinations.
Japanese AI research lab Sakana AI has developed an innovative approach that enables multiple large language models to collaborate, achieving performance gains of up to 30% on complex tasks. This breakthrough technique, known as Multi-LLM AB-MCTS, allows different AI models to pool their strengths, creating a synergistic effect that outperforms any single model working alone.
The method addresses a critical challenge in enterprise AI adoption, no single model excels at every task. Some specialize in coding, others in creative writing or logical reasoning. By forming AI “dream teams,” businesses can dynamically assign the right model to the right subtask, optimizing performance without being tied to a single provider.
The key innovation lies in inference-time scaling, a technique that enhances AI performance after training by strategically allocating computational resources. Unlike traditional methods that focus solely on bigger datasets or larger models, this approach optimizes how models work together in real time.
The algorithm dynamically decides which strategy to prioritize, ensuring efficiency. But Sakana AI took it further by introducing Multi-LLM AB-MCTS, which also selects the best model for each subtask. Initially, the system tests all available models, then gradually shifts workload to the most effective ones.
For businesses, this means reduced hallucinations and higher accuracy, critical for applications like legal analysis, financial forecasting, and technical troubleshooting.
By enabling AI models to collaborate, Sakana AI’s innovation could redefine how enterprises deploy AI, delivering smarter, more reliable results without requiring massive infrastructure upgrades.
(Source: Venturebeat)
