AI & TechArtificial IntelligenceNewswireStartupsTechnology

Thinking Machines Lab Aims for More Consistent AI Models

▼ Summary

– Thinking Machines Lab is researching how to make AI models produce reproducible responses, addressing the randomness in current systems.
– The lab attributes this randomness to GPU kernel orchestration during inference and believes careful control can achieve determinism.
– Reproducible responses could improve reinforcement learning by reducing noise in training data and making the process smoother.
– The lab plans to share research openly through its new blog “Connectionism,” aiming to benefit the public and improve its own culture.
– Thinking Machines Lab faces the challenge of solving these AI problems and developing products to justify its $12 billion valuation.

The newly established Thinking Machines Lab, backed by $2 billion in seed funding and led by former OpenAI researchers, is tackling one of artificial intelligence’s most persistent challenges: the unpredictability of AI model responses. In its inaugural research blog post, the lab outlined a focused effort to create AI systems capable of delivering reproducible and consistent answers, a significant departure from the non-deterministic behavior common in today’s large language models.

Anyone who has repeatedly asked a model like ChatGPT the same question knows the results can vary widely. This randomness has generally been accepted as an inherent trait of contemporary AI systems. Thinking Machines Lab, however, views it as a solvable engineering problem. Their recent publication, titled “Defeating Nondeterminism in LLM Inference,” identifies the underlying technical cause of this variability.

According to researcher Horace He, the inconsistency stems from how GPU kernels, specialized programs within Nvidia’s hardware, are coordinated during inference, the phase when a model generates a response after receiving a prompt. He argues that by refining this orchestration layer, AI models can be made significantly more deterministic, producing the same output given identical inputs.

This pursuit of consistency isn’t merely academic. More reliable model behavior could profoundly benefit enterprises and scientific applications where reproducibility is critical. Additionally, He highlights that reducing randomness could enhance reinforcement learning (RL) training, a method where models learn from feedback. When model outputs vary slightly, training data becomes noisy, complicating the learning process. Smoother, more predictable responses could streamline RL, making it more effective.

Thinking Machines Lab has indicated to investors its intention to use reinforcement learning to develop customized AI solutions for business clients. This aligns with comments from Mira Murati, the lab’s lead and former OpenAI CTO, who hinted that an initial product aimed at researchers and startups will be introduced in the coming months. While it remains unclear whether this product will incorporate the techniques described in the new research, the direction suggests a strong focus on reliability and customization.

The lab has also committed to a culture of openness, promising regular releases of blog posts, code, and research findings through a new series called “Connectionism.” This stands in contrast to the increasingly guarded approach of some larger AI firms, raising questions about whether such transparency can be maintained as the organization grows.

This early disclosure offers a rare look inside one of Silicon Valley’s most closely watched AI startups. While the full scope of their technology remains undisclosed, their focus on foundational issues like determinism signals ambitious goals. The ultimate measure of success will be whether Thinking Machines Lab can translate these research insights into practical, high-value products that justify its substantial $12 billion valuation.

(Source: TechCrunch)

Topics

ai determinism 95% gpu kernels 85% ai randomness 85% mira murati 80% reinforcement learning 80% inference processing 80% product development 75% research transparency 75% research blog 75% openai alumni 75%