tech

Thinking Machines by Mira Murati Innovates with Audio-Video Interaction Models in AI

The company Thinking Machines, founded by Mira Murati, former CTO of OpenAI, is developing interaction models capable of continuously understanding audio, video, and text for natural collaboration with AI. This innovation promises to transform usage and user experience in the artificial intelligence sector.

IA
lundi 11 mai 2026 à 22:236 min
Partager :Twitter/XFacebookWhatsApp
Thinking Machines by Mira Murati Innovates with Audio-Video Interaction Models in AI

Thinking Machines launches interaction models for natural AI collaboration

Thinking Machines, the artificial intelligence startup co-founded by Mira Murati, former Chief Technology Officer of OpenAI, has announced the development of what it calls "interaction models." According to The Verge, these models are designed to allow users to collaborate with AI as they would with a human interlocutor, thanks to continuous consideration of multimodal inputs such as audio, video, and text.

This approach aims to overcome the limitations of traditional voice assistants or chatbots that operate sequentially or in isolation. By integrating multiple real-time data streams, Thinking Machines hopes to create AI agents capable of following complex conversations, understanding non-verbal expressions, and dynamically adapting to context.

Unprecedented capabilities for continuous multimodal interaction

Specifically, these interaction models are designed to capture and analyze multiple types of information simultaneously. For example, an AI could listen to a discussion, analyze the gestures and facial expressions of an interlocutor while responding by voice or displaying relevant visual content. This operation resembles natural collaboration between humans, where verbal language is enriched by non-verbal cues.

This innovation is particularly promising for applications in videoconferencing, real-time virtual assistance, or hybrid work environments where multimodal communication is essential. Compared to classic language models, often limited to textual interaction, Thinking Machines aims for a more immersive and intuitive experience.

At this stage, public demonstrations remain limited, but the promise is of an AI capable of maintaining constant "attention" across multiple information channels and responding coherently and contextually. This represents a significant advance in the field of conversational agents and natural interfaces.

Underlying architecture and technical innovation

According to the information reported, the interaction models rely on deep learning architectures capable of merging multimodal data in real time. This fusion requires fast processing and fine understanding of audio, visual, and textual signals, which demands significant computing resources and sophisticated synchronization algorithms.

The technical complexity also lies in training these models, which must learn to interpret varied contexts and adapt to human nuances such as intonation or facial expressions. The approach adopted by Thinking Machines appears to rely on advanced multimodal neural networks capable of integrating these different information streams into a coherent framework.

This marks an important milestone in the evolution of AI, which until now often remained compartmentalized into a single input-output mode. The ability to process multiple types of data simultaneously opens the way to more natural and richer uses.

Accessibility and envisioned use cases

At this stage, the access modalities to Thinking Machines' interaction models are not fully specified. Information regarding public availability, pricing, or a possible API is not confirmed at this time. However, the startup seems to target its offering towards companies wishing to integrate multimodal AI agents into their products or services.

Potential use cases notably include improving personal assistants in domestic or professional environments, remote collaboration tools, or customer support systems capable of understanding emotions and the overall context of an interaction. This innovation could also benefit the augmented and virtual reality sector, where multimodal understanding is crucial.

A turning point in the global competition of interactive AI

This initiative from Thinking Machines comes amid intensifying competition around multimodal AI technologies. Major players like OpenAI, Google DeepMind, and Meta are also investing in models capable of processing audio, visual, and textual data simultaneously.

Relying on the recognized expertise of Mira Murati, who led significant advances at OpenAI, Thinking Machines positions itself as a player to watch closely. Its approach focused on natural and continuous interaction could differentiate its solutions in a market where user experience is a key success factor.

Historical context and evolution of the multimodal AI concept

The development of multimodal interaction models fits into a broader evolution of artificial intelligence, which has long been segmented into distinct specialties such as natural language processing, computer vision, or speech recognition. Historically, these fields progressed relatively independently, limiting AI's ability to understand and react to complex environments rich in varied signals.

With the advent of deep neural architectures and advances in computing power, research has shifted towards merging these different input modes to design more versatile agents. Thinking Machines thus fits into this dynamic, seeking to realize a vision where AI is no longer confined to a single interaction channel but capable of perceiving and interpreting multiple information streams simultaneously, like human communication.

Tactical challenges and hurdles for successful adoption

Integrating multimodal interaction models into commercial products raises several major tactical challenges. On one hand, it is essential to ensure minimal latency so that the user experience remains smooth and natural. This involves advanced software and hardware optimizations, particularly in real-time processing of audio and video data.

On the other hand, managing data privacy and security becomes crucial, especially when sensitive information is continuously captured. Thinking Machines will therefore need to develop robust data protection mechanisms while ensuring transparency about their use. Finally, adapting to varied cultural and linguistic contexts represents an additional challenge for these interaction models to offer a universal and inclusive experience.

Perspectives and potential impact on the AI market

If Thinking Machines succeeds in realizing its goals, it could represent a paradigm shift in how humans interact with machines. AI would no longer be a simple tool to be called upon occasionally but a true partner capable of continuous and contextual interaction, thus enriching work, learning, and communication processes.

This advance could also stimulate innovation across various sectors, from education to healthcare, including entertainment and personal services. By offering more natural, accessible, and intuitive interactions, interaction models could facilitate the widespread adoption of AI in daily and professional life.

Our analysis: a technical promise to be confirmed

Thinking Machines' project is ambitious and responds to a real need to improve human-machine collaboration. However, the technical complexity and challenges related to robustness, data privacy, and contextual interpretation remain considerable. The next development stages and first practical implementations will need to be closely monitored.

Nevertheless, this innovation marks an important step towards AI more integrated into our daily communication modes, with disruptive potential for many sectors. France, which follows AI advances with interest, could draw inspiration from this approach to accelerate its own initiatives in the field of multimodal artificial intelligence.

Commentaires

Connectez-vous pour laisser un commentaire

Newsletter gratuite

L'actu IA directement dans ta boîte mail

ChatGPT, Anthropic, startups, Big Tech — tout ce qui compte dans l'IA et la tech, chaque matin.

LB
OM
SR
FR

+4 200 supporters déjà abonnés · Gratuit · 0 spam