Thinking Machines Lab is pushing the boundaries of human-AI collaboration with its research preview of interaction models. This new approach embeds interactivity directly into the AI, rather than relying on external systems, aiming to make working with AI as fluid as collaborating with another person. The models are designed to process audio, video, and text continuously, enabling real-time thinking, responding, and acting.
The core idea is to address what the lab calls the 'collaboration bottleneck.' Current AI systems, often optimized for autonomous tasks, struggle with human-in-the-loop workflows. Users can't always specify needs upfront, and interfaces often push humans out, despite their value in clarifying and providing feedback. The goal is to enable AI interfaces that meet humans where they are, facilitating natural interaction through speaking, listening, seeing, and interjecting.