All posts

AI Is Leaving the Era of Turn-Based Chat

Every AI you have used works in turns. You speak, it responds. Thinking Machines wants to build models that listen while they generate, turning conversation into something closer to a real phone call

May 12, 20263 min read
Heavy black zine-style mechanical head with a continuous stream of sound flowing through it, replacing turn-based chat bubbles, with blue signal accents showing real-time listening

The Turn-Based Trap

Every AI you have ever used works the same way. You type or speak. Then you wait. The model thinks. Then it responds. Even the slickest voice mode is essentially a text chain with audio pasted on top. It is polite, sequential, and nothing like an actual human conversation.

A startup called Thinking Machines wants to break that habit. They are building models that process your input while they are still generating a response. The goal is an AI that you can interrupt, redirect, or correct in real time, the same way you would on a phone call with a colleague who suddenly remembers a key detail mid-sentence.

Why This Is Technically Hard

Right now, this is brutal to build. Large language models consume tokens in batches. They predict what comes next based on everything that came before. Changing the prompt while the model is already writing the answer is not a feature you can toggle on. It requires a different architecture, one that treats conversation as a continuous stream instead of a back-and-forth rally.

Thinking Machines showed off an early preview this week. The demo points toward near-realtime voice and video conversations where the model actually listens during its own output. That matters because so much of human communication lives in the overlaps. We cut each other off when we spot a misunderstanding. We change topics when a new priority barges in. Current AI agents miss all of that texture, which is why voice assistants still feel like talking to a voicemail tree with a graduate degree.

What Builders Should Start Designing For

For builders, this shift is bigger than it sounds. The last year has been about strapping voice onto text models and hoping the latency stays low enough that users do not notice the lag. But low latency is not the same as real interaction. If Thinking Machines pulls this off, product teams will need to rethink turn management, interruption handling, and state recovery from scratch. The prompt engineering playbook for chatbots will not transplant cleanly into a world where the user can change the game mid-token.

The enterprise angle is just as interesting. Customer support agents, sales bots, and medical scribes all operate in noisy environments where humans do not wait their turn. A doctor correcting a transcription error while the AI is still dictating, or a customer clarifying an order while the agent is mid-sentence, are not edge cases. They are the normal texture of real work. Turn-based AI forces humans to adapt to the machine. Simultaneous models could finally let the machine adapt to us.

This is still early. Thinking Machines has a preview, not a product you can ship today. But the direction is clear. The next generation of AI-native apps will be built around interaction models, not just foundation models. That means the stack changes too. You need a backend that can handle reactive state, real-time data flows, and instant UI updates without the builder having to hand-roll websockets or write deployment scripts.

At Botflow, we see this as where vibe coding is headed. You should be able to describe a real-time conversational app and have it running live in minutes, with a reactive database handling the state sync and a web or mobile frontend updating instantly. That is the standard we are building toward, because the models are clearly heading there. The gap between what AI can say and how smoothly it can say it with you is about to close. Builders just need the right place to start.