News

Perplexity's Comet Voice Mode Upgrade: Your Browser Now Actually Listens

Perplexity and Comet rolled out voice mode upgrades today, and the timing isn't accidental. With ChatGPT's Advanced Voice Mode winning users and Gemini Live feeling half-baked, Perplexity is betting that voice-first interaction is the next interface battleground. They're not wrong.

What It Is

The voice mode upgrade brings hands-free browsing to Perplexity's Comet AI browser. Users can now navigate sites, ask questions about whatever's on their screen, and maintain context across multiple tabs—all via voice. The assistant can see your current page, understand what you're looking at, and respond conversationally without you typing a word. It's essentially turning your browser into an always-available research assistant that you can talk to like a person.

How This Helps Today

For researchers and knowledge workers, this eliminates the friction of context-switching. You don't have to stop reading to type a query, wait for results, then refind your place. You just ask. The multi-tab context is the killer feature here—Comet remembers what you were reading in Tab 1 when you ask a follow-up in Tab 3. For accessibility, this is obviously huge. But even for power users, voice navigation while reviewing documents or comparing sources across tabs saves cognitive load. The expanded reasoning feature also shows you exactly what actions the assistant is taking, so you're not flying blind.

The Context

Voice has been the next big interface promise since Siri debuted in 2011, but it's mostly disappointed. ChatGPT's Advanced Voice Mode changed the game—suddenly conversations felt natural, not scripted. Google's Gemini Live tried to match it but launched half-finished. Perplexity is now making a land grab for the browser-based voice assistant space, which makes sense given their search-native DNA. The real competition isn't other AI assistants—it's the existing workflow of Google Search + reading + more Google Search.

What to Watch

Accuracy is everything here. If Comet mishears your query or misinterprets what's on your screen, the friction of correcting it undoes the convenience. Early reports suggest it's good but not perfect. Also watch battery life—constant voice listening and screen analysis isn't cheap. The bigger question is whether users actually want to talk to their browser, or if this is a solution looking for a problem outside accessibility use cases. Perplexity's bet is that once you try it, you'll keep using it. Test it on a real research task before deciding.

Stay ahead with the latest news in AI

You will not get replaced by AI, but by someone using AI - Samuel Altman