AI

OpenAI bets big on audio as Silicon Valley declares war on screens

OpenAI is betting big on audio AI, and it’s not just about making ChatGPT sound better. According to new reporting from The Information, the company has spent the past two months uniting several engineering, product and research teams to overhaul its audio models, all in preparation for a personal device expected to hit the market in about a year.

This move reflects where the entire technology industry is heading: towards a future where screens become background noise and audio takes center stage. Smart speakers have already made voice assistants more than a permanent fixture third of American homes. Meta just introduced a feature for its Ray-Ban smart glasses that uses a five-microphone array to help you hear conversations in noisy rooms, essentially turning your face into a directional listening device. Google, meanwhile, started experimenting in June with “Audio Summaries” that turn search results into conversation summaries, and Tesla is integrating xAI’s chatbot Grok into its vehicles to create a conversational voice assistant that handles everything from navigation to climate control via natural dialogue.

It’s not just the tech giants making this bet. A colorful group of startups have emerged with the same conviction, albeit with varying degrees of success. The makers of the Humane AI Pin burned hundreds of millions before their screenless wearable became a cautionary tale. The Friend AI pendant, a necklace that claims to record your life and provide companionship, has sparked privacy concerns and existential dread in equal size. And now at least two companies, including Sandbar and one led by Pebble founder Eric Migicovsky, are building AI rings expected to debut in 2026 that will allow wearers to literally talk to the hand.

See also  Drasi by Microsoft: A New Approach to Tracking Rapid Data Changes

The form factors may differ, but the proposition is the same: audio is the interface of the future. Every space – your home, your car, even your face – becomes a control surface.

OpenAI’s new audio model, slated for early 2026, will reportedly sound more natural, handle interruptions like a real conversationalist, and even speak while you’re talking, something current models can’t handle. The company is also said to be envisioning a family of devices, possibly including glasses or screenless smart speakers, that act less as tools and more as companions.

None of this is hugely surprising. As The Information notes, Apple’s former design chief Jony Ive, who joined OpenAI’s hardware efforts through the $6.5 billion acquisition of his firm io in May, has made reducing device addiction a priority, seeing audio-first design as an opportunity to “right the wrongs” of consumer gadgets of the past.

Source link

Back to top button