OpenAI Invests in Audio Technology as Silicon Valley Shifts Focus from Screens

Posted on

OpenAI is making a significant move into audio AI, and it’s not just about improving how ChatGPT sounds. Recent reports from The Information reveal that the company has consolidated several engineering, product, and research teams over the last two months to revamp its audio models in anticipation of launching an audio-first personal device within the next year.

This strategy aligns with the broader trend in the tech industry, where screens are becoming background elements and audio is taking the spotlight. Smart speakers have already embedded voice assistants into more than a third of homes in the U.S. Meta recently introduced a feature in its Ray-Ban smart glasses that utilizes five microphones to enhance conversations in noisy environments, effectively turning your face into a directional listening device. Meanwhile, Google has been testing “Audio Overviews” since June, which convert search results into conversational summaries. Tesla is also incorporating Grok and other LLMs in its vehicles, enabling conversational voice assistants to manage everything from navigation to climate control through natural dialogue.

It’s not just the big players who are investing in this space. A variety of startups are also jumping in, though with mixed results. The creators of the Humane AI Pin have spent hundreds of millions on their screenless wearable, which has turned into a cautionary tale. The Friend AI pendant— a necklace designed to record life moments and provide companionship— has raised privacy concerns and sparked existential worries. Additionally, at least two companies, including Sandbar and one led by Pebble founder Eric Migicovsky, are working on AI rings that are expected to launch in 2026, allowing users to literally “talk to the hand.”

While the formats may vary, the underlying belief is consistent: audio is the interface of the future. Every environment—your home, your car, even your face—is evolving into an interface.

OpenAI’s upcoming audio model, anticipated for early 2026, is expected to sound more natural, manage interruptions like an actual conversation partner, and even speak while you’re talking—capabilities that current models lack. The company envisions a range of devices, potentially including glasses or screenless smart speakers, that serve more as companions than mere tools.

As The Information highlights, Jony Ive, former design chief at Apple and now part of OpenAI’s hardware division following the $6.5 billion acquisition of his firm io, is prioritizing a reduction in device addiction. He sees audio-first design as an opportunity to “right the wrongs” of past consumer gadgets.