On January 2, according to The InformationOpenAI I'm fully enhancing my voiceAICapability to introduce a voice-centred individual for the future AI DevicesPave the road. A number of sources have revealed that the equipment will beMain form of hearing interactioninstead of relying on the screen。

Currently, ChatGPT does not have the same voice function as the model used behind the text answer. OpenAI internal researchers believe that existing audio modelsThere's a clear lag in accuracy and response speedIn the past two months, the company has been urged to integrate engineering, product and research forces and to focus on the audio model panels。
This adjustment directly points to the hardware target for OpenAI - to create a paragraphConsumer-level equipment that can be operated by natural voice commandI don't know. It was reported earlier that the first productIt'll take at least a yearIt's only possible。
With the introduction of the new architecture, audio models have been generatedMore natural, more emotionalVoice response and availableIt's all the sameCapacity. OpenAI Scheme 2026 First quarterThe model is officially published。
In hardware form, OpenAI has a similar judgement to Google, Amazon, Meta and Apple: the existing mainstream equipment is not intended for future AI interactions. OpenAI team wants users to pass"Speak" instead of "see the screen"Interacting with the device, it is believed that the voice is the closest way to human instinct。
Johnny Ivy of the hardware project with OpenAI also stressed that screenless design is not only more naturalIt also helps to avoid user addictionI don't know. In his view, the new generation of equipment should correct the negative effects of past consumption of electronics and assume responsibility for that。
However, OpenAI still faces real challenges. Internal sources point out that a number of ChatGPT users are not used to voice functions, not only because the audio model is ineffective, but also because of a lack of functional awareness. OpenAI has to change user usage before launching the audio priority AI device。
At the organizational level, OpenAI has formed a dedicated team to advance the Audio AI strategy. Quindan Kumar, a voice researcher from Character.AI, is in charge of the overall direction, and Ben Newhouse is reconfiguring the base structure for audio, with the participation of Jackie Shannon, product manager of ChatGPT。
OpenAI does not intend to launch only one piece of equipment, but rather to plan a product line, includingSmart glasses and non-screen smart speakerI don't know. Within the company, it is envisaged that such equipment will exist in the form of “accompanied assistants”, proactively understanding the environment and user needs and, if authorized, providing ongoing assistance through audio and video。
In support of this long-term set-up, OpenAI has spent nearly $6.5 billion at the beginning of 2025 (note: the current exchange rate is approximately 45.506 billion yuan) on the acquisition of the io co-founded by Johnny Avei and the simultaneous advancement of the supply chain, industrial design and modelling。