“The new voice (and video) mode is the best compute interface I’ve ever used,” said OpenAI CEO Sam Altman yesterday at the company’s launch of its latest AI model, GPT-4o (“omni”). “It feels like AI from the movies; and it’s still a bit surprising to me that it’s real. Getting to human-level response times and expressiveness turns out to be a big change.”
OpenAI’s latest model, GPT-4o, seamlessly blends voice, video, and text to redefine the AI interaction experience, striving for a natural and intuitive engagement. Beyond mere text processing, this model comprehends and reacts to both audio and visual inputs, ushering in a more human-like interaction. Moreover, OpenAI’s efforts in enhancing responsiveness are evident as GPT-4o boasts reaction times in milliseconds, mirroring the pace of a genuine conversation.
The rest of this article is available by subscription only.
Introducing a New Subscription Model from the Future of Work Exchange.
To continue providing valuable insights and resources on the future of work and extended workforce management, we’re transitioning our site to a paid subscription model. While some posts will remain free, subscribing will grant you exclusive access to in-depth analysis, market research, expert interviews, and actionable strategies that will help improve your business. Solution providers and practitioners are invited to join today and gain a competitive edge by tracking the industry’s important innovations, emerging trends, and best practices.