In anticipation of Google I/O’s upcoming presentation, OpenAI has unveiled GPT-4o, an enhanced version of the GPT-4 model that powers the company’s intelligent assistant. During a live stream, CTO Mira Murati announced that GPT-4o is significantly faster and includes improvements in text, visual, and audio capabilities. The new model will be available for free to all users, while paying users will continue to enjoy higher capacity limits compared to free users.

OpenAI CEO Sam Altman provided more details about GPT-4o on X, explaining that the model is “natively multimodal,” meaning it can generate content and understand commands in speech, text, and visual formats. Developers will have access to an API that is half the price and twice as fast as GPT-4 Turbo, opening up new opportunities for innovation.

GPT-4o’s features will be rolled out gradually, with text and visual capabilities already available on ChatGPT starting Monday. Voice mode on ChatGPT will also see significant enhancements added to allow the app to function as a real-time voice assistant. Currently, voice mode only responds to one prompt at a time, limiting its functionality.

Altman reflected on OpenAI’s evolving vision, emphasizing a shift towards making advanced AI models available to developers via paid APIs for creating innovative applications. Prior to the launch of GPT-4o, speculations about various announcements from OpenAI circulated; however, this introduction marks a new phase in AI development and technology use.

OpenAI’s unveiling of GPT-4o sets the stage for future advancements in AI technology just before Google I/O event where more announcements are expected from the Gemini team.