OpenAI, the pioneering research organization, has unveiled GPT-4o, the latest iteration of its renowned GPT-4 model, which powers ChatGPT.
This new release, announced by OpenAI CTO Mira Murati during a livestream event on Monday, promises significant advancements in speed and functionality across text, vision, and audio domains.
Murati highlighted the improvements, stating that GPT-4o is “much faster” and offers enhanced capabilities compared to its predecessors. Notably, the model will be available for free to all users, with paid users benefiting from up to five times the capacity limits of free users.
According to a blog post by OpenAI, GPT-4o’s capabilities will be introduced iteratively, with extended access for testing by a red team starting immediately. The initial rollout will focus on text and image capabilities within ChatGPT, with further enhancements expected in the future.
OpenAI CEO Sam Altman emphasized GPT-4o’s multimodal capabilities, noting that the model can seamlessly process voice commands, text inputs, and images. This natively multimodal approach enables GPT-4o to generate content or interpret commands across various formats, enhancing its versatility and usability.
Developers eager to explore GPT-4o’s capabilities will have access to the API, which Altman described as half the price and twice as fast as its predecessor, GPT-4-turbo.