OpenAI just launched the GPT-4o, an iteration of its GPT-4 model that is known for powering its hallmark product, ChatGPT.
The latest update "is much faster" and improves “capabilities across text, vision, and audio,” OpenAI CTO Mira Murati said in a livestream announcement on Monday, according to The Verge.
It is set to be free for all users, and paid users will continue to “have up to five times the capacity limits" of free users, Murati added.
OpenAI says in a blog post from the company that GPT-4o’s capabilities “will be rolled out iteratively (with extended red team access starting today),” but its text and image capabilities will start to release today in ChatGPT.
Moreover, OpenAI CEO Sam Altman posted that the model is “natively multimodal”. This means that the model could generate content or understand commands in voice, text, or images.
“Developers who want to tinker with GPT-4o will have access to the API, which is half the price and twice as fast as GPT-4-turbo,” Altman added on X.
The features bring speech and video to all users, either free or paid, and will be rolled out over the next few weeks. The important key point is just what a difference using voice and video to interact with ChatGPT-4o brings.
The changes, OpenAI told viewers on the live-stream, are aimed at “reducing the friction” between “humans and machines”, and “bringing AI to everyone”.
"They even made video calls with me, as if I were talking directly to them," says one victim
Intuitive Machines' IM-2 mission, featuring its lander, Athena, to achieve soft landing on March 6
Two Pakistani astronauts will undergo training at the Astronaut Center of China, says statement
DeepSeek's more economical AI model reportedly prompts Meta to ramp up research for Llama AI models
GPT-4.5 comes with improved ability to recognise patterns, generate creative insights without reasoning
Quantum computers could help scientists develop new materials such as batteries and new drugs