At the company’s latest announcement event, OpenAI’s ChatGPT free tier received a significant update known as GPT-4o (omni). GPT-4o, as the name suggests, is an improved version of GPT 4 that includes multimodal features.
The Chief Technology Officer of OpenAI, Mira Murati, claims that this revised model is much faster and improves text, visual, and audio capabilities. Murati made these declarations on Monday during a live broadcast. All users will be able to utilize the upgraded model for free; premium users will be able to use up to five times the capacity restrictions that are available to free users.
At this time, the new GPT-4o only supports word and picture capabilities, but according to an OpenAI blog post, all of its functionality will be “rolled out iteratively.”
According to OpenAI CEO Sam Altman, GPT-4o is completely multimodal, unlike GPT 3.5, which means it can produce content and accept input in the form of text, voice, and images. Additionally, Altman said on X that developers who want to test GPT-4o will have access to an API that is half as expensive and twice as quick as GPT-4 Turbo.
After ChatGPT receives further functionality for its voice mode, it will be able to function as a voice assistant akin to the one in the 2013 film “Her-like,” which is about a conversational AI helper. It will be able to see its surroundings through your camera and react to users in real-time as a result. The voice mode is currently limited because it only accepts audio input and can only react to one command at a time.