OpenAI introduced GPT-4o, a more advanced AI model improving text, audio, and image processing for enhanced voice interactions. The upgrade includes real-time conversation translation and emotion recognition, with broader access options for users.
OpenAI announced upgrades to ChatGPT, introducing a more advanced AI model, GPT-4o, which enhances voice interaction capabilities. The announcement was made during a live stream from their San Francisco headquarters, strategically scheduled a day before Google’s developer conference. GPT-4o can process and respond to text, audio, and image inputs, significantly improving its conversational abilities and emotional range in voice interactions.
The new features were demonstrated using a voice resembling actress Scarlett Johansson, showcasing the AI’s ability to translate conversations in real time and identify emotions. OpenAI also introduced broader access to these enhanced features, offering both free limited access and expanded capabilities for subscribers.
In addition to individual user enhancements, OpenAI revealed that GPT-4o would be integrated into professional applications like the mobile app “Be My Eyes,” which aids visually impaired users. This follows their trend of rapid feature releases, often timed to coincide with or precede announcements by competitors like Google.
Meanwhile, OpenAI’s rival, Anthropic, has expanded availability of its AI assistant, Claude, to the European Union. Claude is known for its reliability and ethical design approach, dubbed “constitutional AI,” aimed at minimizing harm and bias. The app is available in both free and premium versions, with Anthropic having recently secured significant investment from Amazon, further cementing AWS as its primary cloud provider.