OpenAI launches faster and cheaper AI models for all users
ChainCatcher news, OpenAI has released its latest flagship model GPT-4o, which can perform real-time reasoning on audio, visual, and text inputs, focusing on the concept of a personalized, supernatural, ultra-low latency voice interaction assistant.According to information from the official OpenAI website and the official account on the X platform, the "o" in GPT4o stands for Omni (comprehensive), marking a step towards more natural human-computer interaction—accepting any combination of text, audio, and images as input and supporting the generation of any combination of text, audio, and image outputs. It can respond to audio input in 232 milliseconds, averaging 320 milliseconds, which is similar to human response times in conversations. In terms of English and code, it performs comparably to GPT-4 Turbo, with significant improvements in non-English language text, while the API is faster and 50% cheaper. Compared to existing models, GPT-4o excels particularly in visual and audio understanding. Text and image inputs will be launched today in the API and ChatGPT, while voice and video inputs will be rolled out in the coming weeks.