OpenAI launches faster and cheaper AI models for all users

2024-05-14 08:04:33
Collection

ChainCatcher news, OpenAI has released its latest flagship model GPT-4o, which can perform real-time reasoning on audio, visual, and text inputs, focusing on the concept of a personalized, supernatural, ultra-low latency voice interaction assistant.

According to information from the official OpenAI website and the official account on the X platform, the "o" in GPT4o stands for Omni (comprehensive), marking a step towards more natural human-computer interaction—accepting any combination of text, audio, and images as input and supporting the generation of any combination of text, audio, and image outputs. It can respond to audio input in 232 milliseconds, averaging 320 milliseconds, which is similar to human response times in conversations. In terms of English and code, it performs comparably to GPT-4 Turbo, with significant improvements in non-English language text, while the API is faster and 50% cheaper. Compared to existing models, GPT-4o excels particularly in visual and audio understanding. Text and image inputs will be launched today in the API and ChatGPT, while voice and video inputs will be rolled out in the coming weeks.

Related tags
ChainCatcher reminds readers to view blockchain rationally, enhance risk awareness, and be cautious of various virtual token issuances and speculations. All content on this site is solely market information or related party opinions, and does not constitute any form of investment advice. If you find sensitive information in the content, please click "Report", and we will handle it promptly.
banner
Related tags
ChainCatcher Building the Web3 world with innovators