OpenAI announced a new AI model yesterday called GPT-4o that can converse using speech in real time, read emotional cues, and respond to visual input. It will roll out over the next few weeks for free to ChatGPT users and as a service through API. Paid subscribers will have five times the rate limits of free users. The API will feature twice the speed, 50% lower cost, and five times higher rate limits compared to GPT-4 Turbo. A 26-minute long video that introduces GPT-4o and demonstrates its abilities is available in the article.
Tuesday, May 14, 2024The mysterious chat-topping AI chatbot known as 'gpt2-chatbot' that had been undergoing testing on the LMSYS Chatbot Arena has been revealed to be OpenAI's newly announced GPT-4o model. The model had topped the Chatbot Arena leaderboard, achieving the highest documented score ever. OpenAI tested multiple versions of GPT-4o on Chatbot Arena before its announcement yesterday. The models surpass all other models by a significant gap and have become the strongest ever in the Arena.
OpenAI has announced a new model called GPT-4o (o is for omni) that is a natively multimodal model, with superior performance to GPT-4 on text and state-of-the-art performance on a variety of modalities. It also announced a new desktop app, a near real-time audio interface, and a variety of improved reasoning features.
OpenAI has released GPT-4o, an AI model that can reason across audio, vision, and text in real time. Developers can access the GPT-4o API as a text and vision model. It's 2x faster, half the price, and has 5x higher rate limits compared to GPT-4 Turbo.
GPT-4o multimodal abilities, integrating vision and voice, promise significant advances in how AI interacts with the world, paving the way for AI to become a more ubiquitous presence in daily life.
GPT-4o, OpenAI's latest AI model, bridges real-time communication between humans and machines, extending capabilities beyond text to include vision and audio. The AI revolution introduces a new wave of human-to-AI and eventual AI-to-AI interactions, likely impacting the dynamics of our social behaviors and business models. As this technology progresses, its effect on human communication will unfold, potentially catalyzing the creation of innovative companies and software solutions.
AI is leading to a revolution in communication spurred by OpenAI's GPT-4o, which integrates audio, vision, and text in real time. This shift enables more natural interactions with AI, transforming human-to-AI communication into a central mode of digital interaction and potentially leading to significant societal changes and new startups focused on AI-centric communication.
OpenAI's GPT-4o and GPT-4o mini can now be fine-tuned and customized by developers for business use. Developers can now use their own datasets to enhance the model's knowledge base with proprietary information and control how the model responds to specific questions. It costs $25 for every 1 million tokens used to fine-tune GPT-4o - $3 for GPT-4o mini. 1 million tokens is roughly equivalent to 2,500 pages in a standard-size book.
OpenAI has launched fine-tuning for GPT-4o, allowing developers to customize the model for specific use cases with their own datasets. It is offering 1 million free training tokens per day through September 23.