Introducing GPT-4o and more tools to ChatGPT free users
We are launching our newest flagship model and making more capabilities available for free in ChatGPT.
https://openai.com/index/gpt-4o-and-more-tools-to-chatgpt-free/
Commentary
Via Superhuman
“OpenAI unveils a brand-new talking AI model to power ChatGPT |
||
|
||
OpenAI kept us guessing until the very end. On Monday, the company finally showed off the next big thing in AI: Not GPT-5, but a new model called GPT-4o that’s a little different than what most experts were expecting — but no less exciting. | ||
What sets it apart? It’s faster and more efficient, but the biggest changes might come down to its ease of use. It’s reportedly more intuitive, like a jack-of-all-trades that can easily move between different mediums, including voice, text, and images — sort of like an upgraded version of ChatGPT mixed with a superpowered Siri. | ||
The demo showed GPT-4o: | ||
|
||
OpenAI says that GPT-4o is 2x faster and 50% cheaper than GPT-4 Turbo, it’s most capable model until yesterday’s announcement. The company also claims that the model outperforms the competition on several key benchmarks like the ELO ranking system by a significant margin. | ||
The biggest takeaway: GPT-4o is getting eerily good at mimicking the quirks of human communication. As one example, if you want to change the subject of the conversation, you can simply cut it off mid-sentence and move on to something else. | ||
Its ability to recognize human emotions — including the tenor of someone’s voice, their breathing patterns, and their facial expressions — marks a major step forward. | ||
Most users will be able to access GPT-4o for free within the next few weeks. OpenAI is also rolling out an app so users can use it directly on their desktops. But still no word on the supposed Google Search competitor…yet.” |
Via YouTube
Via Unwind AI
“Everything Launched at OpenAI Spring Event 
OpenAI just wrapped its Spring Event. While the hype was at its peak and the half an hour event might have not met some expectations, OpenAI did not reveal a lot of capabilities of the new model in the demos. Here’s EVERYTHING that you can do with it:
- New GPT-4O Model: OpenAI has released a new GPT-4 Omnimodel
- It brings GPT-4-level of intelligence to all but 4x faster. It is also available to free users. Paid users will be 5x higher rate limits.
- It can process text, voice and images end-to-end and can improve on its capabilities. This also means it doesn’t need any other model to process and output these modalities. For instance, if you want to generate an image in ChatGPT, it won’t use DALL.E model, GPT-4O will do it itself.
- It can chat in and translate 50 languages that cover 97% of the world’s population.
- Do check-out their demos in the blog to really understand what this model is capable of.
- GPT-4O API: GPT-4O is not just limited to ChatGPT, it’s also coming to the OpenAI API. It is 2x faster, 50% cheaper than GPT-4 Turbo and comes with 5x higher rate limits.
- ChatGPT Desktop App: The new desktop app makes it easier to take ChatGPT’s assistance while working on something else. With a simple keyboard shortcut (Option + Space), you can instantly ask ChatGPT a question. It will be available soon.
- New Voice Assistant: OpenAI has released a new Voice assistant that is much capable and faster than the existing Voice Mode in ChatGPT. The new Voice Assistant is powered by GPT-4O, which can natively handle text, audio and voice.
- This is different from the current Voice Mode that uses three different models – a transcription model, GPT 3.5 or 4 for intelligence, and a text-to-speech model. Orchestrating these three models together brings a lot of latency.
- Since GPT-4O powers the Voice Assistant single-handedly, it is near real-time without awkward pauses.
- It can be interrupted to change the topic or question easily. It is capable of talking in a wide range of emotions and styles. Just tell it if you want to it to talk with more emotions, it will!
- Video chat with Voice Assistant – It can see while giving voice assistance. You can turn on your camera and show it what you’re doing, and it will help you step-by-step to complete your work. This helps with so many usecases, be it solving math, or guiding in coding, helping in cooking, and more.
- Since GPT-4O is available to free ChatGPT users also, they can now:
- Upload images and files and ask questions about them. Analyze data and charts through advanced data analysis
- Browse the web directly through ChatGPT and get the latest information
- Use GPTs and discover them on the GPT Store
- Use the “Memory” feature where ChatGPT remembers about your preferences that you tell it in chats.”
Via The Neuron
“OpenAI unleashes ChatGPT-4o to everyone. |
|||
|
|||
BREAKING: OpenAI announced, ChatGPT-4o (“o” for “omni”), the fastest, smartest, and most multimodal AI yet—watch the full demo here: | |||
ChatGPT-4o will be available for free for everyone as a desktop app (soon). Yup, everyone will be able to use GPT-4o and GPTs as a result. This means you might want to rethink your Claude Pro or Gemini Ultra subscriptions. | |||
ChatGPT+ users will get first dibs on GPT-4o with 5x more usage. | |||
It’s more multimodal: This new ChatGPT isn’t just higher in IQ (it ranks #1 by far on the LMSYS leaderboard); it can talk and see just like us. | |||
First, talking. We’ve had computers that can talk for a minute, but it never felt like a genuine convo—more a cycle of speak, pause, respond, pause, repeat. | |||
Voice Mode feels like chatting with a real human—it captures your tone, language, and expressions in real-time. Many are describing it as a real-life Her (the voice in all the demos might actually be Scarlett Johansson). | |||
Explore what it can do here: | |||
It’s uncannily human-like, perhaps too much so. But it means that for tasks you’d normally attempt with Siri, you should use ChatGPT instead. And with the new desktop app, Voice Mode will be great for scenarios better explained verbally than typed. Just remember to use inside voices in public spaces! | |||
Here’s a quick guide on adding ChatGPT as a widget to your home screen, courtesy of Google SGE: | |||
|
|||
Second, ChatGPT-4o has live 20/20 vision, meaning it can interpret photos, screenshots, and docs while you work. For example: | |||
|
|||
|
|||
Why it matters: Together, all these new features will unlock new use cases, and we’re pumped about them converging into a super helpful work assistant that can view your screen as you work. | |||
Consider these possibilities: | |||
|
|||
Other updates not in the demo (see here): | |||
|
|||
Catch Pete’s extended analysis on OpenAI’s demos and why the new desktop app is a game-changer (Apple Podcasts, Spotify, YouTube).” |
Via AI Tool Report
” GPT-4 gets an unmissable upgrade |
|
|
|
Our Report: Following mounting speculation of a ChatGPT-powered search engine, a voice assistant, and the release of GPT-5, OpenAI has finally announced the launch of GPT-4o which will bring new and improved features to ChatGPT. | |
![]() |
|
|
|
![]() |
Via The Rundown AI
OPENAI |
||
|
||
|
||
The Rundown: OpenAI just unveiled GPT-4o, a new advanced multimodal model that integrates text, vision and audio processing, setting new benchmarks for performance – alongside a slew of new features. | ||
The new model: | ||
|
||
Voice and other upgrades: | ||
|
||
Free for everyone: | ||
|
||
Why it matters: Real-time voice and multimodal capabilities are shifting AI from a tool, to an intelligence we collaborate, learn, and grow with. Additionally, a whole new group of free users (who might’ve been stuck with a lackluster GPT 3.5) are about to get the biggest upgrade of their lives in the form of GPT-4o. | ||
If you missed it, you can rewatch OpenAI’s full demo here.” |
0 Responses
Stay in touch with the conversation, subscribe to the RSS feed for comments on this post.