Skip to content


Introducing GPT-4o and more tools to ChatGPT free users

Introducing GPT-4o and more tools to ChatGPT free users

We are launching our newest flagship model and making more capabilities available for free in ChatGPT.

https://openai.com/index/gpt-4o-and-more-tools-to-chatgpt-free/

Commentary

Via Superhuman

“OpenAI unveils a brand-new talking AI model to power ChatGPT

OpenAI kept us guessing until the very end. On Monday, the company finally showed off the next big thing in AI: Not GPT-5, but a new model called GPT-4o that’s a little different than what most experts were expecting — but no less exciting.
What sets it apart? It’s faster and more efficient, but the biggest changes might come down to its ease of use. It’s reportedly more intuitive, like a jack-of-all-trades that can easily move between different mediums, including voice, text, and images — sort of like an upgraded version of ChatGPT mixed with a superpowered Siri.
The demo showed GPT-4o:
  • Helping an Italian speaker and an English speaker communicate by acting as their real-time translator
  • Walking a presenter through a math equation he’d written on a piece of paper
  • Analyzing a webpage full of code, then describing weather patterns from a chart of average temperatures
  • Communicating in different styles, from hyper-expressive to sing-songy, to robotic
OpenAI says that GPT-4o is 2x faster and 50% cheaper than GPT-4 Turbo, it’s most capable model until yesterday’s announcement. The company also claims that the model outperforms the competition on several key benchmarks like the ELO ranking system by a significant margin.
The biggest takeaway: GPT-4o is getting eerily good at mimicking the quirks of human communication. As one example, if you want to change the subject of the conversation, you can simply cut it off mid-sentence and move on to something else.
Its ability to recognize human emotions — including the tenor of someone’s voice, their breathing patterns, and their facial expressions — marks a major step forward.
Most users will be able to access GPT-4o for free within the next few weeks. OpenAI is also rolling out an app so users can use it directly on their desktops. But still no word on the supposed Google Search competitor…yet.”

Via YouTube

Via Unwind AI

“Everything Launched at OpenAI Spring Event 🌟

OpenAI just wrapped its Spring Event. While the hype was at its peak and the half an hour event might have not met some expectations, OpenAI did not reveal a lot of capabilities of the new model in the demos. Here’s EVERYTHING that you can do with it:

  1. New GPT-4O Model: OpenAI has released a new GPT-4 Omnimodel
    1. It brings GPT-4-level of intelligence to all but 4x faster. It is also available to free users. Paid users will be 5x higher rate limits.
    2. It can process text, voice and images end-to-end and can improve on its capabilities. This also means it doesn’t need any other model to process and output these modalities. For instance, if you want to generate an image in ChatGPT, it won’t use DALL.E model, GPT-4O will do it itself.
    3. It can chat in and translate 50 languages that cover 97% of the world’s population.
    4. Do check-out their demos in the blog to really understand what this model is capable of.
  1. GPT-4O API: GPT-4O is not just limited to ChatGPT, it’s also coming to the OpenAI API. It is 2x faster, 50% cheaper than GPT-4 Turbo and comes with 5x higher rate limits.
  1. ChatGPT Desktop App: The new desktop app makes it easier to take ChatGPT’s assistance while working on something else. With a simple keyboard shortcut (Option + Space), you can instantly ask ChatGPT a question. It will be available soon.
  2. New Voice Assistant: OpenAI has released a new Voice assistant that is much capable and faster than the existing Voice Mode in ChatGPT. The new Voice Assistant is powered by GPT-4O, which can natively handle text, audio and voice.
    1. This is different from the current Voice Mode that uses three different models – a transcription model, GPT 3.5 or 4 for intelligence, and a text-to-speech model. Orchestrating these three models together brings a lot of latency.
    2. Since GPT-4O powers the Voice Assistant single-handedly, it is near real-time without awkward pauses.
    3. It can be interrupted to change the topic or question easily. It is capable of talking in a wide range of emotions and styles. Just tell it if you want to it to talk with more emotions, it will!
    4. Video chat with Voice Assistant – It can see while giving voice assistance. You can turn on your camera and show it what you’re doing, and it will help you step-by-step to complete your work. This helps with so many usecases, be it solving math, or guiding in coding, helping in cooking, and more.
  1. Since GPT-4O is available to free ChatGPT users also, they can now:
    1. Upload images and files and ask questions about them. Analyze data and charts through advanced data analysis
    2. Browse the web directly through ChatGPT and get the latest information
    3. Use GPTs and discover them on the GPT Store
    4. Use the “Memory” feature where ChatGPT remembers about your preferences that you tell it in chats.”

Via The Neuron

“OpenAI unleashes ChatGPT-4o to everyone.

YouTube video by OpenAI

Introducing GPT-4o
BREAKING: OpenAI announced, ChatGPT-4o (“o” for “omni”), the fastest, smartest, and most multimodal AI yet—watch the full demo here:
ChatGPT-4o will be available for free for everyone as a desktop app (soon). Yup, everyone will be able to use GPT-4o and GPTs as a result. This means you might want to rethink your Claude Pro or Gemini Ultra subscriptions.
ChatGPT+ users will get first dibs on GPT-4o with 5x more usage.
It’s more multimodal: This new ChatGPT isn’t just higher in IQ (it ranks #1 by far on the LMSYS leaderboard); it can talk and see just like us.
First, talking. We’ve had computers that can talk for a minute, but it never felt like a genuine convo—more a cycle of speak, pause, respond, pause, repeat.
Voice Mode feels like chatting with a real human—it captures your tone, language, and expressions in real-time. Many are describing it as a real-life Her (the voice in all the demos might actually be Scarlett Johansson).
Explore what it can do here:
  • live language translation (link).
  • realtime conversational speech (link).
  • lullabies and whispers (link).
  • sarcasm (link).
  • even singing (link)!
It’s uncannily human-like, perhaps too much so. But it means that for tasks you’d normally attempt with Siri, you should use ChatGPT instead. And with the new desktop app, Voice Mode will be great for scenarios better explained verbally than typed. Just remember to use inside voices in public spaces!
Here’s a quick guide on adding ChatGPT as a widget to your home screen, courtesy of Google SGE:
Second, ChatGPT-4o has live 20/20 vision, meaning it can interpret photos, screenshots, and docs while you work. For example:
  • Sal Khan’s son shares his iPad screen, and ChatGPT-4o helps him solve a problem live (link).
  • It can identify objects and teach you how to say them in Spanish (link).
  • It can explain copy & pasted code (link).
  • TOP DEMO: Be My Eyes + GPT-4o helping a blind person “see” what’s in front of him, even flagging down an available taxi (link).

YouTube video by OpenAI

Be My Eyes Accessibility with GPT-4o
Why it matters: Together, all these new features will unlock new use cases, and we’re pumped about them converging into a super helpful work assistant that can view your screen as you work.
Consider these possibilities:
  • Upload a PowerPoint and let ChatGPT-4o suggest layout tweaks, rephrase slide titles, and improve the design.
  • Use ChatGPT-4o to inspect a spreadsheet and highlight trends, anomalies, or discrepancies. Or for tech support.
  • GPT-4o can guide customers through visual step-by-step instructions for installing or setting up products.
Other updates not in the demo (see here):
  1. For developers, GPT-4o is half the price, twice as fast as GPT-4-turbo, and has 5x rate limits.
  2. Way better at writing text correctly in DALL-E 3 images.
  3. It can create fonts.
  4. It can generate 3D visualizations.
Catch Pete’s extended analysis on OpenAI’s demos and why the new desktop app is a game-changer (Apple PodcastsSpotifyYouTube).”

Via AI Tool Report

” GPT-4 gets an unmissable upgrade

GPT-4o launched: Unmissable upgrade
Our Report: Following mounting speculation of a ChatGPT-powered search engine, a voice assistant, and the release of GPT-5, OpenAI has finally announced the launch of GPT-4o which will bring new and improved features to ChatGPT.
🔑 Key Points:
  • GPT-4o is reportedly 2x faster, 50% cheaper, and will provide ChatGPT users with “GPT-4-level” intelligence but with improved text, vision, and audio capabilities.
  • It will also power a new voice assistant feature that can read facial expressions, translate in real-time, be interrupted, and respond to multiple prompts, turning ChatGPT into an assistant.
  • GPT-4o is available now, but OpenAI is launching the new voice feature to “a small group of trusted ChatGPT Plus subscribers” first before a wider rollout begins.
🤔 Why you should care: OpenAI is giving developers access to GPT-4o’s API so they can use it to build their own AI tools, which suggests a company refocus from its original vision, to “create all sorts of benefits for the world” to “we’ll create AI, then other people will use it to create all sorts of things that we all benefit from.””

Via The Rundown AI

OPENAI

🗣️OpenAI unveils GPT-4o and new voice capabilities

Image source: OpenAI
The Rundown: OpenAI just unveiled GPT-4o, a new advanced multimodal model that integrates text, vision and audio processing, setting new benchmarks for performance – alongside a slew of new features.
The new model:
  • GPT-4o provides improved performance across text, vision, audio, coding, and non-English generations, smashing GPT-4T’s performance.
  • The new model is 50% cheaper to use, has 5x higher rate limits than GPT-4T, and boasts 2x the generation speed of previous models.
  • The new model was also revealed to be the mysterious ‘im-also-a-good-gpt2-chatbot’ found in the Lmsys Arena last week.
Voice and other upgrades:
  • New voice capabilities include real-time responses, detecting and responding with emotion, and combining voice with text and vision.
  • The demo showcased feats like real-time translation, two AI models analyzing a live video, and using voice and vision for tutoring and coding assistance.
  • OpenAI’s blog also detailed advances like 3D generation, font creation, huge improvements to text generation within images, sound effect synthesis, and more.
  • OpenAI also announced a new ChatGPT desktop app for macOS with a refreshed UI, integrating directly into computer workflows.
Free for everyone:
  • GPT-4o, GPTs, and features like memory and data analysis are now available to all users, bringing advanced capabilities to the free tier for the first time.
  • The GPT-4o model is currently rolling out to all users in ChatGPT and via the API, with the new voice capabilities expected to arrive over the coming weeks.
Why it matters: Real-time voice and multimodal capabilities are shifting AI from a tool, to an intelligence we collaborate, learn, and grow with. Additionally, a whole new group of free users (who might’ve been stuck with a lackluster GPT 3.5) are about to get the biggest upgrade of their lives in the form of GPT-4o.
If you missed it, you can rewatch OpenAI’s full demo here.”

 

 

 

 

 

0 Shares

Posted on: May 14, 2024, 9:51 am Category: Uncategorized

What Does the TikTok Ban Mean for Your Library Marketing? Now That the Dust Has Settled, Let’s Unpack the Potential Impact

What Does the TikTok Ban Mean for Your Library Marketing? Now That the Dust Has Settled, Let’s Unpack the Potential Impact

0 Shares

Posted on: May 14, 2024, 6:23 am Category: Uncategorized

How the brain chooses which memories are important enough to save and which to let fade away

How the brain chooses which memories are important enough to save and which to let fade away

https://www.today.com/health/sleep/brain-chooses-some-memories-sleep-rcna145642?utm_source=feedingcuriosity.com

0 Shares

Posted on: May 14, 2024, 6:22 am Category: Uncategorized

What Is Personal Branding? Here’s Why It’s So Important

What Is Personal Branding? Here’s Why It’s So Important

Discover the power of personal branding and learn how to establish yourself as an expert in your field. Stand out in a crowded marketplace with a strong online brand identity.

https://www.searchenginejournal.com/what-is-personal-branding-heres-why-its-so-important/512117/

0 Shares

Posted on: May 14, 2024, 6:04 am Category: Uncategorized

Three Ways to Double Your Bookshelf Capacity

This might be useful for smaller libraries (or home collections).

Three Ways to Double Your Bookshelf Capacity

You’ve got more books than bookshelf space—and more coming all the time.
0 Shares

Posted on: May 14, 2024, 6:02 am Category: Uncategorized

BREAKING NEWS: Students are still bored

BREAKING NEWS: Students are still bored

BREAKING NEWS: Students are still bored

“Breaking news! A nationally-representative poll of more than 1,000 teenagers finds that… students are still bored. Here are some key findings:

  • 64% of teenagers think that school is boring
  • 70% of teenagers say that all or most of their classmates are bored in class
  • Only 41% of teenagers like going to school
  • Only 40% of teenagers think that their homework helps them learn
  • 30% of teenagers say that school is a waste of time
  • Only 19% of teenagers say that most of their classmates want to be in school”

 

 

 

 

0 Shares

Posted on: May 13, 2024, 6:59 am Category: Uncategorized

Talk with Me Toolkits

Talk with Me Toolkits

0 Shares

Posted on: May 13, 2024, 6:58 am Category: Uncategorized