OpenAI unveils GPT-4o, a new flagship generative AI model that is faster and natively multimodal, rolling out for free to all ChatGPT users in the coming weeks
There are two things from our announcement today I wanted to highlight. OpenAI : Hello GPT-4o — We're announcing GPT-4o, our new flagship model that can reason across audio, vision, and text in real time. Clare Duffy / CNN : OpenAI unveils newest AI model, GPT-4o Zak Killian / HotHardware : OpenAI's Groundbreaking GPT-4o Teaches ChatGPT Intimate Real Time Conversation The Information : OpenAI's Impressive Engineering Feat with GPT-4o AKA ‘Her’ Nishant N / MarkTechPost : OpenAI Launches ChatGPT Desktop App: Enhancing Productivity for Mac Users OpenAI / OpenAI on YouTube : Math problems with GPT-4o Casey Newton / Platformer : ChatGPT gets an emotional upgrade Ryan Morrison / Tom's Guide : I gave 5 prompts to ChatGPT-4o vs GPT-4 to test the new AI model — here's what happened Kaye Lushington / International Business Times : Cool or cringe? GPT-4o gets remarkable capabilities; revolutionises human-AI interactions [Watch] Jon Martindale / Digital Trends : GPT-4o: What the latest ChatGPT update can do and when you can get it Luke Jones / WinBuzzer : New OpenAI GPT-4o AI Model Now Available on Microsoft Azure OpenAI Service Ben Lovejoy / 9to5Mac : Incredible iPad AI tutor demo points to an incredible new world for students Jacob Ridley / PC Gamer : OpenAI unveils GPT-4o with a video of a man making small talk with his phone and I cannot pretend it's not really weird Cade Metz / New York Times : OpenAI Unveils New ChatGPT That Listens, Looks and Talks Siddharth Jindal / Analytics India Magazine : Top 10 Must Watch OpenAI GPT-4o Demos Mike Young / AIModels.fyi : Losers (and winners) from the GPT-4o announcement Daniel Croft / Cyber Daily : OpenAI's new flagship is faster, more accurate and free Emma Roth / The Verge : ChatGPT is getting a Mac app Aamir Sheikh / Cryptopolitan : OpenAI's ChatGPT-4o Can Show Feeling and Emotions OpenAI / OpenAI on YouTube : Two GPT-4os interacting and singing Matt Swider / The Shortcut : ChatGPT-4o free AI ‘comes alive’ with real-time conversational speech and emotion Mastodon : Mike Elgan (@MikeElgan@mastodon.social) Wes Davis / The Verge : OpenAI opens its GPT Store to all users for free, but the ability to create custom GPTs is still only for paid users; OpenAI debuted GPT Store on January 10 Threads: @xzed87 : $META Now that we know what Chat-GPT-4o is capable of, Meta Ray-Bans will be much more useful in the future. Meta will own their AI, Apple will not, if the Chat-GPT talks are true. At the end of the day, most people wont likely care whether they use Open vs Closed. … Paul Rietschka / @paul_rietschka : One of the problems I have with current LLM mania is the way we're asking too much of transformers. These Rube Goldberg, multi-stage reasoning* monstrosities? Ugh, no. Translation? Why yes. And why? Because transformers were a revolution in NLP and, specifically, algorithmic translation. … Sung Kim / @sung.kim.mw : The most disappointing thing about OpenAI's demo is that they seem to have spent a lot of their efforts in speeding up the inference of their models, instead of pushing the boundaries of LLM or MLLM. I was really hoping for one more thing at the end. Christopher Mims / @technicallymims : I could be totally wrong about this but I am starting to think that, as someone smarter than me put it, advances in the capabilities of today's LLMs are starting to be more of a sigmoid function than an ongoing takeoff to better and better “reasoning” like … Dare Obasanjo / @carnage4life : It's wild living with the certainty that in my lifetime there is no intellectual task a human can do that can't be outperformed by a machine. It reminds me of social media in that the technology will reshape everything and will not be entirely positive despite best intentions. Kylie Robison / @kylie.robison : I asked miramurati about Scarlett Johansson-type voice in today's demo of GPT-4o. She clarified it's not designed to mimic her, and said someone in the audience asked this exact same question! https://www.theverge.com/... Mastodon: @simonbs@mastodon.social : Just updated my app for correcting texts using AI to use the new GPT-4o. I'm excited to see if it makes a difference in real world use! — [image] Bluesky: Hens Breet / @monospace.com : In other words, a far cry from the breathless speculation over the past couple of days. [embedded post] X: Sam Altman / @sama : her @drjimfan : I know your timeline is flooded now with word salads of “insane, HER, 10 features you missed, we're so back”. Sit down. Chill. <gasp> Take a deep breath like Mark does in the demo </gasp>. Let's think step by step: - Technique-wise, OpenAI has figured out a way to map audio to... [video] Sam Altman / @sama : hope you enjoyed! the new voice mode will be live in the coming weeks for plus users. we'll have more stuff to share soon :) Soumith Chintala / @soumithchintala : really exciting demos from @OpenAI establishing new expectations for AI. Lot of work to do on the Llama train — which isn't going to stop until we catch up! My personal feeling with gpt-4o — it feels attainable - contrasting to gpt-4 when released, it felt magically impossible. Dhravya Shah / @dhravyashah : Openai just killed all these YC startups RIP [image] Robert Scoble / @scobleizer : I already gave 1,000 hours of my voice to Elon. :-) Toby Ord / @tobyordoxford : I'm a cheerful person by disposition, but the level of cheerfulness in OpenAI's new GPT-4o is off the charts. It feels like something from Douglas Adams'sSirius Cybernetics Corp, breathing a sigh of contentment every time it makes your day. Lauren Goode / @laurengoode : “Google DeepMind last month released a paper exploring the ethical risks that may arise as AI assistants become more capable...the researchers warn that such assistants could become highly persuasive and addictive.” OpenAI: Here's a chatbot that flirts https://www.wired.com/... @sullyomarr : Man idk what OAI cooked with gpt4o but ain't no way it's only 50% cheaper for them It's: - free (srsly they've been capacity constrained forever) - 4x faster that gpt4 turbo - better at coding - can reason across 3 modalities - realtime They're definitely making killing on Simon Willison / @simonw : OK, I have GLT-4o in my ChatGPT web interface now! The biggest improvement for me looks to be Code Interpreter, having it type out the code faster makes that tool much less frustrating to wait for [image] James Vincent / @jjvincent : OpenAI emphasising voice interaction for gpt-4o is canny and somewhat telling - they're leaning into the masquerade of intelligence, voice is emotive and seductive (and they do it incredibly well) but it doesn't necessarily indicate leaps forward in capability Saman Farid / @samanfarid : It does seem like most of the releases today were engineering “bells and whistles” added on top — not a lot of new fundamental capability breakthrough. - faster - cleaner UI - multi modal - cute voice synthesis Still very far from AGI - and not improving the trajectory. Ben / @olney1ben : OpenAI desktop app for mac will require macOS 14+ with apple silicon (m1, m2, m3). OpenAI just helped Apple with an upgrade push. #ChatGPT #GPT40 [image] @swyx : Vibecheck of GPT4T vs GPT4O from AInews: + summarization quality + more concise, specific, accurate + (unevenly) better topic and highlighting choices - instruction following on links - insturction following on outputting markdown GPT4O (11.5mins) was 74% faster than GPT4T [image] Aaron Levie / @levie : If you have a multimodal AI that can do video and audio in and code out, then you essentially have democratized robotics AI. Truly crazy. @shrihacker : its so over for girls [video] @deliprao : Folks are surprised that they are giving away gpt-4o for free. They are collecting your speech prompt data silly. Remember if you are not on the team plan, unless you go through a complicated opt-out process, all your data is by default used by openai. Not surprising anymore. Gary Marcus / @garymarcus : GPT-4o hot take: • The speech synthesis is terrific, reminds me of Google Duplex (which never took off). but • If OpenAI had GPT-5, they have would shown it. • They don't have GPT-5 after 14 months of trying. • The most important figure in the blogpost is attached below. And... [image] Roon / @tszzl : towards intelligence too cheap to meter @suhail : If there's not a paper evaluating 4o by tomorrow on arxiv, we've decelerated. Patrick Moorhead / @patrickmoorhead : The ChatGPT-4o voice interaction is incredible. So far. I just spent the past half hour asking “it” verbally about the background of the people I will be meeting with, when I should leave the hotel tomorrow, the best way to get there. It nailed every single thing I fed it. Oh... Rohit / @krishnanrohit : a big takeaway for me from today's event is that GPT-5 will be a single, giant, multimodal model. today had clear “this is 3.5, look its cool” vibes, when 4's in the offing. Soumith Chintala / @soumithchintala : @swyx @OpenAI When GPT-4 was released, I think we very vastly unequipped, that was probably the biggest reason. the (Meta) GenAI organization (that is a very focused effort for Generative models, instead of a general AI research org like FAIR) started right around when GPT-4 was released. So, @swyx : @soumithchintala @OpenAI what about gpt4 on release felt impossible? i mean you ended up doing it ~1 year later so you figured it out somehow. just wanted to learn from benefit of hindsight Tanay Jaipuria / @tanayj : Duolingo stock fell 3.5%, wiping out ~$250M in market value, within minutes of OpenAI demoing the real-time translation capabilities of GPT-4o [image] Gary Marcus / @garymarcus : GPT-4o hot take: • The speech synthesis is terrific, reminds me of Google Duplex (which never took off). but • If OpenAI had GPT-5, they have would shown it. • They don't have GPT-5 after 14 months of trying. • The most important figure in the blogpost is attached below. And [image] @benjamindekr : Alright I'm gonna say it... This is essentially AGI. This will be seen as magic to masses. What else do you call it when a virtual “person” can listen, talk, see, and reason almost indistinguishably from an average human? Isn't that AGI? Drew Breunig / @dbreunig : I'd wager making a giggling AI assistant is a mistake. Going after the users you already have and more of a novelty than a utility. Robert Lukoszko / @karmedge : I am 80% sure openAI has extremely low latency low quality model get to pronounce first 4 words in <200ms and then continue with the gpt4o model Just notice, most of the sentences start with “Sure” “Of course” “Sounds amazing” “Let's do it” “Hmm” And then it continues with + gpt4 real answer Grady Booch / @grady_booch : By the very nature of its architecture, no LLM can reason. One might see dim reflections of inductive and deductive reasoning only in so far as the training data set contains such examples, but abductive reasoning - indeed any color of reasoning that requires the production of Benj Edwards / @benjedwards : With the release of GPT-4o and its apparent “artificial emotional intelligence,” you might call it, this seems like a good day to resurface this 2016 tweet by @sama Benjamin Riley / @benjaminjriley : ChatGPT-4o is here and omg...it still can't handle a simple reasoning task that most adult humans can figure out. But it did produce this very wrong answer much faster than it usually takes. (Ongoing shout out to @colin_fraser for identifying this particular task.) [image] Alex Volkov / @altryne : Prompt to diarize is... something else. Check out this thread from VB dissecting the amazing voice capabilities of GPT4 Omni @reach_vb : Okay, GPT4 Omni is pretty rad! 🔥 From an audio-understanding standpoint, it can: 1. Transcribe audio better than Whisper large v3 2. It can diarise audio (meeting notes) 3. Can translate audio from one language to another 4. Summarise audio All of this zero/ few shot. From an [video] Palash Volvoikar / @whipalash : So is this the new Siri then? #GPT4o @sullyomarr : ty sama for saving my startup 50% on llm costs [image] Ben Tossell / @bentossell : searching real-time with my new boo 4o [image] Simon Willison / @simonw : My notes on this morning's OpenAI release of the new GPT-4o model - not a huge leap in “intelligence” (whatever that might mean) but still very significant thanks mainly to the impressive new audio capabilities and the drop in price https://simonwillison.net/... Alexander Doria / @dorialexander : Since it's that time of year for childish AI game, GPT-4o passing the mirror test and depicting itself. [image] Joanna Stern / @joannastern : Also, Plus users get: - Early access to features, including the desktop app for macOS - To create GPTs - Access to DALL-E Joanna Stern / @joannastern : There's still some benefit to paying for Plus. You get 5x higher message limits and access to the Voice Mode with GPT-4o M.G. Siegler / @mgsiegler : Me and 4o are really hitting it off... [image] Aashay Sanghvi / @aashaysanghvi_ : It's fascinating that a bunch of startups launched 4o into prod within an hour of the announcement. Timothy B. Lee / @binarybits : GPT-4o: still bad at clocks. [image] Sophie / @netcapgirl : [image] @8teapi : Gee I wonder where the next big target market is [image] @sullyomarr : GPT-4o is way, way faster than gpt4 Feels like an entirely different model. Insanely fast [video] Alex Volkov / @altryne : OpenAI finally acknowledges the “foreign language tax” by introducing a new tokenizer with significantly improved tokens for other languages! Jong Wook Kim / @_jongwook_kim : 1.7x fewer tokens in Korean, which means GPT-4o feels 3.4x faster to Korean users! [image] Aidan Clark / @_aidan_clark_ : 2x cheaper & faster is for English, but for other languages (especially non-latin-script) expect - thanks to our new tokenizer — even up to 9x cheaper/faster! Andrew Ng / @andrewyng : Congrats to OpenAI for the release of GPT-4o! 2x faster and 50% cheaper tokens will be great for everyone using agentic AI workflows. When an agentic job that used to take 10min now takes 5min just by switching APIs, that's great progress! Boris Power / @borismpower : A lot more demos of GPT-4o at https://www.youtube.com/... ! Marshall Kirkpatrick / @marshallk : High quality inference at twice the speed and half the cost is a big, big deal for organizations building systems that utilize OpenAI technology. Amazing. Caleb Sirak / @calebsirak : gpt-4oh shit some yc startups just went bye bye @manan : waiting for Bloomberg to ask the OpenAI team if GPT-4o is trained based on YouTube François Chollet / @fchollet : The downside of hyping your future V5 so much is that you have to release all of your new models under the V4 brand in order to avoid disappointment — in perpetuity Shubham Saboo / @saboo_shubham_ : GPT-4o is not just limited to ChatGPT, it's also coming to the OpenAI API. GPT-4o is 4x faster and 2x cheaper than GPT-4 Turbo and comes with 5x higher rate limits. New wave of GPT-4o wrappers incoming. [image] Aaron Levie / @levie : OpenAI just made their new GPT4 model 50% cheaper and 2X faster for developers. This is an insane level of improvement for anyone building in AI right now. LinkedIn: Dr. Nicole Wieberneit : Reasoning over vision, text and audio in real-time! How cool is this? — OpenAI just released a new model - gpt-4o. … Tom Hewitson : This might be hyperbolic but I suspect GPT-4o has probably just wiped out half of the conversational AI industry and much of the conversation design discipline. … Ben Saunders : So Monday arrived... 🤯 — If you haven't yet watched the latest OpenAI announcement. Then do it now. — Earlier tonight, they announced GPT-4o. … Jenny O'Leary : Say hello to GPT-4o, our new flagship model which can reason across audio, vision, and text in real time: https://openai.com/... … Lin Yang : It is live via OpenAI API, 5x the rate limit, 2x the speed, half the cost! Online and batch both! Build with us! Shyamal Hitesh Anadkat : GPT-4o: 2x faster, half the price, 5x higher rate limits compared to GPT-4 Turbo. Can't wait to see what you'll build! — https://lnkd.in/... Peter Deng : Incredible work by the entire team bringing GPT-4o to life! Excited to see what you all think of it. … Krithika M. : Hello #GPT4o! — It's hard to explain how much more natural it feels to have a conversation with this model. Check out some of the video demos in the post to see it in action. … Forums: Hacker News : GPT-4o r/technews : Open AI new model-GPT-4o r/technology : Hello GPT-4o r/mlscaling : OpenAI announces GPT-4o (gpt2-chatbot): much higher Elo on hard code/math, low-latency audio/voice, image gen/edit, halved cost (esp foreign language) r/singularity : Sam Altman: Some thoughts on GPT-4o r/DougDoug : OpenAI has just released GPT-4o - A REALTIME voice/video version of GPT4.
$META Now that we know what Chat-GPT-4o is capable of, Meta Ray-Bans will be much more useful in the future. Meta will own their AI, Apple will not, if the Chat-GPT talks are true. At the end of the day, most people wont likely care whether they use Open vs Closed. …
One of the problems I have with current LLM mania is the way we're asking too much of transformers. These Rube Goldberg, multi-stage reasoning* monstrosities? Ugh, no. Translation? Why yes. And why? Because transformers were a revolution in NLP and, specifically, algorithmic…
The most disappointing thing about OpenAI's demo is that they seem to have spent a lot of their efforts in speeding up the inference of their models, instead of pushing the boundaries of LLM or MLLM. I was really hoping for one more thing at the end.
I could be totally wrong about this but I am starting to think that, as someone smarter than me put it, advances in the capabilities of today's LLMs are starting to be more of a sigmoid function than an ongoing takeoff to better and better “reasoning” like …
It's wild living with the certainty that in my lifetime there is no intellectual task a human can do that can't be outperformed by a machine. It reminds me of social media in that the technology will reshape everything and will not be entirely positive despite best intentions.
I asked miramurati about Scarlett Johansson-type voice in today's demo of GPT-4o. She clarified it's not designed to mimic her, and said someone in the audience asked this exact same question! https://www.theverge.com/...
I know your timeline is flooded now with word salads of “insane, HER, 10 features you missed, we're so back”. Sit down. Chill. <gasp> Take a deep breath like Mark does in the demo </gasp>. Let's think step by step: - Technique-wise, OpenAI has figured out a way to map audio to...…
I'm a cheerful person by disposition, but the level of cheerfulness in OpenAI's new GPT-4o is off the charts. It feels like something from Douglas Adams's Sirius Cybernetics Corp, breathing a sigh of contentment every time it makes your day.
“Google DeepMind last month released a paper exploring the ethical risks that may arise as AI assistants become more capable...the researchers warn that such assistants could become highly persuasive and addictive.” OpenAI: Here's a chatbot that flirts https://www.wired.com/...
Man idk what OAI cooked with gpt4o but ain't no way it's only 50% cheaper for them It's: - free (srsly they've been capacity constrained forever) - 4x faster that gpt4 turbo - better at coding - can reason across 3 modalities - realtime They're definitely making killing on
I am 80% sure openAI has extremely low latency low quality model get to pronounce first 4 words in <200ms and then continue with the gpt4o model Just notice, most of the sentences start with “Sure” “Of course” “Sounds amazing” “Let's do it” “Hmm” And then it continues with + gpt4…
OK, I have GLT-4o in my ChatGPT web interface now! The biggest improvement for me looks to be Code Interpreter, having it type out the code faster makes that tool much less frustrating to wait for [image]
@soumithchintala @OpenAI what about gpt4 on release felt impossible? i mean you ended up doing it ~1 year later so you figured it out somehow. just wanted to learn from benefit of hindsight
OpenAI emphasising voice interaction for gpt-4o is canny and somewhat telling - they're leaning into the masquerade of intelligence, voice is emotive and seductive (and they do it incredibly well) but it doesn't necessarily indicate leaps forward in capability
It does seem like most of the releases today were engineering “bells and whistles” added on top — not a lot of new fundamental capability breakthrough. - faster - cleaner UI - multi modal - cute voice synthesis Still very far from AGI - and not improving the trajectory.
OpenAI desktop app for mac will require macOS 14+ with apple silicon (m1, m2, m3). OpenAI just helped Apple with an upgrade push. #ChatGPT #GPT40 [image]
Vibecheck of GPT4T vs GPT4O from AInews: + summarization quality + more concise, specific, accurate + (unevenly) better topic and highlighting choices - instruction following on links - insturction following on outputting markdown GPT4O (11.5mins) was 74% faster than GPT4T [image…
With the release of GPT-4o and its apparent “artificial emotional intelligence,” you might call it, this seems like a good day to resurface this 2016 tweet by @sama
The theme of the age of AI: The demo is extremely impressive and relatively easy. The product is really, really, really hard. It is, perhaps, the perfect VC trap.
Folks are surprised that they are giving away gpt-4o for free. They are collecting your speech prompt data silly. Remember if you are not on the team plan, unless you go through a complicated opt-out process, all your data is by default used by openai. Not surprising anymore.
My notes on this morning's OpenAI release of the new GPT-4o model - not a huge leap in “intelligence” (whatever that might mean) but still very significant thanks mainly to the impressive new audio capabilities and the drop in price https://simonwillison.net/...
GPT-4o hot take: • The speech synthesis is terrific, reminds me of Google Duplex (which never took off). but • If OpenAI had GPT-5, they have would shown it. • They don't have GPT-5 after 14 months of trying. • The most important figure in the blogpost is attached below. And... […
GPT-4o hot take: • The speech synthesis is terrific, reminds me of Google Duplex (which never took off). but • If OpenAI had GPT-5, they have would shown it. • They don't have GPT-5 after 14 months of trying. • The most important figure in the blogpost is attached below. And [ima…
@swyx @OpenAI When GPT-4 was released, I think we very vastly unequipped, that was probably the biggest reason. the (Meta) GenAI organization (that is a very focused effort for Generative models, instead of a general AI research org like FAIR) started right around when GPT-4 was …
Alright I'm gonna say it... This is essentially AGI. This will be seen as magic to masses. What else do you call it when a virtual “person” can listen, talk, see, and reason almost indistinguishably from an average human? Isn't that AGI?
really exciting demos from @OpenAI establishing new expectations for AI. Lot of work to do on the Llama train — which isn't going to stop until we catch up! My personal feeling with gpt-4o — it feels attainable - contrasting to gpt-4 when released, it felt magically impossible.
By the very nature of its architecture, no LLM can reason. One might see dim reflections of inductive and deductive reasoning only in so far as the training data set contains such examples, but abductive reasoning - indeed any color of reasoning that requires the production of
Okay, GPT4 Omni is pretty rad! 🔥 From an audio-understanding standpoint, it can: 1. Transcribe audio better than Whisper large v3 2. It can diarise audio (meeting notes) 3. Can translate audio from one language to another 4. Summarise audio All of this zero/ few shot. From an [vi…
The ChatGPT-4o voice interaction is incredible. So far. I just spent the past half hour asking “it” verbally about the background of the people I will be meeting with, when I should leave the hotel tomorrow, the best way to get there. It nailed every single thing I fed it. Oh...
a big takeaway for me from today's event is that GPT-5 will be a single, giant, multimodal model. today had clear “this is 3.5, look its cool” vibes, when 4's in the offing.
ChatGPT-4o is here and omg...it still can't handle a simple reasoning task that most adult humans can figure out. But it did produce this very wrong answer much faster than it usually takes. (Ongoing shout out to @colin_fraser for identifying this particular task.) [image]
2x cheaper & faster is for English, but for other languages (especially non-latin-script) expect - thanks to our new tokenizer — even up to 9x cheaper/faster!
Congrats to OpenAI for the release of GPT-4o! 2x faster and 50% cheaper tokens will be great for everyone using agentic AI workflows. When an agentic job that used to take 10min now takes 5min just by switching APIs, that's great progress!
High quality inference at twice the speed and half the cost is a big, big deal for organizations building systems that utilize OpenAI technology. Amazing.
The downside of hyping your future V5 so much is that you have to release all of your new models under the V4 brand in order to avoid disappointment — in perpetuity
GPT-4o is not just limited to ChatGPT, it's also coming to the OpenAI API. GPT-4o is 4x faster and 2x cheaper than GPT-4 Turbo and comes with 5x higher rate limits. New wave of GPT-4o wrappers incoming. [image]
OpenAI just made their new GPT4 model 50% cheaper and 2X faster for developers. This is an insane level of improvement for anyone building in AI right now.
That was my read too. Will be interesting how Apple tries to position this, if and should they ink that deal, given their brand arc over the last decade has been “we do it better here” internal verticalisation. …
I guess OpenAI realized that in a world where most models have similar performance, the only moat is the connected experience for end users that brings them back to the platform. They emphasized that GPT-4o is available via API which might be a subtle hint that Siri would fallba…
I can see why there are rumors that Apple has caved and licensed OpenAI's models for Siri (or a replacement). This is the sort of feature Apple used to make compelling ads about the intersection of technology and the liberal arts back in the Jobs days.
OpenAI just dropped the demo of how a modern Siri should work. They are leaving Apple in the dust, given Siri has no such capabilities: and will not have anything close, unless Apple licenses this. Rare to see Apple out-executed with a “magical” phone experience like this. …
GPT-4o is our new state-of-the-art frontier model. We've been testing a version on the LMSys arena as im-also-a-good-gpt2-chatbot 🙂. Here's how it's been doing. [image]
As someone who spent a lot of time making a browser and researching it, I can tell you that this integration of ChatGPT on to the computer belies a greater purpose—one where AI will eat the browser steadily. They will no longer have to be restricted by the Google's platform [imag…
ChatGPT for Mac is a fully native macOS app. Not Electron. Not even Catalyst. This was a heroic effort by the team, and the result looks beautiful, if I may say so myself. And now I can finally say: we're hiring for amazing macOS and Windows engineers! 🧑💻👩💻 My DMs are open.
But the ELO can ultimately become bounded by the difficulty of the prompts (i.e. can't achieve arbitrarily high win rates on the prompt: “what's up"). We find on harder prompt sets — and in particular coding — there is an even larger gap: GPT-4o achieves a +100 ELO over our prior…
This demo is insane. A student shares their iPad screen with the new ChatGPT + GPT-4o, and the AI speaks with them and helps them learn in *realtime*. Imagine giving this to every student in the world. The future is so, so bright. [video]
Microsoft: AI PC!! OpenAI: Nah, MacOS Seriously hoping Windows on ARM brings some much-needed optimizations because the current MS Windows ecosystem is far too fractured for continued success in most AI workloads.
Mac app first. Suspect this is not just about users, but also because the app is a catalyst one. Easier to port an iOS app to the Mac than to build a whole new Windows app from scratch. And I imagine Copilot is getting 4o asap, meaning Windows users won't quite be left in the
With GPT-4o, this product is nearly ready. The next step (which may have already happened) is teaching the LLM to use iPhone apps. Then, when I tell the AI to order sandwiches for lunch, it can automatically launch the DoorDash app, discuss the options, place an order, and keep
OpenAI just dropped the demo of how a modern Siri should work. They are leaving Apple in the dust, given Siri has no such capabilities: and will not have anything close, unless Apple licenses this. Rare to see Apple out-executed with a “magical” phone experience like this.
A couple tidbits I've confirmed as well. 1) The mysterious GPT2-chatbot that showed up on benchmark sites was GPT-4o. 2) OpenAI did desktop version first for Mac because “we're just prioritizing where our users are.” @miramurati told me. https://www.axios.com/...
Now that I have GPT-4o in my ChatGPT iPhone app as well I've been doing a bit more of a vibes check, and the vibes are good Had it write me plot outlines for data journalism themed episodes of “Rick and Morty”, “Riverdale” and “Is It Cake?” https://chat.openai.com/...
OpenAI just announced a ChatGPT App for Mac.. and it looks insane! 💻 ChatGPT will soon be able to see what's on your screen. Then you can ask it, by voice, to explain/summarize what it sees. The outcome is amazing.. 🤯 [video]
Before today: ChatGPT's voice sounded like a young Black woman. Demo video today sounds like she's been replaced by a young white lady voice. :( https://www.techmeme.com/...
I am floored by this new version of ChatGPT. The voice sounds *SO* *NATURAL*! The emotion in the texture of the voice is incredible. Obviously there were issues in the demo, but there's no question we're basically there for voice. Here's an exemple: https://www.youtube.com/...
Looks like OpenAI isn't announcing a search engine, but announcing a refresh of the app as well as GPT-4o, which is an improved model (GPT4 for everyone.)
it is available to all ChatGPT users, including on the free plan! so far, GPT-4 class models have only been available to people who pay a monthly subscription. this is important to our mission; we want to put great AI tools in the hands of everyone.
GPT-4 Omni native integration via desktop apps is much bigger than it seems. It's the removal of dependancy on a browser. This will become the one stop solution for answers.
“Any sufficiently advanced technology is indistinguishable from magic.” Hearing people laugh as ChatGPT switched between voices in real time was really special. [video]
OpenAI just announced ChatGPT's new real-time conversational chat. The model can understand both audio AND video, and can even detect emotion in your voice. This is insane. [video]
Imagine having a fast multimodal AI like we just saw from OpenAI on the Ray-Ban Meta smart glasses. It sees what you see, hears what you hear and whispers in your ear. Magic. Matter of time before Meta is ready.
It is confusing, but users now seem to have access to GPT-4o, the model, that's it. It has the same features at GPT-4 but is faster and smarter. What isn't out yet: -Cool voice features, voice mode still goes to old version -New multimodal features, still DALL-E & old vision.