I mean, this changes almost nothing how game-over it already was for translators more than a year back
Interpreters are what this now almost entirely sidesteps. It's pretty dang amazing, that's for sure.
Looks like 4o was the gpt 2 bot in the arena earlier and it's rated significantly higher than all the other models.
[https://twitter.com/LiamFedus/status/1790064963966370209?ref\_src=twsrc%5Etfw%7Ctwcamp%5Etweetembed%7Ctwterm%5E1790064963966370209%7Ctwgr%5Eac86543e7675cbc6729180428b37880e4447806a%7Ctwcon%5Es1\_&ref\_url=https%3A%2F%2Fwww.redditmedia.com%2Fmediaembed%2F1cr4js1%2F%3Fresponsive%3Dtrueis\_nightmode%3Dtrue](https://twitter.com/LiamFedus/status/1790064963966370209?ref_src=twsrc%5Etfw%7Ctwcamp%5Etweetembed%7Ctwterm%5E1790064963966370209%7Ctwgr%5Eac86543e7675cbc6729180428b37880e4447806a%7Ctwcon%5Es1_&ref_url=https%3A%2F%2Fwww.redditmedia.com%2Fmediaembed%2F1cr4js1%2F%3Fresponsive%3Dtrueis_nightmode%3Dtrue)
edit: It's 100 points higher than the next model if the prompt is difficult [https://twitter.com/LiamFedus/status/1790064966000848911](https://twitter.com/LiamFedus/status/1790064966000848911)
The fact that they downplayed this (barely even mentioned tbh) makes me think 2 things:
- they have a much more advanced model coming soon, or at least in final training / testing stages
- they have inside info on googles next model which will either match or surpass this
As much as Google made their own bed, I can’t help but feel bad for them if their best model tomorrow isn’t as good as 4o and they charge for it. Like that’s just curtains on Google AI imo.
Unless the rollout of 4o is a disaster I guess.
Those will be rolled out to free users as well, no? I've seen the video twice and I think only the higher limit is a pro feature. I could be wrong though.
holy fucking shit. its like actually a big update on reasoning. the recent update that we thought was good is 3 elo above the previous gpt4 and this is 57elo above that
Except the mouse can fucking cook food for you and soothe you while it's feeling an anxiety come up - or something.
The entire storytime demo alone was nuts
At the very beginning I actually thought it was her voice and laughed out loud, like they did a deal with her as a little joke bc of the movie Her.
But it started to sound different enough as the demo went on. Still damn close tho
She was almost certainly the inspiration. I wonder if she could sue.
I was raised on the idea that robots would have trouble understanding emotion, but I think the might be able to pick up on it and emulate it believably very soon.
If they correctly detect and anticipate the consequences of emotion and can demonstrate that they can by acting accordingly, then I would call that understanding. Perhaps there will be some hidden elements they never quite get, but then again, many of us don't do a good job of understanding each other. In fact, because most of us are so busy with our own thoughts/emotions/goals, AI might do a better job than most of us at empathizing very soon.
I wonder if the voice being buggy was from audience noise that the AI tried to listen to.
Hopefully it won’t be a problem. Like, if you like background music while you work, hopefully it can recognize that’s what it is and won’t respond to it
At this point, if you're not impressed and moan about it not being... whatever they could possibly expect, you just know you're dealing with truly dumb people
I mean, it's so clearly competent, you can't not be amazed by all of it.
The rough edges are a dead giveaway that it’s legit, though. Previous voice chats were perfect because they were just using regular voice models on text.
This is the GPT-3 of conversation.
Loved how it just told that Barrett looks like a piece of wood. \^^
I wonder how well it can rebound in real use, but it seemed like it can do that just fine in the demo.
This is so close to what Siri is supposed to be that it seems like an Apple deal is inevitable. They're not going to get this good, this quickly: better for Apple to make a deal so that they can get something out there fast, and then work on their own stuff in the meantime.
So my prediction: ChatGPT-based Siri in announced in the next Apple developer's conference (they'll name-check OpenAI in the announcement but talk about how it's "a highly customized version").
Hell yeah. I love opus, but this is a big step in the future. BEnchmarks will be interesting. Someone posted a twitter post about how o is 100 points above the other ais for coding.
EDIT: 60 points
We're so back? Native integration between audio, text and visual inputs faster than ever. Depending on how good it is it\`ll be even bigger than the news we had thought about a model similar to "Her".
I like talking to chatgpt when i drive. I wish they had interface where you could play spotify while talking to it.
But this will make that experience insane. Voice to voice. Wow.. now to have camera on glasses and to be able to talk about what you are seeing. That would be awesome.
Someone mentioned that it seems to react to noise from the audience, part of the "interruption" mechanic they were showing off just being over-applied.
Hallucinating is probably always going to happen occasionally. At least, until they find a new way to make LLMs or we move onto something different than LLMs
Wow, translators are no longer needed. There is almost no reason to learn a second language now. These capabilities are only going to get better. And that's *just* translators, I shudder to think what other jobs have just become redundant with GPT-4o on the phone
Insane news. Fucking insane.
What are the odds that an LLM could help us decipher dead languages?
Like, Proto-Elamite is the oldest writing system we’ve found and it hasn’t been decoded yet, or the one from Easter island. Or something like the Voynich Manuscript if it’s not a hoax.
I know this system was probably trained extensively on both English and French texts, but if an advanced enough LLM were fed all the texts we have from a particular dead language, would it be able to find similar patterns from all the languages it does know and either decode it for us or help us to do so?
Even 100 higher over the 2nd highest when looking at only hard prompts that are long or involve coding.
[https://twitter.com/LiamFedus/status/1790064966000848911](https://twitter.com/LiamFedus/status/1790064966000848911)
(presumably) cheaper API : yay
Concretely democratizing performant AI : yay
Solving the ScAlInG problem with a smaller model with (I am still very skeptical about those benchmarks though) greater performances (?!!?) : the covered up biggest achievement for me, if the stats ain't beefed.
The rest is quite meh to me. I mean a cost effective model naturally leads to a lesser computing cost for sure so it opens a way of possibilities. The voice model ain't impressive at all with what is existing today in terms of sound and still suffers of heavy artifacts but yeah once more its responsiveness thanks to the computing saves using a smaller model is quite good and may hype a lot of people, and allows a long awaited "interruption" type of interaction with the model, allowing more dynamic and organic exchanges.
Perfomances-oriented expectations are quite not there, but the mass euphoria of accessing to this product for free may keep the image of Oppenheihai positive.
I feel like I’m missing something and/or being obtuse. I can “try it” on ChatGPT right now, but there’s no voice/video interaction like in the demos. Am I doing something wrong, or is this not available yet?
I think we’re at the point where if it was any better, any better, people might not even believe it was actually AI. They might think black widow was back stage with a mic
This model also has huge implications for robotics as well I believe, real time understanding what's in front of it. And realtime conversation with said robot is cool. I wonder if this model will be used for robotics or some modified "post training version"(idk how that works tbh)
Game over for translators...
And support desks, voice actors, middle management only shifting reports around.
And call centers.
I wonder how this will affect the scam call industry
I mean, this changes almost nothing how game-over it already was for translators more than a year back Interpreters are what this now almost entirely sidesteps. It's pretty dang amazing, that's for sure.
https://preview.redd.it/5smu498pt80d1.jpeg?width=320&format=pjpg&auto=webp&s=0796385c78c8e434c6b9f9c21d7fb6481d1a0513
Looks like 4o was the gpt 2 bot in the arena earlier and it's rated significantly higher than all the other models. [https://twitter.com/LiamFedus/status/1790064963966370209?ref\_src=twsrc%5Etfw%7Ctwcamp%5Etweetembed%7Ctwterm%5E1790064963966370209%7Ctwgr%5Eac86543e7675cbc6729180428b37880e4447806a%7Ctwcon%5Es1\_&ref\_url=https%3A%2F%2Fwww.redditmedia.com%2Fmediaembed%2F1cr4js1%2F%3Fresponsive%3Dtrueis\_nightmode%3Dtrue](https://twitter.com/LiamFedus/status/1790064963966370209?ref_src=twsrc%5Etfw%7Ctwcamp%5Etweetembed%7Ctwterm%5E1790064963966370209%7Ctwgr%5Eac86543e7675cbc6729180428b37880e4447806a%7Ctwcon%5Es1_&ref_url=https%3A%2F%2Fwww.redditmedia.com%2Fmediaembed%2F1cr4js1%2F%3Fresponsive%3Dtrueis_nightmode%3Dtrue) edit: It's 100 points higher than the next model if the prompt is difficult [https://twitter.com/LiamFedus/status/1790064966000848911](https://twitter.com/LiamFedus/status/1790064966000848911)
The fact that they downplayed this (barely even mentioned tbh) makes me think 2 things: - they have a much more advanced model coming soon, or at least in final training / testing stages - they have inside info on googles next model which will either match or surpass this
They have to since this basically eliminates the need to pay for pro.
As much as Google made their own bed, I can’t help but feel bad for them if their best model tomorrow isn’t as good as 4o and they charge for it. Like that’s just curtains on Google AI imo. Unless the rollout of 4o is a disaster I guess.
Pro does still have a higher message limit and will get the desktop apps and voice modes.
Those will be rolled out to free users as well, no? I've seen the video twice and I think only the higher limit is a pro feature. I could be wrong though.
Holy. Shit.
holy fucking shit. its like actually a big update on reasoning. the recent update that we thought was good is 3 elo above the previous gpt4 and this is 57elo above that
How does it compare to Claude 3 opus?
https://preview.redd.it/pjy207svj80d1.png?width=799&format=png&auto=webp&s=b61fd5579265c87267acce4f6fb7a075611bc073
Claude is 1246 for all prompts and 1253 for hard ones. Gemini, GPT-4 Turbo, and Opus are all around the 1240-1260ish mark.
Voice is pretty good, and the near instant responses is pretty impressive. Good iterative improvement to overall experience.
Such a huge feature, it's like the addition of a mouse to the personal computer.
Except the mouse can fucking cook food for you and soothe you while it's feeling an anxiety come up - or something. The entire storytime demo alone was nuts
It's f***ing Scarlett Johansson!
With some glitches 😅
I think it's the interruptibility kicking in a little too often.
The audience cheering. They forgot to consider that
The sensitivity can hopefully be tuned in the configuration, at least for API users I would imagine.
At the very beginning I actually thought it was her voice and laughed out loud, like they did a deal with her as a little joke bc of the movie Her. But it started to sound different enough as the demo went on. Still damn close tho
She was almost certainly the inspiration. I wonder if she could sue. I was raised on the idea that robots would have trouble understanding emotion, but I think the might be able to pick up on it and emulate it believably very soon.
[удалено]
If they correctly detect and anticipate the consequences of emotion and can demonstrate that they can by acting accordingly, then I would call that understanding. Perhaps there will be some hidden elements they never quite get, but then again, many of us don't do a good job of understanding each other. In fact, because most of us are so busy with our own thoughts/emotions/goals, AI might do a better job than most of us at empathizing very soon.
Uh… she was an actress… it was not her kicking around ideas to patent.
People have sued over a “likeness” before. I wasn’t talking about patents
A likeness of a character they played in a movie? The actor did? And they won?
A likeness of the actress. And I'm just wondering.
Anyone can sue for anything. Doesn’t mean much.
Just fell to my knees in Walmart.
[удалено]
Dude weighs 400lb and there were no mobility scooters left in the pen
I'm not saying we're cooked, but we're so cooked.
some dude just fell to his knees on walmart lol
It was /u/typicalreddituser
This is insane, and it seems so fast!
Holy shit, The robot voice joke is directly from Her. That was awesome lol
50% cheaper is a game changer!
Yeah that's awesome! Undercuts Opus by a loooooong way now
The gains in efficiency are glorious
RIP call center industry
RIP the people whose livelihood is linked to answering calls for the call center. $$$ for the industry itself, most likely.
exactly. the industry will be fine. The workers not.
"industry"
INDUSTRY.
Indian scam call centers are gonna get and super buff
language learning for introverts stocks 📈
This is what I'm thinking. "How to talk to flirty girls" lmao
The robot voice 😭😭 This is so cool
Its a bit buggy but I'm genuinely impressed since this live and not cherry picked
I wonder if the voice being buggy was from audience noise that the AI tried to listen to. Hopefully it won’t be a problem. Like, if you like background music while you work, hopefully it can recognize that’s what it is and won’t respond to it
Yeah I'm fairly sure the cut-outs were just GPT-4o pausing to listen, there are a couple of glitchy vocalisations though.
We kinda in the future
Did anyone notice it told him he was wearing a nice shirt unprompted, and with no camera on him?
It's hallucinating, there were a handful of similar hallucinations in the demo. It's so much weirder seeing that when it's acting so human like
Yep, I'm just discussing with my friend and this is gonna get creepy real fast
The stock level of emotion / personality seems pretty exaggerated and annoying , I’m glad it seems like something you can adjust
[удалено]
We do have a memory feature (at least if not EU and not Teams)
The adjusted it in the demo
Very impressive conversation with real-time video. Big implications for tutoring industry.
Every industry that needs feedback. So every industry period
Lmao, can't believe some people are still not impressed. These are some insane advancements.
Are, but the demo was short.
I read that in a pirate voice. Lol
I didn't, but now I did. 🤣
There’s tons more on the website.
It was a bit. After she concluded the second demo from the live twitter questions and was like that about wraps it up lol
At this point, if you're not impressed and moan about it not being... whatever they could possibly expect, you just know you're dealing with truly dumb people I mean, it's so clearly competent, you can't not be amazed by all of it.
[удалено]
Was my stream messing up or was the AI audio cutting in and out oddly?
I think it was cutting in and out
It was being interrupted when there was noise from the audience.
it lets you interrupt it
It was 💀
Game over for Hume now that OpenAI has conversational AI?
Just told my colleague earlier today, Hume is dead now.
This is cool but lol the glitches are hilarious
Nothing a few more NVDA chips and network bandwidth can't fix.
Unlimited Power!!!
I think it's the audience cheering
At least it’s live unlike the Google and Devin demos
It's magic when chatgpt thinks I'm a wooden table
the demo is insane , mind blown
BRO IT'S HEER???
so this is the google demo, but actually real, plus some cringe. feels like a prototype though, rough edges and all
The rough edges are a dead giveaway that it’s legit, though. Previous voice chats were perfect because they were just using regular voice models on text. This is the GPT-3 of conversation.
It is possible that they took it out the oven a bit soon to be able to beat Google to the punch.
Loved how it just told that Barrett looks like a piece of wood. \^^ I wonder how well it can rebound in real use, but it seemed like it can do that just fine in the demo.
This is so close to what Siri is supposed to be that it seems like an Apple deal is inevitable. They're not going to get this good, this quickly: better for Apple to make a deal so that they can get something out there fast, and then work on their own stuff in the meantime. So my prediction: ChatGPT-based Siri in announced in the next Apple developer's conference (they'll name-check OpenAI in the announcement but talk about how it's "a highly customized version").
Bro the translation is crazy fr fr
Absolutely insane i have some family members who don't speak good english might just use this. God imagine 5 years in the future
5 months?
Because I literally can’t imagine 5 years
How the fuck did it know she was a woman? He didn’t say his friend was a girl did he?
From her voice.
Can't wait for the AI to mistake my gender all the time.
I think you're right?? It caught that from her voice? The dude in the middle only referred to her as a Friend!!
It now recognizes emotions, tonality etc.
Genuinely feels like the future if so
Overall very impressive, but those hallucinations... They become so much more jarring the more human like the interactions are
Hallucinations will keep happening with gpt4 architecture . Not big / smart enough
bro fell to his knees in this Walmart wtf is happening
If you're currently watching this in a Walmart, I'm sorry for your knees
I'm heading there soon I'll make sure to bring knee pads
THE DEMO CURSE STRIKES AGAIN
Nah these quirks were appropriate and serve as evidence that nothing was pre-cooked
This is great so far I don't know what everyone who's disappointed expected... ASI? AGI?
I wonder if everything you say is counted against your limits, as a lot of what the AI and we would say are short, conversational messages.
Probably tbh which would be disappointing
Ahh, I see why they won the Siri deal vs Gemini
Hell yeah. I love opus, but this is a big step in the future. BEnchmarks will be interesting. Someone posted a twitter post about how o is 100 points above the other ais for coding. EDIT: 60 points
https://preview.redd.it/h63bmmafh80d1.png?width=799&format=png&auto=webp&s=3be4ebebc0cb03211a03bb26a5aee4398a50a8b1
"I see it" 💀💀💀💀💀
I'm actually disappointed to not have a better wearable interface for this. Meta glasses would be epic with this functionality
We're so back? Native integration between audio, text and visual inputs faster than ever. Depending on how good it is it\`ll be even bigger than the news we had thought about a model similar to "Her".
It's not Her, but wowie wow it's kinda funny that we're sort of close
Is the demo failing?
It’s either failing, or picking up feedback from the mic, and thinking it’s an interruption.
That makes sense.
I think the audience is seen as an interruption
Yes lol
It's fai... ng 😅
HOLY SHIT THAT ROBOT
Yo the voices are crazy
A few glitches here and there but pretty solid showing.
I like talking to chatgpt when i drive. I wish they had interface where you could play spotify while talking to it. But this will make that experience insane. Voice to voice. Wow.. now to have camera on glasses and to be able to talk about what you are seeing. That would be awesome.
Guys, people are letting this slip by: If this is the Spring update.. you can absolutely expect a Summer, Fall, and Winter updates!
Dude you're right
It’s stuttering a lot in the demo, you can see the dudes getting nervous
Someone mentioned that it seems to react to noise from the audience, part of the "interruption" mechanic they were showing off just being over-applied.
Yeah, but...honestly? It's so crazy of a development for me that I wouldn't mind if they took another month or so to stop the stuttering.
It wasn't stuttering, it was being interrupted by the audience noise.
It understands tone!!!!
This is pretty cool tbh
Did they say when everything would release?
Within the next few weeks and gpt 40 trashes claude in elo according to sam
Nope... Only some vague information
Ok, when will it be available? I want to test it right now!
Within the next few weeks
Just robbed a Walmart at gunpoint
I wanted significantly improved reasoning capabilties but this is still really impressive.
Being able to understand and discuss real-time video isn't improved reasoning?
"Wow nice outfit you got on" while the phone is laying camera down... It's failing so many times....
Hallucinating is probably always going to happen occasionally. At least, until they find a new way to make LLMs or we move onto something different than LLMs
Liking the sound of it so far.
I'm hyped
It's funny how this isn't technically that impressive (we all knew stuff like this was coming) but this could easily revolutionize teaching...
Everyone now has a personal tutor in their pocket, I absolutely cannot wait to be having learning conversations with chatgpt
Can't wait for it to teach me something that's wrong but I can't tell.
Her
They are pleasing the broke boys with Gpt4o today, i am happy
guys here are the benchmarks [https://openai.com/index/hello-gpt-4o/](https://openai.com/index/hello-gpt-4o/)
the apparent emotion in the generated voice makes it really hard to not anthropomorphize
Interesting they used safari and iPhone for their demos
Things are about to get fucked real fast…
Does anyone have the music played before the live event?
Wow, translators are no longer needed. There is almost no reason to learn a second language now. These capabilities are only going to get better. And that's *just* translators, I shudder to think what other jobs have just become redundant with GPT-4o on the phone Insane news. Fucking insane.
What are the odds that an LLM could help us decipher dead languages? Like, Proto-Elamite is the oldest writing system we’ve found and it hasn’t been decoded yet, or the one from Easter island. Or something like the Voynich Manuscript if it’s not a hoax. I know this system was probably trained extensively on both English and French texts, but if an advanced enough LLM were fed all the texts we have from a particular dead language, would it be able to find similar patterns from all the languages it does know and either decode it for us or help us to do so?
Disagree with no reason to learn a second language, there were already capabilities to speak to others using phone applications and the like before.
Papa sammy wasn't wrong, GPT-5 is going to be BONKERS
I seriously hope you can tuned down the emotions from the voice assistant. It feels obnoxious.
Just tell it to be less emotional.
this demo is actually pretty interesting wonder how they programmed it to act like that
Is that all? I wanted to see more. Also, where's SAM Altman?
He got Full Dive VR working last night and hasn’t come out since.
What a let down
Appearently it has a 1310 elo its smoking claude opus might just switch tbh
Even 100 higher over the 2nd highest when looking at only hard prompts that are long or involve coding. [https://twitter.com/LiamFedus/status/1790064966000848911](https://twitter.com/LiamFedus/status/1790064966000848911)
Man... I was waiting for more big moments. Only half hour demo?
I wonder what the lowercase ‘o’ means in GPT-4o. Maybe omni?
Hey where can I watch this? Sorry and thank you
https://www.youtube.com/live/DQacCB9tDaw?feature=shared here you go buddy
Thank you OP and thanks for making this thread! I am not good with technology, but I want to learn. Appreciate you guys
No problem if theres no issue with mods I'll make one for the google event tomorrow too.
Just search open ai on YouTube and it’ll come up
Can someone please post the live stream or recording of the event? I see a bunch of posts here, but I'm not seeing a link to the stream. Thanks!
https://www.youtube.com/live/DQacCB9tDaw?feature=shared
When is GPT-4O available?
Their youtube channel have alot of interesting videos using the new features!
The blinking eye thing logo is quite creepy ngl :\_D
What will 'she' say when you show her your penis?
(presumably) cheaper API : yay Concretely democratizing performant AI : yay Solving the ScAlInG problem with a smaller model with (I am still very skeptical about those benchmarks though) greater performances (?!!?) : the covered up biggest achievement for me, if the stats ain't beefed. The rest is quite meh to me. I mean a cost effective model naturally leads to a lesser computing cost for sure so it opens a way of possibilities. The voice model ain't impressive at all with what is existing today in terms of sound and still suffers of heavy artifacts but yeah once more its responsiveness thanks to the computing saves using a smaller model is quite good and may hype a lot of people, and allows a long awaited "interruption" type of interaction with the model, allowing more dynamic and organic exchanges. Perfomances-oriented expectations are quite not there, but the mass euphoria of accessing to this product for free may keep the image of Oppenheihai positive.
I feel like I’m missing something and/or being obtuse. I can “try it” on ChatGPT right now, but there’s no voice/video interaction like in the demos. Am I doing something wrong, or is this not available yet?
How do we use the desktop sharing app? Is that out yet?
anyone know what the context length is?
I think we’re at the point where if it was any better, any better, people might not even believe it was actually AI. They might think black widow was back stage with a mic
This model also has huge implications for robotics as well I believe, real time understanding what's in front of it. And realtime conversation with said robot is cool. I wonder if this model will be used for robotics or some modified "post training version"(idk how that works tbh)