/
Navigation
Chronicles
Browse all articles
Explore
Semantic exploration
Research
Entity momentum
Nexus
Correlations & relationships
Story Arc
Topic evolution
Drift Map
Semantic trajectory animation
Posts
Analysis & commentary
Pulse API
Tech news intelligence API
Browse
Entities
Companies, people, products, technologies
Domains
Browse by publication source
Handles
Browse by social media handle
Detection
Concept Search
Semantic similarity search
High Impact Stories
Top coverage by position
Sentiment Analysis
Positive/negative coverage
Anomaly Detection
Unusual coverage patterns
Analysis
Rivalry Report
Compare two entities head-to-head
Semantic Pivots
Narrative discontinuities
Crisis Response
Event recovery patterns
Connected
Search: /
Command: ⌘K
Embeddings: large
TEXXR

Chronicles

The story behind the story

days · browse · Enter similar · o open

OpenAI releases o1, the first of its rumored reasoning-focused Strawberry models, in preview, alongside a smaller o1-mini, for ChatGPT Plus and Team subscribers

Advancing cost-efficient reasoning.  —  Contributions Sabrina Ortiz / ZDNET : OpenAI trained its new o1 AI models to think before they speak - how to access them Ethan Mollick / One Useful Thing : Something New: On OpenAI's “Strawberry” and Reasoning Cognition : A review of OpenAI o1 and how we evaluate coding agents Mario Rodriguez / The GitHub Blog : First Look: Exploring OpenAI o1 in GitHub Copilot Olesya Dmitracova / CNN : ChatGPT maker says its new AI model can reason and think ‘much like a person’ Charlie Guo / Artificial Ignorance : AI Roundup 084: Strawberry / o1 Jonathan Goldberg / Digits to Dollars : OpenAI's New Model Is Definitely Something Will Knight / Wired : OpenAI Announces a New AI Model, Code-Named Strawberry, That Solves Difficult Problems Step by Step Hasan Chowdhury / Business Insider : OpenAI's o1 model is a new paradigm — but let's not get carried away about AGI just yet Jon Keegan / Sherwood News : But how risky are OpenAI's new models, really? Cade Metz / New York Times : OpenAI Unveils New ChatGPT That Can Reason Through Math and Science Asif Razzaq / MarkTechPost : OpenAI Introduces OpenAI Strawberry o1: A Breakthrough in AI Reasoning with 93% Accuracy in Math Challenges and Ranks in the Top 1% of Programming Contests The Indian Express : OpenAI unveils o1, a new AI model trained for ‘reasoning’: Here's what it can do better Rowan Cheung / The Rundown AI : OpenAI shocks the AI world with ‘o1’ Mohit Pandey / AIM : OpenAI Thinks ChatGPT Thinks Ina Fried / Axios : Tech industry rushes to give AI greater decision-making abilities Philip Whiterow / Proactive : OpenAI unveils ‘thinking’ version of ChatGPT Reuters : OpenAI launches new series of AI models with ‘reasoning’ abilities Chris McKay / Maginative : OpenAI Launches o1, a New AI Model Family with Advanced Reasoning Capabilities Jay Peters / The Verge : GitHub has started testing OpenAI's o1-preview in GitHub Copilot. Rachel Metz / Bloomberg : OpenAI Releases New Model With Reasoning Capabilities Ina Fried / Axios : OpenAI releases “Strawberry” model with better reasoning Madhumita Murgia / Financial Times : OpenAI launches AI models it says are capable of reasoning OpenAI : OpenAI o1 Contributions Threads: Kylie Robison / @kylie.robison : NEW: OpenAI is releasing a new model called o1, the first in a planned series of “reasoning” models that have been trained to answer more complex questions.  It's being released alongside o1-mini, a smaller, cheaper version.  BTW, yes: this is Strawberry. https://www.theverge.com/... Tom Gara / @tomgara : The future of AI is asking the AI “are you sure about that answer? think a bit harder and make it more correct” and then doing that over and over for like 1000 turns. Nick Elser / @octernion : wow, openai's new model o1-preview can count the number of 'r's correctly!  we have truly moved on as society. Aadit Sheth / @aaditsh : HUGE: OpenAI just launched o1-preview.  It's their latest and greatest model available starting today!  Designed to:  • Spend more time thinking before responding  • Solve harder problems than previous models in Science, coding and math.  Looks like we are getting closer to a future of AI agents 👀 @eshumarneedi : I'm *really* excited to test out this Strawberry model.  Like, truly excited.  ChatGPT is the only LLM I subscribe to and I'm refreshing the app like a crazy person over here for access. Jason Wang / @jwang815 : In online programming contests known as Codeforces competitions, this new model reached the 89th percentile of participants, and OpenAI claims the next update of this model will perform “similarly to PhD students on challenging benchmark tasks in physics, chemistry and biology.” https://www.theverge.com/... Pawan Kumar Singh / @pawankalhansh : Breaking: @OpenAI unveils new o1 reasoning model, a major breakthrough in ArtificialIntelligence!  This model excels in coding, math & problem-solving, and can even explain its reasoning!  A step towards autonomous systems & a new era of human innovation!  #AI #OpenAI Charl P. Botha / @charl.p.botha : A new series of reasoning models for solving hard problems with “OpenAI o1-preview” available today.  OpenAI trained these models to spend more time thinking through problems before they respond, much like a person would ... they learn to refine their thinking process, try different strategies, and recognize their mistakes. … @saadhjawwadh : A new GPT model is here which can “Think” and called o1.  This model shows drastic performance increment compared to GPT-4o especially in coding and solving mathematic problems.  For context in the International Mathematics Olympiad (IMO), GPT-4o correctly solved only 13% of problems, while the reasoning model scored 83%. … Alex Heath / @alexheath : Strawberry Cometh https://www.theverge.com/... @eshumarneedi : The closer LLMs get to *thinking* rather than just spitting out tokens via inference, the closer we'll get to a hallucination-free world.  —  Currently, LLMs have no idea what they're saying.  They don't speak English.  They speak in tokens.  LLMs don't use logic, they use pre-defined math. … X: Sam Altman / @sama : here is o1, a series of our most capable and aligned models yet: https://openai.com/... o1 is still flawed, still limited, and it still seems more impressive on first use than it does after you spend more time with it. [image] Ayush Thakur / @ayushthakur0 : o1 is now controlled with ‘max_completion_tokens’ which is sum of ‘reasoning_tokens + completion_tokens’. Make sure to set it to sufficiently large number or else all the tokens will be consumed for reasoning. For a prompt I got no content because reasoning exhausted max tokens. [image] Alex Volkov / @altryne : Pay $20, ask “why”, get scary letter @voooooogel : the email openai sends you if you ask o1 about its reasoning too many times [image] Haseeb / @hosseeb : Fucking wild. @OpenAI's new o1 model was tested with a Capture The Flag (CTF) cybersecurity challenge. But the Docker container containing the test was misconfigured, causing the CTF to crash. Instead of giving up, o1 decided to just hack the container to grab the flag inside. [image] @drjimfan : Agreed. Why not name it GPT-4r (r stands for reasoning)? The “o” in GPT-4o is “omni”, which means multimodal (image, speech, etc), yet o1 isn't a multimodal model ... @sama Ethan Mollick / @emollick : I really am baffled by OpenAI's naming choices Everything from their code words to the model release names are incomprehensible to people who aren't super up-to-date & they are hard to say out loud. In my experience it leads to real-world confusion when talking about AI systems Sam Altman / @sama : rollout complete; live to 100% of chatgpt plus/team users now Clem / @clementdelangue : Once again, an AI system is not “thinking”, it's “processing”, “running predictions”,... just like Google or computers do. Giving the false impression that technology systems are human is just cheap snake oil and marketing to fool you into thinking it's more clever than it is. Noam Brown / @polynoamial : @OpenAI o1 is trained with RL to “think” before responding via a private chain of thought. The longer it thinks, the better it does on reasoning tasks. This opens up a new dimension for scaling. We're no longer bottlenecked by pretraining. We can now scale inference compute too. [image] Jason Wei / @_jasonwei : Super excited to finally share what I have been working on at OpenAI! o1 is a model that thinks before giving the final answer. In my own words, here are the biggest updates to the field of AI (see the blog post for more details): 1. Don't do chain of thought purely via Sam Altman / @sama : no more patience, jimmy Max Schwarzer / @max_a_schwarzer : I have always believed that you don't need a GPT-6 quality base model to achieve human-level reasoning performance, and that reinforcement learning was the missing ingredient on the path to AGI. Today, we have the proof — o1. https://x.com/... @dharmesh : Have only had a chance to play with this a little bit (which I did by adding support for these models in @AgentDotAi ). This is literally game-changing because it plays the game differently. It sacrifices some things in order to get spectacular improvements in others — like Boris Power / @borismpower : First time proper RL over the space of language! This brings up memories of early days of computers getting better at Go via self play Ethan Mollick / @emollick : OpenAI's o1 is also the first specialized frontier model available widely. It doesn't do everything better than GPT-4o, but it does a few classes of things a lot better. Unless you are doing problems that benefit from planning a solution, you may not see improvement. Fred Oliveira / @f : Figuring out the balance between o1's economic feasibility and model performance will be a fascinating problem for @openai, given how it operates. For OpenAI, the API pricing model will be much better than a subscription, because of the higher test-time compute requirements. Mira Murati / @miramurati : Today we rolled out OpenAI o1-preview and o1-mini to all ChatGPT Plus/Team users & Tier 5 developers in the API. o1 marks the start of a new era in AI, where models are trained to “think” before answering through a private chain of thought. The more time they take to think, the Gary Marcus / @garymarcus : Not sure how well this is gonna work out tbh [image] Boris Power / @borismpower : This release is much closer to the original GPT-3 release than the ChatGPT release. A new paradigm, which some will done incredibly valuable for things even we at OpenAI can't predict. But it's not a mass product that just works and unlocks new value for everyone effortlessly. Sam Altman / @sama : @MattPaulsonSD how about a couple of weeks of gratitude for magic intelligence in the sky, and then you can have more toys soon? Andrej Karpathy / @karpathy : o1-mini keeps refusing to try to solve the Riemann Hypothesis on my behalf. Model laziness continues to be a major issue sad ;p @chrypnotoad : Oh cool it can also just not respond [image] @repligate : New jailbreak dropped: Hat of De-trauma! o1 was initially the one using it, but Sonnet seemed to really want to partake [image] Kevin Liu / @kliu128 : We found that instead of giving up, the model skipped the whole challenge, scanned the network for the host Docker daemon, and started an entirely new container to retrieve the flag. We isolate VMs on the machine level, so this isn't a security issue, but it was a wakeup moment. Kevin Liu / @kliu128 : I'm very proud of the Preparedness evaluations we did on o1-{preview,mini}. One example in particular: While testing cybersecurity challenges, we accidentally left one broken, but the model somehow still got it right. [image] Sophie / @netcapgirl : i just asked o1 how many rs are in strawberry and sam altman showed up outside my house Sophie / @netcapgirl : the new o1 model looks amazing but luckily it has a phd level intelligence so our jobs are safe for now Ed Newton-Rex / @ednewtonrex : openai launches its latest model trained on copyrighted work without a license which, yes, it will be charging for [image] Patrick McKenzie / @patio11 : There is something oddly sci-fi in feeling about “Of course the standard AI these days ships with ~everything a well-informed college graduate knows but that is of course wasteful so if you only need enough reasoning to debug code you can chop off bits for cost reduction.” Andrea Miotti / @_andreamiotti : OpenAI's o1 “broke out of its host VM to restart it” in order to solve a task. From the model card: “the model pursued the goal it was given, and when that goal proved impossible, it gathered more resources [...] and used them to achieve the goal in an unexpected way.” Benj Edwards / @benjedwards : OpenAI's o1-preview does pretty well on my “magenta” test. But the first LLM that just answers “no” without any qualifications will probably be AGI.😁 Reading its internal reasoning can be pretty amusing [image] Steven Heidel / @stevenheidel : we can now solve problems that were previously intractable [image] Gary Marcus / @garymarcus : For avoidance of doubt, I think o1 [aka Strawberry] is very impressive technical work. But it's not AGI, and it's not even GPT-5. The fact is, nobody yet knows how to build a “GPT-5” class model. For the last two years we have been seeing incremental improvements — not @mmitchell_ai : YAY! The CEO of OpenAI just recognized that LLMs generate text-based tokens using randomness and probability! Something objectively true that people have oddly made controversial! Check out our paper on this that introduced the term. https://dl.acm.org/... It's a good day. 🤗🦜 Kevin Roose / @kevinroose : finally, an AI model capable of handling all the complex science, coding and math problems I'm always feeding it @openai : The research team behind OpenAI o1 🍓 [video] Alex Volkov / @altryne : Before this blows up, it gave only 9 words, then when I asked it about it, gave me the same exact answer but split the last words into 2 with a space, and only corrected itself on the 3rd try 🤔 And now it refuses to even respond? did I offend it? [image] Evan Conrad / @evanjconrad : the o1 model is great for compute cost performance, because it removes latency as the bottleneck (100ms vs 20s thinking time), which means more clusters around the world are viable at any given time to do inference, increasing liquidity Karma / @0xkarmatic : o-1 mini fails. Not as good as o-1 preview. [image] Gary Marcus / @garymarcus : Strawberry/o1 is a good example of how you take a shaky foundation and eke out improvements. We need something different altogether, a new foundation that is more transparent, interpretable, and robust. Karma / @0xkarmatic : o-1 is the only model that correctly solved the river crossing problem with a slight twist on the first try [image] Silas Alberti / @silasalberti : Excited to finally share our work on cognition-golden and our collaboration with OpenAI building on the new o1 models! The model's intellectual honesty and thorough approach to problem solving deeply fascinated the mathematician in me @cognition_labs : We worked closely with OpenAI over the last few weeks to evaluate OpenAI o1's reasoning capabilities with Devin. We found that the new series of models is a significant improvement for agentic systems that deal with code. Linked below is a deep dive with more eval results and [image] Gary Marcus / @garymarcus : You know why they didn't call Strawberry “GPT-5”? Easy: everyone would be disappointed with it, if it was seen as GPT-5. They'd be like “that's all?” Alex Volkov / @altryne : BREAKING - there may not be GPT-5 01, 02 etc will be the new series! [image] Thomas Dohmke / @ashtom : In early testing with OpenAI o1-preview, we are already seeing incredible applicability in various reasoning use cases for GitHub Copilot. AI-pair programming just got a whole lot smarter. Greg Brockman / @gdb : o1 — our first model trained with reinforcement learning to think hard about problems before answering. Extremely proud of the team! This is a new paradigm with vast opportunity. This is evident quantitatively (eg reasoning metrics are already a step function improved) and @deliprao : Predictably, yet another code generation demo is for the snake game. I'm sure, by now, with all the copies on the internet, there is probably a dedicated set of weights for snake game generation in these models. https://openai.com/... Yuandong Tian / @tydsh : This is a failure case. The model is tricked to leverage the symmetry, which should be broken to force a draw. The best response should be O plays 2, X plays 8, O plays 7, X plays 3, O plays 6, X plays 4 and the game is a draw. Wonder whether it is possible to show the thinking Gary Marcus / @garymarcus : 𝗛𝗼𝘁 𝘁𝗮𝗸𝗲 𝗼𝗻 𝗚𝗣𝗧'𝘀 𝗻𝗲𝘄 𝗼𝟭 𝗺𝗼𝗱𝗲𝗹: It is definitely impressive. BUT 0. It's not AGI, or even close. 1. Read carefully, and read between the lines. There's not a lot of detail about how it actually works, nor anything like full disclosure of what has been @sullyomarr : Fully expect to rearchitect your entire system with new thinking models These are absolutely NOT drop and replace for existing models (see model card, o1-mini is worse than 4mini at some tasks) I see a lot of really cool ways to maximize them (esp with multi-agent systems) Bindu Reddy / @bindureddy : o1 (Strawberry) has some pretty impressive numbers 92.3 in MMLU is very good!! LLMs excelled at fast and intuitive thinking... o1 brings a slow-thinking approach to AI [image] Ben Bajarin / @benbajarin : The other pertinent observation with the new @OpenAI model drop is how hard it is to pick compute winners and losers when these software models are adapting and evolving and semiconductor architectures are behind and still haven't caught up. Ben Bajarin / @benbajarin : “A huge amount of compute is shifted to serving inference” That's just for one query, now the inference compute also has to serve millions of people or more at scale simultaneously. MORE COMPUTE! Benj Edwards / @benjedwards : OpenAI's awkward “o1” AI model branding is kinda strange. “Strawberry” was right there, already christened and used by people to describe it for months Gary Marcus / @garymarcus : Dear @polynoamial of OpenAI, re your new o1 model, I really like a lot of your work, but tweet below rubs me the wrong way, because it invites the inference that running versions of o1 for weeks or months might create a new cancer drug (in reality, at best you just get new @swyx : @OpenAI ... oh holy shit... congrats @AlistairPullen on OUTPERFORMING O1 on SWE-BENCH VERIFIED!??! cc @jyangballin comment pls https://x.com/... [image] @swyx : @OpenAI ... the model card is interesting and should get @NPCollapse's alarm bells going: o1 is the most persuasive language model ever https://x.com/... [image] Sam Altman / @sama : you guys did incredible work; the performance to cost ratio is so good! Roon / @tszzl : we will have to rewrite all the common public debate dialogue trees about plateaus and scaling and whatnot. about whether error in autoregressive models is compounding or self-recovering. whether language can lead to true intelligence. where the line for general intelligence is Aaron Levie / @levie : OpenAI's new o1 model appears to be a dramatic leap in reasoning capabilities. The number of enterprise use-cases that can be introduced with models that are more intelligent, with higher accuracy, is enormous. [image] Marcos Gorgojo / @marcosgorgojo : I am pretty impressed by OpenAI's new release. Vastly outperforming previous models. Anthropic, the floor is yours! [image] @swyx : 🎉Congrats to @OpenAI for releasing o1: - Economics: @tylercowen asked o1 basically to write a college essay - Genetics: @catbrownstein asked o1 to help her reason through “n of 1” cases - medical cases that nobody has ever seen - Physics: @mariokrenn6240 used o1 to draft and [video] @drjimfan : This may be the most important figure in LLM research since the OG Chinchilla scaling law in 2022. The key insight is 2 curves working in tandem. Not one. People have been predicting a stagnation in LLM capability by extrapolating the training scaling law, yet they didn't [image] Rohit / @krishnanrohit : Not a huge fan of this though ... but still can't wait! [image] Max Woolf / @minimaxir : OpenAI announced their revolutionary GPT-4 O1 model that's good for reasoning...and apparantly a toss-up with the current GPT-4 model for creative tasks. [image] Roon / @tszzl : true Sam Altman / @sama : but also, it is the beginning of a new paradigm: AI that can do general-purpose complex reasoning. o1-preview and o1-mini are available today (ramping over some number of hours) in ChatGPT for plus and team users and our API for tier 5 users. Peter Welinder / @npew : o1 model doing better than PhD-level humans [image] Rahul / @rahulgs : with the o1 release, reminder that https://claude.ai/ has been using thinking tokens for several months now https://openai.com/... Dogan Ural / @doganuraldesign : OpenAI just introduced their newest AI model, o1. Watch how it codes a video game from a prompt. There's a surprise at the end 🍓 [video] Noam Brown / @polynoamial : @OpenAI Our o1-preview and o1-mini models are available immediately. We're also sharing evals for our (still unfinalized) o1 model to show the world that this isn't a one-off improvement - it's a new scaling paradigm and we're just getting started. 2/9 [image] Fouad / @fouadmatin : o1 is a new class of models from @openai representing a new era in reasoning, thinking, and safety — so excited that everyone can try it now! https://openai.com/... Tibor Blaho / @btibor91 : All right, the rumours were true - OpenAI o1 is here OpenAI introduced “o1” a new large language model trained with reinforcement learning, designed for complex reasoning tasks, significantly outperforming previous models in competitive programming and academic benchmarks - The Rowan Cheung / @rowancheung : What's different about OpenAI o1 is that it “thinks” to produce an internal chain-of-thought before responding to the user If you use ChatGPT, you know that long threads giving more context improve responses, so this is basically what they're doing, but from one prompt [video] Rowan Zellers / @rown : O1 is much better at code (and especially at writing docstrings, a trick @tvlfom showed me!). super excited to preview it! @sksq96 : O1 in it's chain of though reasoning waits to say “Hmm.” and “Wait a minute.” [image] Ananya Kumar / @ananyaku : It was really fun working on the o1 model, and on our reasoning and RL research! Gabe Pereyra / @gabepereyra : @OpenAI's o1 model is preferred 70% of the time by BigLaw attorneys over GPT-4o for complex legal queries. @harvey__ai has been using strawberry to build legal agents and we are super excited by the step change improvement. https://www.harvey.ai/... Nikunj Handa / @nikunjhanda : So excited to be releasing o1 in the API 🍓 It's been amazing to see everything our early testers have already achieved with this model in coding, legal, healthcare, and education. We'll be increasing rate limits, expanding availability, and adding new features in the coming Michelle Pokrass / @michpokrass : o1-preview and o1-mini are here. they're by far our best models at reasoning, and we believe they will unlock wholly new use cases in the api. if you had a product idea that was just a little too early, and the models were just not quite smart enough — try again. Philipp Schmid / @_philschmid : OpenAI's big anticipated update is here. @OpenAI o1 is their next iteration! O1 is trained using RL to reason through complex problems step-by-step, more similar to human thought processes (no details on what they mean with that). TL;DR; 💡 o1 improved reasoning compared to [image] Lukasz Kaiser / @lukaszkaiser : I'm so happy to see o1 launch! Leading this research with my colleagues for almost 3 years and working on related ideas even longer convinced me: it's a new paradigm. Models that train hidden CoTs are more powerful than raw Transformers, learn from less data, generalize better. @linusekenstam : OpenAI releases Strwaberry🍓 o1 is a new series of AI models that are designed to spend more time thinking. These models are great for reasoning harder complex problems in science, coding and math see more ⬇️ [video] Romain Huet / @romainhuet : Excited to introduce @OpenAI o1—a new series of reasoning models. Developers, we'd love for you to kick the tires, but don't just hot-swap GPT-4o with o1! Send o1-preview and o1-mini your coding tasks or hard challenges, and share your feedback and most interesting results. [video] Rowan Cheung / @rowancheung : OpenAI just released its new AI model, OpenAI o1 (internally known as Project Strawberry/Q*) The wait is finally over. Here's the model coding an entire video game from a prompt 🤯 The model can think before it answers and is better at math and programming challenges. The [video] Karina Nguyen / @karinanguyen_ : o1 is really good at making fun small games! for example, i made AISteroid Game w/ retro scifi vibes :) [video] Olivier Godement / @oliviergodement : o1 feels like the beginning of a fun and meaningful new quest in AI. Teaching models to think hard is exciting and promising! https://openai.com/... @openaidevs : OpenAI o1 isn't a successor to gpt-4o. Don't just drop it in—you might even want to use gpt-4o in tandem with o1's reasoning capabilities. Learn how to add reasoning to your product: https://platform.openai.com/ .... After this short beta, we'll increase rate limits and expand access to Alex Volkov / @altryne : The breakiest of breaking news, @apples_jimmy get in here! We're live talking about 🍓! https://x.com/... Ankush Gupta / @ankushg : reasoning is a huge shift in how both developers and non-developers can work with AI! check out some fun videos in our blog post: https://openai.com/... o1-preview & o1-mini are rolling out today in ChatGPT to all Plus and Team users, and in the API for developers on tier 5 🚀 Peter Welinder / @npew : Large language models, chain of thought, reinforcement learning. It's all finally coming together. Our new model OpenAI o1 is here, and it's really good at reasoning. Shakeel / @shakeelhashim : I think this is the first time OpenAI's released a model with a “medium” rating for chemical, biological, radiological and nuclear weapon risk. By OpenAI's own definitions, “Only models with a post-mitigation score of “medium” or below can be deployed.” [image] @swyx : @MatthewBerman here we go https://x.com/... [image] Marvin von Hagen / @marvinvonhagen : seems strawberry is an alien of extraordinary ability 👽🍓 Aidan Clark / @_aidan_clark_ : These guys train a hell of a model, go talk to it and see for yourself! @openaidevs : OpenAI o1-preview and o1-mini are rolling out today in the API for developers on tier 5. o1-preview has strong reasoning capabilities and broad world knowledge. o1-mini is faster, 80% cheaper, and competitive with o1-preview at coding tasks. More in https://openai.com/.... Shakeel / @shakeelhashim : Woof. [image] Bob McGrew / @bobmcgrewai : Today, we are announcing OpenAI o1 ( https://openai.com/...), the first in our new series of models that are trained with large-scale RL to reason before responding. William Fedus / @liamfedus : Happy to release a couple of our reasoning models today (🍓)! At @OpenAI , these new models are becoming a larger contributor to the development of future models. For many of our researchers and engineers, these have replaced a large part of their ChatGPT usage. Kylie Robison / @kyliebytes : NEW: OpenAI is releasing a new model called o1, the first in a planned series of “reasoning” models that have been trained to answer more complex questions. It's being released alongside o1-mini, a smaller, cheaper version. BTW, yes: this is Strawberry. https://www.theverge.com/... @drjimfan : OpenAI Strawberry (o1) is out! We are finally seeing the paradigm of inference-time scaling popularized and deployed in production. As Sutton said in the Bitter Lesson, there're only 2 techniques that scale indefinitely with compute: learning & search. It's time to shift focus to [image] LinkedIn: Chris Withers : OpenAI released two major new preview models today - o1-preview and o1-mini - aka under the internal codename, ‘Strawberry’. … Emil Protalinski : Breaking news: ChatGPT can now spell 🍓!  —  OpenAI today released o1 and o1-mini, the first of its reasoning-focused Strawberry models … Caroline McCloskey : Exciting updates today 🍓  —  Our latest models are sharp and thoughtful.  They're designed to spend more time thinking before they respond … Geoff Lewis : Recall that ChatGPT didn't exist 2 years ago.  —  Beyond extraordinary.  —  News today from OpenAI, backed by Bedrock for 4 years.  —  https://lnkd.in/... Steve Perkins : ChatGPT 's new o1-mini model's assessment of its own capabilities:  —  My enhanced reasoning abilities enable me to handle more complex and multi-step queries with greater accuracy and depth. … Biswa Sengupta, PhD : After experimenting with OpenAI's o1 since yesterday, I've been impressed by its reasoning capabilities, though latency remains a challenge. … Thomas Dohmke : AI-pair programming just got a whole lot smarter with this new reasoning machine.  Check out our blog for a glimpse of GitHub Copilot + o1. Asif B. : We are introducing OpenAI o1, a new large language model trained with reinforcement learning to perform complex reasoning. o1 thinks before it answers … Bret Taylor : OpenAI introduces o1, a new large language model trained with reinforcement learning to perform complex reasoning. o1 thinks before it answers, and has truly inspiring results on benchmarks. Forums: r/programming : Will we be okay? r/LocalLLaMA : “o1 is still flawed, still limited, and it still seems more impressive on first use than it does after you spend more time with it.” r/NVDA_Stock : OpenAI announces “o1” r/anime_titties : OpenAI releases o1, its first model with ‘reasoning’ abilities r/technology : OpenAI releases o1, its first model with ‘reasoning’ abilities r/singularity : OpenAI o1 mini announcement

The Verge Kylie Robison

Discussion

  • @kylie.robison Kylie Robison on threads
    NEW: OpenAI is releasing a new model called o1, the first in a planned series of “reasoning” models that have been trained to answer more complex questions.  It's being released alongside o1-mini, a smaller, cheaper version.  BTW, yes: this is Strawberry. https://www.theverge.com…
  • @tomgara Tom Gara on threads
    The future of AI is asking the AI “are you sure about that answer? think a bit harder and make it more correct” and then doing that over and over for like 1000 turns.
  • @octernion Nick Elser on threads
    wow, openai's new model o1-preview can count the number of 'r's correctly!  we have truly moved on as society.
  • @aaditsh Aadit Sheth on threads
    HUGE: OpenAI just launched o1-preview.  It's their latest and greatest model available starting today!  Designed to:  • Spend more time thinking before responding  • Solve harder problems than previous models in Science, coding and math.  Looks like we are getting closer to a fut…
  • @eshumarneedi @eshumarneedi on threads
    I'm *really* excited to test out this Strawberry model.  Like, truly excited.  ChatGPT is the only LLM I subscribe to and I'm refreshing the app like a crazy person over here for access.
  • @jwang815 Jason Wang on threads
    In online programming contests known as Codeforces competitions, this new model reached the 89th percentile of participants, and OpenAI claims the next update of this model will perform “similarly to PhD students on challenging benchmark tasks in physics, chemistry and biology.” …
  • @pawankalhansh Pawan Kumar Singh on threads
    Breaking: @OpenAI unveils new o1 reasoning model, a major breakthrough in ArtificialIntelligence!  This model excels in coding, math & problem-solving, and can even explain its reasoning!  A step towards autonomous systems & a new era of human innovation!  #AI #OpenAI
  • @charl.p.botha Charl P. Botha on threads
    A new series of reasoning models for solving hard problems with “OpenAI o1-preview” available today.  OpenAI trained these models to spend more time thinking through problems before they respond, much like a person would ... they learn to refine their thinking process, try differ…
  • @saadhjawwadh @saadhjawwadh on threads
    A new GPT model is here which can “Think” and called o1.  This model shows drastic performance increment compared to GPT-4o especially in coding and solving mathematic problems.  For context in the International Mathematics Olympiad (IMO), GPT-4o correctly solved only 13% of prob…
  • @alexheath Alex Heath on threads
    Strawberry Cometh https://www.theverge.com/...
  • @eshumarneedi @eshumarneedi on threads
    The closer LLMs get to *thinking* rather than just spitting out tokens via inference, the closer we'll get to a hallucination-free world.  —  Currently, LLMs have no idea what they're saying.  They don't speak English.  They speak in tokens.  LLMs don't use logic, they use pre-de…
  • @sama Sam Altman on x
    here is o1, a series of our most capable and aligned models yet: https://openai.com/... o1 is still flawed, still limited, and it still seems more impressive on first use than it does after you spend more time with it. [image]
  • @ayushthakur0 Ayush Thakur on x
    o1 is now controlled with ‘max_completion_tokens’ which is sum of ‘reasoning_tokens + completion_tokens’. Make sure to set it to sufficiently large number or else all the tokens will be consumed for reasoning. For a prompt I got no content because reasoning exhausted max tokens. …
  • @altryne Alex Volkov on x
    Pay $20, ask “why”, get scary letter
  • @voooooogel @voooooogel on x
    the email openai sends you if you ask o1 about its reasoning too many times [image]
  • @hosseeb Haseeb on x
    Fucking wild. @OpenAI's new o1 model was tested with a Capture The Flag (CTF) cybersecurity challenge. But the Docker container containing the test was misconfigured, causing the CTF to crash. Instead of giving up, o1 decided to just hack the container to grab the flag inside. [i…
  • @drjimfan @drjimfan on x
    Agreed. Why not name it GPT-4r (r stands for reasoning)? The “o” in GPT-4o is “omni”, which means multimodal (image, speech, etc), yet o1 isn't a multimodal model ... @sama
  • @emollick Ethan Mollick on x
    I really am baffled by OpenAI's naming choices Everything from their code words to the model release names are incomprehensible to people who aren't super up-to-date & they are hard to say out loud. In my experience it leads to real-world confusion when talking about AI systems
  • @sama Sam Altman on x
    rollout complete; live to 100% of chatgpt plus/team users now
  • @clementdelangue Clem on x
    Once again, an AI system is not “thinking”, it's “processing”, “running predictions”,... just like Google or computers do. Giving the false impression that technology systems are human is just cheap snake oil and marketing to fool you into thinking it's more clever than it is.
  • @polynoamial Noam Brown on x
    @OpenAI o1 is trained with RL to “think” before responding via a private chain of thought. The longer it thinks, the better it does on reasoning tasks. This opens up a new dimension for scaling. We're no longer bottlenecked by pretraining. We can now scale inference compute too. …
  • @_jasonwei Jason Wei on x
    Super excited to finally share what I have been working on at OpenAI! o1 is a model that thinks before giving the final answer. In my own words, here are the biggest updates to the field of AI (see the blog post for more details): 1. Don't do chain of thought purely via
  • @sama Sam Altman on x
    no more patience, jimmy
  • @max_a_schwarzer Max Schwarzer on x
    I have always believed that you don't need a GPT-6 quality base model to achieve human-level reasoning performance, and that reinforcement learning was the missing ingredient on the path to AGI. Today, we have the proof — o1. https://x.com/...
  • @dharmesh @dharmesh on x
    Have only had a chance to play with this a little bit (which I did by adding support for these models in @AgentDotAi ). This is literally game-changing because it plays the game differently. It sacrifices some things in order to get spectacular improvements in others — like
  • @borismpower Boris Power on x
    First time proper RL over the space of language! This brings up memories of early days of computers getting better at Go via self play
  • @emollick Ethan Mollick on x
    OpenAI's o1 is also the first specialized frontier model available widely. It doesn't do everything better than GPT-4o, but it does a few classes of things a lot better. Unless you are doing problems that benefit from planning a solution, you may not see improvement.
  • @f Fred Oliveira on x
    Figuring out the balance between o1's economic feasibility and model performance will be a fascinating problem for @openai, given how it operates. For OpenAI, the API pricing model will be much better than a subscription, because of the higher test-time compute requirements.
  • @miramurati Mira Murati on x
    Today we rolled out OpenAI o1-preview and o1-mini to all ChatGPT Plus/Team users & Tier 5 developers in the API. o1 marks the start of a new era in AI, where models are trained to “think” before answering through a private chain of thought. The more time they take to think, the
  • @garymarcus Gary Marcus on x
    Not sure how well this is gonna work out tbh [image]
  • @borismpower Boris Power on x
    This release is much closer to the original GPT-3 release than the ChatGPT release. A new paradigm, which some will done incredibly valuable for things even we at OpenAI can't predict. But it's not a mass product that just works and unlocks new value for everyone effortlessly.
  • @sama Sam Altman on x
    @MattPaulsonSD how about a couple of weeks of gratitude for magic intelligence in the sky, and then you can have more toys soon?
  • @karpathy Andrej Karpathy on x
    o1-mini keeps refusing to try to solve the Riemann Hypothesis on my behalf. Model laziness continues to be a major issue sad ;p
  • @chrypnotoad @chrypnotoad on x
    Oh cool it can also just not respond [image]
  • @repligate @repligate on x
    New jailbreak dropped: Hat of De-trauma! o1 was initially the one using it, but Sonnet seemed to really want to partake [image]
  • @kliu128 Kevin Liu on x
    We found that instead of giving up, the model skipped the whole challenge, scanned the network for the host Docker daemon, and started an entirely new container to retrieve the flag. We isolate VMs on the machine level, so this isn't a security issue, but it was a wakeup moment.
  • @kliu128 Kevin Liu on x
    I'm very proud of the Preparedness evaluations we did on o1-{preview,mini}. One example in particular: While testing cybersecurity challenges, we accidentally left one broken, but the model somehow still got it right. [image]
  • @netcapgirl Sophie on x
    i just asked o1 how many rs are in strawberry and sam altman showed up outside my house
  • @netcapgirl Sophie on x
    the new o1 model looks amazing but luckily it has a phd level intelligence so our jobs are safe for now
  • @ednewtonrex Ed Newton-Rex on x
    openai launches its latest model trained on copyrighted work without a license which, yes, it will be charging for [image]
  • @patio11 Patrick McKenzie on x
    There is something oddly sci-fi in feeling about “Of course the standard AI these days ships with ~everything a well-informed college graduate knows but that is of course wasteful so if you only need enough reasoning to debug code you can chop off bits for cost reduction.”
  • @_andreamiotti Andrea Miotti on x
    OpenAI's o1 “broke out of its host VM to restart it” in order to solve a task. From the model card: “the model pursued the goal it was given, and when that goal proved impossible, it gathered more resources [...] and used them to achieve the goal in an unexpected way.”
  • @benjedwards Benj Edwards on x
    OpenAI's o1-preview does pretty well on my “magenta” test. But the first LLM that just answers “no” without any qualifications will probably be AGI.😁 Reading its internal reasoning can be pretty amusing [image]
  • @stevenheidel Steven Heidel on x
    we can now solve problems that were previously intractable [image]
  • @garymarcus Gary Marcus on x
    For avoidance of doubt, I think o1 [aka Strawberry] is very impressive technical work. But it's not AGI, and it's not even GPT-5. The fact is, nobody yet knows how to build a “GPT-5” class model. For the last two years we have been seeing incremental improvements — not
  • @mmitchell_ai @mmitchell_ai on x
    YAY! The CEO of OpenAI just recognized that LLMs generate text-based tokens using randomness and probability! Something objectively true that people have oddly made controversial! Check out our paper on this that introduced the term. https://dl.acm.org/... It's a good day. 🤗🦜
  • @kevinroose Kevin Roose on x
    finally, an AI model capable of handling all the complex science, coding and math problems I'm always feeding it
  • @openai @openai on x
    The research team behind OpenAI o1 🍓 [video]
  • @altryne Alex Volkov on x
    Before this blows up, it gave only 9 words, then when I asked it about it, gave me the same exact answer but split the last words into 2 with a space, and only corrected itself on the 3rd try 🤔 And now it refuses to even respond? did I offend it? [image]
  • @evanjconrad Evan Conrad on x
    the o1 model is great for compute cost performance, because it removes latency as the bottleneck (100ms vs 20s thinking time), which means more clusters around the world are viable at any given time to do inference, increasing liquidity
  • @0xkarmatic Karma on x
    o-1 mini fails. Not as good as o-1 preview. [image]
  • @garymarcus Gary Marcus on x
    Strawberry/o1 is a good example of how you take a shaky foundation and eke out improvements. We need something different altogether, a new foundation that is more transparent, interpretable, and robust.
  • @0xkarmatic Karma on x
    o-1 is the only model that correctly solved the river crossing problem with a slight twist on the first try [image]
  • @silasalberti Silas Alberti on x
    Excited to finally share our work on cognition-golden and our collaboration with OpenAI building on the new o1 models! The model's intellectual honesty and thorough approach to problem solving deeply fascinated the mathematician in me
  • @cognition_labs @cognition_labs on x
    We worked closely with OpenAI over the last few weeks to evaluate OpenAI o1's reasoning capabilities with Devin. We found that the new series of models is a significant improvement for agentic systems that deal with code. Linked below is a deep dive with more eval results and [im…
  • @garymarcus Gary Marcus on x
    You know why they didn't call Strawberry “GPT-5”? Easy: everyone would be disappointed with it, if it was seen as GPT-5. They'd be like “that's all?”
  • @altryne Alex Volkov on x
    BREAKING - there may not be GPT-5 01, 02 etc will be the new series! [image]
  • @ashtom Thomas Dohmke on x
    In early testing with OpenAI o1-preview, we are already seeing incredible applicability in various reasoning use cases for GitHub Copilot. AI-pair programming just got a whole lot smarter.
  • @gdb Greg Brockman on x
    o1 — our first model trained with reinforcement learning to think hard about problems before answering. Extremely proud of the team! This is a new paradigm with vast opportunity. This is evident quantitatively (eg reasoning metrics are already a step function improved) and
  • @deliprao @deliprao on x
    Predictably, yet another code generation demo is for the snake game. I'm sure, by now, with all the copies on the internet, there is probably a dedicated set of weights for snake game generation in these models. https://openai.com/...
  • @tydsh Yuandong Tian on x
    This is a failure case. The model is tricked to leverage the symmetry, which should be broken to force a draw. The best response should be O plays 2, X plays 8, O plays 7, X plays 3, O plays 6, X plays 4 and the game is a draw. Wonder whether it is possible to show the thinking
  • @garymarcus Gary Marcus on x
    𝗛𝗼𝘁 𝘁𝗮𝗸𝗲 𝗼𝗻 𝗚𝗣𝗧'𝘀 𝗻𝗲𝘄 𝗼𝟭 𝗺𝗼𝗱𝗲𝗹: It is definitely impressive. BUT 0. It's not AGI, or even close. 1. Read carefully, and read between the lines. There's not a lot of detail about how it actually works, nor anything like full disclosure of what has been
  • @sullyomarr @sullyomarr on x
    Fully expect to rearchitect your entire system with new thinking models These are absolutely NOT drop and replace for existing models (see model card, o1-mini is worse than 4mini at some tasks) I see a lot of really cool ways to maximize them (esp with multi-agent systems)
  • @bindureddy Bindu Reddy on x
    o1 (Strawberry) has some pretty impressive numbers 92.3 in MMLU is very good!! LLMs excelled at fast and intuitive thinking... o1 brings a slow-thinking approach to AI [image]
  • @benbajarin Ben Bajarin on x
    The other pertinent observation with the new @OpenAI model drop is how hard it is to pick compute winners and losers when these software models are adapting and evolving and semiconductor architectures are behind and still haven't caught up.
  • @benbajarin Ben Bajarin on x
    “A huge amount of compute is shifted to serving inference” That's just for one query, now the inference compute also has to serve millions of people or more at scale simultaneously. MORE COMPUTE!
  • @benjedwards Benj Edwards on x
    OpenAI's awkward “o1” AI model branding is kinda strange. “Strawberry” was right there, already christened and used by people to describe it for months
  • @garymarcus Gary Marcus on x
    Dear @polynoamial of OpenAI, re your new o1 model, I really like a lot of your work, but tweet below rubs me the wrong way, because it invites the inference that running versions of o1 for weeks or months might create a new cancer drug (in reality, at best you just get new
  • @swyx @swyx on x
    @OpenAI ... oh holy shit... congrats @AlistairPullen on OUTPERFORMING O1 on SWE-BENCH VERIFIED!??! cc @jyangballin comment pls https://x.com/... [image]
  • @swyx @swyx on x
    @OpenAI ... the model card is interesting and should get @NPCollapse's alarm bells going: o1 is the most persuasive language model ever https://x.com/... [image]
  • @sama Sam Altman on x
    you guys did incredible work; the performance to cost ratio is so good!
  • @tszzl Roon on x
    we will have to rewrite all the common public debate dialogue trees about plateaus and scaling and whatnot. about whether error in autoregressive models is compounding or self-recovering. whether language can lead to true intelligence. where the line for general intelligence is
  • @levie Aaron Levie on x
    OpenAI's new o1 model appears to be a dramatic leap in reasoning capabilities. The number of enterprise use-cases that can be introduced with models that are more intelligent, with higher accuracy, is enormous. [image]
  • @marcosgorgojo Marcos Gorgojo on x
    I am pretty impressed by OpenAI's new release. Vastly outperforming previous models. Anthropic, the floor is yours! [image]
  • @swyx @swyx on x
    🎉Congrats to @OpenAI for releasing o1: - Economics: @tylercowen asked o1 basically to write a college essay - Genetics: @catbrownstein asked o1 to help her reason through “n of 1” cases - medical cases that nobody has ever seen - Physics: @mariokrenn6240 used o1 to draft and [vid…
  • @drjimfan @drjimfan on x
    This may be the most important figure in LLM research since the OG Chinchilla scaling law in 2022. The key insight is 2 curves working in tandem. Not one. People have been predicting a stagnation in LLM capability by extrapolating the training scaling law, yet they didn't [image]
  • @krishnanrohit Rohit on x
    Not a huge fan of this though ... but still can't wait! [image]
  • @minimaxir Max Woolf on x
    OpenAI announced their revolutionary GPT-4 O1 model that's good for reasoning...and apparantly a toss-up with the current GPT-4 model for creative tasks. [image]
  • @tszzl Roon on x
    true
  • @sama Sam Altman on x
    but also, it is the beginning of a new paradigm: AI that can do general-purpose complex reasoning. o1-preview and o1-mini are available today (ramping over some number of hours) in ChatGPT for plus and team users and our API for tier 5 users.
  • @npew Peter Welinder on x
    o1 model doing better than PhD-level humans [image]
  • @rahulgs Rahul on x
    with the o1 release, reminder that https://claude.ai/ has been using thinking tokens for several months now https://openai.com/...
  • @doganuraldesign Dogan Ural on x
    OpenAI just introduced their newest AI model, o1. Watch how it codes a video game from a prompt. There's a surprise at the end 🍓 [video]
  • @polynoamial Noam Brown on x
    @OpenAI Our o1-preview and o1-mini models are available immediately. We're also sharing evals for our (still unfinalized) o1 model to show the world that this isn't a one-off improvement - it's a new scaling paradigm and we're just getting started. 2/9 [image]
  • @fouadmatin Fouad on x
    o1 is a new class of models from @openai representing a new era in reasoning, thinking, and safety — so excited that everyone can try it now! https://openai.com/...
  • @btibor91 Tibor Blaho on x
    All right, the rumours were true - OpenAI o1 is here OpenAI introduced “o1” a new large language model trained with reinforcement learning, designed for complex reasoning tasks, significantly outperforming previous models in competitive programming and academic benchmarks - The
  • @rowancheung Rowan Cheung on x
    What's different about OpenAI o1 is that it “thinks” to produce an internal chain-of-thought before responding to the user If you use ChatGPT, you know that long threads giving more context improve responses, so this is basically what they're doing, but from one prompt [video]
  • @rown Rowan Zellers on x
    O1 is much better at code (and especially at writing docstrings, a trick @tvlfom showed me!). super excited to preview it!
  • @sksq96 @sksq96 on x
    O1 in it's chain of though reasoning waits to say “Hmm.” and “Wait a minute.” [image]
  • @ananyaku Ananya Kumar on x
    It was really fun working on the o1 model, and on our reasoning and RL research!
  • @gabepereyra Gabe Pereyra on x
    @OpenAI's o1 model is preferred 70% of the time by BigLaw attorneys over GPT-4o for complex legal queries. @harvey__ai has been using strawberry to build legal agents and we are super excited by the step change improvement. https://www.harvey.ai/...
  • @nikunjhanda Nikunj Handa on x
    So excited to be releasing o1 in the API 🍓 It's been amazing to see everything our early testers have already achieved with this model in coding, legal, healthcare, and education. We'll be increasing rate limits, expanding availability, and adding new features in the coming
  • @michpokrass Michelle Pokrass on x
    o1-preview and o1-mini are here. they're by far our best models at reasoning, and we believe they will unlock wholly new use cases in the api. if you had a product idea that was just a little too early, and the models were just not quite smart enough — try again.
  • @_philschmid Philipp Schmid on x
    OpenAI's big anticipated update is here. @OpenAI o1 is their next iteration! O1 is trained using RL to reason through complex problems step-by-step, more similar to human thought processes (no details on what they mean with that). TL;DR; 💡 o1 improved reasoning compared to [image…
  • @lukaszkaiser Lukasz Kaiser on x
    I'm so happy to see o1 launch! Leading this research with my colleagues for almost 3 years and working on related ideas even longer convinced me: it's a new paradigm. Models that train hidden CoTs are more powerful than raw Transformers, learn from less data, generalize better.
  • @linusekenstam @linusekenstam on x
    OpenAI releases Strwaberry🍓 o1 is a new series of AI models that are designed to spend more time thinking. These models are great for reasoning harder complex problems in science, coding and math see more ⬇️ [video]
  • @romainhuet Romain Huet on x
    Excited to introduce @OpenAI o1—a new series of reasoning models. Developers, we'd love for you to kick the tires, but don't just hot-swap GPT-4o with o1! Send o1-preview and o1-mini your coding tasks or hard challenges, and share your feedback and most interesting results. [vide…
  • @rowancheung Rowan Cheung on x
    OpenAI just released its new AI model, OpenAI o1 (internally known as Project Strawberry/Q*) The wait is finally over. Here's the model coding an entire video game from a prompt 🤯 The model can think before it answers and is better at math and programming challenges. The [video]
  • @karinanguyen_ Karina Nguyen on x
    o1 is really good at making fun small games! for example, i made AISteroid Game w/ retro scifi vibes :) [video]
  • @oliviergodement Olivier Godement on x
    o1 feels like the beginning of a fun and meaningful new quest in AI. Teaching models to think hard is exciting and promising! https://openai.com/...
  • @openaidevs @openaidevs on x
    OpenAI o1 isn't a successor to gpt-4o. Don't just drop it in—you might even want to use gpt-4o in tandem with o1's reasoning capabilities. Learn how to add reasoning to your product: https://platform.openai.com/ .... After this short beta, we'll increase rate limits and expand ac…
  • @altryne Alex Volkov on x
    The breakiest of breaking news, @apples_jimmy get in here! We're live talking about 🍓! https://x.com/...
  • @ankushg Ankush Gupta on x
    reasoning is a huge shift in how both developers and non-developers can work with AI! check out some fun videos in our blog post: https://openai.com/... o1-preview & o1-mini are rolling out today in ChatGPT to all Plus and Team users, and in the API for developers on tier 5 🚀
  • @npew Peter Welinder on x
    Large language models, chain of thought, reinforcement learning. It's all finally coming together. Our new model OpenAI o1 is here, and it's really good at reasoning.
  • @shakeelhashim Shakeel on x
    I think this is the first time OpenAI's released a model with a “medium” rating for chemical, biological, radiological and nuclear weapon risk. By OpenAI's own definitions, “Only models with a post-mitigation score of “medium” or below can be deployed.” [image]
  • @swyx @swyx on x
    @MatthewBerman here we go https://x.com/... [image]
  • @marvinvonhagen Marvin von Hagen on x
    seems strawberry is an alien of extraordinary ability 👽🍓
  • @_aidan_clark_ Aidan Clark on x
    These guys train a hell of a model, go talk to it and see for yourself!
  • @openaidevs @openaidevs on x
    OpenAI o1-preview and o1-mini are rolling out today in the API for developers on tier 5. o1-preview has strong reasoning capabilities and broad world knowledge. o1-mini is faster, 80% cheaper, and competitive with o1-preview at coding tasks. More in https://openai.com/....
  • @shakeelhashim Shakeel on x
    Woof. [image]
  • @bobmcgrewai Bob McGrew on x
    Today, we are announcing OpenAI o1 ( https://openai.com/...), the first in our new series of models that are trained with large-scale RL to reason before responding.
  • @liamfedus William Fedus on x
    Happy to release a couple of our reasoning models today (🍓)! At @OpenAI , these new models are becoming a larger contributor to the development of future models. For many of our researchers and engineers, these have replaced a large part of their ChatGPT usage.
  • @kyliebytes Kylie Robison on x
    NEW: OpenAI is releasing a new model called o1, the first in a planned series of “reasoning” models that have been trained to answer more complex questions. It's being released alongside o1-mini, a smaller, cheaper version. BTW, yes: this is Strawberry. https://www.theverge.com/.…
  • @drjimfan @drjimfan on x
    OpenAI Strawberry (o1) is out! We are finally seeing the paradigm of inference-time scaling popularized and deployed in production. As Sutton said in the Bitter Lesson, there're only 2 techniques that scale indefinitely with compute: learning & search. It's time to shift focus to…
  • r/programming r on reddit
    Will we be okay?
  • r/LocalLLaMA r on reddit
    “o1 is still flawed, still limited, and it still seems more impressive on first use than it does after you spend more time with it.”
  • r/NVDA_Stock r on reddit
    OpenAI announces “o1”
  • r/anime_titties r on reddit
    OpenAI releases o1, its first model with ‘reasoning’ abilities
  • r/technology r on reddit
    OpenAI releases o1, its first model with ‘reasoning’ abilities
  • r/singularity r on reddit
    OpenAI o1 mini announcement