/
Navigation
Chronicles
Browse all articles
Explore
Semantic exploration
Research
Entity momentum
Nexus
Correlations & relationships
Story Arc
Topic evolution
Drift Map
Semantic trajectory animation
Posts
Analysis & commentary
Pulse API
Tech news intelligence API
Browse
Entities
Companies, people, products, technologies
Domains
Browse by publication source
Handles
Browse by social media handle
Detection
Concept Search
Semantic similarity search
High Impact Stories
Top coverage by position
Sentiment Analysis
Positive/negative coverage
Anomaly Detection
Unusual coverage patterns
Analysis
Rivalry Report
Compare two entities head-to-head
Semantic Pivots
Narrative discontinuities
Crisis Response
Event recovery patterns
Connected
Search: /
Command: ⌘K
Embeddings: large
TEXXR

Chronicles

The story behind the story

days · browse · Enter similar · o open

OpenAI debuts a research preview of GPT-5.3-Codex-Spark, a smaller version of GPT-5.3-Codex that it claims generates code 15 times faster, for ChatGPT Pro users

ZDNET's key takeaways  — OpenAI targets “conversational” coding, not slow batch-style agents.  — Big latency wins: 80% faster roundtrip, 50% faster time-to-first-token.

ZDNET David Gewirtz

Discussion

  • @sama Sam Altman on x
    GPT-5.3-Codex-Spark is launching today as a research preview for Pro. More than 1000 tokens per second! There are limitations at launch; we will rapidly improve.
  • @_simonsmith Simon Smith on x
    We've seen how much speed affects people's model preferences recently (e.g. the arena @swyx is running), so I think Codex Spark will be well-received. Also interesting that this initial release is a step towards combining long-horizon and real-time agents, including delegating to…
  • @derrickcchoi Derrick Choi on x
    One of the top pieces of feedback we get about @OpenAI Codex: “make it faster”. We addressed it in a big way with our ultra fast Codex-Spark model (research preview for Pro). Available in the latest Codex app, CLI, and IDE extension. Here's Spark vs 5.3-Codex-Low side-by-side [vi…
  • @romainhuet Romain Huet on x
    Hello GPT-5.3-Codex-Spark! ✨ Our first real-time coding model. It is... FAST. 1,000+ tokens per second. Once you experience latency this low, it's hard to go back. This is an exciting first milestone in our partnership with @Cerebras. [video]
  • @kevinweil Kevin Weil on x
    Coding at 1000 tokens/sec is a mind-expanding experience. You have to try this.
  • @openai @openai on x
    Rolling out today to ChatGPT Pro users in the Codex app, CLI, and IDE extension. https://openai.com/...
  • @skirano Pietro Schirano on x
    Been using this model for a bit now, the combination of speed and intelligence is insane. It genuinely feels like a new paradigm shift. Excited to plug it into more specialized coding pipelines.
  • @danshipper Dan Shipper on x
    BREAKING: @OpenAI just launched a new Codex model, Spark—it serves at 1,000 tokens per second.  It's blow your hair back fast.  It's their first model publicly released on Cerebras hardware, and you can see the difference.  We've been testing internally @every for the last week o…
  • @scaling01 @scaling01 on x
    GPT-5.3-Codex-Spark size: ~700B@30B OpenAI's new GPT-5.3-Codex-Spark is the first model for which we can somewhat reliably estimate its size... GPT-5.3-Codex-Spark gets “over 1000 tokens/s”, so probably 1000-1100 tokens/s... Based on that GPT-5.3-Codex-Spark should be: ~30B activ…
  • @heccbrent Brent Schooley on x
    This was so fun to work on. GPT-5.3-Codex-Spark built a snake game so fast that I was able to start setting high scores in about 9 seconds. If you have ChatGPT Pro you're going to want to check this out today! [video]
  • @embirico Alexander Embiricos on x
    ✨GPT-5.3-Codex-Spark✨ We're rolling out our first @cerebras model to Pro users today. It's fast! Rollout will be slow and very capacity constrained. Excited to roll out to more folks, and improve it with your feedback.
  • @openai @openai on x
    GPT-5.3-Codex-Spark is now in research preview. You can just build things—faster. [video]
  • @andrewdfeldman Andrew Feldman on x
    Just one month after announcing our partnership with @OpenAI, we're launching our first model together: OpenAI Codex-Spark, powered by @cerebras. Codex-Spark is built for real-time software development. In coding, responsiveness is the product. It is not a nice to have. [image]
  • @mweinbach Max Weinbach on x
    GPT 5.3 Codex Spark! It's a smaller version of GPT 5.3 Codex running at over 1000 tokens per second on Cerebras hardware GPT 5.3 Codex was trained for GB200, on GB200. I wonder what this is? Maybe GPT 5.3 Codex that used Cerebras Reap?
  • @dimillian Thomas Ricouard on x
    I've been playing with GPT-5.3-Codex-Spark this week, and it's a really a✨ experience. Basically, using it for smaller tasks, context scanning, quick analysis, and smaller code edits. It feels so natural and instant; it's really hard to go back to other models.
  • @eliebakouch Elie on x
    ok this is very interesting, this is not the same perf than gpt5.3, and might not be the same arch as well? > Codex-Spark marks the first milestone in our partnership with Cerebras. Codex-Spark is optimized to feel near-instant when served on ultra-low latency hardware (from [ima…
  • @sungkim Sung Kim on bluesky
    OpenAI has released GPT-5.3-Codex-Spark: 1000 tokens per second  —  openai.com/index/introd...  [embedded post]
  • r/singularity r on reddit
    Introducing GPT-5.3-Codex-Spark.  An ultra-fast model for real-time coding in Codex
  • r/OpenAI r on reddit
    Introducing GPT-5.3-Codex-Spark
  • @kylebrussell Kyle Russell on x
    I thought this was going to come like next year, not now
  • @openaidevs @openaidevs on x
    GPT-5.3-Codex-Spark is the first milestone in our partnership with @cerebras. It provides a faster tier on the same production stack as our other models, complementing GPUs for workloads where low latency is critical. https://openai.com/...
  • @mweinbach Max Weinbach on x
    Codex Spark was trained on GPUs for Cerebras hardware but OpenAI added support to their inference framework for Cerebras meaning they're reading to load future models onto it too GPUs are still foundational for inference and training, though [image]
  • @cerebras @cerebras on x
    OpenAI Codex-Spark powered by Cerebras You can now just build things faster—at 1,000 tokens/s. [video]
  • @openaidevs @openaidevs on x
    Introducing GPT-5.3-Codex-Spark, our ultra-fast model purpose built for real-time coding. We're rolling it out as a research preview for ChatGPT Pro users in the Codex app, Codex CLI, and IDE extension. [video]
  • @benbajarin Ben Bajarin on x
    As the world moves to inference, dedicated inference designs will be prominant. Great customer case for @cerebras