OpenAI says GPT-5.3-Codex-Spark is its first AI model that runs on Cerebras chips, after they signed a $10B+ deal in January; Codex has 1M+ weekly active users
at 1,000 tokens/s. [video]@openaidevs:Introducing GPT-5.3-Codex-Spark, our ultra-fast model purpose built for real-time coding. We're rolling it out as a research preview for ChatGPT Pro users in the Codex app, Codex CLI, and IDE extension. [video]Ben Bajarin /@benbajarin:As the world moves to inference, dedicated inference designs will be prominant. Great customer case for @cerebras
GPT-5.3-Codex-Spark is the first milestone in our partnership with @cerebras. It provides a faster tier on the same production stack as our other models, complementing GPUs for workloads where low latency is critical. https://openai.com/...
Codex Spark was trained on GPUs for Cerebras hardware but OpenAI added support to their inference framework for Cerebras meaning they're reading to load future models onto it too GPUs are still foundational for inference and training, though [image]
Introducing GPT-5.3-Codex-Spark, our ultra-fast model purpose built for real-time coding. We're rolling it out as a research preview for ChatGPT Pro users in the Codex app, Codex CLI, and IDE extension. [video]