Nvidia announces the Nvidia Groq 3 LPX, an inference server rack featuring 256 Groq 3 LPUs, 128GB of SRAM, and 40 PBps SRAM bandwidth, available in H2 2026
Nvidia announced Monday at GTC 2026 that its new Groq-based inference server rack will be available alongside the Vera Rubin NVL72 rack …
Nvidia announces the Nvidia Groq 3 LPX, an inference server rack featuring 256 Groq 3 LPUs and 128GB of on-chip SRAM, available in H2 2026
Nvidia announced Monday at GTC 2026 that its new Groq-based inference server rack will be available alongside the Vera Rubin NVL72 rack …
OpenAI releases gpt-oss-120b and gpt-oss-20b, its first open-weight models since GPT-2; the smaller gpt-oss-20b can run locally on a device with 16GB+ of RAM
gpt-oss-120b and gpt-oss-20b push the frontier of open-weight reasoning models Simon Willison / Simon Willison's Weblog : OpenAI's new open weight (Apache 2) models are really good OpenAI on GitHub : ...
A look at efforts by startups, such as Positron and Groq, to develop inference-focused chips that aim to be more energy efficient and performant than Nvidia's
Christopher Mims / Wall Street Journal :
Meta says its Llama models were downloaded almost 350M times, are used by AT&T and others, and usage via cloud providers more than doubled from May to July 2024
we just published a bunch of updates on the adoption we're seeing. And yes, we have a lot more work to do on dev tools and resources which we're bringing online as quickly as we can. https://ai.meta....
An interview with Groq's CEO about its AI chips that let chatbots answer queries almost instantly, its cease and desist to X.ai over Groq's trademark, and more
Steven Levy / Wired : X: @groqinc , @evankirstel , @stevenlevy , @glynmoody , and @nobbelty X: @groqinc : With accelerated growth at @GroqInc we're excited to announce the acquisition of @DefinitiveI...
An interview with Groq's CEO about its AI chips that let chatbots answer queries almost instantly, its cease and desist to X.ai over Groq's trademark, and more
AI chips from startup Groq allow chatbots to answer queries almost instantly. That could open up whole new use cases for generative AI helpers.
Demos from AI chipmaker Groq go viral after the startup's inference engine shows lightning-fast speeds when running LLMs, including for real-time conversations
Two AI companies are claiming the science fiction term, “Grok,” as their own, but only one is turbocharging the AI industry.