/
Navigation
C
Chronicles
Browse all articles
C
E
Explore
Semantic exploration
E
R
Research
Entity momentum
R
N
Nexus
Correlations & relationships
N
~
Story Arc
Topic evolution
S
Drift Map
Semantic trajectory animation
D
P
Posts
Analysis & commentary
P
Browse
@
Entities
Companies, people, products, technologies
Domains
Browse by publication source
Handles
Browse by social media handle
Detection
?
Concept Search
Semantic similarity search
!
High Impact Stories
Top coverage by position
+
Sentiment Analysis
Positive/negative coverage
*
Anomaly Detection
Unusual coverage patterns
Analysis
vs
Rivalry Report
Compare two entities head-to-head
/\
Semantic Pivots
Narrative discontinuities
!!
Crisis Response
Event recovery patterns
Connected
Nav: C E R N
Search: /
Command: ⌘K
Embeddings: large
VOICE ARCHIVE

@cerebrassystems

@cerebrassystems
12 posts
2025-10-07
On Friday, Cerebras withdrew our S-1. It had become stale and no longer reflected the current state of our business. Our business and financial position have evolved significantly for the better since our initial filing in 2024: • In 2024, we achieved record revenues. • In
2025-10-07 View on X
CNBC

Cerebras CEO Andrew Feldman says the AI chipmaker pulled its IPO filing to share updated financials, admitting it was a mistake not to explain the move sooner

So I want to give you the full picture. …

2025-08-02
Cerebras Code: 20x faster than Claude, 1x the price Today we are launching two monthly coding plans: ➡️Cerebras Code Pro: $50/m - for indie developers ➡️Cerebras Code Max: $200/m - for power users with 5x rate limits Both plans get: Qwen3-Coder at 2,000 tokens/s, 131K context, [video]
2025-08-02 View on X
Cerebras

Cerebras announces the $50/month Code Pro and the $200/month Code Max plans, offering users access to Qwen3-Coder at speeds of up to 2,000 tokens per second

Two interesting examples of inference speed as a flagship feature of LLM services today. Bluesky: Tim Kellogg / @timkellogg.me : Cerebras Code — use models hosted on Cerebras with ...

2024-08-28
Verified by @ArtificialAnlys, Cerebras Inference achieves 1,850 tokens/sec on Llama 3.1 8B and 450 tokens/sec on Llama 3.1 70B! By dramatically reducing processing time, we're enabling more complex AI workflows and enhancing real-time LLM intelligence. This includes a new class [image]
2024-08-28 View on X
SiliconANGLE

Cerebras launches the “world's fastest” AI inference service with “GPU-impossible performance”, with costs starting at $0.10 per million tokens, to rival Nvidia

Mike Wheatley / SiliconANGLE :

Cerebras Inference is the fastest Llama3.1 inference API by far: 1,800 tokens/s for 8B and 450tokens/s for 70B. We are ~20x faster than NVIDA GPUs and ~2x faster than Groq. [image]
2024-08-28 View on X
SiliconANGLE

Cerebras launches the “world's fastest” AI inference service with “GPU-impossible performance”, with costs starting at $0.10 per million tokens, to rival Nvidia

Mike Wheatley / SiliconANGLE :

Lastly, we didn't just build a fast demo - we have capacity to serve hundreds of billions of tokens per day to developers and enterprises. We will be adding new models (eg. Llama3.1-405B) and ramp even greater capacity in the coming months. [image]
2024-08-28 View on X
SiliconANGLE

Cerebras launches the “world's fastest” AI inference service with “GPU-impossible performance”, with costs starting at $0.10 per million tokens, to rival Nvidia

Mike Wheatley / SiliconANGLE :

Cerebras Inference is just 10c per million tokens for 8B and 60c per million tokens for 70B. Our price-performance is so strong, we practically broke the chart on Artificial Analysis. [image]
2024-08-28 View on X
SiliconANGLE

Cerebras launches the “world's fastest” AI inference service with “GPU-impossible performance”, with costs starting at $0.10 per million tokens, to rival Nvidia

Mike Wheatley / SiliconANGLE :

Introducing Cerebras Inference ‣ Llama3.1-70B at 450 tokens/s - 20x faster than GPUs ‣ 60c per M tokens - a fifth the price of hyperscalers ‣ Full 16-bit precision for full model accuracy ‣ Generous rate limits for devs Try now: https://inference.cerebras.ai/ [video]
2024-08-28 View on X
SiliconANGLE

Cerebras launches the “world's fastest” AI inference service with “GPU-impossible performance”, with costs starting at $0.10 per million tokens, to rival Nvidia

Mike Wheatley / SiliconANGLE :

2023-03-29
🎉 Exciting news! Today we are releasing Cerebras-GPT, a family of 7 GPT models from 111M to 13B parameters trained using the Chinchilla formula. These are the highest accuracy models for a compute budget and are available today open-source! (1/5) Press: https://www.businesswire.com/ ...
2023-03-29 View on X
SiliconANGLE

Cerebras open sources seven GPT-based LLMs, ranging from 111M to 13B parameters and trained using its Andromeda supercomputer for AI, on GitHub and Hugging Face

Artificial intelligence chipmaker Cerebras Systems Inc. today announced it has trained and now released seven GPT-based large language models …

One notable output of Cerebras-GPT is a new scaling law that predicts model performance for a given compute budget. This is the first scaling law derived using a public dataset. (3/5) https://twitter.com/...
2023-03-29 View on X
SiliconANGLE

Cerebras open sources seven GPT-based LLMs, ranging from 111M to 13B parameters and trained using its Andromeda supercomputer for AI, on GitHub and Hugging Face

Artificial intelligence chipmaker Cerebras Systems Inc. today announced it has trained and now released seven GPT-based large language models …

2021-08-22
.@NewYorker looks back at our history & how fearless engineering enabled us to overcome the impossible, creating the world's first wafer-scale engine. It powers our leading CS-2 system - addressing the most important workload of this decade. Read here: https://www.newyorker.com/... #AI
2021-08-22 View on X
New Yorker

Profile of Cerebras, which made the world's largest chip by using a “wafer-scale” approach that offers one possibility for AI chips to keep up with Moore's law

In the race to accelerate A.I., the Silicon Valley company Cerebras has landed on an unusual strategy: go big.

2021-08-21
.@NewYorker looks back at our history & how fearless engineering enabled us to overcome the impossible, creating the world's first wafer-scale engine. It powers our leading CS-2 system - addressing the most important workload of this decade. Read here: https://www.newyorker.com/... #AI
2021-08-21 View on X
New Yorker

Profile of Cerebras, which made the world's largest chip by using a “wafer-scale” approach that offers one possibility for AI chips to keep up with Moore's law

In the race to accelerate A.I., the Silicon Valley company Cerebras has landed on an unusual strategy: go big.

2019-08-20
Thanks @DannyCrichton and @TechCrunch for this article on the many engineering challenges we overcame on the way to #waferscale integration-a first in the industry! https://techcrunch.com/...
2019-08-20 View on X
Fortune

Cerebras Systems unveils the world's biggest semiconductor chip that is the size of a large mousepad, with 400K cores, 1.2T transistors, and 18GB of SRAM memory

In the late 1970s, I sat down with technologist … Colm Gorey / Silicon Republic : World's largest semiconductor with 1.2trn transistors could supercharge AI Andy Hock / Cerebras : ...