DeepSeek releases its new flagship models V4 Pro and V4 Flash in preview, saying V4 Pro trails the performance of state-of-the-art models by about 3 to 6 months
DeepSeek rolled out preview versions of a new flagship artificial intelligence model a year after upending Silicon Valley …
Bloomberg
Related Coverage
- DeepSeek-V4: Towards Highly Efficient Million-Token Context Intelligence DeepSeek on Hugging Face
- China's DeepSeek rolls out a long-anticipated update of its AI model Associated Press
- China's AI upstart DeepSeek drops new model. Will it make waves like last year? CNN · John Liu
- China's DeepSeek Launches Long-Awaited AI Model Wall Street Journal · Tracy Qu
- China's DeepSeek releases long-awaited new AI model RTÉ
- China's DeepSeek unveils V4 AI model in fresh challenge to US rivals Nikkei Asia · Cissy Zhou
- DeepSeek is back: China's AI claims to surpass ChatGPT and Gemini in key benchmarks Livemint · Aman Gupta
- DeepSeek Launches New-Generation V4 Models The Information · Juro Osawa
- Nvidia Chips Are Losing Out in One Key Market. DeepSeek Lays Bare a Missed Opportunity. Barron's Online · Adam Clark
- DeepSeek returns with V4-Pro and V4-Flash, a year after its ‘Sputnik moment’ The Next Web · Ana Maria Constantin
- US-China AI race intensifies as DeepSeek releases ‘reduced’ cost model France 24
- China's DeepSeek unveils latest models a year after upending global tech Al Jazeera · John Power
- China's DeepSeek releases preview of long-awaited V4 model as AI race intensifies CNBC · Dylan Butts
- DeepSeek's Sequel Set to Extend China's Reach in Open-Source A.I. New York Times
- China's DeepSeek returns with new model, a year after viral rise Reuters
- China's DeepSeek releases new AI model it claims beats all open-source competitors The Independent · Vishwam Sankaran
- China's DeepSeek releases long-awaited new AI model Hong Kong Free Press HKFP
- DeepSeek v4 Hacker News
- DeepSeek V4 goes live with massive 1.6T parameters and 1M context support Neowin · Paul Hill
- DeepSeek-V4: Towards Highly Efficient Million-Token Context Intelligence Hacker News
- China's DeepSeek unveils long-awaited V4 AI model Silicon Republic · Suhasini Srinivasaragavan
- China's DeepSeek makes a comeback with V4 model; here's all you need to know Fortune India · Tamanna Mohanty
- China's DeepSeek V4 AI model released with 1M-token context window Crypto Briefing · Estefano Gomez
- DeepSeek Is Turning China's Chip Problem Into a Compute-Shaping Bet Hello China Tech · Poe Zhao
- As agentic AI pushes rivals to raise prices and cap usage, Deepseek ships a good-enough model for almost nothing The Decoder · Maximilian Schreiner
- Trump ally visits China, DeepSeek releases V4, EV earnings South China Morning Post
- Investors rotate to China's chipmakers as DeepSeek intensifies AI competition South China Morning Post · Yulu Ao
- China's DeepSeek previews new AI model a year after jolting US rivals The Verge · Robert Hart
- DeepSeek Unveils V4: The Latest Open-Source AI Model Challenging Big Tech Giants Blockonomi · Trader Edge
- After eyeing $20B in valuation, DeepSeek launches its long-awaited model: What we know so far Tech Funding News · Abhinaya Prabhu
- China's DeepSeek Debuts Flagship AI Model As Compute Race Intensifies ZeroHedge News · Tyler Durden
- DeepSeek Unveils V4 AI Model Optimized For Huawei Chips Amid US-China AI Rivalry Forbes Middle East · Khadijah Khogeer
- China's DeepSeek launches new AI model Semafor · Jeronimo Gonzalez
- DeepSeek says its new V4 models trail OpenAI and Google by months, not years crypto.news · Rony Roy
- DeepSeek Launches V4 AI Model on Huawei Chips Modern Diplomacy · Sana Khan
- DeepSeek-V4-Pro — Introduction We present a preview version of DeepSeek-V4 series … DeepSeek on Hugging Face
- DeepSeek promises its new AI model has ‘world-class’ reasoning Engadget · Mariella Moon
- Nvidia (NVDA) Faces Setback as DeepSeek Opts for Huawei Over American Chips Blockonomi · Trader Edge
- DeepSeek Releases V4 Pro and Flash, Undercutting OpenAI Pricing by Up to 10x Implicator.ai · Marcus Schuler
- DeepSeek previews new AI model that ‘closes the gap’ with frontier models TechCrunch · Ram Iyer
- China's DeepSeek rolls out a long-anticipated update of its AI model Associated Press
- DeepSeek launches V4 AI model to rival OpenAI Tech in Asia · Naomi Li Gan
- Huawei backs DeepSeek V4 with Ascend AI chips Tech in Asia · Minh Le
- DeepSeek returns with new AI model amid US-China tech rivalry Türkiye Today
- DeepSeek releases new V4 series models highlighting efficiency and long context Sherwood News · Jon Keegan
- DeepSeek-V4, the Chinese AI model adapted for Huawei chips Reuters
- DeepSeek rolls out V4 preview models with efficiency upgrades Proactive · Emily Jarvie
- DeepSeek previews V4 models with Huawei integration, signaling shift in China's AI stack DigiTimes
- DeepSeek V4 Is Here, but Don't Expect Another ‘DeepSeek Moment’ Morningstar · Ivan Su
- DeepSeek unveils its newest model at rock-bottom prices and with ‘full support’ from Huawei chips Fortune · Nicholas Gordon
- Huawei, DeepSeek strengthen China's AI self-reliance with collaboration on V4 model South China Morning Post
- DeepSeek's Newest Models Take on Silicon Valley at a Fraction of the Cost Gizmodo · Bruce Gil
- Five things to know about DeepSeek Oman Observer
- DeepSeek V4 Finally Drops After a Year of Anticipation TechEBlog · Jackson Chung
- DeepSeek's New V4 Model Signals Shift Toward Hardware Independence Techstrong.ai · Jon Swartz
- DeepSeek-V4 Preview launches with open weights and API access GSMArena.com
- DeepSeek unveils V4 preview model with Huawei chip support Crypto Briefing · Estefano Gomez
- DeepSeek's Long-Awaited New Model Fails to Narrow US Lead in AI Bloomberg · Seth Fiegerman
- DeepSeeek V4 is out, touting some disruptive wins over Gemini, ChatGPT, and Claude Digital Trends · Shikhar Mehrotra
- DeepSeek Releases V4 on Huawei Chips the Same Day the White House Accused China of Stealing US AI crypto.news · Peace Longe
- Tencent's New Hy3 AI Model Is the Most Efficient Chinese LLM No One's Talking About Decrypt · Jose Antonio Lanz
- DeepSeek's new models are so efficient they'll run on a toaster ... by which we mean Huawei's NPUs The Register · Tobias Mann
- DeepSeek's Much-Awaited AI Model Boosts Chinese Semiconductor Stocks MarketWatch · Sherry Qin
- DeepSeek open-sources V4 large language model series SiliconANGLE · Maria Deutscher
- DeepSeek-V4 preview now available with open-source access TechNode
- DeepSeek V4 is here: How it compares to ChatGPT, Claude, Gemini Mashable · Timothy Beck Werth
- Summary of Qwen3.6 GGUF Evals (Updating...) The Kaitchup · Benjamin Marie
- DeepSeek AI Releases DeepSeek-V4: Compressed Sparse Attention and Heavily Compressed Attention Enable One-Million-Token Contexts MarkTechPost · Asif Razzaq
Discussion
-
@deepseek_ai
@deepseek_ai
on x
🚀 DeepSeek-V4 Preview is officially live & open-sourced! Welcome to the era of cost-effective 1M context length. 🔹 DeepSeek-V4-Pro: 1.6T total / 49B active params. Performance rivaling the world's top closed-source models. 🔹 DeepSeek-V4-Flash: 284B total / 13B active params. [ima…
-
@suchenzang
Susan Zhang
on x
so that explains the delay... deepseek could not fix training instabilities, after doubling from ~15T tokens in v3 to ~33T tokens in v4 the 10+ mentions of “stability” tricks seem to be wildly lacking if these two were the main bandages (mismatched routing + clamping) but [image]
-
@eliebakouch
Elie
on x
Deepseek V4 Pro is the biggest open model ever with 1.6T total 49B active, trained on 33T tokens, 1M context, with 2 new attention mechanisms, Muon, mHC, open source kernels, FP4 QAT, MIT license and with one of the best tech repot of the year [image]
-
@arena
@arena
on x
Exciting news - DeepSeek V4 Pro is in the Arena with 1.6T parameters (49B activated) alongside V4 Flash at 284B parameters (13B activated). Both support 1M token context. It's a major leap over DeepSeek V3.2! Code Arena: - DeepSeek V4 Pro (thinking): #3 open model (#14 overall), …
-
@emollick
Ethan Mollick
on x
My first two TiKZ Sparks unicorns from DeepSeek v4. (Expert mode, from the DeepSeek site, which is supposed to be v4 Pro according to the release) [image]
-
@iamgingertrash
@iamgingertrash
on x
A Chinese company facing chip restrictions can train this But xAI can't even get SOTA With a million H100 equivalents
-
@mweinbach
Max Weinbach
on x
Here are the deepseek v4 benchmarks vs. US frontier models [image]
-
@basedjensen
@basedjensen
on x
I can't state enough how big of a deal this is and they have made everything open source. Whale bros might have saved local inference [image]
-
@emollick
Ethan Mollick
on x
Kimi K2.6, for comparison. Honestly don't know why the gap is so large. [image]
-
@amasad
Amjad Masad
on x
While US politicians/lobbyists are scaremongering about “Chinese distillation,” Chinese scientists are actually sharing real AI breakthroughs in the open. These kind of advances have nothing to do with data and benefit everyone, including small (and possibly big) US labs.
-
@dee_bosa
Deirdre Bosa
on x
What does frontier pricing power look like when theres a permanent, well-funded, open-source shadow market 3-6 months behind?
-
@yuchenj_uw
Yuchen Jin
on x
I'm still amazed that DeepSeek, Kimi, and Qwen can train very strong LLMs with far fewer and often nerfed NVIDIA GPUs, or even Huawei chips. DeepSeek V4 report shows they invent new attention architectures to make training/inference more efficient. Creativity loves constraints.
-
@mtslive
@mtslive
on x
DeepSeek promising price reductions for V4-Pro later in the year as Huawei's next generation Atlas 950 SuperPoD multi-rack clusters come online.
-
@simonw
Simon Willison
on x
These pelicans are kind of angry looking! Left is deepseek-v4-flash, right is deepseek-v4-pro - both generated using OpenRouter via my LLM tool [image]
-
@jukan05
Jukan
on x
What kind of magic did DeepSeek pull off this time? With V4, they seem to be back at SOTA again. Their coding performance also looks pretty serious. [image]
-
@haider1
Haider
on x
finally, deepseek v4 is here and the crazy part is that it is now easily comparable to opus 4.7 and gpt-5.5 level models, especially in coding, reasoning, and long-context work deepseek v4-pro has a 1m context window and costs around $1.74/m input and $3.48/m output how they [ima…
-
@victor207755822
Deli Chen
on x
DeepSeek-V3: Dec 26, 2024 DeepSeek-V4: Apr 24, 2026 484 days later, we humbly share our labor of love. As always, we stay true to long-termism and open source for all. AGI belongs to everyone. ❤️🌍 #DeepSeekV4 #AGIforEveryone #OpenSource
-
@valsai
@valsai
on x
The 🐳 has surfaced and it's a powerhouse on the Vals leaderboards, dominating on coding. DeepSeek V4 just landed #2 on the Vals Index, nearly tying Kimi K2.6 (only 0.07% behind). [image]
-
@chrisrmcguire
Chris McGuire
on x
DeepSeek v4 just dropped. At first glance it does not appear to be the kind of leap that v3 claimed to be in January 2025, nor does it challenge the consensus regarding the state of the U.S.-China AI competition: U.S. models lead by ~7 months, and leading Chinese models remain
-
@_arohan_
Rohan Anil
on x
DeepSeek Pro is a 1e25 flop run. Flash is a 2.5e24 flops in 6ND.
-
@scobleizer
Robert Scoble
on x
DeepSeek 4 is out. My AI says: +++++ The Numbers That Matter V4 Pro costs $3.48 per million output tokens. Claude Opus 4.6 costs $25. GPT-5.4 costs $15. Same benchmark tier. One fifth the price. ValsAI ran independent tests. V4 is now number 1 on their Vibe Code Benchmark.
-
@arena
@arena
on x
DeepSeek v4 lands in the Text Arena: - DeepSeek V4 Pro (thinking): #2 open model (#14 overall), matching Kimi-2.6 - DeepSeek V4 Flash (thinking): #10 open model (#47 overall) Top 10 Text categories: - #1 Medicine & Healthcare (v4 Pro) - #8 Legal & Government (v4 Pro) - #8 Math [i…
-
@jukan05
Jukan
on x
Very interesting. DeepSeek added the following comment with V4: “Due to constraints in high-end compute capacity, the current service capacity for Pro is very limited. After the 950 supernodes are launched at scale in the second half of this year, the price of Pro is expected [im…
-
@mweinbach
Max Weinbach
on x
Yea deepseek v4 flash/pro don't really perform that well compared to any of the major US models, even 1-2 revisions old. Looks like it's slightly behind Opus 4.5 in practice, and on par or slightly behind Kimi K2.6 Some good optimization techniques there, but overall, eh
-
@zephyr_z9
@zephyr_z9
on x
Massive compute gap At least 40x-80x
-
r/DeepSeek
r
on reddit
Deepseek-v4 flash and v4 pro
-
r/DeepSeek
r
on reddit
Deepseek v4 Released!
-
r/singularity
r
on reddit
DeepSeek V4 has released
-
r/LocalLLaMA
r
on reddit
Buried lede: Deepseek v4 Flash is incredibly inexpensive from the official API for its weight category
-
r/LocalLLaMA
r
on reddit
Deepseek V4 Flash and Non-Flash Out on HuggingFace
-
@simonw
Simon Willison
on x
More of my notes on DeepSeek V4 - the really big news is the pricing: both DeepSeek-V4-Flash and DeepSeek-V4-Pro are the cheapest models in their categories while benchmarking close to the frontier models from other providers https://simonwillison.net/... [image]
-
@simonwillison.net
Simon Willison
on bluesky
DeepSeek V4 just dropped - two models, Flash and Pro, both benchmarking well, decent pelicans and prices that put them both as the cheapest in their respective categories by a solid margin simonwillison.net/2026/Apr/24/ ... [images]
-
@sino_market
@sino_market
on x
HUAWEI ASCEND SUPERNODE TO SUPPPORT DEEPSEEK V4 : HUAWEI STATEMENT #CHINA #TECH #AI #HUAWEI #DEEPSEEK (https://mktnews.com/...)
-
@macrobombastic
@macrobombastic
on x
@jukan05 Bro, that's a pretty transparent statement from DeepSeek. Good on 'em for being upfront about their limitations.
-
@teortaxestex
@teortaxestex
on x
DeepSeek is trolling They don't say *what* they trained on, only “eh MXFP4” Hawks will now go on a hunt for Blackwells in Inner Mongolia In a year they'll publish “insights into V4” with details of an experimental Huawei cluster man I love this lab so much [image]
-
@teortaxestex
@teortaxestex
on x
> The library supports SM90 (Hopper) and SM100 (Blackwell), with no Huawei Ascend support I wanted to say “TileLang doesn't even support Ascend” but apparently there is a variant still, I repeat that all this “DeepSeek spent 9000 years adapting for Huawei” is most likely bullshit…
-
@poezhao0605
Poe Zhao
on x
Buried in the fine print: DeepSeek says V4-Pro throughput is currently limited by high-end compute supply. Prices will drop significantly once Huawei Ascend 950 super nodes ship at scale in H2. DeepSeek is publicly tying its API economics to domestic chip infrastructure. That's […
-
@zichenwanghere
Zichen Wang
on x
DeepSeek 4 upon Huawei's Ascend chipsis apparently being released in a live event via Bilibili, China's equivalent to Youtube, today at 7pm Beijing Time on April 24. [image]
-
@sgg_trader
@sgg_trader
on x
@jukan05 Chinese from China said they are “forced” to use Huawei .
-
@sheriyuo
Xiuyu Li
on x
This is the inevitable outcome of locally deploying domestically produced (localized) cards, and it will definitely bring the prices down.
-
@himself65
@himself65
on x
@jukan05 They have to use hw, for political propaganda.
-
@ErikJonker@mastodon.social
Erik Jonker
on mastodon
Do not only look at benchmarks of AI models. Costs are also very important and the differences are big. In the end that is very important for businesses using AI at scale. Picture is from this excellent blog/post from @simon — https://simonwillison.net/... #AI #deepseekv4 #…
-
Houwei Rui
Houwei Rui
on linkedin
DeepSeek V4 marks a clear shift in AI: — performance is catching up, but cost is dropping faster. …
-
Daniel Langkilde
Daniel Langkilde
on linkedin
DeepSeek v4 is out ❤️ I'm trying to slow-read papers right now. It's tempting to just push them to Claude and ask “what's new”, but that's a slippery slope towards AI-brain-rot. …
-
Merve Noyan
Merve Noyan
on linkedin
if you were in a cave, DeepSeek v4 is out, and it's groundbreaking, here's a short post on why: — it's the first open model to have solved long context …
-
Angus CHENG Zhe
Angus CHENG Zhe
on linkedin
The long-awaited, cost-efficient DeepSeek V4 unveiled its preview model today, with fully support by Huawei's Ascend supernode. …
-
Jing Conan Wang
Jing Conan Wang
on linkedin
DeepSeek v4 might be a turning point for the entire agentic AI stack. — Not just because it's a strong model — but because it shifts two important dynamics: …
-
@deanwball
Dean W. Ball
on x
This is actually the most salient fact about V4, rather than “da bar charts look gud!!!” simpleton logic. It's not actually that competitive of a model, but it is useful substrate for all manner of upstarts, worldwide. This is pro-social, and I wish the US had an equivalent.
-
@deanwball
Dean W. Ball
on x
V4's benchmarks don't tell you much, but the revealing thing is DeepSeek can barely serve the model. The fact that a model came out and “da charts looked good!!!” is not a sign of much; thinking otherwise is the logic of apes or, in Arnaud's case, obvious assets.
-
@synthwavedd
Leo
on x
my first impressions on deepseek v4: - little disappointing it's not SoTA after all this time, but it's close - new pareto frontier - a lot cheaper than 5.4/opus 4.6 for comparable performance! - new favourite model for creative writing - pretty noticeable big model smell -
-
@kyleichan
Kyle Chan
on x
DeepSeek V4 is impressive because it's a near-SOTA model with highly efficient 1 million token context that can run on Huawei's new Ascend 950PR chips. But equally notable is what V4 didn't do: - No mention of training on Chinese AI chips - Still lags behind US frontier models
-
@dorialexander
@dorialexander
on x
Didn't expect to be hit with the most significant AI paper of the year on breakfast. Longer read coming but what's obvious already: turning a sovereignty play (using Huawei ascend) into an opportunity to reshape hardware. Intercoms, memory, power: wishlist everywhere. [image]
-
@antirez
@antirez
on x
First impressions on DeepSeek v4 pro used via Claude Code. It is great but not so cheap compared to how much tokens you get with the OpenAI 200$ subscription. More or less I burned 1$ per hour of intense usage.
-
@deanwball
Dean W. Ball
on x
This prediction from a Chinese AI researcher—that the gap between US and Chinese AI is growing larger—matches my own, from well over a year ago now, that DeepSeek r1 marked ~the closest Chinese models would get to the US frontier absent a change in China's access to compute.
-
@teortaxestex
@teortaxestex
on x
anon do you realize that V4-Pro is straight up the strongest pretrained model we have? Like... 1.6T@49AB (≈280B dense), 33T - even by meme formula it's > LLaMA 3. Add Muon, mHC, most steps 64K context + extended to 1M... No excuses now. Every “unicorn” can have its brand AGI. [im…
-
@artificialanlys
@artificialanlys
on x
DeepSeek V4 Pro is the #1 open weights model on GDPval-AA, our agentic real-world work tasks evaluation @deepseek_ai has released V4 Pro (1.6T total / 49B active) and V4 Flash (284B total / 13B active). V4 is DeepSeek's first new size since V3, with all intermediate models [image…
-
@sn@mastodon.ping.de
@sn@mastodon.ping.de
on mastodon
#deepseek V4 Flash should run ok on 2x Strix Halo being around 158GB with 13b active parameters. Looking forward to see their innovations in action, in particular the Hybrid attention architecture for FLOPs and KV cache savings. Plus it's only a preview with multi-modality comi…
-
r/LocalLLaMA
r
on reddit
No Multimodality yet in DeepSeek-V4. But I'll wait.
-
@thdxr
Dax
on x
deepseek v4 can run on huawei chips another step towards self reliance thanks to our own policies can't believe how many times we'll make this mistake
-
Denys Linkov
Denys Linkov
on linkedin
DeepseekV4 is out, retaining it's MIT licence! We got GPT 5.5 and Deepseek v4 on the same day, some initial thoughts from the technical report: …
-
@timkellogg.me
Tim Kellogg
on bluesky
Today is DeepSeek V4 Day — Truly hangs with Opus 4.6 & GPT-5.4 — Pro: 1.4T (49B active) — Flash: 248B / 13B — 1M context using DSA — Optimized for Claude Code — Pro price: $0.14 in / $3.48 out — Flash price: $0.028 in / $0.28 out — Open source has caught the frontier. …
-
@mukund
M Mohan
on x
OMG DeepSeek is turning model efficiency into a weapon. Its 30 - 44% cheaper DeepSeek AI is making LLM it look like an economic design problem: sparse attention, compressed KV cache, fewer activated parameters, FP4/FP8 weights, and reasoning modes that let users trade speed,
-
M Mohan
M Mohan
on linkedin
DeepSeek is turning model efficiency into a weapon. Its 30 - 44% cheaper — DeepSeek AI is making LLM it look like an economic design problem …