/
Navigation
Chronicles
Browse all articles
Explore
Semantic exploration
Research
Entity momentum
Nexus
Correlations & relationships
Story Arc
Topic evolution
Drift Map
Semantic trajectory animation
Posts
Analysis & commentary
Pulse API
Tech news intelligence API
Browse
Entities
Companies, people, products, technologies
Domains
Browse by publication source
Handles
Browse by social media handle
Detection
Concept Search
Semantic similarity search
High Impact Stories
Top coverage by position
Sentiment Analysis
Positive/negative coverage
Anomaly Detection
Unusual coverage patterns
Analysis
Rivalry Report
Compare two entities head-to-head
Semantic Pivots
Narrative discontinuities
Crisis Response
Event recovery patterns
Connected
Search: /
Command: ⌘K
Embeddings: large
TEXXR

Chronicles

The story behind the story

days · browse · Enter similar · o open

Google introduces its eighth generation of TPUs, including the TPU 8t for training and the TPU 8i for inference, generally available later this year

Alphabet Inc.'s Google Cloud division unveiled the latest generation of its tensor processing unit, or TPU, a homegrown chip that's designed …

Bloomberg Ian King

Discussion

  • @sundarpichai Sundar Pichai on x
    TPU 8t, optimized for training and TPU 8i, optimized for inference. Looking good! [image]
  • @scaling01 @scaling01 on x
    Google also said they can now scale to a million TPUs within a single cluster with TPU8t [image]
  • @koraykv Koray Kavukcuoglu on x
    This week at Google Cloud Next, we introduced 8th gen TPUs, a critical milestone in our accelerator roadmap. TPUs enable us to optimize the entire stack for AI (with 8t for massive-scale training and 8i for low-latency inference). Exciting breakthrough from our hardware teams!
  • @officiallogank Logan Kilpatrick on x
    Huge improvements over Ironwood TPU's (2-3x), can't wait to see Gemini in action with these : ) [image]
  • @beffjezos @beffjezos on x
    We're a long way from the days of TPUv2 🥲 they grow up so fast
  • @elonmusk Elon Musk on x
    @sundarpichai TPUs are underrated
  • @chanduthota Chandu Thota on x
    Indeed, they are a thing of beauty! [image]
  • @officiallogank Logan Kilpatrick on x
    TPU's are a core part of the Google secret sauce, excited to see our 8th generation TPU see the light of day : ) [image]
  • @benbajarin Ben Bajarin on x
    GCP embracing full multi-cloud support is smart. Many CIO/CTOs we talk to emphasize they will use multiple vendors and increasingly so, so this makes GCP a great second or third source, if not primary. [image]
  • @schrep Mike Schroepfer on x
    It is staggering the advances in compute, memory, and bandwidth for AI training and inference. https://blog.google/...
  • @scaling01 @scaling01 on x
    TPU 8i/8t seems to be a massive upgrade in compute [image]
  • @officiallogank Logan Kilpatrick on x
    @sundarpichai @samsheffer nice, I'll take both : )
  • @usgraphics @usgraphics on x
    omfg hnggg.... [image]
  • @mweinbach Max Weinbach on x
    This is for all of the nerds who want to know about TPU v8 Anyone who thinks that TPUs aren't good or viable vs GPUs should think very carefully about their position on this https://cloud.google.com/...
  • @patrickmoorhead Patrick Moorhead on x
    May not seem sexy but I love love love Google going all in on MCP.  Even dummies like me can create amazing apps that leverage MCP...This is the most consequential piece as it turns the entire Google Cloud surface area into addressable tools for any MCP-capable agent, including n…
  • @mweinbach Max Weinbach on x
    TPU 8i pods are no longer 3D torus 64 chip pods, but 36 chip Boardfly Here's some info on why they switched and the benefits TPU 8t still uses 3D torus [image]
  • @mweinbach Max Weinbach on x
    TPU 8t ASIC block diagram & TPU 8t rack level connectivity to Virgo fabric [image]
  • @danielnewmanuv Daniel Newman on x
    $GOOGL drops two new TPUs. The TPU 8t and TPU 8i. Strong generation over generation performance improvements. Google continues to lead the home grown full stack narrative. 💪🏻🚀 [image]
  • @mweinbach Max Weinbach on x
    Google also announces TorchTPU so you can take training code written for GPUs with PyTorch and easily run it on TPUs https://developers.googleblog.com/ ... [image]
  • @mweinbach Max Weinbach on x
    Google also announced a new Virgo networking stack with TPU v8. A big part: Massive fabric scale: Virgo Network can link 134,000 chips (TPU 8t) with up to 47 petabits/sec of non-blocking bi-sectional bandwidth in a single fabric. [image]
  • @mweinbach Max Weinbach on x
    Big part of TPU v8T “TPU 8t can provide near-linear scaling for up to a million chips in a single logical cluster.” [image]
  • @mweinbach Max Weinbach on x
    It seems like the previous TPUs were using x86 for the host system, not Axiom (Google's custom Arm CPU). TPU v8 swaps to Axiom and 2 TPU per CPU. Previously it was 4 TPU per CPU
  • @rihardjarc Rihard Jarc on x
    The chip company $GOOGL just announced a new version of TPUs.  TPU 8t for training and TPU 8i for inference.  - TPU 8t 124% better performance per watt vs last generation - TPU 8i 117% better performance per watt vs last generation - TPU 8i is also 80% better performance per doll…
  • @benbajarin Ben Bajarin on x
    I was at a small preview event last night on Google's new TPU 8t and 8i. Have a post coming shortly, but got some eye candy! Board pics of 8t and 8i. [image]
  • @benbajarin Ben Bajarin on x
    Inference is getting its own architectural focus because reasoning and agents place different demands on the system.  That is one of many signals in Google's latest TPU strategy.  I also like new words.  “Boardfly”.  Which may have been the most revealing detail in Google's TPU u…
  • @patrickmoorhead Patrick Moorhead on x
    Two new TPUs, one for training and one for inference. TPU 8t is the training box: 9,600 chips per superpod, 2+ PB of shared HBM, 121 exaflops, 2.8x the prior generation and 2x better perf/watt vs. prior gen, native FP4 in the MXUs, and Axion Arm hosts. With Pathways and JAX, a [i…
  • Jim Fairweather Jim Fairweather on linkedin
    $750m to help our ISV clients and partners meet this Agentic moment.  Let's go!
  • Umesh Vaghela Umesh Vaghela on linkedin
    🚀 Huge news from Google Cloud Next '26!  —  We just announced a massive $750 million investment to accelerate the development …
  • Patrick Musau, Ph.D. Patrick Musau, Ph.D. on linkedin
    Proud to be part of the CodeMender team!  Today Sundar called us out by name in his Cloud Next '26 keynote post as an example of how Google uses AI agents …
  • Manuel Greisinger Manuel Greisinger on linkedin
    Live from #GoogleCloudNext in Las Vegas.  —  Sundar's message in one line: we're firmly in the agentic Gemini era.  —  16 billion tokens/minute. …
  • Oliver Parker Oliver Parker on linkedin
    The Agentic Enterprise has arrived.  At #GoogleCloudNext, we're showcasing more than 500 organizations already putting these technologies into production to fundamentally transform their operations. …
  • Anna Berenberg Anna Berenberg on linkedin
    Gemini Enterpise Agent Platfom is vast.  Governance layer is where GCP infrastructure for AI shines the most and obviously dear to my heart : …
  • LinkedIn Sundar Pichai on linkedin
    Sundar Pichai's Post
  • r/Bard r on reddit
    Our eighth generation TPUs: two chips for the agentic era
  • @googledeepmind @googledeepmind on x
    We're launching Gemini Enterprise Agent Platform with @GoogleCloud: a platform for businesses to develop, scale, govern and optimize agents. It's the evolution of Vertex AI, bringing together model selection and agent building with new features for integration, security and [imag…
  • @googledeepmind @googledeepmind on x
    It gives access to 200+ of the world's leading models through the Model Garden. This includes our latest breakthroughs: Gemini 3.1 Pro, Gemini 3.1 Flash Image, and Lyria 3, alongside our open models like Gemma 4. [image]
  • @nvidia @nvidia on x
    @googlecloud .@CrowdStrike uses NVIDIA NeMo open libraries on Gemini Enterprise Agent Platform to generate synthetic data and is fine-tuning Nemotron for domain-specific cybersecurity.
  • @advaitonline Advait Bopardikar on x
    Good morning from Las Vegas. Today is all about evolution and how we are addressing the toughest challenges Enterprises face when moving into their Agentic era. We are excited to announce the Gemini Enterprise Agent Platform!
  • @sundarpichai Sundar Pichai on x
    Google Cloud has incredible momentum: our models now process 16B+ tokens /min via direct API use by our customers (up from 10B last quarter). This week at Cloud Next we're sharing an extraordinary range of new partnerships and innovations, including our new Gemini Enterprise [ima…
  • @thomasortk Thomas Kurian on x
    Gemini Enterprise app includes a new Agent Designer, Inbox for managing agent activity, long-running agents, Skills, Projects and more.
  • @oracledatabase @oracledatabase on x
    We're expanding our partnership with Google Cloud to bring new AI capabilities to Oracle AI Database@Google Cloud customers. With Oracle AI Database Agent for Gemini Enterprise, you can ask natural-language questions about your Oracle data—no SQL or data model expertise [video]
  • Guillaume Roques Guillaume Roques on linkedin
    The era of the helpful chatbot is officially over.  We have entered the era of the Agentic Enterprise. …
  • Sujay Solomon Sujay Solomon on linkedin
    Gemini Enterprise Agent Platform is here.  You can build, scale, optimize and most importantly govern agents running anywhere all in one place: https://lnkd.in/... …
  • Michael Bachman Michael Bachman on linkedin
    Agents are now at the center of enterprise productivity and operations but it has to start with a foundation of trust.  Introducing our Gemini Enterprise Agent Platform. …
  • @chicagomike Mike on bluesky
    Finally.  Very interested.  [embedded post]
  • r/Bard r on reddit
    Introducing Gemini Enterprise Agent Platform
  • @deedydas Deedy on x
    In ~2yrs, Google has gone from 0% code written by AI to 75% code written by AI. What an incredible transformation of how software is created. [image]
  • @hesamation @hesamation on x
    but which AI are you using, Google? [video]
  • @rekdt @rekdt on x
    We can tell
  • @pcshipp @pcshipp on x
    UPDATE: Google says AI now generates 75% of its new code I can bet that Gemini isn't writing all of that 75% they're probably using other LLM models too Gemini too bad for coding
  • @tunguz Bojan Tunguz on x
    Translation: GDM is responsible for 75% of all new code at Google.
  • @kimmonismus @kimmonismus on x
    75% of all new code at Google is now AI-generated and approved by engineers, up from 50% last fall. 2027 90%, and 2028...? [image]
  • @lauridonahue Lauri Donahue on bluesky
    Taking notes is often the only way to keep yourself awake in a boring meeting... Also, real time annotation of those notes can help make them actionable.  [embedded post]
  • @treblaw Treb on bluesky
    taking notes for people who won't read the notes anyway or send them out, nice [embedded post]