Google Cloud on Wednesday announced that its eighth generation of custom-built AI chips, or tensor processing units (TPUs), will be split in two. One chip, named the TPU 8t, will be geared for model training and another, the TPU 8i, is aimed at inference. Inference is the ongoing usage of models, aka what happens after users submit prompts. As you might expect, the company touts some impressive performance specs for these new TPUs compared to the previous generations: up to 3x faster AI model t…
Why this update matters
This developing story is relevant for readers tracking technology because it reflects fresh changes from the original source and signals where attention is shifting next.
Key details
The report was collected automatically and prepared for publication with a newsroom workflow that focuses on clarity, search visibility, and quick understanding.
Readers should review the original source for direct statements, official notices, and any later corrections or additions as the story evolves.
Related coverage
Continue reading with more reporting from the same topic cluster.