Home / Technology / Google Splits AI Chips for Training and Inference
Google Splits AI Chips for Training and Inference
23 Apr
Summary
- Google Cloud launched new TPUs for training and inference.
- New chips offer up to 3x faster AI model training.
- Google's chips supplement, not replace, Nvidia's infrastructure.

Google Cloud announced its eighth generation of custom-built AI chips, known as tensor processing units (TPUs), will be offered in two specialized versions. The TPU 8t is designed for AI model training, while the TPU 8i is optimized for inference, the process of using trained models with new data.
These new TPUs reportedly deliver substantial improvements over previous generations. Google claims they provide up to three times faster AI model training and an 80% enhancement in performance per dollar. Additionally, the infrastructure can support over one million TPUs working together in a single cluster, promising increased compute efficiency with reduced energy consumption and costs for customers.