TAAFT
Free mode
100% free
Freemium
Free Trial
Deals
Create tool

Tensor Processing Units (TPUs)

[ˈtɛnsər prəˈsɛsɪŋ ˈjunɪts]
AI Infrastructure
Last updated: December 9, 2024

Definition

Custom-designed AI accelerator application-specific integrated circuits (ASICs) developed by Google specifically for neural network machine learning.

Detailed Explanation

TPUs are custom-developed chips optimized specifically for Google's TensorFlow machine learning framework. They use a systolic array matrix processor design that makes them highly efficient at the matrix multiplication operations central to deep learning. TPUs operate on tensors (multi-dimensional arrays) with baked-in support for quantization and other ML-specific optimizations, offering 15-30x higher performance and 30-80x higher performance-per-watt than contemporary GPUs and CPUs for machine learning workloads.

Use Cases

Large-scale machine learning model training and inference, powering Google services like Search, Maps, Photos, and Cloud TPU services for enterprise customers

Related Terms