1
Inputs
Chips, power, memory, networking
Learn
The basic resource behind training and running AI models.
AI compute is the capacity used to train and run artificial-intelligence models. It includes the chips doing the work, the memory and networking that connect them, the data centers and power that support them, and the cloud or infrastructure access that makes them usable.
Compute includes chips, memory, networking, power, facilities, and access to them.
Training builds models. Inference runs them after deployment.
Example
Think of AI compute as the factory capacity behind AI. GPUs do the core work, but the factory also needs memory, networking, power, cooling, and data-center space to turn chips into usable output.
1
Chips, power, memory, networking
2
Training or inference
3
A model learned or a model response served
Market map
Workloads
Build
Training uses large amounts of compute to teach a model from data. It is usually concentrated, expensive, and capacity-intensive.
Run
Inference uses compute after deployment to answer prompts, generate outputs, or serve users repeatedly at scale.
Common mistake
Two companies can own or rent the same number of GPUs and still have very different usable compute. Chip generation, memory, networking, power, software, and utilization all affect how much real work the system can deliver.
Why it matters
Keep learning
Concept
Why chips, power, and capacity are becoming economic constraints.
Unit
The basic unit behind compute pricing.
Compare
How accelerator generations affect performance, supply, and cost.