Everything in AI runs on matrix math, and compute is the hardware that executes it. Understanding GPU architecture, cloud vs on-prem tradeoffs, and cost structures is the foundation. You do not need to buy GPUs to understand this layer — but you need to know what drives the cost of every layer above it.
01.1 Silicon
GPU vs TPU vs custom silicon. Memory bandwidth is the bottleneck, not FLOPS. CUDA dominates.
Key concepts
- Hardware tradeoffs
- Memory bandwidth
- CUDA lock-in
- Chip generations
Practical skills
- Read GPU specs
- Compare silicon options
01.2 GPU Clouds
CoreWeave, Lambda, RunPod, Modal. Reserved vs on-demand vs spot pricing.
Key concepts
- GPU cloud vs hyperscaler
- Serverless GPU
- Marketplace models
Practical skills
- Compare cloud GPU pricing
- Estimate compute costs
01.3 Training Infra
Distributed training software. Data/model/pipeline parallelism.
Key concepts
- Parallelism strategies
- NCCL
- Distributed frameworks
Practical skills
- Set up multi-GPU training
- Choose parallelism strategy
See this layer in the value chain → L01 Compute