Thunder Compute logo

What are CUDA Cores?

Learn the basics regarding CUDA cores, plus a simple example of how they enable parallel work on NVIDIA GPUs.

CUDA Cores are the individual processing units inside NVIDIA GPUs that execute floating-point and integer arithmetic operations. A single GPU can contain thousands of CUDA cores working in parallel.

What Is CUDA Programming

CUDA programming is how developers write code that runs on NVIDIA GPUs instead of only on the CPU. It breaks work into many parallel threads so thousands of CUDA cores can execute at once, which is ideal for vector math and ML workloads.

Basic CUDA Programming Example

Below is a tiny PyTorch example that runs on the GPU. Each element-wise add is split across many CUDA cores, so the work happens in parallel.

# A simple vector add — each CUDA core handles one element
import torch
a = torch.randn(1_000_000, device="cuda")
b = torch.randn(1_000_000, device="cuda")
c = a + b  # distributed across CUDA cores

NVIDIA GPU CUDA Core Comparison

Graphics CardArchitectureCUDA Cores
RTX 3090Ampere10,496
RTX 4090Ada Lovelace16,384
RTX 5090Blackwell21,760
RTX A6000Ampere10,752
A100Ampere6,912
H100Hopper14,592
Recommended article

Want a deeper, practical take? Read about how Cuda core counts translate to training speed and cost.

Read more

See Also