Thunder Compute logo

What is VRAM?

Short definition of VRAM and why GPU memory capacity matters for real workloads.

VRAM (Video RAM) is the dedicated memory on a GPU used to store data that the GPU is actively processing, such as model weights, activations, and intermediate tensors. If you are asking what is vram, think of it as the working memory for GPU tasks.

Why VRAM Matters

VRAM capacity determines how large a model or dataset you can fit on the GPU and how much work you can do without slow CPU offloading, so it is a practical limit for many workloads.

NVIDIA GPU VRAM & Memory Comparison

Graphics CardVRAM CapacityMemory TypeMemory Bandwidth
RTX 309024 GBGDDR6X936 GB/s
RTX 409024 GBGDDR6X1,008 GB/s
RTX 509032 GBGDDR71,792 GB/s
RTX A600048 GBGDDR6 (ECC)768 GB/s
A10080 GBHBM2 / HBM2eUp to 2,039 GB/s
H10080 GBHBM3Up to 3,352 GB/s

How to Check VRAM

On NVIDIA GPUs, you can check VRAM capacity and usage with nvidia-smi, or inside Python with torch.cuda.get_device_properties(0).total_memory.

How to Increase VRAM

You cannot increase physical VRAM on a GPU, but you can reduce usage with smaller models, lower batch sizes, or offloading to CPU memory. You can also choose a GPU with more VRAM.

Recommended article

Keep reading for practical guidance on choosing a GPU for AI workloads.

Read more

See Also