Create an Instance
Interactive Mode
Run the create command to launch an interactive menu:One-Line Creation
Specify all options in a single command:Configuration Options
| Flag | Description |
|---|---|
--mode | prototyping or production |
--gpu | a6000, a100, or h100 (prototyping); a100 or h100 (production) |
--num-gpus | 1-8 (production); 1-2 for H100 prototyping |
--vcpus | CPU cores (prototyping only). Options vary by GPU type and count. RAM: 8GB per vCPU |
--template | base, comfy-ui, ollama, or a snapshot name |
--disk-size-gb | 100-400 GB (prototyping), 100-1000 GB (production) |
Mode Selection
Choose between optimized development pricing or full compatibility:- Prototyping (default): Lower cost with CUDA-level optimizations. Best for development.
- Production: Standard VM with full compatibility. Best for long-running jobs and production workloads.
GPU Options
| GPU | VRAM | Availability |
|---|---|---|
| A6000 | 48GB | Prototyping only |
| A100 | 80GB | Both modes |
| H100 | 80GB | Both modes |
CPU and RAM
Prototyping mode: Configure vCPUs with 8GB RAM per vCPU. Options vary by GPU type and count:| GPU | GPU Count | vCPU Options |
|---|---|---|
| A6000 | 1 | 4, 8 |
| A100 | 1 | 4, 8, 12 |
| H100 | 1 | 4, 8, 12, 16 |
| H100 | 2 | 8, 12, 16, 20, 24 |
- 2 GPUs = 36 vCPUs, 180GB RAM
- 4 GPUs = 72 vCPUs, 360GB RAM
- 8 GPUs = 144 vCPUs, 720GB RAM
By default, 4 vCPUs and 32GB of memory are included with prototyping instances. Additional vCPUs are billed hourly at the rates shown on the pricing page.
Templates
Templates pre-configure your instance for common AI workflows:| Template | Description |
|---|---|
base | Ubuntu with PyTorch + CUDA |
ollama | Ollama server environment |
comfy-ui | ComfyUI for AI image generation |