Provision and manage GPU pods on RunPod for long-running experiments. Use when the user needs persistent GPU compute with SSH access, large datasets, or multi-step experiments.
Use runpodctl CLI for persistent GPU pods with SSH access.
brew install runpod/runpodctl/runpodctl # macOS
runpodctl config --apiKey=YOUR_KEY
| Command | Description |
|---|---|
runpodctl create pod --gpuType "NVIDIA A100 80GB PCIe" --imageName "runpod/pytorch:2.4.0-py3.11-cuda12.4.1-devel-ubuntu22.04" --name experiment | Create a pod |
runpodctl get pod | List all pods |
runpodctl stop pod <id> | Stop (preserves volume) |
runpodctl start pod <id> | Resume a stopped pod |
runpodctl remove pod <id> | Terminate and delete |
runpodctl gpu list |
| List available GPU types and prices |
runpodctl send <file> | Transfer files to/from pods |
runpodctl receive <code> | Receive transferred files |
ssh root@<IP> -p <PORT> -i ~/.ssh/id_ed25519
Get connection details from runpodctl get pod <id>. Pods must expose port 22/tcp.
NVIDIA GeForce RTX 4090, NVIDIA RTX A6000, NVIDIA A40, NVIDIA A100 80GB PCIe, NVIDIA H100 80GB HBM3
command -v runpodctl