Flash provides access to a wide range of NVIDIA GPUs through both pool-based and specific GPU selection. This page lists all available GPU types and explains how to use them.Documentation Index
Fetch the complete documentation index at: https://runpod-b18f5ded-promptless-remove-flash-beta-notification.mintlify.app/llms.txt
Use this file to discover all available pages before exploring further.
GPU selection methods
Flash offers two ways to specify GPU hardware:- GPU pools (
GpuGroup): Select from predefined pools of similar GPUs grouped by architecture and VRAM. - Specific GPU types (
GpuType): Target exact GPU models when you need precise hardware characteristics.
GPU pools
TheGpuGroup enum provides access to GPU pools. Each pool contains specific GPU models grouped by architecture and VRAM capacity.
Available GPU pools
| GpuGroup | GPUs Included | VRAM | Best For |
|---|---|---|---|
GpuGroup.ANY | Any available GPU | Varies | Fast provisioning, prototyping |
GpuGroup.AMPERE_16 | RTX A4000, RTX 4000 Ada, RTX 2000 Ada | 16GB | Small models, basic inference |
GpuGroup.AMPERE_24 | RTX A4500, RTX A5000, RTX 3090 | 20-24GB | General ML, mid-size models |
GpuGroup.ADA_24 | L4, RTX 4090 | 24GB | Cost-effective inference |
GpuGroup.ADA_32_PRO | RTX 5090 | 32GB | Latest consumer flagship |
GpuGroup.AMPERE_48 | A40, RTX A6000 | 48GB | Large models, fine-tuning |
GpuGroup.ADA_48_PRO | L40S, L40, RTX 6000 Ada | 48GB | Professional inference |
GpuGroup.AMPERE_80 | A100 80GB PCIe, A100-SXM4-80GB | 80GB | XL models, intensive training |
GpuGroup.ADA_80_PRO | H100 80GB HBM3 | 80GB | Cutting-edge inference |
GpuGroup.BLACKWELL_96 | RTX PRO 6000 Blackwell (Server, Workstation, Max-Q) | 96GB | Professional Blackwell workloads |
GpuGroup.HOPPER_141 | H200 | 141GB | Largest models, maximum VRAM |
GpuGroup.BLACKWELL_180 | B200 | 180GB | Maximum VRAM, next-gen training |
Using GPU pools
Specific GPU types
TheGpuType enum provides access to specific GPU models. Use these when you need exact hardware characteristics.
Available GPU types
| GpuType | GPU Model | VRAM | Architecture |
|---|---|---|---|
GpuType.ANY | Any available GPU | Varies | Any |
GpuType.NVIDIA_RTX_A4000 | NVIDIA RTX A4000 | 16GB | Ampere |
GpuType.NVIDIA_RTX_A4500 | NVIDIA RTX A4500 | 20GB | Ampere |
GpuType.NVIDIA_RTX_4000_ADA_GENERATION | NVIDIA RTX 4000 Ada | 16GB | Ada Lovelace |
GpuType.NVIDIA_RTX_2000_ADA_GENERATION | NVIDIA RTX 2000 Ada | 16GB | Ada Lovelace |
GpuType.NVIDIA_RTX_A5000 | NVIDIA RTX A5000 | 24GB | Ampere |
GpuType.NVIDIA_L4 | NVIDIA L4 | 24GB | Ada Lovelace |
GpuType.NVIDIA_GEFORCE_RTX_3090 | NVIDIA GeForce RTX 3090 | 24GB | Ampere |
GpuType.NVIDIA_GEFORCE_RTX_4090 | NVIDIA GeForce RTX 4090 | 24GB | Ada Lovelace |
GpuType.NVIDIA_GEFORCE_RTX_5090 | NVIDIA GeForce RTX 5090 | 32GB | Blackwell |
GpuType.NVIDIA_A40 | NVIDIA A40 | 48GB | Ampere |
GpuType.NVIDIA_RTX_A6000 | NVIDIA RTX A6000 | 48GB | Ampere |
GpuType.NVIDIA_RTX_6000_ADA_GENERATION | NVIDIA RTX 6000 Ada | 48GB | Ada Lovelace |
GpuType.NVIDIA_A100_80GB_PCIe | NVIDIA A100 80GB PCIe | 80GB | Ampere |
GpuType.NVIDIA_A100_SXM4_80GB | NVIDIA A100-SXM4-80GB | 80GB | Ampere |
GpuType.NVIDIA_H100_80GB_HBM3 | NVIDIA H100 80GB HBM3 | 80GB | Hopper |
GpuType.NVIDIA_RTX_PRO_6000_BLACKWELL_SERVER_EDITION | NVIDIA RTX PRO 6000 Blackwell Server Edition | 96GB | Blackwell |
GpuType.NVIDIA_RTX_PRO_6000_BLACKWELL_WORKSTATION_EDITION | NVIDIA RTX PRO 6000 Blackwell Workstation Edition | 96GB | Blackwell |
GpuType.NVIDIA_RTX_PRO_6000_BLACKWELL_MAX_Q_WORKSTATION_EDITION | NVIDIA RTX PRO 6000 Blackwell Max-Q Workstation Edition | 96GB | Blackwell |
GpuType.NVIDIA_H200 | NVIDIA H200 | 141GB | Hopper |
GpuType.NVIDIA_B200 | NVIDIA B200 | 180GB | Blackwell |
Using specific GPU types
Advanced fallback strategies
CombineGpuGroup and GpuType for robust availability:
GPU selection behavior
Single GPU type: Flash waits for this specific GPU to become available. Jobs stay in queue until capacity is available.Multi-GPU workers
Request multiple GPUs per worker usinggpu_count:
Handling unavailability
If requested GPUs are unavailable, jobs stay in queue:-
Add fallback options: Use multiple GPU types.
-
Use broader selection: Switch to
GpuGroup.ANY. - Contact support: For capacity guarantees, contact Runpod support.