GPU
AMD Instinct MI300X
Available from RunPod, Lambda Labs, DigitalOcean, Vultr
5
Listings
$0.50
From /hr
Save 97% vs avg
$15.95
Avg /hr
$365
From /mo
1536 GB
VRAM
up to 8×
GPUs/server
Price trend (30 days)
Accumulating data…Specifications
Architecture
CDNA3
VRAM
192 GB HBM3
Memory BW
5.3 TB/s
FP16 TFLOPs
1,307
TDP
750 W
NVLink
No
192 GB unified HBM3 fits the largest open-source models on one GPU. ROCm software stack.
MI300X Price Alert
Get notified when prices drop. No spam — one email per significant price change.
5 results
Density:
| Provider | GPU | GPUs↕ | VRAM↕ | Cores↕ | RAM↕ | $/mo↑ | $/hr↕ | $/GB·hr | Location | |
|---|---|---|---|---|---|---|---|---|---|---|
| MI300X | 1× | 192 GB | — | — | $365/mo | $0.50 | $0.0026 | Global | View | |
| MI300X | 1× | 192 GB | 24 | 384 GB | $2548/mo | $3.49 | $0.018 | 🇺🇸US | View | |
| MI300X | 8× | 192 GB | 160 | 1920 GB | $11,621.6/mo | $15.92 | $0.010 | 🇺🇸US | View | |
| MI300X | 8× | 192 GB | 192 | 3072 GB | $20,381.6/mo | $27.92 | $0.018 | 🇺🇸US | View | |
| MI300X | 1× | 1536 GB | 128 | 2048 GB | $21,450.24/mo | $31.92 | $0.021 | — | View |