RESERVED MI350X

RESERVED MI350X

RESERVED MI350X

RESERVED MI350X

AMD Instinct MI350X 288GB

The MI350X is AMD's CDNA 4 inference flagship, pairing 288GB of HBM3e with native FP4 support and 8.0 TB/s memory bandwidth. Reserved capacity is the primary path to MI350X access in 2026 on-demand availability is limited and rates have ranged $14–$18/hr at early launch. Memory advantage over H200 (288GB vs 141GB) makes it compelling for 400B-plus models on a single device.

AVAILABLE TERM LENGTH

USED V100 32GB — Indicative Range (Q1 2026)

1MO

3MO

6MO

12MO

24MO

36MO

Supply is constrained through Q2 2026 due to ramp limitations. Most reserved availability sits with DigitalOcean and select AMD-aligned neoclouds. Short terms (1-month) typically not offered providers prioritize 12-month-plus commits.

TECHNICAL SPECIFICATIONS

VRAM

VRAM

288 GB HBM3e

288 GB HBM3e

MEMORY BANDWIDTH

MEMORY BANDWIDTH

8.0 TB/s

8.0 TB/s

FP 16 TENSOR

FP 16 TENSOR

2,300 TFLOPS (dense)

2,300 TFLOPS (dense)

FP 8 TENSOR

FP 8 TENSOR

4,500 TFLOPS (dense)

4,500 TFLOPS (dense)

TDP

TDP

1000W

1000W

FORM FACTOR

FORM FACTOR

OAM

OAM

INTERCONNECT

INTERCONNECT

Infinity Fabric 4

Infinity Fabric 4

ARCHITECTURE

ARCHITECTURE

CDNA 4

CDNA 4

Partner Network

AGGREGATED ACROSS LEADING NEOCLOUDS

Compute Exchange aggregates reserved capacity from a verified network of leading AI-native cloud providers and hyperscalers. All partners undergo identity, capacity, SLA, and operational verification before quotes surface on the network.

You receive a normalized comparison across providers in a single quote response rather than evaluating each neocloud's contract structure, billing model, and SLA terms in isolation. Compute Exchange stays neutral; we do not operate compute capacity ourselves.

WORKLOAD FIT

RESERVED MI350X

USE CASES

01

Frontier-scale inference

02

405B-class single-device LLM serving

03

Memory-bound multi-modal inference

04

Long-context generative AI workloads

WHY RESERVE

RESERVED MI350X

VS ON-DEMAND

MI350X on-demand availability is exceptionally limited through Q2 2026 with launch-era rates of $14–$18 per GPU-hour. A reserved commit is effectively required to secure capacity at scale and to lock in pricing before the rate compression that typically follows initial supply ramp. For organizations serving 405B-parameter models on a single GPU, the 288GB HBM3e memory advantage versus H200 delivers material cost-per-token improvements.

FREQUENTLY ASKED QUESTIONS

RESERVED MI350X

RESERVED MI350X

KEY QUESTIONS

What drives reserved MI350X pricing?

Why is reserved MI350X supply allocation-constrained in Q1 2026?

How does MI350X compare to NVIDIA B200?

What term lengths and commitment structures are available for reserved MI350X?

READY TO RESERVE?

GET A LIVE

GET A LIVE

RESERVED MI350X

RESERVED MI350X

QUOTE

QUOTE

Compute Exchange returns indicative pricing within 24 hours, anchored to your specific quantity, region, and condition. We do not publish active counterparty listings.

COMPUTE

EXCHANGE

The transparent GPU marketplace for AI infrastructure. Built for builders.

ALL SYSTEMS OPERATIONAL

© 2026 COMPUTE EXCHANGE

BUILT FOR THE AI ERA

COMPUTE

EXCHANGE

The transparent GPU marketplace for AI infrastructure. Built for builders.

ALL SYSTEMS OPERATIONAL

© 2026 COMPUTE EXCHANGE

BUILT FOR THE AI ERA

COMPUTE

EXCHANGE

The transparent GPU marketplace for AI infrastructure. Built for builders.

ALL SYSTEMS OPERATIONAL

© 2026 COMPUTE EXCHANGE

BUILT FOR THE AI ERA