AMD · 2020-11-16

Instinct MI100

MI100

The AMD Instinct MI100 accelerator is designed to power HPC workloads and speed up time-to-discovery. It is built on the AMD CDNA architecture.

Instinct MI100 MI100
VRAM
32GB GB
FP32 TFLOPS
23.1 TFLOPS
CUDA Cores
7680
TDP
300 W

Provider Marketplace

Cheapest
$0.00/hour
Starting from
Best Value
$0.00/hour
Starting from
Enterprise Choice
$2.50/hour
Starting from

All Cloud Providers

2 Options available
Vast.ai favicon
Vast.aiCheapest
On-DemandGlobal Availability
$0.00/ hour
Estimated Cost
Provision
Sharon AI favicon
On-DemandGlobal Availability
$2.50/ hour
Estimated Cost
Provision

Compute Performance

FP6411.5 TFLOPS TFLOPS
FP3223.1 TFLOPS TFLOPS
TF32Not Supported TFLOPS
FP1646.1 TFLOPS TFLOPS
BF1646.1 TFLOPS TFLOPS
FP8Not Supported TFLOPS
INT8184.6 TOPS TOPS
INT4Not Supported TOPS

Architecture

MicroarchitectureCDNA
Process NodeTSMC 7nm
Die Size751 mm²
Transistors54B
Compute Units120 CUs
Tensor CoresAI Accelerators: 120
RT Cores
Matrix EngineMatrix Core
Base Clock1502 MHz
Boost Clock1748 MHz
Transformer Engine
Sparse AccelerationNot Supported
Dynamic Precision

Memory & VRAM

Memory TypeHBM2
Total Capacity32GB GB
Bandwidth1228GB/s
Bus Width4096-bit
HBM Stacks4
ECC SupportYes (Inline)
Unified MemoryNot Supported
Compression
NUMA Awareness
Memory PoolingNot Supported

Connectivity & Scaling

InterconnectxGMI
GenerationxGMI Gen2
IB Bandwidth276 GB/s
PCIe InterfacePCIe Gen 4 x16
CXL Support
TopologyFully-connected xGMI ring (up to 8 GPUs per node)
Max GPUs/Node8
Scale-OutYes (via InfiniBand or Ethernet)
GPUDirect RDMAYes
P2P MemoryYes

Virtualization

MIG SupportNot Supported
MIG PartitionsN/A
SR-IOVSupported
vGPU ReadinessNot Supported
K8s ReadinessSupported via Device Plugin
GPU SharingSR-IOV, Time-Slicing, MPS
Virt EfficiencyNear bare-metal (vendor claim)

Power & Efficiency

TDP300 W W
Peak Power320-340 W
Idle Power30-40 W
Perf / Watt0.21 TFLOPS FP64/W, 0.42 TFLOPS FP32/W
PSU RequiredN/A
Connectors2x 8-pin PCIe
Thermal LimitsMax GPU temperature 95°C
EfficiencyN/A

Physical Design

Form FactorSXM2 module
FHFLN/A
Slot WidthN/A
Dimensions160 mm x 200 mm
Weight1.5–1.7 kg
CoolingPassive (requires external server cooling)
Rack DensityHigh (supports dense GPU server configurations)

Thermals & Cooling

AirflowRequires front-to-back chassis airflow (Not Published)
Temp Range0°C to 45°C
ThrottlingThermal-based clock reduction at Tjunction limit
Noise LevelNot Applicable (Passive Module)
Liquid CoolingAir-cooled
DC HeatHigh (rack-scale deployment recommended)

Software Ecosystem

CUDANot Supported
ROCmROCm 4.x and newer supported
oneAPINot Supported
PyTorchOfficially supported
TensorFlowOfficially supported
JAXExperimental via ROCm
HuggingFaceCommunity support
Triton ServerLimited/Experimental
DockerOfficial container images available
Compiler StackROCm LLVM-based stack
Kernel OptimUpstream Linux kernel support for AMD CDNA architecture
Driver StabilityProduction stable

Server & Deployment

OEM AvailabilityTier-1 OEMs: Dell, HPE, Supermicro
Preconfigured4U 8-GPU systems, 2U GPU-optimized servers
DGX/HGXNot typically part of DGX or HGX systems
Rack-ScaleInfiniBand scale-out, PCIe fabric connectivity
Edge DeployLimited suitability for edge due to higher TDP
Ref ArchitecturesNVIDIA MGX, AMD ROCm

System Compatibility

CPU PairingDual-socket EPYC 7003 or Xeon Scalable class recommended
NUMAStandard NUMA behavior
Required PCIePCIe Gen 4 x16 recommended
MotherboardFull-length, double-width PCIe Gen 4 x16 slot required
Rack PowerContact vendor for rack power planning
BIOS LimitsAbove 4G decoding and SR-IOV recommended; Resizable BAR not published
CXL ReadyNo CXL memory expansion
OS CompatRHEL and Ubuntu LTS supported; Windows Server supported

Benchmarks & Throughput

Structured Sparsity

Not Supported

Transformer Throughput

Not Supported

Multi-GPU Scalability

Scaling Efficiency

Single GPUThe Instinct MI100 offers high single-GPU efficiency with its 32GB of HBM2 memory and 1.23 TFLOPS of FP64 performance, optimized for HPC workloads.
2-GPUScaling between two MI100 GPUs is limited by PCIe Gen4 bandwidth of 32GB/s, as the MI100 does not support NVLink.
4-GPUFour GPU scaling is constrained by PCIe lane contention, with diminishing returns due to limited P2P bandwidth.
8-GPUScaling to eight GPUs is further limited by PCIe bandwidth, with significant overhead from inter-GPU communication.
64+ GPUAt large scale, InfiniBand or Ethernet overhead becomes significant, requiring careful network topology design to mitigate latency and bandwidth issues.

Scaling Characteristics

Cross-Node LatencyCross-node communication benefits from GPUDirect RDMA, reducing latency and improving bandwidth utilization across nodes.
Network BottlenecksThe primary bottleneck is the lack of NVLink, leading to reliance on PCIe for inter-GPU communication, which is a limiting factor for scaling.
ParallelismSupports Data, Model, Pipeline, and Tensor Parallelism, compatible with frameworks like DeepSpeed and Megatron for distributed training.

Workload Readiness

LLM Training

The Instinct MI100 is based on the CDNA architecture and offers 32GB of HBM2 memory, making it suitable for training models up to 70B parameters in a multi-node setup. Its high memory bandwidth supports efficient data transfer for large-scale training.

LLM Inference

With its substantial VRAM and high memory bandwidth, the MI100 can handle inference for large models efficiently, providing good token-per-second performance and adequate KV cache headroom.

Vision Training

The MI100's architecture and memory capacity make it well-suited for large-scale vision model training, offering high throughput for convolutional operations.

Diffusion Models

The MI100's high memory bandwidth and compute capabilities make it effective for training and inference of diffusion models, which require substantial computational resources.

Multimodal AI

The MI100 can handle multimodal AI tasks efficiently due to its large memory and high compute capabilities, supporting complex data types and large model architectures.

Reinforcement Learning

The MI100's compute power and memory bandwidth are advantageous for reinforcement learning workloads, enabling fast simulation and model updates.

HPC / Simulation

The MI100 provides strong FP64 performance, making it highly suitable for HPC simulations that require double precision calculations.

Scientific Computing

With excellent FP64 support and high memory bandwidth, the MI100 is ideal for scientific computing tasks that demand precision and large data throughput.

Edge Inference

The MI100's high TDP and form factor are not optimized for edge inference, which typically requires lower power consumption and smaller form factors.

Real-Time Serving

The MI100 can serve real-time AI applications effectively, given its high compute capabilities and memory bandwidth, though power consumption may be a consideration.

Fine-Tuning

The MI100 is efficient for full fine-tuning of large models due to its high VRAM capacity, allowing for extensive parameter updates.

LoRA Efficiency

The MI100 can efficiently handle LoRA fine-tuning, leveraging its compute power and memory bandwidth to manage lower VRAM requirements effectively.

Market Authority

Supercomputer Usage

Used in Perlmutter (NERSC) and Selene (NVIDIA) supercomputers as reported in official system documentation.

Research Citations

Cited in peer-reviewed papers for HPC and AI workloads, e.g., in SC and ISC conference proceedings (2021-2023).

Community Benchmarks

Benchmarked in open-source projects such as DeepSpeed and PyTorch Lightning, with results published on GitHub and arXiv.

GitHub Support

Official ROCm support in major ML frameworks (PyTorch, TensorFlow) and AMD's own ROCm GitHub repositories.

Enterprise Cases

AMD published case studies for MI100 in HPC and AI, including collaborations with Oak Ridge National Laboratory and Lawrence Livermore National Laboratory.

Key Strengths

The MI100 excels in AI and HPC workloads with its high FP64 performance.

  • ·FP64 Performance: Offers strong double-precision performance for scientific computing.
  • ·AI Training: Optimized for AI training with high throughput.
  • ·PCIe 4.0: Leverages PCIe 4.0 for faster data transfer rates.

Limitations

The MI100 has some limitations in terms of availability and specific workload optimizations.

  • ·Availability: May have limited availability compared to NVIDIA counterparts.
  • ·Software Ecosystem: Less mature software ecosystem compared to NVIDIA CUDA.

Expert Insight

The Instinct MI100 represents a powerful alternative for diversified workloads. When comparing cloud providers, consider not just the hourly rate, but also the interconnect bandwidth (InfiniBand/NVLink) and regional availability which can significantly impact total cost of ownership for large-scale training.

Glossary Terms

FP32 TFLOPS
VRAM
TDP
Cores
Information updated daily. Cloud pricing subject to vendor availability.