NVIDIA · October 2022

GeForce RTX 4090

Founders Edition

The NVIDIA GeForce RTX 4090 Founders Edition is a high-end consumer GPU designed for enthusiasts and professionals. It is part of the Ada Lovelace architecture, offering significant performance improvements over previous generations. Targeted at gamers and content creators, it features advanced ray tracing and AI capabilities, making it ideal for demanding applications and next-gen gaming experiences.

GeForce RTX 4090 Founders Edition
VRAM
24GB GB
FP32 TFLOPS
82.6 TFLOPS
CUDA Cores
16384

Provider Marketplace

Cheapest
$0.55/hour
Starting from
Best Value
$0.59/hour
Starting from
Enterprise Choice
$1.60/month
Starting from

All Cloud Providers

3 Options available
Genesis Cloud favicon
On-DemandGlobal Availability
$0.55/ hour
Estimated Cost
Provision
RunPod favicon
On-DemandGlobal Availability
$0.59/ hour
Estimated Cost
Provision
Vast.ai favicon
On-DemandGlobal Availability
$1.60/ month
Estimated Cost
Provision

Compute Performance

FP640.66 TFLOPS TFLOPS
FP3282.6 TFLOPS TFLOPS
TF32Not Supported TFLOPS
FP16165.2 TFLOPS (Dense) TFLOPS
BF16Not Supported TFLOPS
FP8Not Supported TFLOPS
INT8329.8 TOPS (Dense) TOPS
INT4Not Supported TOPS

Architecture

MicroarchitectureAda Lovelace
Process NodeTSMC 4N
Die Size608 mm²
Transistors76.3B
Compute Units128 SMs
Tensor Cores4th Gen, 512 Tensor Cores
RT Cores3rd Gen, 128 RT Cores
Matrix EngineTensor Core
Base Clock2235 MHz
Boost Clock2520 MHz
Transformer Engine
Sparse AccelerationSupported (2:4 structured sparsity)
Dynamic PrecisionSupported (FP8/FP16/BF16/TF32)

Memory & VRAM

Memory TypeGDDR6X
Total Capacity24GB GB
Bandwidth1008GB/s
Bus Width384-bit
HBM Stacks
ECC Support
Unified MemoryYes (CUDA Unified Memory)
Compression
NUMA Awareness
Memory PoolingNot Supported

Connectivity & Scaling

InterconnectPCIe
GenerationPCIe Gen 4
IB Bandwidth32 GB/s
PCIe InterfacePCIe Gen 4 x16
CXL Support
TopologyPCIe peer-to-peer
Max GPUs/Node4
Scale-Out
GPUDirect RDMA
P2P Memory

Virtualization

MIG SupportNot Supported
MIG PartitionsN/A
SR-IOVNot Supported
vGPU ReadinessNot Supported
K8s ReadinessSupported via Device Plugin
GPU SharingTime-Slicing, MPS
Virt EfficiencyNear bare-metal (vendor claim)

Power & Efficiency

TDP450 W W
Peak Power480-500 W
Idle Power15-25 W
Perf / Watt0.53 TFLOPS FP32/W
PSU Required850 W (minimum recommended)
Connectors1x 16-pin (12VHPWR) or 3x 8-pin via adapter
Thermal LimitsGPU temperature limit: 83°C
EfficiencyN/A

Physical Design

Form FactorPCIe card
FHFLFull Height, Full Length
Slot Width3–3.5 slots
Dimensions304 mm x 137 mm x 61 mm
Weight2186 g
CoolingActive air (dual axial fans)
Rack DensityNot optimized for rack density

Thermals & Cooling

AirflowActive cooling (vendor-specific CFM)
Temp Range0°C to 45°C
ThrottlingThermal-based clock reduction at Tjunction limit
Noise Level
Liquid CoolingAir-cooled
DC HeatLow (workstation class)

Software Ecosystem

CUDACUDA 12.x supported
ROCmNot Supported
oneAPINot Supported
PyTorchCommunity supported
TensorFlowCommunity supported
JAXSupported via CUDA backend
HuggingFaceCommunity support
Triton ServerSupported
DockerCommunity images available
Compiler StackMature CUDA compiler stack
Kernel OptimStandard driver-based support
Driver StabilityRapid-release cadence

Server & Deployment

OEM AvailabilityTier-1 OEMs: Dell, HPE, Lenovo, Supermicro
PreconfiguredProfessional workstations and specialized rack-mount kits
DGX/HGXNot applicable for DGX or HGX systems
Rack-ScaleTypically used in PCIe-based systems with standard networking options
Edge DeployLimited due to high TDP; more suitable for workstation environments
Ref ArchitecturesNVIDIA RTX Studio, Omniverse Enterprise

System Compatibility

CPU PairingHigh-performance desktop or workstation CPU recommended (e.g., Intel Core i9, AMD Ryzen 9, Threadripper class)
NUMAStandard NUMA behavior
Required PCIePCIe Gen 4 x16 recommended
MotherboardFull-length PCIe x16 slot, ATX/E-ATX form factor with adequate clearance
Rack PowerTDP up to 450W; low rack density advised; ensure sufficient PSU and cooling
BIOS LimitsResizable BAR and Above 4G decoding recommended; SR-IOV Not Supported
CXL ReadyNot Supported
OS CompatWindows 10/11 supported; major Linux distributions (RHEL, Ubuntu LTS) supported

Benchmarks & Throughput

Structured Sparsity

Not Supported

Transformer Throughput

Not Supported

Multi-GPU Scalability

Scaling Efficiency

Single GPUThe GeForce RTX 4090 Founders Edition offers high single-GPU performance with its substantial CUDA core count and high memory bandwidth.
2-GPUScaling is limited by PCIe Gen4 bandwidth, with potential contention on the PCIe lanes affecting P2P communication.
4-GPUScaling is further constrained by PCIe bandwidth limitations, with increased contention and reduced efficiency in P2P transfers.
8-GPUScaling efficiency is significantly impacted by PCIe lane contention, with diminishing returns due to the lack of NVLink support.
64+ GPUInfiniBand or high-speed Ethernet is necessary to mitigate overhead, but PCIe limitations and lack of NVLink severely hinder scalability at this level.

Scaling Characteristics

Cross-Node LatencyCross-node communication relies on GPUDirect RDMA, but PCIe bottlenecks can introduce latency, especially without NVLink.
Network BottlenecksThe primary bottleneck is the Host-to-Device bridge due to PCIe bandwidth limitations and the absence of NVLink.
ParallelismSupports Data and Model Parallelism, but Pipeline and Tensor Parallelism are limited by PCIe bandwidth and lack of NVLink.

Workload Readiness

LLM Training

The GeForce RTX 4090, based on the Ada Lovelace architecture, is suitable for training models up to 70B parameters in a single-node setup due to its high VRAM capacity of 24GB. Multi-node setups can extend this capability further.

LLM Inference

With its 4th-gen Tensor cores, the RTX 4090 is highly efficient for inference tasks, providing excellent token-per-second performance and sufficient KV cache headroom for large models.

Vision Training

The RTX 4090 excels in vision training tasks due to its high CUDA core count and advanced architecture, making it ideal for complex image processing and deep learning workloads.

Diffusion Models

The GPU's substantial VRAM and Tensor core capabilities make it well-suited for diffusion model training and inference, handling large datasets and complex computations efficiently.

Multimodal AI

The RTX 4090's robust architecture supports multimodal AI tasks effectively, leveraging its high computational power and memory bandwidth to process diverse data types simultaneously.

Reinforcement Learning

The GPU's high throughput and efficient parallel processing make it a strong candidate for reinforcement learning workloads, enabling rapid simulation and model updates.

HPC / Simulation

While the RTX 4090 offers limited FP64 performance, it can still handle HPC simulations that do not heavily rely on double precision, benefiting from its overall computational power.

Scientific Computing

For scientific computing tasks that require high precision, the RTX 4090 may not be optimal due to its limited FP64 capabilities, but it can still perform well in less precision-critical applications.

Edge Inference

With a TDP of 450W, the RTX 4090 is not ideal for edge inference due to its high power consumption and large form factor, better suited for data center environments.

Real-Time Serving

The RTX 4090's high performance and low latency make it suitable for real-time AI serving, providing rapid inference capabilities for demanding applications.

Fine-Tuning

The GPU's 24GB VRAM supports full fine-tuning of large models efficiently, allowing for comprehensive updates to model parameters.

LoRA Efficiency

The RTX 4090 is highly efficient for LoRA fine-tuning, leveraging its advanced architecture to optimize parameter updates with lower VRAM requirements.

Market Authority

Research Citations

Moderate; GeForce RTX 4090 is cited in a growing number of arXiv and peer-reviewed research papers for deep learning and computer vision experiments, especially in academic and hobbyist contexts.

Community Benchmarks

Extensive; widely benchmarked in community forums (Reddit, Hacker News), enthusiast sites (AnandTech, Tom's Hardware), and YouTube channels for AI, gaming, and rendering workloads.

GitHub Support

High; numerous open-source repositories provide scripts, Dockerfiles, and configuration guides for optimizing deep learning frameworks (PyTorch, TensorFlow) on RTX 4090.

Key Strengths

Excels in high-resolution gaming and content creation tasks.

  • ·4K Gaming: Delivers exceptional performance in 4K gaming with ray tracing.
  • ·Content Creation: Accelerates rendering and video editing tasks.
  • ·AI Workloads: Supports AI development with Tensor Cores.

Limitations

High cost and power requirements are key considerations.

  • ·Cost: Premium pricing limits accessibility to enthusiasts.
  • ·Power Demand: High power draw necessitates a robust power supply.
  • ·Size: Large physical size may not fit in all cases.

Expert Insight

The GeForce RTX 4090 represents a strategic leap in AI compute. When comparing cloud providers, consider not just the hourly rate, but also the interconnect bandwidth (InfiniBand/NVLink) and regional availability which can significantly impact total cost of ownership for large-scale training.

Glossary Terms

FP32 TFLOPS
VRAM
TDP
Cores
Information updated daily. Cloud pricing subject to vendor availability.