The NVIDIA DGX B200 server, priced at $500,000, delivers unmatched AI performance but comes with a hefty price tag. Why is it so expensive, and is there a smarter, more affordable way to access its power?
Fortunately, innovative solutions like Novita AI now provide a more accessible and cost-effective way to harness the power of top-tier GPUs, including the B200, without the prohibitive upfront costs.
The price of the first NVIDIA DGX B200 8X bare metal server has been announced, with a selling price of $500,000.

What Makes the B200 Price so High?
Outstanding Performance

You can collect the specific details directly from the table above!
| Feature / Metric | NVIDIA L40 | NVIDIA L40S | NVIDIA A40 | NVIDIA A6000 | NVIDIA RTX 4090 | NVIDIA H100 | NVIDIA B200 |
| Architecture | Ada Lovelace | Ada Lovelace | Ampere | Ampere | Ada Lovelace | Hopper | Blackwell |
| CUDA Cores | 18,176 | 18,176 | 10,752 | 10,752 | 9,728 | 16,896 | System is 8-card |
| Tensor Cores | 568 (4th generation) | 568 (4th generation) | 336 (3rd generation) | 336 (3rd generation) | 4th generation (supports FP8) | 528 (4th generation + Transformer Engine) | 5th generation |
| RT Cores | 142 (3rd generation) | 142 (3rd generation) | 84 (2nd generation) | 84 (2nd generation) | Same as L40 | No RT Core | 4th generation |
| Memory | 48 GB ECC GDDR6 | 48 GB ECC GDDR6 | 48 GB GDDR6 ECC | 48 GB GDDR6 | 24 GB | 80–98 GB HBM3 | System with 8 cards, total 1,440 GB HBM3e |
| Memory Bandwidth | 864 GB/s | 864 GB/s | 696 GB/s | 768 GB/s | 1008 GB/s | 3.35 TB/s (HBM3) | System total 64 TB/s HBM3e |
| FP32 | 90 TFLOPS | 91.6 TFLOPS | 37.4 TFLOPS | 38.71 TFLOPS | 82‑100 TFLOPS | 66.9 TFLOPS | 600 TFLOPS (whole machine) |
| TF32 (Sparse) | 181-362 TFLOPS | 183-366 TFLOPS | 75-150 TFLOPS | 77.4-155 TFLOPS | 83-165 TFLOPS | 989-1.979 PFLOPS | 9,000-18,000 TFLOPS |
| FP8 | 362-724 TFLOPS (Sparse) | 733 TFLOPS (Sparse) | Not supported | Not supported | Supported | 3.958-7.91 PFLOPS | 20 PFLOPS (per card) |
| FP64 | 1.4 TFLOPS | 1.4 TFLOPS | – | 0.605 TFLOPS | 1.29 TFLOPS | 26‑34 TFLOPS (scalar) | Approx. 40 TFLOPS per card |
| Power (TDP) | 300 W | 350 W | 300 W | 300 W | ≈450 W | SXM5 up to 700 W | System ~14.3 kW |
| NVLink | No (PCIe Card) | No | Supported (dual card) | Supported | No | Supported (3.9 TB/s) | System 14.4 TB/s |
| MIG Support | No | No | No | No | No | Yes | No |
High-cost Energy
| GPU / System | Power Consumption (TDP / System Power) |
|---|---|
| L40 | 300 W |
| L40S | 350 W |
| A40 | 300 W |
| A6000 | 300 W |
| RTX 4090 | 450 W |
| H100 (SXM5) | 400–700 W (depending on form factor) |
| DGX B200 System | 14.3 kW |
Broad Applications
- AI Training & Inference
The B200 is optimized for large-scale AI, excelling in both training and inference.
- Training: With FP4 support and a second-gen Transformer Engine, it accelerates massive model training while reducing cost.
- Inference: Benchmarks (e.g., Llama 4 Maverick) show over 1,000 tokens/sec per user, enabling fast, multi-user inference.
- Graphics & Visualization
Though AI-focused, the B200 delivers strong graphics performance:
- Real-time Ray Tracing via 4th-gen RT Cores.
- SER 2.0 Shader Optimization improves complex rendering efficiency.
- Precision Computing
The B200 handles precision-intensive workloads across science and engineering:
- Supports FP4 to TF32 via 5th-gen Tensor Cores.
- Ideal for simulations, digital twins, and real-time analytics.
How to Run B200 at a very Good Price? Novita AI at $4.77/hour.
Novita AI is an AI cloud platform that offers developers an easy way to deploy AI models using our simple API, while also providing a affordable and reliable GPU cloud for building and scaling.
Novita AI offers good price of H200 SXM 141GB, H100 SXM 80GB, A100 SXM 80GB, RTX 6000 Ada 48GB, RTX 3090 24GB, RTX 4090 24GB, RTX 4090 24GB, L40S 48GB, RTX 5090 32GB.
Step 1: Log In and Access the GPU Bare Metal
Log in to your account and click on the GPU Bare Metal button.

Step2: Chooose Your GPU

Select the Device
- Device Name: Choose H100 SXM or B200 SXM.
- Region: United States of America.
- Configuration (for H100 SXM):
- 8 GPUs
- 2048 GB Memory
- 104vCPU/Node
- 15.36 TB Storage
- at $1.7./hour.
- Configuration (for B200 SXM):
- 8 GPUs
- 2304 GB Memory
- 144vCPU/Node
- 30.8 TB Storage
- at $4.77./hour.
Set the Quantity and Rental Duration
- Adjust the GPU Quantity field to match your needs. For example, select 8 GPUs.
- Choose the rental duration. For instance, set it to 1 month.
While the B200 offers top-tier AI capabilities, its cost is prohibitive for most. Cloud services like Novita AI let you use B200 GPUs for as little as $4.77 per hour—making advanced AI accessible to everyone, no purchase required.
Frequently Asked Questions
Cutting-edge hardware, massive memory, and industry-leading AI performance drive up the price.
The B200 is built for large-scale AI training and inference, high-performance graphics and visualization, and precision computing tasks like scientific simulations and digital twins. Its flexible architecture supports many advanced workloads.
Yes! Instead of buying the hardware outright, you can rent access to B200 GPUs through Novita AI’s cloud platform at just $4.77 per hour.
Novita AI is an AI cloud platform that offers developers an easy way to deploy AI models using our simple API, while also providing the affordable and reliable GPU cloud for building and scaling.
Recommended Reading
- L40S vs RTX 4090: The Right GPU Depends on Your Needs
- L40S on Novita AI: A Versatile GPU for AI, Graphics, and HPC
- B200 on Novita AI: Only $4.77/hr for Running DeepSeek R1!
Discover more from Novita
Subscribe to get the latest posts sent to your email.





