Talk to sales

Bare Metal NVIDIA HGX H200 GPUs: Dedicated machines for advanced AI workloads

Seriously powerful compute, purpose-built for your most demanding AI/ML projects. Talk to an expert to reserve capacity today.

© DigitalOcean, LLC.

Reserve Bare Metal GPU Capacity

Power your projects with the flexibility of Bare Metal GPUs

Platform Reliability

Get more reliable, high-performance compute for demanding workloads.

Support

Receive high-touch engineering support, helping to ensure smooth operations and peak server performance.

Secure, reliable infrastructure

Our ultra reliable high-performance cloud compute is monitored 24/7 so you can focus on scaling your AI model training to an entire cluster powered by the latest NVIDIA GPUs.

Pricing

Enjoy cost-efficient pricing that includes storage, and favorably compares to other providers.

Availability

Bare Metal GPUs are available in New York, USA and Amsterdam, Netherlands, with more data centers coming soon.

Have a complex setup or additional questions around pricing?

Contact our experts for more information and personalized guidance on how we can meet your specific workload requirements.

Reserve Bare Metal Capacity

NVIDIA HGX H200

GPU Count
8

GPU Ram
1,128 GB

CPU (x2)
Intel® Xeon® Platinum 8592+

System RAM
2,048 GiB

NVMe Storage
61.44 TiB

Frequently asked questions


A Bare Metal GPU is a physical server equipped with powerful GPUs, like the NVIDIA H200. This gives you full control over the hardware so you can install your preferred operating system. Bare Metal GPUs are isolated from other servers, providing excellent security. The server can be used for AI/ML tasks like model training, fine-tuning, and inference.


Full control: Bare Metal provides dedicated, single-tenant hardware, offering full control over the operating system, software, and configurations.
High performance: It delivers peak GPU performance, ideal for high-scale AI/ML model training, complex computations, and real-time inference.
Customizability: You can deeply customize the hardware and software environment, enabling tailored setups like Kubernetes clusters or other custom orchestration needs.
Resource isolation: It offers complete isolation from other users, eliminating the risks of 'noisy neighbors' and helping to ensure security and privacy.
Scalability: Bare Metal can be configured for single-node or multi-node setups, allowing for large-scale distributed training and high-performance workloads.


Bare Metal works by providing a physical, dedicated server where you can install and run your applications. This setup is particularly useful if you need stable, high-performance infrastructure for demanding workloads like AI/ML model training or custom orchestration.


Bare Metal supports a wide range of use-cases, including:

AI/ML workloads: Model training, fine-tuning, and inference, especially for large-scale data processing.
Custom orchestrations: Technologies like Kubernetes for containerized environments or other custom setups for complex applications.
High-performance computing: Applications that require dedicated resources, such as simulations, scientific computations, and real-time data processing.

Take advantage of our promotional prices! Try out our on-demand H100 GPUs at our promotional prices. Save more by reserving instances for prices as low as $2.50/GPU/hr

Build epic AI/ML applications with NVIDIA H200 GPUs

Whether you’re training large-scale models, fine-tuning AI systems, or running real-time inference, NVIDIA H200 GPUs provide the power and flexibility to accelerate AI applications.

These machines deliver cutting-edge performance needed for modern AI workloads. NVIDIA H200 GPUs provide faster memory bandwidth, enhanced Tensor Core performance, and more efficient real-time inference, making them the ideal choice for large-scale AI deployments.

Key benefits of NVIDIA H200 GPUs include:

  • Faster Model Training: Handle larger batch sizes and improve memory efficiency, speeding up training times.

  • Efficient Fine-Tuning: Optimize foundation models with lower compute overhead and better latency.

  • Enhanced Real-Time Inference: Run transformer models with faster inference and better power efficiency.

Reserve capacity

From the creator of Tabnine

Supermaven's AI code completion tool boosts coding speed and efficiency with high-quality suggestions, in-editor AI chat, and a large context window of up to 1 million tokens. Powered by their advanced LLM, Babble, it's 2.5x larger than their previous model.

Supermaven

Moonvalley AI specializes in generative media, creating cinematic videos and animations from text and image prompts. Their technology enables fast, high-definition, 16:9 video production, reducing typical production time and effort. With over 300,000 users, they operate through a Discord server and Easy With AI.

Moonvalley

How our customers use Bare Metal GPUs