AI GPU Cloud Infrastructure

Gcore bare metal servers and virtual machines powered by NVIDIA A100 and H100 GPUs. Boost the productivity of your AI tasks with breakthrough performance!

See pricing
top image

Configurations and prices

H100 Configuration with Infiniband (3.2 Tbit/s): 2 Intel Xeon 8480+ / 2TB RAM / 8x3.84 TB NVMe / 8x Nvidia H100, IB 3.2 Tbit/s, 2x100Gbit/s Ethernet

H100 Price per GPU 1 Month Rolling 6 Months Reserved 12 Months Reserved 36 Months Reserved
Greater than 512 GPUs €3.30/hour €3.15/hour €3.05/hour €2.90/hour
128 to 512 GPUs €3.60/hour €3.30/hour €3.25/hour €3.10/hour
32 to 128 GPUs €3.70/hour €3.50/hour €3.40/hour €3.25/hour
Less than 32 GPUs €3.75/hour €3.60/hour €3.50/hour €3.35/hour

H100 Configuration without Infiniband: 2 Intel Xeon 8480+ / 2TB RAM / 8x3.84 TB NVMe / 8x Nvidia H100, 2x100Gbit/s Ethernet

H100 Price per GPU 1 Month Rolling 6 Months Reserved 12 Months Reserved 36 Months Reserved
Greater than 512 GPUs €2.71/hour €2.56/hour €2.46/hour €2.31/hour
128 to 512 GPUs €3.01/hour €2.71/hour €2.66/hour €2.51/hour
32 to 128 GPUs €3.11/hour €2.91/hour €2.81/hour €2.66/hour
Less than 32 GPUs €3.16/hour €3.01/hour €2.91/hour €2.76/hour

Specifically crafted for AI and HPC tasks, it boasts fourth-generation Tensor Cores and the Transformer Engine

Bare Metal

Flavor ID Server config GPUs GPU Memory Infiniband Interconnect (Gbit/s) Quantity Price
bm3-ai-large-h100-80-82 Intel Xeon 8480+ / 2TB RAM / 8x3.84 TB NVMe / 8x Nvidia H100 (No IB) 8xH100 80 GB without IB
€ 18.4 / 1 hour
bm3-ai-large-a100-40-42 Intel Xeon 8468 / 2 TB RAM / 4x3.84 TB NVMe / 4x Nvidia A100 / 200Gbit/s Infiniband 4xA100 40 GB 200
€ 15.52 / 1 hour
€ 11,174.4 / 1 month
€ 127,388.16 / 1 year
bm3-ai-large-a100-80-82 Intel Xeon 8468 / 2 TB RAM / 8x3.84 TB NVMe / 8x Nvidia A100 / 800Gbit/s Infiniband 8xA100 80 GB 800
€ 16.48 / 1 hour
€ 11,865.6 / 1 month
€ 135,267.84 / 1 year
bm3-ai-large-h100-80-82 Intel Xeon 8468 / 2TB RAM / 8x3.84 TB NVMe / 8x Nvidia H100 / 3200Gbit/s Infiniband 8xH100 80 GB 3200
€ 30.01 / 1 hour
€ 21,608.00 / 1 month
€ 246,331.2 / 1 year

Virtual Instance

Flavor ID Server config GPUs GPU Memory Infiniband Interconnect (Gbit/s) Quantity Price
g3-ai-24-232-1100-a100-80-124 vCPU / 232 GB RAM / 1100 GB NVMe / A100-1GPU 1xA100 80 GB 800
€ 2.06 / 1 hour
€ 1,483.2 / 1 month
€ 16,908.48 / 1 year
g3-ai-48-464-2200-a100-80-248 vCPU / 464 GB RAM / 2200 GB NVMe / A100-2GPU 2xA100 80 GB 800
€ 4.12 / 1 hour
€ 2,966.4 / 1 month
€ 33,816.96 / 1 year
g3-ai-96-1856-8800-a100-80-896 vCPU / 1856 GB RAM / 8800 GB NVMe / A100-8GPU 8xA100 80 GB 800
€ 16.48 / 1 hour
€ 11,865.6 / 1 month
€ 135,267.84 / 1 year
g3-ai-24-232-1100-h100-80-124 vCPU / 232 GB RAM / 1100 GB NVMe / H100-1GPU 1xH100 80 GB 3200
€ 3.75 / 1 hour
€ 2,701.0 / 1 month
€ 30,791.4 / 1 year
g3-ai-48-464-2200-h100-80-248 vCPU / 464 GB RAM / 2200 GB NVMe / H100-2GPU 2xH100 80 GB 3200
€ 7.5 / 1 hour
€ 5,402.0 / 1 month
€ 61,582.8 / 1 year
g3-ai-96-1856-8800-h100-80-896 vCPU / 1856 GB RAM / 8800 GB NVMe / H100-8GPU 8xH100 80 GB 3200
€ 30.01 / 1 hour
€ 21,608.0 / 1 month
€ 246,331.2 / 1 year

L40s Configuration: 2x Intel Xeon 8468 / 2TB RAM / 4x7.68TB NVMe RI SSD + 2x960SSD OS / 8x Nvidia L40S / 2x25Gbit/s Ethernet

L40s 1 Month Rolling 6 Months Reserved 12 Months Reserved 36 Months Reserved
Greater than 512 GPUs €1.64/hour €1.50/hour €1.41/hour €1.28/hour
128 to 512 GPUs €1.91/hour €1.64/hour €1.59/hour €1.46/hour
32 to 128 GPUs €2.00/hour €1.82/hour €1.73/hour €1.59/hour
Less than 32 GPUs €2.05/hour €1.91/hour €1.82/hour €1.69/hour

Scroll horizontally to view the table

*Prices do not include VAT.

Designed for AI and compute-intensive workloads

AI training

With thousands of processing cores, a graphics processing unit (GPU) can perform multiple matrix operations and calculations in parallel. As a result, GPUs complete AI training tasks much faster than traditional CPUs.

Deep learning

GPUs easily handle the high computational demands of deep neural networks and ​​recurrent neural networks, which are fundamental to developing complex deep learning models, including generative AI.

 

High-performance computing

Superior GPU performance is well suited for compute-intensive workloads, including dynamic programming algorithms, video rendering, and scientific simulations.

Data analytics

GPUs provide high memory bandwidth and efficient data transfer capabilities. This improves the processing and manipulation of large data sets, enabling faster analysis.

GPU champs

The NVIDIA A100 and latest H100 GPUs are at the forefront of the enterprise GPU market. Both are powerful
and versatile accelerators for a wide range of AI and high-performance computing (HPC) workloads.

gpu

A100 specs

  • Up to 249x higher AI inference performance over CPUs
  • Up to 20x higher performance than the previous generation of the NVIDIA GPU, V100
  • Tensor Core 3rd generation
  • Up to 80GB of HBM2e memory

H100 specs

  • Up to 4x higher performance than the A100 GPU for AI training on GPT-3
  • ​​Up to 7x higher performance than the A100 GPU for HPC applications
  • Tensor Core 4th generation
  • Up to 100GB of HBM3 memory
A100
H100

Ideal for AI frameworks

NVIDIA GPUs are great for running AI frameworks and tools that help to build, train, and deploy AI models.

tensor-flow
pytorch
keras
paddle
onnx
hf
chainer
tensor-rt
rapids
apache
jupyter
sci

Dedicated bare metal GPU servers or virtual GPU instances?

Сhoose what works for you!

Bare metal GPU servers

Bare metal servers provide direct access to the physical hardware, including the GPU. This means that all GPU resources are dedicated to you. Bare metal GPU gives you optimal performance for AI and compute-intensive workloads.

Virtual GPU instances

For the same configuration, GPUs on VMs may perform slightly slower than those on bare metal servers. But VMs offer easier management, scalability, and lower prices than bare metal GPU servers.

Managed Kubernetes with GPU worker nodes

Features like autoscaling and autohealing make Kubernetes ideal for dynamic workloads, including machine learning, video processing, and other compute-intensive tasks. With Gcore’s Managed Kubernetes, you can use Bare Metal and VMs with GPU as worker nodes (A100 and H100.) Simply utilize GPUs in your containers by requesting the custom GPU resource, just like you would request CPU or memory.

kubernetes
kuberneteskubernetes
Learn more about Managed Kubernetes

Take advantage of
Gcore Cloud solutions

AI IPU

Use Gcore’s AI cloud infrastructure powered by Graphcore IPUs to accelerate machine learning.

Bare metal servers

Deploy resource-intensive applications and services on high-performance physical servers.

Virtual machines

Leverage production-grade VMs designed for a wide range of workloads and predictable performance.

Managed Kubernetes

Provision, manage, and scale Kubernetes clusters with 99.9% SLA and support for bare metal nodes.

GCORE PARTNERS

  • nvidia-partner
  • dell-partner
  • intel-partner

Frequently Asked Questions

A graphics processing unit (GPU) is a specialized electronic circuit designed to improve the rendering of computer graphics. GPUs are used in various applications, including video games, 3D modeling, and AI training.

GPUs are designed for parallel processing, which means that they can execute multiple instructions at the same time. This is the main difference between GPUs and central processing units (CPUs); the latter executes instructions one at a time.

You will be charged for a specific configuration that you choose. If you purchase a separate GPU instance that is not part of a Kubernetes cluster, you will be charged for the corresponding VM or bare metal configuration. See the Configuration and pricing section above to learn more about our pricing.

Contact our sales team at sales@gcore.com with your desired new instance configuration. If you need help choosing a configuration, they’ll get back to you with the best solution for your request.

It depends on the type of instances you choose, bare metal or VMs. If you choose a bare metal server, all of its resources are dedicated to you.

If you choose a VM, you get virtual computing resources, including those of a GPU. The physical resources of the instance (server) are shared, but the virtual resources are not. You get access to the full amount of resources that you purchased.

After you purchase the GPU instance, it is up and running:

  • Within 3–5 minutes if it is a virtual machine
  • Within 15–20 minutes if it is a bare metal server

Yes. Fill out this form, and our sales team will contact you to discuss this option. Please note that at the end of your trial period, you will be switched to the standard pay-as-you-go plan.

Yes. Fill out this form and our sales team will contact you to discuss this option.

Yes. Fill out this form and our sales team will contact you to discuss the details and add you to our waiting list.

Contact us to get personalized offer

For more information about AI GPU Cloud Infrastructure, please fill out the form