Gaming industry under DDoS attack. Get DDoS protection now. Start onboarding
  1. Home
  2. Blog
  3. Introducing GPU VMs on NVIDIA AI infrastructure in Sines (EU): flexible, cost-efficient compute for AI workloads
AI
News

Introducing GPU VMs on NVIDIA AI infrastructure in Sines (EU): flexible, cost-efficient compute for AI workloads

  • March 30, 2026
  • 3 min read
Introducing GPU VMs on NVIDIA AI infrastructure in Sines (EU): flexible, cost-efficient compute for AI workloads

Some AI jobs require the full power and predictability of dedicated bare metal clusters. Others need something more agile: compute that can be sized up or down quickly, used for a burst of experimentation, powered down when idle, and spun back up when the next training run begins.

To support agile AI workloads, we’re introducing a new addition to Gcore’s AI infrastructure: GPU Virtual Machines on NVIDIA Hopper H100, launching first in Sines-3, our sovereign AI region in Portugal. This brings a new level of flexibility to the same infrastructure with NVIDIA Quantum InfiniBand fabric that is already available as Bare Metal on GPU Cloud.

Like with GPU Cloud bare metal and Gcore Everywhere Inference, you can deploy GPU VMs in just three clicks.

A more adaptable way to use NVIDIA Hopper GPUs

GPU VMs give teams a way to tap into Hopper performance without committing to long-running hardware. Instead of provisioning a full server, you can start with a single NVIDIA H100 GPU, scale to two or four, or jump to an eight-GPU VM when your workload requires it.

This makes it easier to match GPU capacity to the stage of your project, whether you're testing a training script, running a few fine-tuning cycles, or pushing through a heavier training window.

Everything runs on the same AI-optimized infrastructure as Gcore bare metal GPU instances, including high-bandwidth NVIDIA Quantum InfiniBand networking. The result is a VM environment that still feels like working on serious AI hardware, just with more elasticity built in.

Cutting idle costs without complexity

One of the biggest advantages of GPU VMs is how they behave when they’re not in use. Many AI workloads come in waves: bursts of experimentation, rapid iteration, then a quiet period while teams review results or prepare the next dataset.

With GPU VMs, when you power off the instance, GPU billing pauses automatically. Your volumes, IPs, and configuration remain intact, but the GPU meter stops running. You only pay for optional additional volume storage and IPs while paused. When you’re ready to start work again, simply restart the VMs as required without needing to set up or reconfigure.

For teams experimenting, fine-tuning, or running training jobs on a schedule, this can significantly reduce operational overhead. You don’t need to redesign your workflows to fit spot markets or preemptible machines; you simply power the VM off when it’s not needed and resume when you’re ready, subject to available capacity.

When GPU VMs make the most sense

Teams already using our bare metal GPU Cloud often tell us they handle two distinct types of workloads. On one side are long, uninterrupted training cycles that need maximum consistency and dedicated hardware. On the other are the day-to-day pieces of AI development: testing a hypothesis, trying out a new architecture, running a few epochs to validate an idea, or preparing a model for deployment.

GPU VMs are designed for the latter. They’re ideal when:

  • You want to iterate quickly without reserving hardware for days at a time
  • Your training or fine-tuning runs aren’t continuous
  • You’re validating training code before scaling it to dedicated nodes
  • You need short-term GPU capacity that fits within an existing workflow
  • You want to serve or test inference traffic before moving it to Everywhere Inference

Whether you’re an early/growth-phase AI startup looking for performant GPUs without the high fixed costs, an EU R&D lab needing sovereign infrastructure for burst PoCs and experiments, or a research institution seeking to run short-term, high-intensity fine-tuning runs on a budget, Gcore GPU VMs deliver the flexibility and cost efficiency you need.

Many teams use both Bare Metal and VM options side by side: VMs for experimentation and agile development, and bare metal for the heavy lifting once the model is ready for large-scale training. Gcore makes this easy with 3-click BM and VM deployments in a single platform.

Part of a larger AI infrastructure roadmap

Alongside this new VM solution, we are also now offering spot bare metal GPUs. Spot instances are ideal for interruptible, high-throughput jobs that require uncompromising performance at a cost-effective price. With Gcore, you can combine GPU solutions.

If your team is exploring how to mix VMs and bare metal, needs help calibrating the right GPU shapes, or wants to test workloads ahead of a migration, we're here to help.

Deploy a workload on Gcore GPU VMs

Try Gcore AI

Gcore all-in-one platform: cloud, AI, CDN, security, and other infrastructure services.

Related articles

Gcore Everywhere AI evolves to full-lifecycle management with Slurm, Jupyter, and token-based inference integrations

AI adoption has a fragmentation problem. Organizations routinely stitch together separate tools for development, training, and serving, each with its own infrastructure, access controls, and operational overhead. The result is a patchwork t

Introducing faster, lower-cost LLM inference with NVIDIA Dynamo

Imagine if you could click a button and suddenly your GPUs increase their throughput by 6x. Or reduce latency by 2x. Or route inference requests seamlessly across different GPU types.That's the experience we're bringing to our inference cus

New AI inference models on Application Catalog: translation, agents, and flagship reasoning

We’ve expanded our AI inference Application Catalog with three new state-of-the-art models, covering massively multilingual translation, efficient agentic workflows, and high-end reasoning. All models are live today via Everywhere Inference

New AI inference models available now on Gcore

We’ve expanded our Application Catalog with a new set of high-performance models across embeddings, text-to-speech, multimodal LLMs, and safety. All models are live today via Everywhere Inference and Everywhere AI, and are ready to deploy i

Introducing Gcore Everywhere AI: 3-click AI training and inference for any environment

For enterprises, telcos, and CSPs, AI adoption sounds promising…until you start measuring impact. Most projects stall or even fail before ROI starts to appear. ML engineers lose momentum setting up clusters. Infrastructure teams battle to b

Introducing AI Cloud Stack: turning GPU clusters into revenue-generating AI clouds

Enterprises and cloud providers face major roadblocks when trying to deploy GPU infrastructure at scale: long time-to-market, operational inefficiencies, and difficulty bringing new capacity to market profitably. Establishing AI environment

Subscribe to our newsletter

Get the latest industry trends, exclusive insights, and Gcore updates delivered straight to your inbox.