- Bare Metal GPU: Dedicated physical servers without virtualization, offering maximum performance and full hardware control.
- Spot Bare Metal GPU: Discounted servers suitable for batch processing, experiments, and testing. Spot clusters provide the same hardware access as standard Bare Metal GPUs and may be reclaimed with 24 hours’ notice.
Cluster architecture
Each cluster consists of one or more dedicated bare-metal GPU servers. When creating a multi-node cluster, all servers are placed in the same private network and share an identical configuration, including the image, network settings, and file shares. For flavors with InfiniBand cards, high-speed inter-node networking is configured automatically. No manual network configuration is required for distributed training. The platform provides the infrastructure layer: GPU servers, networking, storage options, and secure access. This allows installing and running preferred frameworks for distributed training, job scheduling, or container orchestration. For multi-node workloads, configure SSH trust between nodes to enable distributed training frameworks. File shares provide shared storage for datasets and checkpoints across all nodes.Create a GPU cluster
To create a Bare Metal GPU cluster, complete the following steps in the Gcore Customer Portal.- In the Gcore Customer Portal, navigate to GPU Cloud.
- In the sidebar, expand GPU Clusters and select Bare Metal GPU Clusters.
- Click Create Cluster.
Step 1. Select region
In the Region section, select the data center location for the cluster.
GPU model availability and pricing vary by region. If a specific GPU model is required, check multiple regions for stock availability.
Step 2. Configure cluster capacity
Cluster capacity determines the hardware specifications for each node in the cluster. The available options depend on the selected region.-
In the Cluster capacity section, select the GPU Cluster type:
- Bare Metal GPU for dedicated physical servers
- Spot Bare Metal GPU for discounted, interruptible instances (available in select regions)
- Select the GPU Model. Available models (such as A100, H100, or H200) depend on the region.
- Enable or disable Show out of stock to filter available flavors.
- Select a flavor. Each flavor card displays GPU configuration, CPU type, RAM capacity, storage, network connectivity, pricing, and stock availability.

Step 3. Set the number of instances
In the Number of Instances section, specify how many servers to provision in the cluster.
After creation, the cluster can be resized. Scaling up adds nodes with the same configuration used at creation. Scaling down removes a random node—to delete a specific node, use the per-node delete action in the cluster details. Deleting the last node in a cluster deletes the entire cluster.
Step 4. Select image
The image defines the operating system and pre-installed software for cluster nodes.
- In the Image section, choose the operating system:
- Public: Pre-configured images with NVIDIA drivers and CUDA toolkit (recommended)
- Custom: Custom images uploaded to the account
- Note the default login credentials displayed below the image selector: username
ubuntu, SSH port22. These credentials are used to connect to the cluster after creation.
Step 5. Configure file share integration
File shares provide shared storage accessible from all cluster nodes simultaneously, allowing access to shared datasets, checkpoints, and outputs even if a cluster is deleted. They use NFS with a minimum size of 100 GiB, and the creation form displays this option only in regions where file shares are available. Full configuration details, including manual mounting procedures, are described in the file share documentation.
- Enable the File Share integration checkbox.
- Select an existing file share, or create a new one by specifying its name, size (minimum 100 GiB), and optional settings such as Root squash or Slurm compatibility.

- Specify the mount path for the file share on cluster nodes (default:
/home/ubuntu/mnt/nfs). Additional file shares can be attached by clicking Add File Share.
Step 6. Configure network settings
Network settings define how the cluster communicates with external services and other resources. At least one interface is required.
- In the Network settings section, configure the network interface:
| Type | Access | Use case |
|---|---|---|
| Public | Direct internet access with dynamic public IP | Development, testing, quick access to cluster |
| Private | Internal network only, no external access | Production workloads, security-sensitive environments |
| Dedicated public | Reserved static public IP | Production APIs, services requiring stable endpoints |
Step 7. Configure firewall settings (conditional)
Firewall settings appear only in regions where the hardware supports this feature (servers with Bluefield network cards). If this section does not appear, proceed to the next step.

Step 8. Configure SSH key
In the SSH key section, select an existing key from the dropdown or create a new one. Keys can be uploaded or generated directly in the portal. If generating a new key pair, save the private key immediately as it cannot be retrieved later.
Step 9. Set additional options
The Additional options section provides optional settings: user data scripts for automated configuration and metadata tags for resource organization.
Step 10. Name and create the cluster
The final step assigns a name to the cluster and initiates provisioning.
- In the GPU Cluster Name section, enter a name or use the auto-generated one.
- Review the Estimated cost panel on the right.
- Click Create Cluster.
Connect to the cluster
After the cluster is created, use SSH to access the nodes. The default username isubuntu.
<instance-ip-address> with the public or floating IP shown in the cluster details.
For instances with only private interfaces, connect through a bastion host or VPN, or use the Gcore Customer Portal console.
Verify cluster status
After connecting, verify that GPUs are available and drivers are loaded:Automating cluster management
The Customer Portal is suitable for creating and managing individual clusters. For automated workflows—such as CI/CD pipelines, infrastructure-as-code, or batch provisioning—use the GPU Bare Metal API. The API allows:- Creating and deleting clusters programmatically
- Scaling the number of instances in a cluster
- Querying available GPU flavors and regions
- Checking quota and capacity before provisioning