Gcore named a Leader in the GigaOm Radar for AI Infrastructure!Get the report
  1. Home
  2. Blog
  3. NVIDIA GPUs: H100 vs. A100 | a detailed comparison

NVIDIA GPUs: H100 vs. A100 | a detailed comparison

  • By Gcore
  • 6 min read
NVIDIA GPUs: H100 vs. A100 | a detailed comparison

In 2022, NVIDIA released the H100, marking a significant addition to its GPU lineup. Designed to both complement and compete with the A100 model, the H100 received major updates in 2024, including expanded memory configurations with HBM3, enhanced processing features like the Transformer Engine for accelerated AI training, and broader cloud availability.

While the H100 leads in performance, market shifts in 2025 have made it an increasingly popular choice for AI workloads, particularly since H100 cloud pricing has dropped significantly due to enhanced availability. This pricing shift reduces the A100’s former cost advantage, making the H100’s superior performance (2-3x faster for most workloads) a critical consideration for many organizations.

However, the A100 still plays a vital role in broader, mixed-use environments and legacy deployments where compatibility and diverse workload support remain essential. Businesses are increasingly adopting a hybrid GPU strategy, leveraging both H100 and A100 instances to optimize for cost, availability, and performance.

Both GPUs remain highly capable for computation-intensive tasks like machine learning and scientific calculations. This article provides a detailed comparison of the H100 and A100, focusing on their performance metrics and suitability for specific workloads so you can decide which is best for your use case.

What are the performance differences between A100 and H100?

According to benchmarks by NVIDIA and independent parties, the H100 offers double the computation speed of the A100. This performance boost has two major implications:

  • Engineering teams can iterate faster if workloads take half the time to complete.
  • Even though the H100 costs about twice as much as the A100, the overall expenditure via a cloud model could be similar if the H100 completes tasks in half the time because the H100’s price is balanced by its processing time.

To compare the A100 and H100, we need to first understand what the claim of “at least double” the performance means. Then, we’ll discuss how it’s relevant to specific use cases, and finally, turn to whether you should pick the A100 or H100 for your GPU workloads.

Interpreting NVIDIA’s benchmarks

Let’s start by looking at NVIDIA’s own benchmark results, which you can see in Figure 1. They compare the H100 directly with the A100.

Figure 1: NVIDIA GPU performance comparison showing that H100 exceeds A100 performance by a factor of 1.5x to 6x.

The benchmarks comparing the H100 and A100 are based on artificial scenarios, focusing on raw computing performance or throughput without considering specific real-world applications. In reality, different data formats may experience varying levels of speed improvements, so it’s essential to work with your engineering team or software vendor to determine how your specific workload might benefit from the H100’s enhancements.

The charts in Figure 2 show a practical example of training GPT-3 with an A100 compared to an H100.

Figure 2: GPT-3 training performance comparison: NVIDIA A100 least performant, H100 moderate, and H100 + NVLink Switch best.

In this example, clusters equipped with the A100 and H100 were used to train two LLMs (large language models.) The results showed notable speed improvements, especially when the software was optimized for the H100, such as by using the FP8 data format. However, the standout feature was the new NVLink Switch System, which enabled the H100 cluster to train these models up to nine times faster than the A100 cluster. This significant boost suggests that the H100’s advanced scaling capabilities could make training larger LLMs feasible for organizations previously limited by time constraints.

These numbers are impressive, but they come from NVIDIA, which has a vested interest in promoting its latest (and more expensive) GPU. To get a complete picture, we should also look at what independent sources say.

What independent benchmarks reveal

NVIDIA sells GPUs, so they want them to look as good as possible. The GPT-3 training example above is impressive and likely accurate, but the amount of time spent optimizing the training software for these data formats is unknown. That’s why checking what independent sources say is always a good idea—you’ll get a better idea of how the comparison applies in a real-life, out-of-the-box scenario.

MosaicML compared the training of multiple LLMs on A100 and H100 instances. MosaicML is a managed LLM training and inference service; it doesn’t sell GPUs but rather a service, so it doesn’t care which GPU runs its workload as long as it is cost-effective. That means it has every reason to run realistic test cases, and therefore, its benchmarks could be more directly transferrable than NVIDIA’s own.

Table 1 shows the results for the different models.

ModelOptimized for H100Speedup over A100
1BNo2.2x
1BYes2.7x
3BNo2.2x
3BYes2.8x
7BYes3.0x
30BYes3.3x
Table 1: MosaicML benchmark results

The smaller, unoptimized models achieved a respectable 2.2x speedup on the H100. However, the larger models that were optimized for the H100 showed more significant gains. Notably, the 30B model experienced a 3.3x increase in speed compared to the A100. Another LLM training benchmark for the H100 shows at least doubled performance compared to the A100.

While these numbers aren’t as impressive as NVIDIA claims, they suggest that you can get a speedup of two times using the H100 compared to the A100, without investing in extra engineering hours for optimization. If your goal is to increase the size of your LLMs, and you have an engineering team ready to optimize your code base, you can get even more performance from an H100.

What does the H100 offer that the A100 doesn’t?

The H100 introduces a new chip design and several additional features, setting it apart from its predecessor. Let’s explore these updates to assess whether your use case requires the new model.

Confidential computing

An exciting new privacy feature is the confidential computing (CC) environment. In addition to data encryption at rest (i.e., on a hard drive) and data encryption in transit (i.e., on a network), CC allows data encryption in use. If you’re handling private or confidential information and security compliance is of concern—like in the healthcare and financial industries—the H100’s CC feature could make it the preferred choice.

Tensor Memory Accelerator

The Tensor Memory Accelerator (TMA) is a new part of the H100 Hopper architecture that frees GPU threads from memory management tasks. Its introduction primarily enhances performance, representing a significant architectural shift rather than just an incremental improvement like adding more cores.

With the ever-increasing volume of training data required for reliable models, the TMA’s capability to seamlessly transfer large data sets without overloading the computation threads could prove to be a crucial advantage, especially as training software begins to fully use this feature. Thanks to its TMA, the H100 may prove itself to be a more futureproof option and a superior choice for large-scale AI model training.

Transformer Engine support

The H100’s Transformer Engine processes both floating-point and integer data simultaneously using mixed-precision calculations, combining FP8, FP16, and INT8 operations. This significantly reduces memory usage while improving computational efficiency. By dynamically adjusting precision levels, the H100 delivers up to 6x faster training of models like GPT-style transformers compared to the A100, enabling efficient large-scale model training.

How much more does the H100 cost?

The H100 is more expensive than the A100. Let’s look at a comparable on-demand pricing example created with the Gcore pricing calculator to see what this means in practice.

SpecsA100 ServerH100 Server
CPUs2x Intel Xeon 84682x Intel Xeon 8468
Memory2TB2TB
Block storage8x 3.84TB NVMe8x 3.84 TB NVMe
GPUs8x NVIDIA A100 80GB 800Gbit/s Infiniband8x NVIDIA H100 80GB 3200Gbit/s Infiniband
Cost16.483 €/h30.013 €/h
Table 2: Cloud GPU price comparison

The H100 is 82% more expensive than the A100: less than double the price. However, considering that billing is based on the duration of workload operation, an H100—which is between two and nine times faster than an A100—could significantly lower costs if your workload is effectively optimized for the H100.

Should you pick the A100 or the H100?

Picking the right GPU is clearlyn’t simple. Here are the factors worth considering when weighing up the A100 and H100 for your business’ workloads.

Cost efficiency

While the A100 typically costs about half as much to rent from a cloud provider compared to the H100, this difference may be offset if the H100 can complete your workload in half the time. Consult with your engineers or vendors to make sure that your specific GPU software won’t suffer any performance regressions, which could negate the cost benefits of the speedups.

Licensing costs

The software you plan to use with the GPUs has licensing terms that bind it to a specific GPU model. Licensing for software compatible with the A100 can be considerably less expensive than for the H100.

Use cases

The H100 is NVIDIA’s first GPU specifically optimized for machine learning, while the A100 offers more versatility, handling a broader range of tasks like data analytics effectively. If your primary focus is on training large language models, the H100 is likely to be the most cost-effective choice. If it’s anything other than LLMs, the A100 is worth serious consideration.

Power consumption

For on-premises operations, the H100 can consume up to 700W, compared to the A100’s maximum of 400W. Increased performance comes with higher energy demands and heat output, so check whether your infrastructure can support such requirements if you’re considering buying GPUs outright.

Availability

Not all cloud providers offer every GPU model. Due to overwhelming demand, H100 models have had availability issues. If your provider only offers one of these GPUs, your choice may be predetermined. However, depending on your relationship with the provider, you might find more competitive pricing for the A100. Gcore has both A100 and H100 in stock right now.

Access the H100 and A100 for training and inference with Gcore’s convenient, scalable AI solution

The H100 offers undisputed improvements in machine learning and scientific computing, including enhanced scaling through NVLink 4.0 and significant AI-specific upgrades. If you’re ready to optimize your workloads, the H100 will deliver better performance and ROI. However, if you require broader versatility, the A100 remains a reliable, cost-effective alternative.

With Gcore, you can access the industry’s most advanced GPUs—the NVIDIA H100, A100, and H200—on a scalable cloud platform. Experience unmatched flexibility, real-time deployment, and cost transparency.

Talk to us about your GPU requirements

Related articles

Deploy GPT-OSS-120B privately on Gcore

OpenAI’s release of GPT-OSS-120B is a turning point for LLM developers. It’s a 120B parameter model trained from scratch, licensed for commercial use, and available with open weights. This is a serious asset for serious builders.Gcore now supports private GPT-OSS-120B deployments via our Everywhere Inference platform. That means you can stand up your own endpoint in minutes, run inference at scale, and control the full stack, without API limits, vendor lock-in, or hidden usage fees. Just fast, secure, controlled deployment on your terms. Deploy now in three clicks or read on to learn more.Why GPT-OSS-120B is big news for buildersThis model changes the game for anyone developing AI apps, platforms, or infrastructure. It brings GPT-3-level reasoning to the open-source ecosystem and frees developers from closed APIs.With GPT-OSS-120B, you get:Full access to model weights and architectureSelf-hosting for maximum data control and privacySupport for fine-tuning and model editingOffline deployment for secure or air-gapped useMassive cost savings at scaleYou can deploy in any Gcore region (or leverage Gcore’s three-click serverless inference on your own infrastructure), route traffic through your own stack, and fully control load, latency, and logs. This is LLM deployment for real-world apps, not just playground prompts.How to deploy GPT-OSS-120B with Gcore Everywhere InferenceGcore Everywhere Inference gives you a clean path from open model to production endpoint. You can spin up a dedicated deployment in just three clicks. We offer configuration options to suit your business needs:Choose your location (cloud or on-prem)Integrate via standard APIs (OpenAI-compatible)Control usage, autoscale, and costsDeploying GPT-OSS-120B on Gcore takes just three clicks in the Gcore Customer Portal.There are no shared endpoints. You get dedicated compute, low-latency routing, and full control and observability.You can also bring your own trained variant if you’ve fine-tuned GPT-OSS-120B elsewhere. We’ll help you host it reliably, close to your users.Use cases: where GPT-OSS-120B fits bestCommercial GPTs still outperform OSS models on some general tasks, but GPT-OSS-120B gives you control, portability, and flexibility where it counts. Most importantly, it gives you the ability to build privacy-sensitive applications.Great fits include:Internal dev tools and copilotsRetrieval-augmented generation (RAG) pipelinesSecure, private enterprise assistantsData-sensitive, on-prem AI workloadsModels requiring full customization or fine-tuningIt’s especially relevant for finance, healthcare, government, and legal teams operating under strict compliance rules.Deploy GPT-OSS-120B todayWant to learn more about GPT-OSS-120B and why Gcore is an ideal provider for deployment? Get all the information you need on our dedicated page.And if you’re ready to deploy in just three clicks, head on over to the Gcore Customer Portal. GPT-OSS-120B is waiting for you in the Application Catalog.Learn more about deploying GPT-OSS-120B on Gcore

Announcing new tools, apps, and regions for your real-world AI use cases

Three updates, one shared goal: helping builders move faster with AI. Our latest releases for Gcore Edge AI bring real-world AI deployments within reach, whether you’re a developer integrating genAI into a workflow, an MLOps team scaling inference workloads, or a business that simply needs access to performant GPUs in the UK.MCP: make AI do moreGcore’s MCP server implementation is now live on GitHub. The Model Context Protocol (MCP) is an open standard, originally developed by Anthropic, that turns AI models into agents that can carry out real-world tasks. It allows you to plug genAI models into everyday tools like Slack, email, Jira, and databases, so your genAI can read, write, and reason directly across systems. Think of it as a way to turn “give me a summary” into “send that summary to the right person and log the action.”“AI needs to be useful, not just impressive. MCP is a critical step toward building AI systems that drive desirable business outcomes, like automating workflows, integrating with enterprise tools, and operating reliably at scale. At Gcore, we’re focused on delivering that kind of production-grade AI through developer-friendly services and top-of-the-range infrastructure that make real-world deployment fast and easy.” — Seva Vayner, Product Director of Edge Cloud and AI, GcoreTo get started, clone the repo, explore the toolsets, and test your own automations.Gcore Application Catalog: inference without overheadWe’ve upgraded the Gcore Model Catalog into something even more powerful: an Application Catalog for AI inference. You can still deploy the latest open models with three clicks. But now, you can also tune, share, and scale them like real applications.We’ve re-architected our inference solution so you can:Run prefill and decode stages in parallelShare KV cache across pods (it’s not tied to individual GPUs) from August 2025Toggle WebUI and secure API independently from August 2025These changes cut down on GPU memory usage, make deployments more flexible, and reduce time to first token, especially at scale. And because everything is application-based, you’ll soon be able to optimize for specific business goals like cost, latency, or throughput.Here’s who benefits:ML engineers can deploy high-throughput workloads without worrying about memory overheadBackend developers get a secure API, no infra setup neededProduct teams can launch demos instantly with the WebUI toggleInnovation labs can move from prototype to production without reconfiguringPlatform engineers get centralized caching and predictable scalingThe new Application Catalog is available now through the Gcore Customer Portal.Chester data center: NVIDIA H200 capacity in the UKGcore’s newest AI cloud region is now live in Chester, UK. This marks our first UK location in partnership with Northern Data. Chester offers 2000 NVIDIA H200 GPUs with BlueField-3 DPUs for secure, high-throughput compute on Gcore GPU Cloud, serving your training and inference workloads. You can reserve your H200 GPU immediately via the Gcore Customer Portal.This launch solves a growing problem: UK-based companies building with AI often face regional capacity shortages, long wait times, or poor performance when routing inference to overseas data centers. Chester fixes that with immediate availability on performant GPUs.Whether you’re training LLMs or deploying inference for UK and European users, Chester offers local capacity, low latency, and impressive capacity and availability.Next stepsExplore the MCP server and start building agentic workflowsTry the new Application Catalog via the Gcore Customer PortalDeploy your workloads in Chester for high-performance UK-based computeDeploy your AI workload in three clicks today!

Gcore recognized as a Leader in the 2025 GigaOm Radar for AI Infrastructure

Gcore recognized as a Leader in the 2025 GigaOm Radar for AI Infrastructure

We’re proud to share that Gcore has been named a Leader in the 2025 GigaOm Radar for AI Infrastructure—the only European provider to earn a top-tier spot. GigaOm’s rigorous evaluation highlights our leadership in platform capability and innovation, and our expertise in delivering secure, scalable AI infrastructure.Inside the GigaOm Radar: what’s behind the Leader statusThe GigaOm Radar report is a respected industry analysis that evaluates top vendors in critical technology spaces. In this year’s edition, GigaOm assessed 14 of the world’s leading AI infrastructure providers, measuring their strengths across key technical and business metrics. It ranks providers based on factors such as scalability and performance, deployment flexibility, security and compliance, and interoperability.Alongside the ranking, the report offers valuable insights into the evolving AI infrastructure landscape, including the rise of hybrid AI architectures, advances in accelerated computing, and the increasing adoption of edge deployment to bring AI closer to where data is generated. It also offers strategic takeaways for organizations seeking to build scalable, secure, and sovereign AI capabilities.Why was Gcore named a top provider?The specific areas in which Gcore stood out and earned its Leader status are as follows:A comprehensive AI platform offering Everywhere Inference and GPU Cloud solutions that support scalable AI from model development to productionHigh performance powered by state-of-the-art NVIDIA A100, H100, H200 and GB200 GPUs and a global private network ensuring ultra-low latencyAn extensive model catalogue with flexible deployment options across cloud, on-premises, hybrid, and edge environments, enabling tailored global AI solutionsExtensive capacity of cutting-edge GPUs and technical support in Europe, supporting European sovereign AI initiativesChoosing Gcore AI is a strategic move for organizations prioritizing ultra-low latency, high performance, and flexible deployment options across cloud, on-premises, hybrid, and edge environments. Gcore’s global private network ensures low-latency processing for real-time AI applications, which is a key advantage for businesses with a global footprint.GigaOm Radar, 2025Discover more about the AI infrastructure landscapeAt Gcore, we’re dedicated to driving innovation in AI infrastructure. GPU Cloud and Everywhere Inference empower organizations to deploy AI efficiently and securely, on their terms.If you’re planning your AI infrastructure roadmap or rethinking your current one, this report is a must-read. Explore the report to discover how Gcore can support high-performance AI at scale and help you stay ahead in an AI-driven world.Download the full report

Protecting networks at scale with AI security strategies

Network cyberattacks are no longer isolated incidents. They are a constant, relentless assault on network infrastructure, probing for vulnerabilities in routing, session handling, and authentication flows. With AI at their disposal, threat actors can move faster than ever, shifting tactics mid-attack to bypass static defenses.Legacy systems, designed for simpler threats, cannot keep pace. Modern network security demands a new approach, combining real-time visibility, automated response, AI-driven adaptation, and decentralized protection to secure critical infrastructure without sacrificing speed or availability.At Gcore, we believe security must move as fast as your network does. So, in this article, we explore how L3/L4 network security is evolving to meet new network security challenges and how AI strengthens defenses against today’s most advanced threats.Smarter threat detection across complex network layersModern threats blend into legitimate traffic, using encrypted command-and-control, slow drip API abuse, and DNS tunneling to evade detection. Attackers increasingly embed credential stuffing into regular login activity. Without deep flow analysis, these attempts bypass simple rate limits and avoid triggering alerts until major breaches occur.Effective network defense today means inspection at Layer 3 and Layer 4, looking at:Traffic flow metadata (NetFlow, sFlow)SSL/TLS handshake anomaliesDNS request irregularitiesUnexpected session persistence behaviorsGcore Edge Security applies real-time traffic inspection across multiple layers, correlating flows and behaviors across routers, load balancers, proxies, and cloud edges. Even slight anomalies in NetFlow exports or unexpected east-west traffic inside a VPC can trigger early threat alerts.By combining packet metadata analysis, flow telemetry, and historical modeling, Gcore helps organizations detect stealth attacks long before traditional security controls react.Automated response to contain threats at network speedDetection is only half the battle. Once an anomaly is identified, defenders must act within seconds to prevent damage.Real-world example: DNS amplification attackIf a volumetric DNS amplification attack begins saturating a branch office's upstream link, automated systems can:Apply ACL-based rate limits at the nearest edge routerFilter malicious traffic upstream before WAN degradationAlert teams for manual inspection if thresholds escalateSimilarly, if lateral movement is detected inside a cloud deployment, dynamic firewall policies can isolate affected subnets before attackers pivot deeper.Gcore’s network automation frameworks integrate real-time AI decision-making with response workflows, enabling selective throttling, forced reauthentication, or local isolation—without disrupting legitimate users. Automation means threats are contained quickly, minimizing impact without crippling operations.Hardening DDoS mitigation against evolving attack patternsDDoS attacks have moved beyond basic volumetric floods. Today, attackers combine multiple tactics in coordinated strikes. Common attack vectors in modern DDoS include the following:UDP floods targeting bandwidth exhaustionSSL handshake floods overwhelming load balancersHTTP floods simulating legitimate browser sessionsAdaptive multi-vector shifts changing methods mid-attackReal-world case study: ISP under hybrid DDoS attackIn recent years, ISPs and large enterprises have faced hybrid DDoS attacks blending hundreds of gigabits per second of L3/4 UDP flood traffic with targeted SSL handshake floods. Attackers shift vectors dynamically to bypass static defenses and overwhelm infrastructure at multiple layers simultaneously. Static defenses fail in such cases because attackers change vectors every few minutes.Building resilient networks through self-healing capabilitiesEven the best defenses can be breached. When that happens, resilient networks must recover automatically to maintain uptime.If BGP route flapping is detected on a peering session, self-healing networks can:Suppress unstable prefixesReroute traffic through backup transit providersPrevent packet loss and service degradation without manual interventionSimilarly, if a VPN concentrator faces resource exhaustion from targeted attack traffic, automated scaling can:Spin up additional concentratorsRedistribute tunnel sessions dynamicallyMaintain stable access for remote usersGcore’s infrastructure supports self-healing capabilities by combining telemetry analysis, automated failover, and rapid resource scaling across core and edge networks. This resilience prevents localized incidents from escalating into major outages.Securing the edge against decentralized threatsThe network perimeter is now everywhere. Branches, mobile endpoints, IoT devices, and multi-cloud services all represent potential entry points for attackers.Real-world example: IoT malware infection at the branchMalware-infected IoT devices at a branch office can initiate outbound C2 traffic during low-traffic periods. Without local inspection, this activity can go undetected until aggregated telemetry reaches the central SOC, often too late.Modern edge security platforms deploy the following:Real-time traffic inspection at branch and edge routersBehavioral anomaly detection at local points of presenceAutomated enforcement policies blocking malicious flows immediatelyGcore’s edge nodes analyze flows and detect anomalies in near real time, enabling local containment before threats can propagate deeper into cloud or core systems. Decentralized defense shortens attacker dwell time, minimizes potential damage, and offloads pressure from centralized systems.How Gcore is preparing networks for the next generation of threatsThe threat landscape will only grow more complex. Attackers are investing in automation, AI, and adaptive tactics to stay one step ahead. Defending modern networks demands:Full-stack visibility from core to edgeAdaptive defense that adjusts faster than attackersAutomated recovery from disruption or compromiseDecentralized detection and containment at every entry pointGcore Edge Security delivers these capabilities, combining AI-enhanced traffic analysis, real-time mitigation, resilient failover systems, and edge-to-core defense. In a world where minutes of network downtime can cost millions, you can’t afford static defenses. We enable networks to protect critical infrastructure without sacrificing performance, agility, or resilience.Move faster than attackers. Build AI-powered resilience into your network with Gcore.Check out our docs to see how DDoS Protection protects your network

Introducing Gcore for Startups: created for builders, by builders

Building a startup is tough. Every decision about your infrastructure can make or break your speed to market and burn rate. Your time, team, and budget are stretched thin. That’s why you need a partner that helps you scale without compromise.At Gcore, we get it. We’ve been there ourselves, and we’ve helped thousands of engineering teams scale global applications under pressure.That’s why we created the Gcore Startups Program: to give early-stage founders the infrastructure, support, and pricing they actually need to launch and grow.At Gcore, we launched the Startups Program because we’ve been in their shoes. We know what it means to build under pressure, with limited resources, and big ambitions. We wanted to offer early-stage founders more than just short-term credits and fine print; our goal is to give them robust, long-term infrastructure they can rely on.Dmitry Maslennikov, Head of Gcore for StartupsWhat you get when you joinThe program is open to startups across industries, whether you’re building in fintech, AI, gaming, media, or something entirely new.Here’s what founders receive:Startup-friendly pricing on Gcore’s cloud and edge servicesCloud credits to help you get started without riskWhite-labeled dashboards to track usage across your team or customersPersonalized onboarding and migration supportGo-to-market resources to accelerate your launchYou also get direct access to all Gcore products, including Everywhere Inference, GPU Cloud, Managed Kubernetes, Object Storage, CDN, and security services. They’re available globally via our single, intuitive Gcore Customer Portal, and ready for your production workloads.When startups join the program, they get access to powerful cloud and edge infrastructure at startup-friendly pricing, personal migration support, white-labeled dashboards for tracking usage, and go-to-market resources. Everything we provide is tailored to the specific startup’s unique needs and designed to help them scale faster and smarter.Dmitry MaslennikovWhy startups are choosing GcoreWe understand that performance and flexibility are key for startups. From high-throughput AI inference to real-time media delivery, our infrastructure was designed to support demanding, distributed applications at scale.But what sets us apart is how we work with founders. We don’t force startups into rigid plans or abstract SLAs. We build with you 24/7, because we know your hustle isn’t a 9–5.One recent success story: an AI startup that migrated from a major hyperscaler told us they cut their inference costs by over 40%…and got actual human support for the first time. What truly sets us apart is our flexibility: we’re not a faceless hyperscaler. We tailor offers, support, and infrastructure to each startup’s stage and needs.Dmitry MaslennikovWe’re excited to support startups working on AI, machine learning, video, gaming, and real-time apps. Gcore for Startups is delivering serious value to founders in industries where performance, cost efficiency, and responsiveness make or break product experience.Ready to scale smarter?Apply today and get hands-on support from engineers who’ve been in your shoes. If you’re an early-stage startup with a working product and funding (pre-seed to Series A), we’ll review your application quickly and tailor infrastructure that matches your stage, stack, and goals.To get started, head on over to our Gcore for Startups page and book a demo.Discover Gcore for Startups

Announcing a new AI-optimized data center in Southern Europe

Good news for businesses operating in Southern Europe! Our newest cloud regions in Sines, Portugal, give you faster, more local access to the infrastructure you need to run advanced AI, ML, and HPC workloads across the Iberian Peninsula and wider region. Sines-2 marks the first region launched in partnership with Northern Data Group, signaling a new chapter in delivering powerful, workload-optimized infrastructure across Europe. And Sines-3 expands capacity and availability for the region.Strategically positioned in Portugal, Sines-2 and Sines-3 enhance coverage in Southern Europe, providing a lower-latency option for customers operating in or targeting this region. With the explosive growth of AI, machine learning, and compute-intensive workloads, these new regions are designed to meet escalating demand with cutting-edge GPU and storage capabilities.You can activate Sines-2 and Sines-3 for GPU Cloud or Everywhere Inference today with just a few clicks.Built for AI, designed to scaleSines-2 and Sines-3 bring with them next-generation infrastructure features, purpose-built for today's most demanding workloads:NVIDIA H100 GPUs: Unlock the full potential of AI/ML training, high-performance computing (HPC), and rendering workloads with access to H100 GPUs.VAST NFS (file sharing protocol) support: Benefit from scalable, high-throughput file storage ideal for data-intensive operations, research, and real-time AI workflows.IaaS portfolio: Deploy Virtual Machines, manage storage, and scale infrastructure with the same consistency and reliability as in our flagship regions.Organizations operating in Portugal, Spain, and nearby regions can now deploy workloads closer to end users, improving application performance. For finance, healthcare, public sector, and other organisations running sensitive workloads that must stay within a country or region, Sines-2 and Sines-3 are easy ways to access state-of-the-art GPUs with simplified compliance. Whether you're building AI models, running simulations, or managing rendering pipelines, Sines-2 and Sines-3 offer the performance, capacity, availability, and proximity you need.And best of all, servers are available and ready to deploy today.Run your AI workloads in Portugal todayWith these new Sines regions and our partnership with Northern Data Group, we're making it easier than ever for you to run AI workloads at scale. If you need speed, flexibility, and global reach, we're ready to power your next AI breakthrough.Unlock the power of Sines-2 and Sines-3 today

Subscribe to our newsletter

Get the latest industry trends, exclusive insights, and Gcore updates delivered straight to your inbox.