Explore Gcore Cloud Edge Services
Service group
- Service group
- Learn more →
Virtual instances
Deploy projects faster with ready-made virtual instance images.
- Learn more →
Application marketplace
Get access to ready-made systems and applied services or suggest your own product.
Powerful computing resources and services in 50+ locations around the world
Deploy instances in optimal locations around the world within 15 minutes. No need to search for Tier IV data centers, purchase advanced hardware, and worry about upgrading.
Test your Latency
Solutions for all sectors
Gaming
Create games and integrate any online entertainment solution of any complexity.
A simple way to control your Cloud
The Cloud control panel is integrated with other infrastructure products: CDN, Streaming, Storage, DDoS Protection, and DNS Hosting.

Manage all Cloud services with one simple and efficient panel.
Read about new locations, configurations, and features in our blog
More articles →- Outpacing cloud‑native threats: How to secure distributed workloads at scale
The cloud never stops. Neither do the threats.Every shift toward containers, microservices, and hybrid clouds creates new opportunities for innovation…and for attackers. Legacy security, built for static systems, crumbles under the speed, scale, and complexity of modern cloud-native environments.To survive, organizations need a new approach: one that’s dynamic, AI-driven, automated, and rooted in zero trust.In this article, we break down the hidden risks of cloud-native architectures and show how intelligent, automated security can outpace threats, protect distributed workloads, and power secure growth at scale.The challenges of cloud-native environmentsCloud-native architectures are designed for maximum flexibility and speed. Applications run in containers that can scale in seconds. Microservices split large applications into smaller, independent parts. Hybrid and multi-cloud deployments stretch workloads across public clouds, private clouds, and on-premises infrastructure.But this agility comes at a cost. It expands the attack surface dramatically, and traditional perimeter-based security can’t keep up.Containers share host resources, which means if one container is breached, attackers may gain access to others on the same system. Microservices rely heavily on APIs to communicate, and every exposed API is a potential attack vector. Hybrid cloud environments create inconsistent security controls across platforms, making gaps easier for attackers to exploit.Legacy security tools, built for unchanging, centralized environments, lack the real-time visibility, scalability, and automated response needed to secure today’s dynamic systems. Organizations must rethink cloud security from the ground up, prioritizing speed, automation, and continuous monitoring.Solution #1: AI-powered threat detection forsmarter defensesModern threats evolve faster than any manual security process can track. Rule-based defenses simply can’t adapt fast enough.The solution? AI-driven threat detection.Instead of relying on static rules, AI models monitor massive volumes of data in real time, spotting subtle anomalies that signal an attack before real damage is done. For example, an AI-based platform can detect an unauthorized process in a container trying to access confidential data, flag it as suspicious, and isolate the threat within milliseconds before attackers can move laterally or exfiltrate information.This proactive approach learns, adapts, and neutralizes new attack vectors before they become widespread. By continuously monitoring system behavior and automatically responding to abnormal activity, AI closes the gap between detection and action, critical in cloud-native, regulated environments where even milliseconds matter.Solution #2: Zero trust as the new security baseline“Trust but verify” no longer cuts it. In a cloud-native world, the new rule is “trust nothing, verify everything”.Zero-trust security assumes that threats exist both inside and outside the network perimeter. Every request—whether from a user, device, or application—must be authenticated, authorized, and validated.In distributed architectures, zero trust isolates workloads, meaning even if attackers breach one component, they can’t easily pivot across systems. Strict identity and access management controls limit the blast radius, minimizing potential damage.Combined with AI-driven monitoring, zero trust provides deep, continuous verification, blocking insider threats, compromised credentials, and advanced persistent threats before they escalate.Solution #3: Automated security policies for scalingprotectionManual security management is impossible in dynamic environments where thousands of containers and microservices are spun up and down in real time.Automation is the way forward. AI-powered security policies can continuously analyze system behavior, detect deviations, and adjust defenses automatically, without human intervention.This eliminates the lag between detection and response, shrinks the attack window, and drastically reduces the risk of human error. It also ensures consistent security enforcement across all environments: public cloud, private cloud, and on-premises.For example, if a system detects an unusual spike in API calls, an automated security policy can immediately apply rate limiting or restrict access, shutting down the threat without impacting overall performance.Automation doesn’t just respond faster. It maintains resilience and operational continuity even in the face of complex, distributed threats.Unifying security across cloud environmentsSecuring distributed workloads isn’t just about having smarter tools, it’s about making them work together. Different cloud platforms, technologies, and management protocols create fragmentation, opening cracks that attackers can exploit. Security gaps between systems are as dangerous as the threats themselves.Modern cloud-native security demands a unified approach. Organizations need centralized platforms that pull real-time data from every endpoint, regardless of platform or location, and present it through a single management dashboard. This gives IT and security teams full, end-to-end visibility over threats, system health, and compliance posture. It also allows security policies to be deployed, updated, and enforced consistently across every environment, without relying on multiple, siloed tools.Unification strengthens security, simplifies operations, and dramatically reduces overhead, critical for scaling securely at cloud-native speeds. That’s why at Gcore, our integrated suite of products includes security for cloud, network, and AI workloads, all managed in a single, intuitive interface.Why choose Gcore for cloud-native security?Securing cloud-native workloads requires more than legacy firewalls and patchwork solutions. It demands dynamic, intelligent protection that moves as fast as your business does.Gcore Edge Security delivers robust, AI-driven security built for the cloud-native era. By combining real-time AI threat detection, zero-trust enforcement, automated responses, and compliance-first design, Gcore security solutions protect distributed applications without slowing down development cycles.Discover why WAAP is essential for cloud security in 2025
26 Jun 2025 - Edge Cloud news: more regions and volume options available
At Gcore, we’re committed to delivering high-performance, globally distributed infrastructure that adapts to your workloads—wherever they run. This month, we’re excited to share major updates to our Edge Cloud platform: two new cloud IaaS regions in Europe and expanded storage options in São Paulo.New IaaS regions in Luxembourg and Portugal available nowLuxembourg‑3 and Sines‑2 mark the next step in the Gcore mission to bring compute closer to users. From compliance-focused deployments in Central Europe to GPU‑powered workloads in the Iberian Peninsula, these new regions are built to support diverse infrastructure needs at scale.Luxembourg‑3: expanding connectivity in Central EuropeWe’re expanding our European footprint by opening an additional IaaS point of presence (PoP) in Luxembourg. Strategically located in the heart of Europe, this region offers low-latency connectivity across the EU and is a strong compliance choice for data residency requirements.Here’s what’s available in Luxembourg‑3:Virtual Machines: High-performance, reliable, and scalable compute power for a wide range of workloads - with free egress traffic and pay-as-you-go billing for active instances only.Volumes: Standard, High IOPS, and Low Latency block storage for any workload profile.Load Balancers: Distribute traffic intelligently across instances to boost availability, performance, and fault tolerance.Managed Kubernetes: Fully managed Kubernetes clusters with automated provisioning, scaling, and updates optimized for production-ready deployments.Sines‑2, Portugal: a new hub for Southern Europe and a boost for AI workloadsWe’re also opening a brand-new location: Sines‑2, Portugal. This location enhances coverage across Southern Europe and boosts our AI and compute capabilities with more GPU availability.In addition to offering the same IaaS services as Luxembourg‑3, Sines‑2 also includes:H100 NVIDIA GPUs for AI/ML, high-performance computing, and rendering workloads.New VAST NFS Fileshare support for scalable, high-throughput file storage.This new region is ideal for organizations looking to deploy close to the Iberian Peninsula, reducing latency for regional users while gaining access to powerful GPU resources.Enhanced volume types in São PauloVolumes are the backbone of any cloud workload. They store the OS, applications, and essential data for your virtual machines. Developers and businesses building latency-sensitive or I/O-intensive applications now have more options in the São Paulo-2 region, thanks to two newly added volume types optimized for speed and responsiveness:Low-latency volumesDesigned for applications where every millisecond matters, Low Latency Volumes are non-redundant block storage ideal for:ETCD clustersTransactional databasesOther real-time, latency-critical workloadsBy minimizing overhead and focusing on speed, this volume type delivers faster response times for performance-sensitive use cases. This block storage offers IOPS up to 5000 and an average latency of 300 microseconds.High-IOPS volumesFor applications that demand both speed and resilience, High IOPS Volumes offer a faster alternative to our Standard Volumes:Higher IOPS and increased throughputSuitable for high-traffic web apps, analytics engines, and demanding databasesThis volume type accelerates data-heavy workloads and keeps performance consistent under peak demand by delivering significantly higher throughput and IOPS. The block storage offers IOPS up to 9,000 and a 500 MB/s bandwidth limit.Ready to deploy with Gcore?These new additions help to fine-tune your performance strategy, whether you're optimizing for throughput, latency, or both.From scaling in LATAM to expanding into the EU or pushing performance at the edge, Gcore continues to evolve with your needs. Explore our new capabilities in Luxembourg‑3, Sines‑2, and São Paulo‑2.Discover more about Gcore Cloud Edge Services
07 May 2025 - 5 ways to keep gaming customers engaged with optimal performance
Nothing frustrates a gamer more than lag, stuttering, or server crashes. When technical issues interfere with gameplay, it can be a deal breaker. Players know that the difference between winning and losing should be down to a player’s skill, not lag, latency issues, or slow connection speed—and they want gaming companies to make that possible every time they play.And gamers aren’t shy about expressing their opinion if a game hasn’t met their expectations. A game can live or die by word-of-mouth, and, in a highly competitive industry, gamers are more than happy to spend their time and money elsewhere. A huge 78% of gamers have “rage-quit” a game due to latency issues.That’s why reliable infrastructure is crucial for your gaming offering. A solid foundation is good for your bottom line and your reputation and, most importantly, provides a great gaming experience for customers, keeping them happy, loyal, and engaged. This article suggests five technologies to boost player engagement in real-world gaming scenarios.The technology powering seamless gaming experiencesHaving the right technology behind the scenes is essential to deliver a smooth, high-performance gaming experience. From optimizing game deployment and content delivery to enabling seamless multiplayer scalability, these technologies work together to reduce latency, prevent server overloads, and guarantee fast, reliable connections.Bare Metal Servers provide dedicated compute power for high-performing massive multiplayer games without virtualization overhead.CDN solutions reduce download times and minimize patch distribution delays, allowing players to get into the action faster.Managed Kubernetes simplifies multiplayer game scaling, handling sudden spikes in player activity.Load Balancers distribute traffic intelligently, preventing server overload during peak times.Edge Cloud reduces latency for real-time interactions, improving responsiveness for multiplayer gaming.Let’s look at five real-world scenarios illustrating how the right infrastructure can significantly enhance customer experience—leading to smooth, high-performance gaming, even during peak demand.#1 Running massive multiplayer games with bare metal serversImagine a multiplayer FPS (first-person shooter gaming) game studio that’s preparing for launch and needs low-latency, high-performance infrastructure to handle real-time player interactions. They can strategically deploy Gcore Bare Metal servers across global locations, reducing ping times and providing smooth gameplay.Benefit: Dedicated bare metal resources deliver consistent performance, eliminating lag spikes and server crashes during peak hours. Stable connections and seamless playing are assured for precision gameplay.#2 Seamless game updates and patch delivery with CDN integrationLet’s say you have a game that regularly pushes extensive updates to millions of players worldwide. Instead of overwhelming origin servers, they can use Gcore CDN to cache and distribute patches, reducing download times and preventing bottlenecks.Benefit: Faster updates for players, reduced server tension, and seamless game launches and updates.#3 Scaling multiplayer games with Managed KubernetesAfter a big update, a game may experience a sudden spike in the number of players. With Gcore Managed Kubernetes, the game autoscales its infrastructure, dynamically adjusting resources to meet player demand without downtime.Benefit: Elastic, cost-efficient scaling keeps matchmaking fast and smooth, even under heavy loads.#4 Load balancing for high-availability game serversAn online multiplayer game with a global base requires low latency and high availability. Gcore Load Balancers distribute traffic across multiple regional server clusters, reducing ping times and preventing server congestion during peak hours.Benefit: Consistent, lag-free gameplay with improved regional connectivity and failover protection.#5 Supporting live events and seasonal game launchesIn the case of a gaming company hosting a global in-game event, attracting millions of players simultaneously, leveraging Gcore CDN, Load Balancers, and autoscaling cloud infrastructure can prevent crashes and provide a seamless and uninterrupted experience.Benefit: Players enjoy smooth, real-time participation while the infrastructure is stable under extreme load.Building customer loyalty with reliable gaming infrastructureIn a challenging climate, focusing on maintaining customer happiness and loyalty is vital. The most foolproof way to deliver this is by investing in reliable and secure infrastructure behind the scenes. With infrastructure that’s both scalable and high-performing, you can deliver uninterrupted, seamless experiences that keep players engaged and satisfied.Since its foundation in 2014, Gcore has been a reliable partner for game studios looking to deliver seamless, high-performance gaming experiences worldwide, including Nitrado, Saber, and Wargaming. If you’d like to learn more about our global infrastructure and how it provides a scalable, high-performance solution for game distribution and real-time games, get in touch.Talk to our gaming infrastructure experts
19 Mar 2025
What we guarantee
Premium technical support
Our experts will help with integration
GDPR
Compliance
24/7 availability
Real time monitoring and maintenance
DDoS protection
At the network and transport layers
SLA
99.95% with financial guarantees
Recent case studiesMore case studies
Funcom delivers the successful launch of Dune: Awakening in South America with Gcore
Founded in 1993, Funcom is a leading developer and publisher of online multiplayer and open-world games. Known for its rich storytelling and immersive universes, Funcom has developed acclaimed titles like Conan Exiles, The Secret World, and Anarchy Online. With its latest and most ambitious project, Dune: Awakening, Funcom is building an expansive open-world multiplayer survival game on a massive scale set in the iconic sci-fi universe of Dune.Launching Dune: Awakening with low-latency performance for South American playersIn preparation for the global launch of Dune: Awakening, Funcom faced a critical challenge: delivering a smooth, high-performance multiplayer experience for players in South America, a region often underserved by traditional infrastructure providers.With a large and passionate LATAM player base, the stakes were high. Funcom needed to deploy compute-intensive workloads capable of powering real-time gameplay and matchmaking with minimal latency, all while providing resilience against potential DDoS attacks during the launch window.Choosing Gcore for high-frequency compute power and managed orchestrationTo meet these infrastructure demands, Funcom partnered with Gcore to deploy:Bare Metal Servers configured with AMD Ryzen 9 9950x CPUs for high single-threaded performanceManaged Kubernetes clusters to orchestrate scalable multiplayer backend services on bare metal serversBuilt-in advanced DDoS Protection to secure critical launch infrastructureThe robust presence of Gcore in Latin America, supported by its global backbone and edge PoPs, made it possible for Funcom to deliver a high-quality experience to South American players comparable to what’s typically available in North America or Europe.The Gcore infrastructure in South America is purpose-built to support latency-sensitive workloads like online multiplayer gaming. With multi-terabit capacity in São Paulo, participation in IX.br (the region’s largest internet exchange), and private peering agreements with major ISPs such as Claro and TIM, Gcore ensures stable, low-latency connectivity across the region. Crucially, DDoS mitigation is handled locally, eliminating the need for long-haul traffic rerouting and enabling faster, more reliable protection at scale.The ability to directly deploy high-frequency bare metal nodes in the region has been a cornerstone of our South American launch strategy. Gcore allows us to reach players in regions where performance at this level is not usually possible.Stian Drageset, CFO & COO, FuncomGuaranteeing smooth operations with Kubernetes and low-latency infrastructureWith Gcore Managed Kubernetes, Funcom was able to dynamically manage containers across a cluster of powerful bare metal nodes, crucial for maintaining game state, matchmaking, and multiplayer interactions in real time. This setup enables flexible scaling in response to player demand, whether it spikes on launch day or ramps up as more players join.Thanks to Gcore’s managed services, our team can focus on game logic and player experience, not orchestration or hardware.Rui Casais, CEO, FuncomProving performance at scale during beta—and beyondAnticipation was already high leading up to the launch. During the invite-only beta weekend in May 2025, the game attracted nearly 40,000 concurrent players—a strong early signal of the momentum behind the title. Behind the scenes, Gcore supported Funcom with high-performance Bare Metal servers and Managed Kubernetes to provide uninterrupted performance at scale during this critical milestone. That success laid the groundwork for a smooth and stable full launch in South America.Monitoring results post-launchAs Dune: Awakening prepared for its launch, Funcom and Gcore closely monitored infrastructure performance and prepared for a high-concurrency environment. Post-launch data included:Reached the top ten most-played games on Steam globally within 24 hours of launch, climbing to number two within the first weekPeak of 142,000 concurrent players in the first couple of days, and 189,000 by the end of the weekExpanding into underserved gaming regionsThis deployment showcases how Gcore’s infrastructure helps game studios expand into emerging regions like South America, where consistent low-latency, high-frequency compute has traditionally been harder to access.South America is often seen as a “blue ocean” market in the gaming industry—vast, underserved, and perceived as difficult to serve due to infrastructure limitations. With a population of over 400 million, the region holds immense potential. Gcore makes it easy for publishers like Funcom to unlock that opportunity, delivering a seamless experience to players across LATAM without compromise.Gcore’s ability to deliver high-frequency compute in South America gives us a real advantage in reaching players where latency and infrastructure have long been challenges for online multiplayer gaming.Stian Drageset, CFO & COO, FuncomPowering next-gen multiplayer survival games globallyBy choosing Gcore Bare Metal servers and Managed Kubernetes, Funcom is positioned to deliver a high-performance multiplayer experience to players in South America and beyond. The flexibility of Gcore infrastructure ensures optimal resource usage, rapid scaling, and reliable DDoS protection—foundational components for a smooth multiplayer survival game launch.Scale your multiplayer experience—everywhereLooking to launch your next multiplayer title in regions others can’t reach? Gcore offers flexible, high-performance infrastructure tailored for real-time gaming. Contact us to learn more about how we can help you reach every corner of the globe.Contact us
Saber delivers record-breaking launch for Warhammer 40,000: Space Marine 2 with Gcore
Founded in 2001, Saber Interactive is renowned for developing games across major platforms. Their portfolio includes hugely popular titles like Warhammer 40,000: Space Marine 2, World War Z, SnowRunner, and RoadCraft. With a commitment to delivering exceptional gaming experiences, Saber continually pushes the boundaries of game development and innovation.Preparing for massive player surges during a hotly anticipated title launchIn September 2024, Saber prepared to launch Warhammer 40,000: Space Marine 2, one of the most eagerly awaited releases in the franchise's history. Given the overwhelming excitement surrounding the game, the team anticipated an enormous surge in player activity, particularly during launch week. Achieving a smooth and uninterrupted experience for millions of players requires careful planning—choosing the right infrastructure was critical to preventing performance bottlenecks, latency issues, or downtime.For the launch of Space Marine 2, Saber developed Hydra, an advanced multiplayer game services middleware that enables cross-platform play, matchmaking, and dedicated server management. This proprietary technology allows for a hybrid approach to server hosting, combining bare-metal servers with cloud-based infrastructure to handle peak concurrent user (CCU) demands efficiently. Saber needed to select the optimal infrastructure providers that could integrate with Hydra’s architecture and deliver reliable performance with low latency and fast load times, leveraging edge delivery.During our major game launches, we see a substantial increase in concurrent players, requiring robust server support to maintain optimal performance and the best gaming experience for our players.Janna Goranskaya, Head of Business Development CIS&EE, Saber InteractiveDeploying 300+ bare metal servers and high-performance virtual machines for a seamless global launchTo meet the launch demand and deliver a smooth experience, Saber turned to Gcore as their primary provider for bare metal servers in most regions worldwide, continuing a trusted partnership from previous game releases like World War Z. Gcore provided over 300 bare metal servers optimized for the most latency-sensitive workloads, delivering minimal lag and uninterrupted gameplay. High-performance virtual machines complemented this foundation by supporting additional gaming infrastructure, including testing and development environments.Hydra enabled Space Marine 2 to dynamically utilize this hybrid infrastructure, orchestrating resources efficiently between Gcore Bare Metal and cloud services to address fluctuating player demands while maintaining seamless cross-platform multiplayer experiences. Built-in DDoS protection—available by default for both bare metal and virtual machines from Gcore—played a critical role in safeguarding Saber’s game servers from malicious attacks, supporting a secure and stable launch.Gcore’s extensive range of bare metal configurations allowed us to select the ideal setup for each workload, providing maximum performance across all aspects of our game infrastructure. Their ability to customize servers to meet the specific requirements of our game engine was essential in optimizing performance and delivering a seamless experience for players worldwide.Kirill Igumnov, Lead Backend Infrastructure Developer at Hydra Team, Saber InteractiveFlexible and scalable server solutions for ultimate efficiencyThe Gcore pay-as-you-go model with hourly billing for bare metal servers offered Saber the flexibility to scale server usage in line with player demand. During the initial launch phase, server capacity was ramped up to accommodate the influx of players, with the ability to scale down as demand normalized over time.The flexibility of Gcore Bare Metal servers allowed us to manage resources efficiently, scaling up during peak times and reducing capacity as needed without long-term commitments. This is perfect for the fluctuating demands of a new title launch during its first weeks.Dmitri Brevdo, Head of Game Services, Saber InteractivePowering a record-breaking game launchWith Gcore support, Saber successfully launched Space Marine 2, which set new franchise records with over 400,000 concurrent players. The robust hybrid server infrastructure enabled a seamless experience for players worldwide, even during peak usage periods. The game attracted more than two million players within the first few days of its release.Our Hydra platform integrated seamlessly with Gcore’s infrastructure, giving us the perfect combination of performance and flexibility. The Gcore global network allows us to deploy servers exactly where our players are, providing minimal latency and a smooth gaming experience.Vladislav Nazaruk, Senior Backend Developer at Hydra Team, Saber InteractiveCelebrating six years of collaborationOver the past six years, Saber and Gcore have cultivated a strong partnership focused on flexibility, trust, and performance.Our long-standing relationship with Gcore has been pivotal in our ability to deliver high-quality games to a global audience. Their dedication to innovation and performance aligns perfectly with our commitment to gamer experience.Janna Goranskaya, Head of Business Development CIS&EE, Saber InteractiveBy choosing Gcore as the main provider of bare metal servers for integration with their Hydra middleware’s hybrid approach, Saber has effectively managed the challenges of several large-scale game launches. Powered by high-frequency CPUs/vCPUs, these servers provide the dedicated compute resources needed for latency-sensitive applications, guaranteeing optimal performance and stable connections for real-time gaming experiences. This partnership has resulted in seamless performance during peak periods, contributing to the success and popularity of their titles.Achieving seamless, scalable, and successful game launches with GcoreBy leveraging an extensive global network with 180+ points of presence (PoPs) and continuously evolving to deliver cutting-edge infrastructure, Gcore is well-equipped to meet the gaming industry’s demands and pace of innovation, no matter the use case.If you’re looking for high-performance, flexible infrastructure that can scale with your plans, contact us to talk through your bare metal or virtual machine needs.Contact us
Riga Technical University accelerates genomic research with Gcore GPU Cloud
We saw a 95% reduction in processing time, but more than speed, we also gained flexibility. We could scale from 2 to 8 GPUs instantly, and because usage was on-demand, we only paid for what we needed.Andris Locāns, Head of RTU HPCCompany backgroundRiga Technical University High-Performance Computing Center (RTU HPC) is Latvia’s largest supercomputing resource provider, supporting scientific and technological advancements across the Baltic region. RTU HPC has collaborated with multiple research institutions, including the Latvian Biomedical Research and Study Centre (BMC), a leader in molecular biology and biomedical research. BMC’s genomic research focuses on analyzing thousands of human genomes as part of European initiatives.Accelerating AI-powered genomic processing without compromising controlGenomic research is essential for understanding human health and disease origins, but like any activity that requires sizeable data-set processing, its computational demands are immense. In Latvia, the Riga Technical University High Performance Computing Center (RTU HPC) is leading a shift from traditional scientific computing toward an AI-first model of innovation.Working alongside the Latvian Biomedical Research and Study Centre (BMC), the team set out to solve a critical challenge: rapidly process thousands of human genomes using AI, without losing time or control to hardware bottlenecks or foreign cloud vendors.Essentially, what we wanted was to accelerate variant calling, the computational process of identifying genetic variations.Edgars Liepa, Scientific Assistant, BMCTraditional CPU-based computing often struggles with large-scale genome sequencing and analysis, leading to extended processing times. As a result, the RTU HPC faced several key challenges:The need for faster genome sequencing to support biomedical research.High compute requirements for analyzing large datasets efficiently.The difficulty of sourcing high-performance GPU hardware within a short timeframe.Ensuring cost-effective and scalable computing solutions without major upfront investments.GPU-as-a-service (GPUaaS) for genomic researchRTU HPC turned to Gcore, provisioning Cloud GPUs for immediate access to high-performance computing. Instead of waiting months for on-premises GPU hardware, they gained on-demand access to NVIDIA’s most advanced GPUs—including the H100, designed for AI inference at scale, and located on Gcore’s European cloud infrastructure.“We didn’t want to offload sensitive health data to platforms outside our legal jurisdiction,” Andris Locāns, Head of RTU HPC explains. “Gcore extensive infrastructure enabled us to maintain data sovereignty and compliance by keeping our data in-region, while still delivering the AI acceleration we needed.”This immediately unlocked the following benefits for the team:Instant access to powerful GPUs: Avoiding long procurement cycles for physical infrastructure.Scalability & cost-efficiency: Gcore’s pay-as-you-go model allowed RTU HPC to allocate resources flexibly based on research demands.Data sovereignty: Ensuring genomic data remains within a secure, compliant cloud infrastructure in the Baltic region.Optimized performance: Benchmarking multiple GPU configurations (V100, A100, L40S, H100) for genomic analysis with NVIDIA Clara Parabricks software.“With Gcore, we had near-instant access to compute that would have taken us six months to deploy internally,” says Edgars Liepa, Scientific Assistant at BMC. “That completely changed the pace of our work.”Benchmarking performance for maximum efficiencyRTU HPC and BMC collaborated with Gcore to conduct extensive performance tests on various GPU configurations. By leveraging Cloud GPUs, they identified optimal setups for accelerating genomic workflows.Comparison of CPU vs. GPU: Genome sequencing that previously took over 650 minutes on CPUs was reduced to under 30 minutes with GPU-powered processing.Testing NVIDIA GPUs: Experiments with GPU configurations provided insights into computational efficiency, with findings indicating that scaling up GPUs did not always equate to faster processing.Future discussions with NVIDIA: Gcore’s collaboration enabled further optimizations in GPU usage for genomic analysis.A comparison of CPU and GPU computing times: while CPU processing time exceeded 650 minutes, it could be significantly reduced to under 30 minutes for all tested configurations when using fq2bam H100Faster, scalable, and cost-effective genomic researchNVIDIA H100 GPUs, provided as-a-service by Gcore, delivered graphics processing units with a compute performance that is revolutionizing cloud infrastructure. They are also specifically designed with the power required for high-performance computing tasks such as computational genomics. “It was important for us to see how fast inference runs on the H100,” says Edgars Liepa, “We didn’t customize the model but instead used one developed by NVIDIA, which was already well suited for our task.”The collaboration between RTU HPC, BMC, and Gcore delivered significant benefits to the research program, including:Significant reduction in processing time: Variant calling tasks were completed up to 50x faster.Cost savings with on-demand GPUs: Eliminating upfront hardware investments while optimizing computing costs.Scalable infrastructure: The ability to dynamically allocate resources based on real-time needs.Data sovereignty and security: Genomic data was processed within a compliant, secure cloud environment.The measured processing times of the tested H100 GPUs in detail: the processing time ranged from 13 minutes for haplotypecaller H100 with 2 GPUs used to just under 90 minutes for Deepvariant H100 with 8 GPUs used, indicating that GPU overhead can slow down the processing time when the effective memory limit is reached“It’s not just about going faster,” Liepa adds. “It’s about enabling analysis at a national scale. The AI models are there—but without the right compute power, they’re just theory.”Advancing high-performance genomics in the Baltics and beyondBy leveraging Gcore’s Cloud GPUs, RTU HPC has established a model for scalable, cost-effective genomic research, and this is only the beginning. Now that the speed and flexibility has been proven on genomics processing, RTU HPC plan to broaden AI applications even further.Wider adoption of Cloud GPUs in genomics: RTU HPC is considering Cloud GPUs expansion for broader research applications.Future collaboration with Gcore: Optimization of GPU configurations will continue, and RTU HPC plans to explore AI Inference opportunities with Gcore Everywhere Inference for genomic workloads.Global implications: BMC’s work with the 1+ Million Genomes Project, an EU-wide initiative to make genomic information more accessible for diagnosis and treatment, contributes to international research efforts.The benchmark test results for H100 GPUs using 2, 4, or 8 GPUs: fq2bam and haplotypecaller achieved the shortest processing times on average and comparatively consistent results across 2, 4, and 8 GPUsPioneering AI-powered genomics with sovereign cloud infrastructure“This is the future of AI in healthcare: fast, flexible, sovereign,” says Liepa. “Gcore gave us the infrastructure to make it real—not just for today, but for what comes next.”As AI continues to transform life sciences, the ability to combine cutting-edge GPU performance, regional data compliance, and on-demand scalability is emerging as the key to competitive advantage—not just for companies, but for countries.“We’re proud to support Latvia’s vision for AI-powered genomics,” says Vsevolod Vayner, Product Director of Edge & AI Cloud at Gcore. “This project is a blueprint for how nations can lead in biotech innovation without giving up digital sovereignty.”Find out more about how Gcore Cloud GPUs can enhance your high-performance computing projects.Try Gcore Cloud GPUs
LetzAI rapidly scales AI-powered image generation for global users
Everywhere Inference reduces the latency of our output and enhances the performance of AI-enabled apps, allowing us to optimize our workflows for more accurate, real-time results.Misch Strotz, CEO and co-founder, LetzAIA futuristic tram journey, blending reality and imagination, created with LetzAI. Destination: To the MoonLaunched as a skunkworks project from within Neon Internet, LetzAI is quickly becoming a go-to platform for high-quality AI-generated images. With a mission to democratize and personalize AI-powered image generation, it has emerged as one of the most popular and high-quality options on the market. To support its rapid growth and scale seamlessly, LetzAI partnered with Gcore for advanced AI and cloud infrastructure.Global GPU shortages threaten to derail a new AI image generation ideaIn 2023, Neon Internet CEO and co-founder Misch Strotz was struck by a clever idea: give Luxembourg residents the power to easily generate local images using AI. Within a month, Luxembourg-focused LetzAI V1 went live.Encouraged by strong local demand, Strotz and his team began working on a global version of the platform. The vision? An opt-in AI platform empowering brands, creators, artists, and individuals to unlock endless creative possibilities by adding their own images, art styles, and products. “Other AI platforms scrape the internet, incorporating people and their content without permission. We wanted to put the choice and power in each person’s hands,” Strotz explains.Before long, the team began working on V2. In addition to generating higher quality and more personalized AI-generated images, V2 would drive consistency across objects, characters, and styles. After uploading their own photos and creating their own models, users can blend them with other models created by the community to create an endless number of unique images.However, LetzAI faced a significant hurdle in training and launching V2—a global GPU shortage. With limited resources to train its models, LetzAI needed a reliable partner to help evolve its AI-driven platform and keep it operating smoothly.Finding a trusted, Europe-based AI partnerIn the search for a fitting partner, Strotz spoke to major vendors including hyperscalers and various Europe-based providers. Meeting Gcore’s product leadership team made the decision clear. “It was amazing to meet executives who were so knowledgeable about technology and took us seriously,” recalls Strotz.Gcore’s approach to data security and sovereignty further solidified the decision. “We needed a trusted partner who shared our commitment to European data protection principles, which we incorporated into the development of our platform” he continues.Maximizing AI efficiency by training on local GPUsLetzAI opted for Gcore’s state-of-the-art NVIDIA H100 GPUs in Luxembourg. “This was the perfect option, allowing us to keep our model training and development local. With Gcore, we can rent GPUs rather than entire servers, making it a far more cost-effective solution by avoiding unnecessary costs like excess storage and idle server capacity,” Strotz explains. This approach provided flexibility, efficiency, and high performance, tailored specifically for AI workloads.LetzAI was able to adapt its app to run in containers, configure model training tasks to run on GPU Cloud, and use Everywhere Inference for image generation and upscaling. “Everywhere Inference reduces the latency of our output and enhances the performance of AI-enabled apps, allowing us to optimize our workflows for more accurate, real-time results,” Strotz says.In just two months, LetzAI V2 launched to serve users around the world. And Strotz and team were already developing its successor.Empowering creativity with scalable, high-performance AI infrastructureWith Gcore’s continued support, LetzAI quickly deployed V3. “The Gcore team was incredibly responsive to our needs, guiding us to the best solution for our evolving requirements. This has given us a powerful and efficient infrastructure that can flex according to demand,” says Strotz.Running V3 on Gcore means LetzAI users experience fast, reliable performance. Artists, individuals, and brands are already putting V3 to use in interesting ways. For example, in response to what LetzAI calls its ‘AI Challenges’, a Luxembourg restaurant chain prompted residents to create thousands of images using its model of a pizza.In another example, LetzAI teamed with digital agency LOOP to dress PUMA’s virtual influencer and avatar, Laila, in a Moroccan soccer jersey. According to Strotz, “PUMA had struggled to make clothing look realistic on Laila. When they saw our images, they said the result was 1,000 times better than anything they had tried.”That wasn’t the only brand intrigued by V3’s possibilities. After LetzAI posted V3-generated images of models wearing Sloggi underwear, Sloggi’s creative agency STAN Studios asked LetzAI to generate more images for market testing.Always looking for new ways to support creators, LetzAI also launched its Image Upscaler feature, which enhances images and doubles their resolution. “Our creators can now resolve common AI image issues around quality and resolution. Everywhere Inference is pivotal in delivering the power and speed needed for these dynamic image enhancements,” notes Strotz.Platform evolution and AI innovation without limitsAs its models exceed user expectations worldwide, LetzAI can rely on Gcore to handle a high volume of requests. Confident about generating a limitless number of high-quality images on the fly, LetzAI can continue to scale rapidly to become a sustainable, innovation-driven business.“As we further evolve—such as by adding video features to our platform—our partnership with Gcore will be central to LetzAI’s continued success,” Strotz concludes.
Customers that trust Gcore to power their business and infrastructure
Cloud services
Virtual data center with 50+ Cloud services.
Content Delivery Network
Next-gen CDN for dynamic and static content delivery worldwide.
DDoS Protection
Reliable infrastructure protection against L3, L4 & L7 DDoS attacks.