Radar has landed - discover the latest DDoS attack trends. Get ahead, stay protected.Get the report
Under attack?

Products

Solutions

Resources

Partners

Why Gcore

  1. Home
  2. Blog
  3. Managed Kubernetes with GPU Worker Nodes for Faster AI/ML Inference

Managed Kubernetes with GPU Worker Nodes for Faster AI/ML Inference

  • By Gcore
  • November 23, 2023
  • 6 min read
Managed Kubernetes with GPU Worker Nodes for Faster AI/ML Inference

Currently, 48% of organizations use Kubernetes for AI/ML workloads, and the demand for such workloads also drives usage patterns on Kubernetes. Let’s look at the key technical reasons behind this trend, how AI/ML workloads benefit from running on GPU worker nodes in managed K8s clusters, and some considerations regarding GPU vendors and scheduling.

Why Kubernetes is Good for AI/ML

A number of features make Kubernetes popular and effective in the AI/ML realm:

  • Scalability. K8s enables seamless, on-demand scalability of AI/ML workloads. This is especially critical for inference workloads because they are more dynamic regarding resource utilization than training workloads, and can be resource-intensive. The latter means they often require frequent scaling up or down based on the volume of data being processed.
  • Automated scheduling. The ability to automatically schedule AI/ML workloads reduces the operational overhead for MLOps teams. It also improves the performance of AI/ML applications by ensuring they are scheduled to the nodes that have the required resources.
  • Resource utilization. K8s can help to optimize physical resource utilization for AI/ML workloads. It can dynamically and automatically allocate the required amounts of CPU, GPU, and RAM resources. This is critical due to the resource-intensive nature of these workloads and the potential for cost reduction.
  • Flexibility. With K8s, you can deploy AI/ML workloads across multiple infrastructures, including on-premises, public cloud, and edge cloud. This feature also makes Kubernetes a good option for organizations that need to deploy AI/ML workloads in hybrid or multicloud environments.
  • Portability. You can easily migrate Kubernetes-based AI/ML applications between different environments and installations. This is critical for deploying and managing AI/ML workloads in a hybrid infrastructure.

Use Cases

Here are some examples of how companies have adopted Kubernetes (K8s) for their AI/ML projects:

  • OpenAI has been an early adopter of K8s. In 2017, the company was running machine learning experiments on K8s clusters. With the K8s autoscaler, OpenAI could deploy such a project in a few days and scale it up to hundreds of GPUs in a week or two. Without the Kubernetes autoscaler, such a process would take months. As a result, OpenAI increased the number of AI experiments tenfold. In 2021, the company expanded its K8s infrastructure to 7,500 nodes for large ML models such as GPT-3, DALL-E and CLIP.
  • Shell uses a K8s-based platform Kubeflow to run tests and quickly experiment with ML models on laptops. Engineers can move these workloads from the test environment to production, and the workloads will function just the same. With Kubernetes, Shell builds thousands of ML models in two hours instead of a month. The time to write the underlying code is reduced from two weeks to four hours.
  • IKEA has developed an internal MLOps platform based on K8s to train ML models on-premises and get inference in the cloud. This allows the MLOps team to orchestrate different types of trained models and, ultimately, improve the customer experience.

Of course, these examples are not broadly representative. Most companies are not fully AI-focused like OpenAI and are not as large as IKEA. They can’t afford to train large AI/ML models from scratch, which takes time and money, but instead run pretrained models and integrate them with other internal services. In other words, these companies use AI/ML inference, not training.

Inference workloads tend to be more dynamic regarding resource utilization than training workloads because production clusters are more likely to experience user and traffic spikes. In such cases, the infrastructure needs to scale up and down quickly, whereas AI/ML training typically requires gradual scaling. Therefore, for AI/ML models that are already trained and deployed, the scalability and dynamic resource utilization of K8s are especially beneficial.

Why GPU Is Better than CPU for Worker Nodes

GPU worker nodes are a better fit for containerized AI/ML workloads than CPU worker nodes for the same reasons as for non-containerized workloads: GPU offers parallel processing capabilities and higher performance for AI/ML than CPUs.

Inference for AI/ML workloads running on GPU worker nodes can be faster than those running on CPU worker nodes due to the following factors:

  • The GPU’s memory architecture is specifically optimized for AI/ML processing, enabling higher memory bandwidth than CPUs.
  • GPUs often provide better computational performance than CPUs for AI/ML training and inference because they have more transistors to process data.

Kubernetes adds its own performance benefits to those of GPUs. In addition to hardware acceleration, AI/ML workloads running on GPU worker nodes get scalability and dynamic resource allocation. Kubernetes also includes plugins for GPU vendor support, making it easy to configure GPU resources for use by AI/ML workloads.

Figure 1. The simplified K8s cluster architecture with GPU worker node

With Kubernetes, you can manage GPU resources across multiple worker nodes. Containers consume GPU resources in essentially the same way as they consume CPU resources.

GPU Vendors Comparison

There are three GPU vendors available for Kubernetes: NVIDIA, AMD, and Intel. When choosing GPU vendors for worker nodes, it’s important to keep in mind that their compatibility with Kubernetes, tool ecosystem, performance, and cost can vary.

 NVIDIA GPU worker nodesAMD GPU worker nodesIntel GPU worker nodes
Compatibility with K8sExcellentGoodGood
Tools ecosystemExcellentGoodFair
PerformanceExcellentGoodFair
CostHighMediumMedium

Let’s compare the three vendors.

  • Compatibility with Kubernetes: NVIDIA is the most compatible with K8s. The company provides CUDA drivers, various container runtimes, and other tools and features that simplify GPU integration and management. AMD and Intel support for K8s is less mature and often requires custom configuration.
  • Tools ecosystem: NVIDIA has the best ecosystem of tools for AI/ML, thanks to software such as the GPU Operator and Container Toolkit, and ML frameworks adapted for NVIDIA GPUs, such as TensorFlow, PyTorch, and MXNet. AMD and Intel also have tools for AI/ML, but they are not as comprehensive as NVIDIA’s.
  • Performance: NVIDIA GPUs are known for their high performance on AI workloads, outperforming the competition on most MLPerf benchmarks. NVIDIA GPUs are ideal for demanding tasks such as deep learning and high-performance computing.
  • Cost: NVIDIA GPUs are the most expensive type of GPU worker node.
  • Flexibility: NVIDIA offers several features that make its GPU-based K8s clusters highly flexible in terms of management and resource utilization compared to its competitors:
    • Multi-instance GPU (MIG) mechanism for NVIDIA A100 GPU to allow a GPU to be securely partitioned into up to seven separate instances for better GPU utilization
    • Multicloud GPU clusters, which can be seamlessly managed and scaled as if deployed in a single cloud
    • Heterogeneous GPU and CPU clusters to simplify the training and management of distributed deep learning models
    • GPU metrics monitoring with Prometheus and visualization with Grafana
    • Support for multiple container runtimes, including Docker, CRI-O, and containers

In summary, NVIDIA GPU worker nodes are the best choice for AI/ML workloads in Kubernetes. They offer the best compatibility with K8s, the best tools ecosystem, and the best performance. That’s why we chose NVIDIA GPUs for Gcore Managed Kubernetes. Our customers get all the benefits of NVIDIA, including the highest performance level for faster training and inference of their AI/ML workloads.

Important Specifics of GPU Scheduling in Kubernetes

To enable GPU scheduling and allow pods to access its resources, you need to install a vendor-specific device plugin from your chosen GPU vendor — NVIDIA, AMD, or Intel.

Pods request GPU resources in the same way they request CPU resources. However, Kubernetes is less flexible with GPU than with CPU when it comes to configuring `limits` and `requests`. With `requests`, you set the amount of resources that a pod is guaranteed to get, such as a minimum quantity. With `limits`, you set the amount of resources that won’t be exceeded, for instance, a maximum quantity. When configuring a pod manifest for GPU requests, `limits` and `requests` should be equal, meaning that a pod won’t get more resources than guaranteed if, for example, the application needs them.

Also, by default, you can’t allocate part of a GPU or multiple GPUs to a container because of the way CPU allocation works. You can only allocate one full GPU per container. This limitation doesn’t help with resource economics. But NVIDIA has managed to overcome this. With its GPU, you can use either use:

  • Time-sharing GPUs, which work by sequentially assigning time intervals to shared containers on a physical GPU. This works for all NVIDIA GPUs.
  • Multi-instance GPUs, which allow a GPU to be divided into up to seven instances for better GPU utilization. This only works with the NVIDIA A100 GPU.

These two features help you to use NVIDIA GPU resources more efficiently and save money on renting GPU instances in the cloud. This is also a significant advantage over other GPU vendors.

Managed Kubernetes vs. Vanilla Kubernetes with GPU

A managed Kubernetes service can offer several advantages over vanilla (open source) Kubernetes for AI/ML workloads running on GPU worker nodes:

  • Flexible choice of GPUs. Managed K8s services typically provide support for GPU instances with various specifications. This makes it easier to choose the appropriate level of GPU acceleration for your AI/ML workloads.
  • Reduced operational overhead. Managed Kubernetes handles the everyday responsibilities of overseeing a Kubernetes cluster, like managing the control plane and implementing K8s updates. This enables you to focus on creating, deploying and managing AI/ML applications.
  • Scalability and reliability. Managed K8s services are typically designed with a strong focus on scalability and reliability, ensuring that your AI/ML workloads can adeptly handle fluctuating traffic and spikes in resource demand.

Gcore Managed Kubernetes with NVIDIA GPU Workers

Gcore Managed Kubernetes helps you to deploy Kubernetes clusters fast, without the need to maintain the underlying infrastructure and Kubernetes backend. The Gcore team controls the master nodes while you control only the worker nodes, reducing your operational burden. Worker nodes can be Gcore Virtual Machines or Bare Metal servers in various configurations, including those with NVIDIA GPU modules.

Conclusion

Managed Kubernetes with GPU worker nodes is a powerful and flexible combination for accelerating AI/ML inference. By taking advantage of both Kubernetes and GPUs, managed Kubernetes with GPU worker nodes can help you improve the performance and efficiency of your AI/ML workloads. The service also frees you from the need to maintain the underlying GPU infrastructure and most Kubernetes components.

Gcore Managed Kubernetes can boost your AI/ML workloads with GPU worker nodes on Bare Metal for faster inference and operational efficiency. We offer a 99.9% SLA with free production management and free egress traffic—at outstanding value for money.

Explore Managed Kubernetes

Related articles

Edge Cloud news: more regions and volume options available

At Gcore, we’re committed to delivering high-performance, globally distributed infrastructure that adapts to your workloads—wherever they run. This month, we’re excited to share major updates to our Edge Cloud platform: two new cloud IaaS regions in Europe and expanded storage options in São Paulo.New IaaS regions in Luxembourg and Portugal available nowLuxembourg‑3 and Sines‑2 mark the next step in the Gcore mission to bring compute closer to users. From compliance-focused deployments in Central Europe to GPU‑powered workloads in the Iberian Peninsula, these new regions are built to support diverse infrastructure needs at scale.Luxembourg‑3: expanding connectivity in Central EuropeWe’re expanding our European footprint by opening an additional IaaS point of presence (PoP) in Luxembourg. Strategically located in the heart of Europe, this region offers low-latency connectivity across the EU and is a strong compliance choice for data residency requirements.Here’s what’s available in Luxembourg‑3:Virtual Machines: High-performance, reliable, and scalable compute power for a wide range of workloads - with free egress traffic and pay-as-you-go billing for active instances only.Volumes: Standard, High IOPS, and Low Latency block storage for any workload profile.Load Balancers: Distribute traffic intelligently across instances to boost availability, performance, and fault tolerance.Managed Kubernetes: Fully managed Kubernetes clusters with automated provisioning, scaling, and updates optimized for production-ready deployments.Sines‑2, Portugal: a new hub for Southern Europe and a boost for AI workloadsWe’re also opening a brand-new location: Sines‑2, Portugal. This location enhances coverage across Southern Europe and boosts our AI and compute capabilities with more GPU availability.In addition to offering the same IaaS services as Luxembourg‑3, Sines‑2 also includes:H100 NVIDIA GPUs for AI/ML, high-performance computing, and rendering workloads.New VAST NFS Fileshare support for scalable, high-throughput file storage.This new region is ideal for organizations looking to deploy close to the Iberian Peninsula, reducing latency for regional users while gaining access to powerful GPU resources.Enhanced volume types in São PauloVolumes are the backbone of any cloud workload. They store the OS, applications, and essential data for your virtual machines. Developers and businesses building latency-sensitive or I/O-intensive applications now have more options in the São Paulo-2 region, thanks to two newly added volume types optimized for speed and responsiveness:Low-latency volumesDesigned for applications where every millisecond matters, Low Latency Volumes are non-redundant block storage ideal for:ETCD clustersTransactional databasesOther real-time, latency-critical workloadsBy minimizing overhead and focusing on speed, this volume type delivers faster response times for performance-sensitive use cases. This block storage offers IOPS up to 5000 and an average latency of 300 microseconds.High-IOPS volumesFor applications that demand both speed and resilience, High IOPS Volumes offer a faster alternative to our Standard Volumes:Higher IOPS and increased throughputSuitable for high-traffic web apps, analytics engines, and demanding databasesThis volume type accelerates data-heavy workloads and keeps performance consistent under peak demand by delivering significantly higher throughput and IOPS. The block storage offers IOPS up to 9,000 and a 500 MB/s bandwidth limit.Ready to deploy with Gcore?These new additions help to fine-tune your performance strategy, whether you're optimizing for throughput, latency, or both.From scaling in LATAM to expanding into the EU or pushing performance at the edge, Gcore continues to evolve with your needs. Explore our new capabilities in Luxembourg‑3, Sines‑2, and São Paulo‑2.Discover more about Gcore Cloud Edge Services

How AI is improving L7 DDoS protection solutions

How AI is improving L7 DDoS protection solutionsDDoS attacks have always been a concern for organizations, but with the recent rise of AI and machine learning, the threat has grown. Layer 7 attacks are particularly damaging, as they focus on the application layer that users utilize to interact with your system. Unlike traditional DDoS attacks, which aim to overwhelm the servers with sheer traffic, these advanced threats imitate real user behavior, making it incredibly difficult for defenses to identify and block malicious traffic.While this challenge is complex, it is far from insurmountable. In this situation, the mantra "fight fire with fire" really applies. By using machine learning and AI against AI-based attacks, organizations can then retaliate with equally advanced Layer 7 protection. These newer technologies can offer something beyond what more traditional techniques could hope to achieve, including significantly faster response times, smarter threat detection, and precision. Here’s how AI and ML are redefining how businesses stay online and secure.Why L7 DDoS attacks are dangerous and hard to stopL7 DDoS attacks are sneaky. Unlike network-based attacks that flood your bandwidth, these attacks go after your application logic. Picture thousands of fake users trying to log in, search for products, or complete transactions all at once. Your systems become overwhelmed, not because they’re receiving a massive amount of data, but because they’re handling what looks like genuine traffic.The big challenge is filtering out the bad traffic while letting legitimate users through. After all, if you accidentally block real customers, you’re essentially doing the attackers’ job for them.Manual defenses used in the past, such as rate limiting with static thresholds, can result in a lose-lose situation. When the threshold is set too high, attackers can enter, often in place of traditional users. If the threshold is set too low, legitimate users are left unable to access the application. This acts as a collective punishment, blocking users out of fear of a few malicious actors rather than an accurate solution that can identify the malicious activity and block it without compromising users’ experience. Traditional defenses, based on static rules or human intervention, simply cannot scale at the speed and intricacy of a modern attack. They’re reactive when they need to be proactive.Filtering traffic without blocking customersAI and ML avoid the pitfalls of traditional security systems by continuously analyzing traffic and identifying anomalies dynamically. One of the biggest pain points in DDoS defense is false positives, which block legitimate users because their behavior looks suspicious.Traditional solutions relying on static rules simply block any IPs displaying suspicious behavior, while AI and ML track the activity of IPs over time, building a detailed profile of legitimate traffic. Sometimes referred to as IP profiling, this process groups together the IP addresses that interact predictably and legitimately with your systems. By analyzing both current and historical data, these systems can differentiate suspicious IPs from legitimate users. In the event of an attack, “safe” IPs are automatically allowed through, while suspicious ones are challenged or blocked.These AI systems learn over time from previous attacks they’ve encountered, adapting for greater accuracy without any manual updating or intervention to counter-changing tactics. This allows the systems to correlate current traffic with historical profiles and continuously reassess the safety of certain profiles. This ensures that legitimate accounts can continue to access services unimpeded while malicious traffic is contained.Traditional systems cannot achieve this level of precision, and instead tend to shut down applications during attacks, essentially allowing the attackers to win. With advanced AI and ML based defenses, businesses can maintain their service undisturbed for real users, even during an attack.Fighting AI attacks with AI defensesDDoS attacks are becoming increasingly adaptive, using AI to mimic real users, leaving the static rules in traditional solutions unable to identify the subtle signs of attack traffic. Attackers constantly change their methods to avoid fixed security rules. Manually updating defenses each time a new attack method pops up is time-consuming and inefficient.AI-powered solutions overcome this limitation by using the same strategy as attackers, continuously learning from data input to adapt to increasingly convincing DDoS traffic in real time. This can stop even zero-day and self-evolving AI cyberattacks.Staying Ahead of Attackers With Smarter DefensesOur AI-driven WAAP solution delivers intelligent, interconnected protection, enabling businesses to stay ahead of even the most advanced and evolving threats, including L7 DDoS attacks. By leveraging deep traffic analysis, heuristic tagging, and adaptive learning, it provides a proactive defense strategy. With cross-domain capabilities and actionable security insights, Gcore WAAP is an essential asset for security architects and key decision-makers, seamlessly blending innovation with practicality to meet the demands of today’s digital landscape.Interested in exploring WAAP further? Download our ebook to discover cybersecurity best practices, the most prevalent threats, and how WAAP can protect your business’s digital infrastructure. Or, reach out to our team to learn more about Gcore WAAP.Discover why WAAP is a must-have for modern businesses—get your free ebook

How AI is transforming gaming experiences

AI is reshaping how games are played, built, and experienced. Although we are in a period of flux where the optimal combination of human and artificial intelligence is still being ironed out, the potential for AI to greatly enhance both gameplay and development is clear.PlayStation CEO Hermen Hulst recently emphasized the importance of striking the right balance between the handcrafted human touch and the revolutionary advances that AI brings. AI will not replace the decades of design, storytelling, and craft laid down by humans—it will build on that foundation to unlock entirely new possibilities. In addition to an enhanced playing experience, AI is shaking up gaming aspects such as real-time analytics, player interactions, content generation, and security.In this article, we explore three specific developments that are enriching gaming storyworlds, as well as the technology that’s bringing them to life and what the future might hold.#1 Responsive NPC behavior and smarter opponentsAI is evolving to create more realistic, adaptive, and intelligent non-player characters (NPCs) that can react to individual player choices with greater depth and reasoning. The algorithms allow NPCs to respond dynamically to players’ decisions so they can adjust their strategies and behaviors in real time. This provides a more immersive and dynamic gameplay environment and means gamers have endless opportunities to experience new gaming adventures and write their own story every time.A recent example is Red Dead Redemption 2, which enables players to interact with NPCs in the Wild West. Players were impressed by its complexity and the ability to interact with fellow cowboys and bar patrons. Although this is limited for now, eventually, it could become like a video game version of the TV series Westworld, in which visitors pay to interact with incredibly lifelike robots in a Wild West theme park.AI also gives in-game opponents more “agency,” making them more reactive and challenging for players to defeat. This means smarter, more unpredictable enemies who provide a heightened level of achievement, novelty, and excitement for players.For example, AI Limit, released in early 2025, is an action RPG incorporating AI-driven combat mechanics. While drawing comparisons to Soulslike games, the developers emphasize its unique features, including the Sync Rate system, which adds depth to combat interactions.#2 AI-assisted matchmaking and player behavior predictionsAI-powered analytics can identify and predict player skill levels and playing styles, leading to more balanced and competitive matchmaking. A notable example is the implementation of advanced matchmaking systems in competitive games such as Apex Legends and Call of Duty: Modern Warfare III. These titles use AI algorithms to analyze not just skill levels but also playstyle preferences, weapon selections, and playing patterns to create matches optimized for player retention and satisfaction. The systems continuously learn from match outcomes to predict player behavior and create more balanced team compositions across different skill levels.By analyzing a player’s past performance, AI can also create smarter team formations. This makes for fairer and more rewarding multiplayer games, as players are matched with others who complement their skill and strategy.AI can monitor in-game interactions to detect and mitigate toxic behavior. This helps encourage positive social dynamics and foster a more collaborative and friendly online environment.#3 Personalized gaming experiencesMultiplayer games can use AI to analyze player behavior in real time, adjusting difficulty levels and suggesting tailored missions, providing rich experiences unique to each player. This creates personalized, player-centric gameplay that evolves dynamically and can change over time as a player’s knowledge and ability improve.Games like Minecraft and Skyrim already use AI to adjust difficulty and offer dynamic content, while Oasis represents a breakthrough as an AI-generated Minecraft-inspired world. The game uses generative AI to predict and render gameplay frames in real time, creating a uniquely responsive environment.Beyond world generation, modern games are also incorporating AI chatbots that give players real-time coaching and personalized skill development tips.How will AI continue to shape gaming?In the future, AI will continue to impact not just the player experience but also the creation of games. We anticipate AI revolutionizing game development in the following areas:Procedural content generation: AI will create vast, dynamic worlds or generate storylines, allowing for more expansive and diverse game worlds than are currently available.Game testing: AI will simulate millions of player interactions to help developers find bugs and improve gameplay.Art and sound design: AI tools will be used to a greater extent than at present to create game art, music, and voiceovers.How Gcore technology is powering AI gaming innovationIn terms of the technology behind the scenes, Gcore Everywhere Inference brings AI models closer to players by deploying them at the edge, significantly reducing latency for training and inference. This powers dynamic features like adaptive NPC behavior, personalized gameplay, and predictive matchmaking without sacrificing performance.Gcore technology differentiates itself with the following features:Supports all major frameworks, including PyTorch, TensorFlow, ONNX, and Hugging Face Transformers, making deploying your preferred model architecture easy.Offers multiple deployment modes, whether in the cloud, on-premise, or across our distributed edge network with 180+ global locations, allowing you to place inference wherever it delivers the best performance for your users.Delivers sub-50ms latency for inference workloads in most regions, even during peak gaming hours, thanks to our ultra-low-latency CDN and proximity to players.Scales horizontally, so studios can support millions of concurrent inferences for dynamic NPC behavior, matchmaking decisions, or in-game voice/chat moderation, without compromising gameplay speed.Keeps your models and training data private through confidential computing and data sovereignty controls, helping you meet compliance requirements across regions including Europe, LATAM, and MENA.With a low-latency infrastructure that supports popular AI frameworks, Gcore Everywhere Inference allows your studio to deploy custom models and deliver more immersive, responsive player experiences at scale. With our confidential computing solutions, you retain full control over your training assets—no data is shared, exposed, or compromised.Deliver next-gen gaming with Gcore AIAI continues to revolutionize industries, and gaming is no exception. The deployment of artificial intelligence can help make games even more exciting for players, as well as enabling developers to work smarter when creating new games.At Gcore, AI is our core and gaming is our foundation. AI is seamlessly integrated into all our solutions with one goal in mind: to help grow your business. As AI continues to evolve rapidly, we're committed to staying at the cutting edge and changing with the future. Contact us today to discover how Everywhere Inference can enhance your gaming offerings.Get a customized consultation about AI gaming deployment

How to choose the right CDN provider in a turbulent marketplace

In a CDN marketplace marked by provider shutdowns, price hikes, and shifting priorities, reliability is survival. If your current provider folds, you're not just facing downtime—you're losing revenue and customer trust. For the world’s top 2,000 companies, the total annual downtime cost is $400 billion, eroding 9% of profits. Choosing the right CDN partner isn’t just about performance, it’s about protecting your business from disruption.In this guide, we show you how to identify early warning signs, evaluate providers, and switch before your business takes the hit.Red flags: signs that it’s time to consider a new CDN providerIf you’re experiencing any of the following issues with your current CDN provider, it might be time to reconsider your current setup.Slower load times: If you’ve noticed lagging performance, your CDN provider may be running on outdated infrastructure or not investing in upgrades.Rising costs: Increasing prices without additional value? A higher bill and the same service is a major red flag.Uncertainty about long-term service: Look for clear communication and a demonstrated commitment to infrastructure investment, essential a market where providers frequently consolidate and shift focus.Your CDN should scale with you, not hold you back. Prioritize a partner who can evolve with your needs and support your long-term success.5 must-haves when choosing a CDN partnerNot all CDNs are created equal. Before switching, compare providers across these five key factors.Performance: Check real-world performance benchmarks and global coverage maps to understand how a CDN will serve your audience in key regions. Throughput (the amount of data that can be successfully delivered from a server to an end user over a specific period of time) and low latency are non-negotiable when choosing a CDN provider.Pricing: Before signing up, it’s essential to know what is and isn’t included in the price in case there are hidden fees. Look for predictable billing, volume-based tiers, and transparent overage charges to avoid surprise costs. Avoid vendors who lure you in with low rates, then add hidden overage fees.Security: Choose a CDN that offers built-in protection out of the box: DDoS mitigation, TLS, WAF, and API security. Bonus points for customizable policies that fit your stack. Strong security features should be standard for CDNs to combat advanced cyber threats.Edge computing: When it comes to Edge computing, understanding the power of this strategic CDN add-on can give you a significant advantage. Look for CDN providers that offer flexible edge compute capabilities, so you can process data closer to users, reduce latency, and improve response times.Future-proofing: The CDN market’s volatility makes partnering with providers with long-term stability vital. Pick a provider that’s financially solid, tech-forward, and committed to innovation—not just sticking around to get acquired.Choosing a new provider may feel like a challenge, but the long-term payoff—improved performance, lower risk, and a future-ready infrastructure—makes it well worth it. By picking a CDN partner that meets your needs now and for the future, you’ll receive fast, personalized, and secure experiences that truly stand out.What makes Gcore CDN different?Gcore CDN isn’t just another CDN, we’re your long-term performance partner. Here’s what we offer:Global scale, blazing speed: Our network spans 180+ edge locations across 6 continents, optimized for low-latency delivery no matter where your users are.Transparent, flexible pricing: No hidden fees. No lock-in. Just fair, flexible pricing models designed to scale with your growth.A stable partner in a shaky market: While others pivot or fold, Gcore is doubling down. We’re investing in infrastructure, expanding globally, and building for the next wave of content and edge use cases.If you’re ready to make the switch, we’re here to help. Get in touch for a free consultation to discuss your specific needs and tailor a transition plan that suits your business. For more insights about choosing the right CDN for your business, download our free CDN buyer's guide for a more in-depth look at the CDN landscape.Get your free CDN buyers guide now

Introducing Super Transit for outstanding DDoS protection performance

We understand that security and performance for your online services are non-negotiables. That’s why we’re introducing Super Transit, a cutting-edge DDoS protection and acceleration feature designed to safeguard your infrastructure while delivering lightning-fast connectivity. Read on to discover the benefits of Super Transit, who can benefit from the feature, and how it works.DDoS mitigation meets exceptional network performanceSuper Transit intelligently routes your traffic via Gcore’s 180 point-of-presence global network, proactively detecting, mitigating, and filtering DDoS attacks. When an attack occurs, your customers don’t notice any difference: Their connection remains stable and secure. Plus, they get an enhanced end-user experience, as the delay between the end user and the server is significantly reduced, cutting down latency.“Super Transit allows for fast, worldwide access to our DDoS protection services,” explains Andrey Slastenov, Head of Security at Gcore. “This is particularly important for real-time services such as online gaming and video streaming, where delay can significantly impact user experience.”Who needs Super Transit?Super Transit is designed for enterprises that require both high-performance connectivity and strong DDoS protection. Here’s how it helps different roles in your organization:CISOs and security teams: Reduce risks and help ensure compliance by integrating seamless DDoS protection into your network.CTOs and IT leaders: Optimize traffic performance and maintain uninterrupted business operations.Network engineers and security architects: Simplify security management with API, automated attack mitigation, and secure GRE tunneling.How Super Transit worksSuper Transit optimizes performance and security by performing four steps.Traffic diversion: Incoming traffic is automatically routed through Gcore’s global anycast network, where it undergoes real-time analysis. Malicious traffic is blocked before it can reach your infrastructure.Threat detection and mitigation: Using advanced filtering, Super Transit identifies and neutralizes DDoS attacks.Performance optimization: Legitimate requests are routed through the optimal path within Gcore’s high-performance backbone, minimizing latency and maximizing speed.Secure tunneling to your network: Traffic is securely forwarded to your origin via stable tunneling protocols, providing a smooth, uninterrupted, and secure connection.Get Super Transit today for high-performance securitySuper Transit is available now to all Gcore customers. To get started, get in touch with our security experts who’ll guide you through how to get Super Transit up and running. You can also explore our product documentation, which provides a clear and simple guide to configuring the feature.Our innovations are driven by cutting-edge research, enabling us to stay one step ahead of attackers. We release the latest DDoS attack trends twice yearly, so you can make informed decisions about your security needs. Get the H1 2024 report free.Discover the latest DDoS attack trends with Gcore Radar

How gaming studios can use technology to safeguard players

Online gaming can be an enjoyable and rewarding pastime, providing a sense of community and even improving cognitive skills. During the pandemic, for example, online gaming was proven to boost many players’ mental health and provided a vital social outlet at a time of great isolation. However, despite the overall benefits of gaming, there are two factors that can seriously spoil the gaming experience for players: toxic behavior and cyber attacks.Both toxic behavior and cyberattacks can lead to players abandoning games in order to protect themselves. While it’s impossible to eradicate harmful behaviors completely, robust technology can swiftly detect and ban bullies as well as defend against targeted cyberattacks that can ruin the gaming experience.This article explores how gaming studios can leverage technology to detect toxic behavior, defend against cyber threats, and deliver a safer, more engaging experience for players.Moderating toxic behavior with AI-driven technologyToxic behavior—including harassment, abusive messages, and cheating—has long been a problem in the world of gaming. Toxic behavior not only affects players emotionally but can also damage a studio’s reputation, drive churn, and generate negative reviews.The online disinhibition effect leads some players to behave in ways they may not in real life. But even when it takes place in a virtual world, this negative behavior has real long-term detrimental effects on its targets.While you can’t control how players behave, you can control how quickly you respond.Gaming studios can implement technology that makes dealing with toxic incidents easier and makes gaming a safer environment for everyone. While in the past it may have taken days to verify a complaint about a player’s behavior, today, with AI-driven security and content moderation, toxic behavior can be detected in real time, and automated bans can be enforced. The tool can detect inappropriate images and content and includes speech recognition to detect derogatory or hateful language.In gaming, AI content moderation analyzes player interactions in real time to detect toxic behavior, harmful content, and policy violations. Machine learning models assess chat, voice, and in-game media against predefined rules, flagging or blocking inappropriate content. For example, let’s say a player is struggling with in-game harassment and cheating. With AI-powered moderation tools, chat logs and gameplay behavior are analyzed in real time, identifying toxic players for automated bans. This results in healthier in-game communities, improved player retention, and a more pleasant user experience.Stopping cybercriminals from ruining the gaming experienceAnother factor negatively impacting the gaming experience on a larger scale is cyberattacks. Our recent Radar Report showed that the gaming industry experienced the highest number of DDoS attacks in the last quarter of 2024. The sector is also vulnerable to bot abuse, API attacks, data theft, and account hijacking.Prolonged downtime damages a studio’s reputation—something hackers know all too well. As a result, gaming platforms are prime targets for ransomware, extortion, and data breaches. Cybercriminals target both servers and individual players’ personal information. This naturally leads to a drop in player engagement and widespread frustration.Luckily, security solutions can be put in place to protect gamers from this kind of intrusion:DDoS protection shields game servers from volumetric and targeted attacks, guaranteeing uptime even during high-profile launches. In the event of an attack, malicious traffic is mitigated in real-time, preventing zero downtime and guaranteeing seamless player experiences.WAAP secures game APIs and web services from bot abuse, credential stuffing, and data breaches. It protects against in-game fraud, exploits, and API vulnerabilities.Edge security solutions reduce latency, protecting players without affecting game performance. The Gcore security stack helps ensure fair play, protecting revenue and player retention.Take the first steps to protecting your customersGaming should be a positive and fun experience, not fraught with harassment, bullying, and the threat of cybercrime. Harmful and disruptive behaviors can make it feel unsafe for everyone to play as they wish. That’s why gaming studios should consider how to implement the right technology to help players feel protected.Gcore was founded in 2014 with a focus on the gaming industry. Over the years, we have thwarted many large DDoS attacks and continue to offer robust protection for companies such as Nitrado, Saber, and Wargaming. Our gaming specialization has also led us to develop game-specific countermeasures. If you’d like to learn more about how our cybersecurity solutions for gaming can help you, get in touch.Speak to our gaming solutions experts today

Subscribe to our newsletter

Get the latest industry trends, exclusive insights, and Gcore updates delivered straight to your inbox.