Radar has landed - discover the latest DDoS attack trends. Get ahead, stay protected.Get the report
Under attack?

Products

Solutions

Resources

Partners

Why Gcore

  1. Home
  2. Blog
  3. Evaluating Server Connectivity with Looking Glass: A Comprehensive Guide

Evaluating Server Connectivity with Looking Glass: A Comprehensive Guide

  • By Gcore
  • November 1, 2023
  • 8 min read
Evaluating Server Connectivity with Looking Glass: A Comprehensive Guide

When you’re considering purchasing a dedicated or virtual server, it can be challenging to assess how it will function before you buy. Gcore’s free Looking Glass network tool allows you to assess connectivity, giving you a clear picture of whether a server will meet your needs before you make a financial commitment. This article will explore what the Looking Glass network tool is, explain how to use it, and delve into its key functionalities like BGP, PING, and traceroute.

What Is Looking Glass?

Looking Glass is a dedicated network tool that examines the routing and connectivity of a specific AS (autonomous system) to provide real-time insights into network performance and route paths, and show potential bottlenecks. This makes it an invaluable resource for network administrators, ISPs, and end users considering purchasing a virtual or dedicated server. With Looking Glass, you can:

  • Test connections to different nodes and their response times. This enables users to initiate connection tests to various network nodes—crucial for evaluating response times, and thereby helpful in performance tuning or troubleshooting.
  • Trace the packet route from the router to a web resource. This is useful for identifying potential bottlenecks or failures in the network.
  • Display detailed BGP (Border Gateway Protocol) routes to any IPv4 or IPv6 destination. This feature is essential for ISPs to understand routing patterns and make informed peering decisions.
  • Visualize BGP maps for any IP address. By generating a graphical representation or map of the BGP routes, Looking Glass provides an intuitive way to understand the network architecture.

How to Work with Looking Glass

Let’s take a closer look at Looking Glass’s interface.

Under Gcore’s AS number (AS199524), header, and introduction, you will see four fields to complete. Operating Looking Glass is straightforward:

  1. Diagnostic method: Pick from three available diagnostic methods (commands.)
    • BGP: The BGP command in Looking Glass gives you information about all autonomous systems (AS) traversed to reach the specified IP address from the selected router. Gcore Looking Glass can present the BGP route as an SVG/PNG diagram.
    • Ping: The ping command lets you know the round trip time (RRT) and Time to Live (TTL) for the route between the IP address and the router.
    • Traceroute: The traceroute command displays all enabled router hops encountered in the path between the selected router and the destination IP address. It also marks the total time for the request fulfillment and the intermediate times for each AS router that it passed.
  2. Region: Choose a location from the list of locations where our hosting is available. You can filter server nodes by region to narrow your search.
  3. Router: Pick a router from the list of Gcore’s available in the specified region.
  4. IP address: Enter the IP address to which you want to connect. You can type in both IPv4 and IPv6 addresses.
  5. Click Run test.

After you launch the test, you will see the plain text output of the command and three additional buttons in orange, per the image below:

  • Copy to clipboard: Copies command output to your clipboard.
  • Open results page: Opens the output in a separate tab. You can share results via a link with third parties, as this view masks tested IP addresses. The link will remain live for three days.
  • Show BGP map: Provides a graphical representation of the BGP route and shows which autonomous systems the data has to go through on the way from the Gcore’s node to the IP address. This option is only relevant for the BGP command.

In this article, we will examine each command (BGP, ping, and traceroute) using 93.184.216.34 (example.com) as a target IP address and Gcore’s router in Luxembourg (capital of the Duchy of Luxembourg), where our HQ is located. So our settings for all three commands will be the following:

  • Region: Europe
  • Router: Luxembourg (Luxembourg)
  • IP address: 93.184.216.34. We picked example.com as a Looking Glass target server.

Now let’s dive deeper into each command: BGP, ping, and traceroute.

BGP

The BGPLooking Glass shows the best BGP route (or routes) to the destination point. BGP is a dynamic routing protocol that connects autonomous systems (AS)—systems of routers and IP networks with a shared online routing policy. BGP allows various sections of the internet to communicate with one another. Each section has its own set of IP addresses, like a unique ID. BGP captures and maintains these IDs in a database. When data has to be moved from one autonomous system to another over the internet, BGP consults this database to determine the most direct path.

Based on this data, the best route for the packets is built; and this is what the Looking Glass BGP command can show. Here’s an example output:

The data shows two possible paths for the route to travel. The first path has numbers attached; here’s what each of them means:

  1. 93.184.216.0/24: CIDR notation for the network range being routed. The “/24” specifies that the first 24 bits identify the network, leaving 8 bits for individual host addresses within that network. Thus it covers IP addresses from 93.184.216.0 to 93.184.216.255.
  2. via 92.223.88.1 on eno1: The next hop’s IP address and the network interface through which the data will be sent. In this example, packets will be forwarded to 92.223.88.1 via the eno1 interface.
  3. BGP.origin: IGP: Specifies the origin of the BGP route. The IGP (interior gateway protocol) value implies that this route originated within the same autonomous system (AS.)
  4. BGP.as_path: 174 15133: The AS path shows which autonomous systems the data has passed through to reach this point. Here, the data traveled through AS 174 and then to AS 15133.
  5. BGP.next_hop: 92.223.112.66: The next router to which packets will be forwarded.
  6. BGP.med: 84040: The Multi-Exit Discriminator (MED) is a metric that influences how incoming traffic should be balanced over multiple entry points in an AS. Lower values are generally preferred; here, the MED value is 84040.
  7. BGP.local_pref: 80: Local preference, which is used to choose the exit point from the local AS. A higher value is preferred when determining the best path. The local preference of 80 in the route output indicates that this route is more preferred than other routes to the same destination with a lower local preference.
  8. BGP.community: These are tags or labels that can be attached to a route. Output (174,21001) consists of pairs of ASNs and custom values representing a specific routing policy or action to be taken. Routing policies can use these communities as conditions to apply specific actions. The meaning of these values depends on the internal configurations of the network and usually requires documentation from the network provider for interpretation.
  9. BGP.originator_id: 10.255.78.64: This indicates the router that initially advertised the route. In this context, the route originated from the router with IP 10.255.78.64.
  10. BGP.cluster_list: This is used in route reflection scenarios. It lists the identifiers of the route reflectors that have processed the route. Here, it shows that this route has passed through the reflector identified by 10.255.8.68 or 10.255.8.69 depending on the path.

Both routes are part of AS 15133 and pass through AS 174, but they have different next hops (92.223.112.66 and 92.223.112.67.) This allows for redundancy and load balancing.

BGP map

When you run the BGP command, the Show BGP map button will become active. Here’s what we will see for our IP address:

Let’s take this diagram point by point:

  • AS199524 | GCORE, LU: This is the autonomous system belonging to Gcore, based in Luxembourg. The IP 92.223.88.1 is the part of this AS, functioning as a gateway or router.
  • AS174 | COGENT-174, US: This is Cogent Communications’ autonomous system, based in the United States. Cogent is a major ISP.
  • AS15133 | EDGECAST, US: This AS belongs to Edgecast, also based in the United States. Edgecast is generally involved in content delivery network (CDN) services.
  • 93.184.216.0/24: This CIDR notation indicates a network range where example.com (93.184.216.34) is located. It might be a part of Edgecast’s CDN services or another network associated with one of the listed AS.

In summary, Gcore’s BGP Looking Glass command is an essential tool for understanding intricate network routes. By offering insights into autonomous systems, next hops, and metrics like MED and local preference, it allows for a nuanced approach to network management. Whether you’re an ISP peered with Gcore or a network administrator seeking to optimize performance, the data generated by this command offers a roadmap for strategic decision making.

Ping

The ping command is a basic, essential network troubleshooting tool that measures the round-trip time for sending a packet of data from the source to a destination and back. Ping shows the packet transfer speed and can also be used to check the node’s overall availability.

The command utilizes the ICMP protocol. It works as follows:

  • The router sends a packet from the IP address to the node.
  • The node sends it back.

In our case, this command shows how much time it takes to transfer a packet from the specified IP address to the node.

Let’s break down our output:

Main part:

  1. Target IP: You pinged 93.184.216.34, which is the example.com IP address we are testing.
  2. Packet Size: 56(84) bytes of data were sent. The packet consists of 56 bytes of data and 28 bytes of header, totaling 84 bytes.
  3. Individual pings: Each line indicates a single round trip of a packet, detailing:
    • icmp_seq: Sequence number of the packet.
    • ttl: Time-to-Live, showing how many more hops the packet could make before being dropped.
    • time: Round-trip time (RTT) in milliseconds.

Statistics:

  1. 5 packets transmitted, 5 received: All packets were successfully transmitted and received, indicating no packet loss
  2. 0% packet loss: No packets were lost during the transmission
  3. time 4005ms: Total time taken for these five pings
  4. rtt min/avg/max/mdev: Round-trip times in milliseconds:
    • min: minimum time
    • avg: average time
    • max: maximum time
    • mdev: mean deviation time

To summarize, the average round-trip time here is 87.138 ms, and the TTL is 52. RTT of less than 100 ms is generally considered acceptable for interactive applications, and TTL of 50 is considered a good value. No packet loss suggests a stable connection to the IP address 93.184.216.34.

The ping function provides basic, vital metrics for assessing network health. By offering details on round-trip times, packet loss, and TTL, this command allows for a quick yet comprehensive evaluation of network connectivity. For any network stakeholder—whether ISP or end user—understanding these metrics is crucial for effective network management and troubleshooting.

Traceroute

The Looking Glass traceroutecommand is a diagnostic tool that maps out the path packets take from the source to the destination, enabling you to identify potential bottlenecks or network failures. Traceroute relies on the TTL (Time-to-Live) parameter, which basically determines how long this packet can stay in the network. Every router along the packet’s path decrements the TTL by 1 and forwards the packet to the next router in the path. The process works as follows:

  1. The traceroute sends a packet to the destination host with TTL value of 1.
  2. The first router that receives the packet decrements the TTL value by 1 and forwards the packet.
  3. When the TTL reaches zero, the router drops the packet and sends an ICMP Time Exceeded message back to the source host.
  4. The traceroute records the IP address of the router that sent back the ICMP Time Exceeded message.
  5. The traceroute then sends another packet to the destination host with a TTL value of 2.
  6. Steps 2-4 are repeated until the traceroute routine reaches the destination host or until it exceeds the maximum number of hops.

Now let’s apply this command to the address we used earlier. The traceroute command will test our target IP address with 60-byte packets and a maximum of fifteen hops. Here’s what we get as output:

Apart from the header, each output line consists of the following information, labeled on the image below:

  1. IP and hostname: e.g., vrrp.gcore.lu (92.223.88.2)
  2. AS information: Provided in square brackets, e.g., [AS199524/AS202422]
  3. Latency: Time in milliseconds for the packet to reach the hop and return, e.g., 0.202 ms

In our example, traceroute traverses through three different autonomous systems (AS):

  1. AS199524 (GCORE, LU): The first two hops are within this AS, representing the initial part of the route.
  2. Hops 3 and 4 fall under the private IPv4 address space (10.255.X.X), meaning the hops are within a private network. This could be an internal router or other networking device not directly accessible over the public Internet. Private addresses like this are often used for internal routing within an organization or service provider’s network.
  3. AS174 (COGENT, US): Hops 5 to 9 are within Cogent’s network.
  4. AS15133 (EDGECAST, US): The final hops are within EdgeCast’s network, where the destination IP resides.

Example Hop: ae-66.core1.bsb.edgecastcdn.net (152.195.233.131) [AS15133] 82.450 ms

To sum up, the traceroute command offers a comprehensive view of the packet journey across multiple autonomous systems. Providing latency data and AS information at each hop, it aids in identifying potential bottlenecks or issues in the network. This insight is invaluable for anyone looking to understand or troubleshoot a network path.

Conclusion

Looking Glass is a tool for pre-purchase network testing, covering node connectivity, response times, packet paths, and BGP routes. Its user-friendly interface requires just a few inputs—location, target IP address, and the command of your choice—to deliver immediate results.

Based on your specific needs, such as connectivity speeds and location, and the insights gained from Looking Glass test results, you can choose between Gcore Virtual or Dedicated servers, both boasting outstanding connectivity. Want to learn more? Contact our team.

Related articles

Introducing FastEdge Triggers: real-time edge logic

When you're building real-time applications, whether for streaming platforms, SaaS dashboards, or security-sensitive services, you need content that adapts on the fly. Blocking suspicious IPs, injecting personalized content, transforming media on the edge—these should be fast, scalable, and reliable.Until now, they weren't.Developers and technical teams often had to work across multiple departments to create brittle, hardcoded solutions. Each use case, like watermarking video or rewriting headers, required a custom integration. There was no easy way to run logic dynamically at the edge. That changes with FastEdge Triggers.Real-time logic, built into the edgeFastEdge Triggers let you execute custom serverless logic at key moments in the HTTP lifecycle:on_request_headerson_request_bodyon_response_headerson_response_bodyFastEdge is built on the proxy-wasm standard, making it easy to adapt existing proxy-wasm applications (e.g., for Envoy or Kong) for use with Gcore. These trigger types align directly with proxy-wasm conventions, meaning less friction for developers familiar with modern proxy architectures.This means that you can now:Authenticate users' tokens, such as JWTBlock access by IP, region, or user agentInject CSS, HTML, or JavaScript into responsesTransform images or convert markdown to HTML before deliveryAdd security tokens or watermarks to video contentRewrite or sanitize request headers and bodiesNo backend round-trips. No manual routing. Just real-time, programmable edge behavior, backed by Gcore's global infrastructure.While FastEdge enables instant logic execution at the edge, response-stage triggers (on_response_headers and on_response_body) naturally depend on receiving data from the origin before acting. Even so, transformations happen at the edge, reducing backend load and improving overall efficiency.Our architecture means that FastEdge logic is executed in ultra-low-latency environments, tightly coupled with CDN. Triggers can be layered across multiple stages of a request without performance degradation.Built for developersFastEdge Triggers were built to solve three core pain points for technical teams:Hard to scale: Custom logic used to require bespoke, team-specific workaroundsHard to maintain: Even single-team solutions became brittle without proper edge infrastructureLimited flexibility: Legacy CDN logic couldn't support complex, dynamic behaviorWith FastEdge, developers have full control: no DevOps bottlenecks, no workarounds, no compromises. Logic runs at the edge, not your origin, minimizing backend exposure. FastEdge apps execute in isolated, sandboxed environments, reducing the risk of vulnerabilities that might otherwise be introduced when logic runs on central infrastructure.How it works behind the scenesEach FastEdge application is written in Rust or AssemblyScript and connected to the HTTP request lifecycle through Gcore's configuration interface. Apps are linked to trigger types through the CDN resource settings page in the Gcore Customer Portal.Configuring FastEdge Triggers from the CDN resource settings screen in the Gcore Customer PortalHere's what happens under the hood:You assign a FastEdge app to a trigger point.Our Core Proxy detects that trigger and automatically routes execution through your custom logic.The result is returned before hitting cache or origin, modified, enriched, and secured.This flow is deeply integrated with our CDN, delivering minimal latency with zero friction.A sequence diagram showing how FastEdge Triggers works under the hood A real-life use case: markdown to HTML at the edgeHere's a real-world example that shows how FastEdge Triggers can power multi-step content transformation without a single backend server.One customer wanted to serve Markdown-based documentation as styled HTML, without spinning up infrastructure. Using this FastEdge app written in Rust, they achieved just that.The app listens at three trigger points: on_request_headers, on_response_headers, and on_response_bodyIt detects requests for .md files and converts them on the flyThe HTML is served directly via CDN, no origin compute requiredYou can see it live here:README renderedTerraform docs renderedThis use case showcases FastEdge's ability to orchestrate multi-stage logic at the edge: ideal for serverless documentation, lightweight rendering, or content transformation pipelines.Ready to build smarter at the edge?FastEdge Triggers are available now for all FastEdge customers. If you're looking to modernize your edge logic, simplify architecture, and ship faster with fewer backend dependencies, FastEdge is built for you.Reach out to your account manager or contact us to activate FastEdge Triggers in your environment.Try Fastedge Triggers

Gcore and Orange Business launch innovation program piloting joint solution to deliver sovereign inference as a service

Gcore and Orange Business have kicked off a strategic co-innovation program with the mission to deliver a scalable, production-grade AI inference service that is sovereign by design. By combining Orange Business’ secure, trusted cloud infrastructure and Gcore’s AI inference private deployment service, the collaboration empowers European enterprises and public sector organizations to run inference workloads at scale, without compromising on latency, control, or compliance.Gcore’s AI inference private deployment service is already live on Orange Business’ Cloud Avenue infrastructure. Selected enterprises across industries are actively testing it in real-world scenarios. These pilot customers are exploring how fast, secure, and compliant inference can accelerate their AI projects, cut deployment times, and reduce infrastructure overhead.The prototype will be demonstrated at NVIDIA GTC Paris, at the Taiga Cloud booth G26. Stop by any time to see it in action.The inference supercycle is underwayBy 2030, inference will comprise 70% of enterprise AI workloads. Telcos are well positioned to lead this shift due to their dense edge presence, licensed national data infrastructure, and long-standing trust relationships.Gcore’s inference solution provides a sovereign, edge-native inference layer. It enables users to serve real-time, GPU-intensive applications like agentic AI, trusted LLMs, computer vision, and predictive analytics, all while staying compliant with Europe’s evolving data and AI governance frameworks.From complexity to three clicksEnterprise AI doesn’t need to be hard. Deploying inference workloads at scale used to demand Kubernetes fluency, large MLOps teams, and costly trial-and-error.Now? It’s just three clicks:Pick a model: Choose from NVIDIA NIMs, open source, or proprietary libraries.Choose a region: Select one of Orange Business’ accredited EU data centers.Deploy: See your workloads go live in under 10 seconds.Enterprises can launch inference projects faster, test ideas more quickly, and deliver production-ready AI services without spending months on ML plumbing.Explore our blog to watch a demo showing how enterprises can deploy inference workloads in just three clicks and ten seconds.Sovereign by designAll model data, logs, and inference results are stored exclusively within Orange Business’ own data centers in France, Germany, Norway, and Sweden. Cross-border data transfer is opt-in only, helping ensure alignment with GDPR, sector-specific regulations, and the forthcoming EU AI Act.This platform is built for trust, transparency, and sovereignty by default. Customers maintain full control over their data, with governance baked into every layer of the deployment.Performance without trade-offsGcore’s AI inference solution avoids the latency spikes, cold starts, and resource waste common in traditional cloud AI setups. Key design features include:Smart GPU routing: Directs each request to the nearest in-region GPU, delivering real-time performance with sub-50ms latency.Pre-loaded models: Reduces cold start delays and improves response times.Secure multi-tenancy: Isolates customer data while maximizing infrastructure efficiency.The result is a production-ready inference platform optimized for both performance and compliance.Powering the future of AI infrastructureThis partnership marks a step forward for Europe’s sovereign AI capabilities. It highlights how telcos can serve as the backbone of next-generation AI infrastructure, hosting, scaling, and securing workloads at the edge.With hundreds of edge POPs, trusted national networks, and deep ties across vertical industries, Orange Business is uniquely positioned to support a broad range of use cases, including real-time customer service AI, fraud detection, healthcare diagnostics, logistics automation, and public sector digital services.What’s next: validating real-world performanceThis phase of the Gcore and Orange Business program is focused on validating the solution through live customer deployments and performance benchmarks. Orange Business will gather feedback from early access customers to shape its future sovereign inference service offering. These insights will drive refinements and shape the roadmap ahead of a full commercial launch planned for later this year.Gcore and Orange Business are committed to delivering a sovereign inference service that meets Europe’s highest standards for speed, simplicity, and trust. This co-innovation program lays the foundation for that future.Ready to discover how Gcore and Orange Business can deliver sovereign inference as a service for your business?Request a preview

Why on-premises AI is making a comeback

In recent years, cloud AI infrastructure has soared in popularity. With its scalability and ease of deployment, it’s no surprise that organizations rushed to transfer their data to the cloud in a bid to become “cloud-first.”But now, the tide is turning.As AI workloads grow more complex and regulatory pressures increase, many companies are reconsidering their reliance on cloud and turning back toward on-premises AI infrastructure.Rather than doubling down on the cloud, organizations are diversifying—adopting multi-cloud models, sovereign cloud environments, and even hybrid or fully on-prem setups. The era of a single cloud provider handling everything is coming to an end. Why? Control, security, and performance are hard to find in the public cloud.Here’s why more businesses are bringing AI back in-house.#1 Enhanced data security and controlData security remains one of the most urgent concerns driving the return to on-prem infrastructure.For sensitive or high-priority workloads—common in sectors like finance, healthcare, and government—keeping data off the cloud is often non-negotiable. Cloud computing inherently increases risk by exposing data to shared environments, wider attack surfaces, and complex supply chains.Choosing a trusted cloud provider can mitigate some of those risks. But it can’t replace the peace of mind that comes from keeping sensitive data in-house.With on-premises AI, organizations gain fine-grained access control. Encryption keys remain internal and breach exposure shrinks dramatically. It’s also much easier to stay compliant with privacy laws when data never leaves your own secure perimeter.For industries where trust and confidentiality are everything, on-prem solutions offer full visibility into where and how data is stored and processed.#2 Performance enhancement and latency reductionLatency matters—especially in AI.On-premises AI systems excel in environments that require real-time performance and heavy compute loads. Processing data locally avoids the physical delays caused by transferring it across the internet to a cloud data center.By eliminating long-haul network hops, companies get near-instant access to computing resources. They also get to fine-tune their internal networks—using private fiber, low-hop switching, and other low-latency optimizations that cloud customers can’t control.Unlike multi-tenant cloud platforms, on-prem resources aren’t shared. That means consistently low, predictable latency.This is vital for use cases where milliseconds—or even microseconds—make a difference: autonomous vehicles, real-time analytics, robotic control systems, and high-speed trading. Fast feedback loops and localized processing enable better outcomes, tighter control, and faster decision-making at the edge.#3 Regulatory compliance and data sovereigntyAround the world, data privacy regulations are tightening. For most organizations, compliance isn’t optional.On-premises infrastructure helps keep data safely inside the organization’s network. This supports data sovereignty, ensuring that sensitive information remains subject only to local laws—not the policies of another country’s cloud provider.It's also a powerful hedge against geopolitical instability.While hyperscalers operate globally, they’re always headquartered somewhere. That makes their infrastructure vulnerable to political shifts, sanctions, or changes in international data law. Governments may require them to restrict access, share data, or cut off services entirely—especially to organizations in sanctioned or adversarial jurisdictions.Businesses relying on these providers risk disruption when regulations change. On-premises infrastructure, by contrast, offers reliable continuity and greater control—especially in uncertain times.#4 Cost control and operational benefitsCloud pricing may look flexible, but costs can escalate quickly.Data transfers, storage, and compute spikes all add up—fast. In contrast, on-premises infrastructure provides a predictable Total Cost of Ownership (TCO). Although upfront CapEx is higher, OpEx remains more stable over time.Organizations can invest in high-performance hardware tailored to their specific needs and amortize those costs across years. That means no surprise bills, no sudden price hikes, and no dependence on vendor pricing models.Of course, running on-prem infrastructure comes with its own challenges. It demands specialized teams for deployment, maintenance, and support. These experts are costly to recruit and retain—but they’re critical to ensure uptime, security, and performance.Still, for companies with relatively stable compute and storage needs, the long-term savings often outweigh the initial setup effort. On-prem also integrates more smoothly into existing IT workflows, without the need for internet access or additional network setup—another operational bonus.#5 Proactive threat detection and automated responsesOn-premises AI sometimes enables smarter, more customized security.Advanced platforms can continuously analyze live data streams using machine learning to detect anomalies and predict threats. When something suspicious is flagged, the system can respond instantly by quarantining data, blocking traffic, and alerting security teams.That kind of automation is essential for minimizing damage and downtime.With full infrastructure control, organizations can deploy bespoke monitoring systems that align with their threat models. Deep packet inspection, real-time anomaly detection, and behavioral analytics can be easier to configure and maintain on-prem than in shared cloud environments.These systems can also work seamlessly with WAAP and DDoS tools to detect and neutralize threats before they spread. The key is flexibility: whether on-prem or cloud-based, AI-driven security should adapt to your architecture and threat landscape, not the other way around.End-to-end visibility can give security teams a clearer picture and faster response options than generic, one-size-fits-all public cloud security tools.How to combine eon-premises control with cloud scalabilityLet’s be clear: on-premises AI isn’t perfect. It demands upfront investment. It requires skilled personnel to deploy and manage systems. And integrating AI into legacy environments takes thoughtful planning.But today’s tools are helping bridge those gaps. Modern platforms reduce the need for constant manual intervention. They support real-time updates to threat models and detection logic. As a result, security teams can spend more time on strategy and less on maintenance.Meanwhile, the cloud still plays an important role. It offers faster access to new tools, software updates, and next-gen GPU hardware.That’s why many organizations are opting for a hybrid model.Our recommendation: Keep your sensitive, high-priority workloads on-prem. Use the cloud for elastic scale and innovation. Together, they deliver the best of both worlds: performance, control, compliance, and flexibility.Secure your digital infrastructure with Gcore on-premises AI inferenceWhether you’re protecting sensitive data or running high-demand workloads, on-premises AI gives you the control and confidence you need. Securing sensitive data and managing high-demand workloads requires a level of control, performance, and predictability that only on-premises AI infrastructure delivers.Gcore Everywhere Inference Private Deployment makes it easier than ever to bring powerful serverless AI inference capabilities directly into your physical environment. Designed for scalable global performance, Everywhere Inference enables robust and secure multi-tenant AI inference deployments across on-prem and cloud environments, helping you meet data sovereignty requirements, reduce latency, and streamline deployment.Talk to us about your on-prem AI plans

3 clicks, 10 seconds: what real serverless AI inference should look like

Deploying a trained AI model could be the easiest part of the AI lifecycle. After the heavy lifting of data collection, training, and optimization, pushing a model into production is where “the rubber hits the road”, meaning the business expects to see the benefits of invested time and resources. In reality, many AI projects fail in production because of poor performance stemming from suboptimal infrastructure conditions.There are, broadly speaking, two paths developers can take when deploying inference: DIY, which is time and resource-consuming and requires domain expertise from several teams within the business, or the ever-so-popular “serverless inference” solution. The latter is supposed to simplify the task at hand and deliver productivity, cutting down effort to seconds, not hours. Yet most platforms offering “serverless” AI inference still feel anything but effortless. They require containers, configs, and custom scripts. They bury users in infrastructure decisions. And they often assume your data scientists are also DevOps engineers. It’s a far cry from what serverless was meant to be.At Gcore, we believe real serverless inference means this: three clicks and ten seconds to deploy a model. That’s not a tagline—it’s the experience we built. And it’s what infrastructure leaders like Mirantis are now enabling for enterprises through partnerships with Gcore.Why deployment UX matters more than you thinkServerless inference isn’t just a backend architecture choice. It’s a business enabler, a go-to-market accelerator, an ROI optimizer, a technology democratizer—or, if poorly executed, a blocker.The reality is that inference workloads are a key point of interface between your AI product or service and the customer. If deployment is clunky, you’re struggling to keep up with demand. If provisioning takes too long, latency spikes, performance is inconsistent, and ultimately your service doesn’t scale. And if the user experience is unclear or inconsistent, customers end up frustrated—or worse, they churn.Developers and data scientists don’t want to manage infrastructure. They want to bring a model and get results without becoming cloud operators in the process.Dom Wilde, SVP Marketing, MirantisThat’s why deployment UX is no longer a nice-to-have. It’s the core of your product.The benchmark: 3 clicks, 10 secondsWe built Gcore Everywhere Inference to remove every unnecessary step between uploading a model and running it in production. That includes GPU provisioning, routing, scaling, isolation, and endpoint generation, all handled behind the scenes.The result is what we believe should be the default:Upload a modelConfirm deployment parametersClick deployAnd within ten seconds, you’re serving live inference.For platform teams supporting AI workloads, this isn’t just a better workflow. It’s a transformation.With Gcore, our customers can deliver not just self-service infrastructure but also inference as a product. End users can deploy models in seconds, and customers don’t have to micromanage the backend to support that.Dom Wilde, MirantisSimple frontend, powerful backendIt’s worth saying: simplifying the frontend doesn’t mean weakening the backend. Gcore’s platform is built for scale and performance, offering the following:Multi-tenant GPU isolationSmart routing based on location and loadAuto-scaling based on demandA unified API and UI for both automation and accessibilityWhat makes this meaningful isn’t just the tech, it’s the way it vanishes behind the scenes. With Gcore, Mirantis customers can deliver low-latency inference, maximize GPU efficiency, and meet data privacy requirements without touching low-level infrastructure.Many enterprises and cloud customers worry about underutilized GPUs. Now, every cycle is optimized. The platform handles the complexity so our customers can focus on building value.Dom Wilde, MirantisIf it’s not 3 clicks and 10 seconds, it’s not really serverlessThere’s a growing gap between what serverless inference promises and what most platforms deliver. Many cloud providers are focused on raw compute or orchestration, but overlook the deployment layer. That’s a mistake. Because when it comes to customer experience, ease of deployment is the product.Mirantis saw that early on and partnered with Gcore to bring inference-as-a-service to CSP and enterprise customers, fast. Now, customers can launch new offerings more quickly, reduce operational overhead, and improve the user experience with a simple, elegant deployment path.Redefine serverless AI with GcoreIf it takes a config file, a container, and a support ticket to deploy a model, it’s not serverless—it’s server-less-ish. With Gcore Everywhere Inference, we’ve set a new benchmark: three clicks and ten seconds to deploy AI. And, our model catalog offers a variety of popular models so you can get started right away.Whether you’re frustrated with slow, inefficient model deployments or looking for the most effective way to start using AI for your company, you need Gcore Everywhere Inference. Give our experts a call to discover how we can simplify your AI so you can focus on scaling and business logic.Let’s talk about your AI project

Run AI inference faster, smarter, and at scale

Training your AI models is only the beginning. The real challenge lies in running them efficiently, securely, and at scale. AI and reality meet in inference—the continuous process of generating predictions in real time. It is the driving force behind virtual assistants, fraud detection, product recommendations, and everything in between. Unlike training, inference doesn’t happen once; it runs continuously. This means that inference is your operational engine rather than just technical infrastructure. And if you don’t manage it well, you’re looking at skyrocketing costs, compliance risks, and frustrating performance bottlenecks. That’s why it’s critical to rethink where and how inference runs in your infrastructure.The hidden cost of AI inferenceWhile training large models often dominates the AI conversation, it’s inference that carries the greatest operational burden. As more models move into production, teams are discovering that traditional, centralized infrastructure isn’t built to support inference at scale.This is particularly evident when:Real-time performance is critical to user experienceRegulatory frameworks require region-specific data processingCompute demand fluctuates unpredictably across time zones and applicationsIf you don’t have a clear plan to manage inference, the performance and impact of your AI initiatives could be undermined. You risk increasing cloud costs, adding latency, and falling out of compliance.The solution: optimize where and how you run inferenceOptimizing AI inference isn’t just about adding more infrastructure—it’s about running models smarter and more strategically. In our new white paper, “How to Optimize AI Inference for Cost, Speed, and Compliance”, we break it down into three key decisions:1. Choose the right stage of the AI lifecycleNot every workload needs a massive training run. Inference is where value is delivered, so focus your resources on where they matter most. Learn when to use pretrained models, when to fine-tune, and when simple inference will do the job.2. Decide where your inference should runFrom the public cloud to on-prem and edge locations, where your model runs, impacts everything from latency to compliance. We show why edge inference is critical for regulated, real-time use cases—and how to deploy it efficiently.3. Match your model and infrastructure to the taskBigger models aren’t always better. We cover how to choose the right model size and infrastructure setup to reduce costs, maintain performance, and meet privacy and security requirements.Who should read itIf you’re responsible for turning AI from proof of concept into production, this guide is for you.Inference is where your choices immediately impact performance, cost, and customer experience, whether you’re managing infrastructure, developing models, or building AI-powered solutions. This white paper will help you cut through complexity and focus on what matters most: running smarter, faster, and more scalable inference.It’s especially relevant if you’re:A machine learning engineer or AI architect deploying models across environmentsA product manager introducing real-time AI featuresA technical leader or decision-maker managing compute, cloud spend, or complianceOr simply trying to scale AI without sacrificing controlIf inference is the next big challenge on your roadmap, this white paper is where to start.Scale AI inference seamlessly with GcoreEfficient, scalable inference is critical to making AI work in production. Whether you’re optimizing for performance, cost, or compliance, you need infrastructure that adapts to real-world demand. Gcore Everywhere Inference brings your models closer to users and data sources—reducing latency, minimizing costs, and supporting region-specific deployments.Our latest white paper, “How to optimize AI inference for cost, speed, and compliance”, breaks down the strategies and technologies that make this possible. From smart model selection to edge deployment and dynamic scaling, you’ll learn how to build an inference pipeline that delivers at scale.Ready to make AI inference faster, smarter, and easier to manage?Download the white paper

How to comply with NIS2: practical tips and key requirements

The European Union is boosting cybersecurity legislation with the introduction of the NIS2 Directive. The new rules represent a significant expansion in how organizations across the continent approach digital security. NIS2 establishes specific and clear expectations that impact not just technology departments but also legal teams and top decision-makers. It refines old protocols while introducing additional obligations that companies must meet to operate within the EU.In this article, we explain the role and scope of the NIS2 Directive, break down its key security requirements, analyze the anticipated business impact, and provide a checklist of actions that businesses can take to remain in compliance with continually evolving regulatory demands.Who needs to comply with NIS2?The NIS2 Directive applies to essential and important organizations operating within the European Union in sectors deemed critical to society and the economy. NIS2 also applies to non-EU companies offering services within the EU, requiring non-EU companies that offer covered services in the EU without a local establishment to appoint a representative in one of the member states where they operate.In general, organizations with 50 or more employees and an annual turnover above €10M fall under NIS2. Smaller entities can also be included if they provide key services, including energy, transport, banking, healthcare, water supply, digital infrastructure, and public administration.4 key security requirements of NIS2Under the NIS2 Directive, organizations are required to have an integrated approach to cybersecurity. There are 10 basic measures that companies subject to this legislation must follow: risk policies, incident handling, supply-chain security, MFA, cryptography, backups, BCP/DRP, vulnerability management, security awareness, crypto-control, and “informational hygiene”. In this article, we will cover the four most important of them.These four are necessary steps for limiting disruptions and achieving full compliance with stringent regulatory demands. They include incident response, risk management, corporate accountability, and reporting obligations.#1 Incident responseUnder NIS2, a solid incident response is required. Companies must document processes for the detection, analysis, and management of cyber incidents. Additionally, organizations must have a trained team ready to respond quickly when there's a breach, reducing damage and downtime. Having the right plan in place can make the difference between a minor issue and a major disruption.#2 Risk managementContinuous risk evaluation is paramount within NIS2. Businesses should constantly be scouting out internal vulnerabilities and external dangers while following a clear, defined risk management protocol. Regular audits and monitoring help businesses stay a step ahead of future threats.#3 Corporate accountabilityNIS2 emphasizes corporate accountability by requiring clear cybersecurity responsibilities across all management levels, placing direct oversight on executive leadership. Additionally, due to the dependency of most organizations on third-party suppliers, supply chain security is paramount. Executives need to check the security measures of their partners. One weak link in the chain can destroy the entire system, making stringent security measures a prerequisite for all partners to reduce risks.#4 Reporting obligationsTransparency lies at the heart of NIS2. Serious incidents need to be reported promptly to maintain the culture of accountability the directive encourages. Good reporting mechanisms ensure that vital information is delivered to the concerned authorities in a timely manner, akin to formal channels in data protection legislation such as the GDPR.What NIS2 means for applicable organizationsSome of the potential implications of NIS2 include an increased regulatory burden, financial and reputational risks, and operational challenges. These apply to all businesses that are already established in the European Union. With compliance now becoming mandatory in all member states, businesses that have lagged behind in implementing effective cybersecurity measures will be put under increased pressure to improve their processes and systems.Increased regulatory burdenFor most firms, the new directive means a huge increase in their regulatory burden. The broadened scope of the directive applies to more industries, and this may lead to additional administrative tasks. Legal personnel and compliance officers will need to sift through current cybersecurity policies and ensure all parts of the organization are in line with the new requirements. This exercise can entail considerable coordination between different departments, including IT, risk management, and supply chain management.Financial and reputational risksThe penalty for non-compliance is steep. The fines for failure to comply with the NIS2 Directive are comparable to the GDPR fines for non-compliance, up to €10 million or 2% of a company's worldwide annual turnover for critical entities, while important organizations face a fine of up to €7M or 1.4% of their global annual turnover. Financial fines and reputational damage are significant risks that organizations must take into account. A single cybersecurity incident can lead to costly investigations, legal battles, and a loss of trust among customers and partners. For companies that depend on digital infrastructure for their day-to-day operations, the cost of non-compliance can be crippling.Operational challengesNIS2 compliance requires more than administrative change. Firms may have to make investments into new technology when trying to meet the directive's requirements, such as expanded monitoring, expanded protection of data, and sophisticated incident response protocols. Legacy system firms can be put at a disadvantage with the need for rapid cybersecurity improvements.NIS2 compliance checklistDue to the comprehensive nature of the NIS2 Directive, organizations will need to adopt a systematic compliance strategy. Here are 5 practical steps organizations can take to comply:Start with a thorough audit. Organizations must review their current cybersecurity infrastructure and identify areas of vulnerability. This kind of audit helps reveal areas of weakness and makes it easier to decide where to invest funds in new tools and training employees.Develop a realistic incident response plan. It is essential to have a short, actionable plan in place when things inevitably go wrong. Organizations need to develop step-by-step procedures for handling breaches and rehearse them through regular training exercises. The plan needs to be constantly updated as new lessons are learned and industry practices evolve.Sustain continued risk management. Risk management is not a static activity. Organizations need to keep their systems safe at all times and update risk analyses from time to time to combat new issues. This allows for timely adjustments to their approach.Check supply chain security. Organizations need to find out how secure their third-party vendors are. They need to have clear-cut security standards and check periodically to help ensure that all members of the supply chain adhere to those standards.Establish clear reporting channels. Organizations must have easy ways of communicating with regulators. They must establish proper reporting schedules and maintain good records. Training reporting groups to report issues early can avoid delays and penalties.Partner with Gcore for NIS2 successGcore’s integrated platform helps organizations address key security concerns relevant to NIS2 and reduce cybersecurity risk:WAAP: Real-time bot mitigation, API protection, and DDoS defense support incident response and ongoing threat monitoring.Edge Cloud: Hosted in ISO 27001 and PCI DSS-compliant EU data centers, offering scalable, resilient infrastructure that aligns with NIS2’s focus on operational resilience and data protection.CDN: Provides fast, secure content delivery while improving redundancy and reducing exposure to availability-related disruptions.Integrated ecosystem: Offers unified visibility across services to strengthen risk management and simplify compliance.Our infrastructure emphasizes data and infrastructure sovereignty, critical for EU-based companies subject to local and cross-border data regulation. With fully-owned data centers across Europe and no reliance on third-party hyperscalers, Gcore enables businesses to maintain full control over where and how their data is processed.Explore our secure infrastructure overview to learn how Gcore’s ecosystem can support your NIS2 compliance journey with continuous monitoring and threat mitigation.Please note that while Gcore’s services support many of the directive’s core pillars, they do not in themselves guarantee full compliance.Ready to get compliant?NIS2 compliance doesn’t have to be overwhelming. We offer tailored solutions to help businesses strengthen their security posture, align with key requirements, and prepare for audits.Interested in expert guidance? Get in touch for a free consultation on compliance planning and implementation. We’ll help you build a roadmap based on your current security posture, business needs, and regulatory deadlines.Get a free NIS2 consultation

Subscribe to our newsletter

Get the latest industry trends, exclusive insights, and Gcore updates delivered straight to your inbox.