Gaming industry under DDoS attack. Get DDoS protection now. Start onboarding
  1. Home
  2. Blog
  3. Gcore Launches AI Cloud Based on Graphcore IPU in Manassas, US
AI
Cloud
News

Gcore Launches AI Cloud Based on Graphcore IPU in Manassas, US

  • October 27, 2023
  • 2 min read
Gcore Launches AI Cloud Based on Graphcore IPU in Manassas, US

We are pleased to announce the expansion of the Gcore AI Cloud powered by Graphcore Bow IPU to our data center in Manassas, Virginia. Strategically located near Washington DC, Manassas is at the heart of the US’s Eastern Seaboard region, a hub of tech innovation. The AI Cloud provides virtual vPODs with multiple configurations, suitable for running all types of AI/ML workloads—including those that are particularly demanding.

What Is the Basis of the Infrastructure?

The AI cluster is built on the Graphcore Bow Pod64, an IPU system that supports multiple users and various AI workloads across virtual vPODs. The Bow Pod64 contains 94,208 IPU cores and delivers 22.4 petaFLOPS of total compute power. These capabilities allow us to deliver powerful AI computing for training and inference, enabling our customers to develop and deploy their AI/ML models in the cloud and at the edge.

Bow Pod is compatible with all popular ML frameworks, including TensorFlow, PyTorch, and Paddle Paddle, and with cloud-native applications such as Kubernetes, Prometheus, and Grafana.

Manassas Virtual vPOD Configurations

Here are the virtual vPOD configurations we offer at the new Manassas location:

ProductInstance configurationNumber of IPUsPrice per hour
BOW-vPOD460 vCPU / 116GB RAM / 1100GB NVMe (ephemeral) / 100Gbit/s Interconnect1$3.68
BOW-vPOD16120 vCPU / 232GB RAM / 2200GB NVMe (ephemeral) / 100Gbit/s Interconnect16$13.70
BOW-vPOD16240 vCPU / 464GB RAM / 4400GB NVMe (ephemeral) / 100Gbit/s Interconnect16$15.80
BOW-vPOD64240 vCPU / 464GB RAM / 4400GB NVMe (ephemeral) / 100Gbit/s Interconnect64$64.30

All configurations are available now via our Cloud platform.

You can also try out vPOD4 for free for 24 hours. To get the trial, contact our sales team.

What Is a Virtual vPOD?

A virtual vPOD is a flavor in which a Poplar® server—a server that manages Bow IPU servers in a cluster—is deployed on a virtual machine. This is in contrast to a dedicated vPOD, which deploys a Poplar® server on a dedicated bare metal server.

A virtual vPOD provides direct access to a host server and fast connectivity to physical IPU accelerators, offering a cost-effective solution to those who need fast and dynamic access to IPU computing power. You can provision virtual vPODs in just five minutes, compared to fifteen minutes for dedicated physical vPODs.

Benefits of Virtual vPODs

Virtual vPODs let you:

  • Deploy your development environment seamlessly.
  • Install and run code at lightning speed.
  • Enhance your experience with popular ML frameworks like TensorFlow, PyTorch, and Keras.
  • Leverage ephemeral storage for stateless workloads.
  • Use external volumes to connect to external block storage and add new data volumes.
  • Use suspension mode to pause your cluster, useful for unpredictable workloads, resource management, and cost optimization; you can resume the cluster in minutes.

Learn More

Visit our AI IPU product page to explore all the features and benefits of our AI Cloud.

Watch our webinar to learn how IPUs enable you to train AI models for various applications.

Check out our blog to find out how the Gcore AI IPU Cloud can help you with your AI/ML development, training, and inference needs.

Try Gcore AI

Gcore all-in-one platform: cloud, AI, CDN, security, and other infrastructure services.

Related articles

New AI inference models available now on Gcore

We’ve expanded our Application Catalog with a new set of high-performance models across embeddings, text-to-speech, multimodal LLMs, and safety. All models are live today via Everywhere Inference and Everywhere AI, and are ready to deploy i

Introducing Gcore Everywhere AI: 3-click AI training and inference for any environment

For enterprises, telcos, and CSPs, AI adoption sounds promising…until you start measuring impact. Most projects stall or even fail before ROI starts to appear. ML engineers lose momentum setting up clusters. Infrastructure teams battle to b

Introducing AI Cloud Stack: turning GPU clusters into revenue-generating AI clouds

Enterprises and cloud providers face major roadblocks when trying to deploy GPU infrastructure at scale: long time-to-market, operational inefficiencies, and difficulty bringing new capacity to market profitably. Establishing AI environment

Edge AI is your next competitive advantage: highlights from Seva Vayner’s webinar

Edge AI isn’t just a technical milestone. It’s a strategic lever for businesses aiming to gain a competitive advantage with AI.As AI deployments grow more complex and more global, central cloud infrastructure is hitting real-world limits: c

From budget strain to AI gain: Watch how studios are building smarter with AI

Game development is in a pressure cooker. Budgets are ballooning, infrastructure and labor costs are rising, and players expect more complexity and polish with every release. All studios, from the major AAAs to smaller indies, are feeling t

How AI-enhanced content moderation is powering safe and compliant streaming

How AI-enhanced content moderation is powering safe and compliant streaming

As streaming experiences a global boom across platforms, regions, and industries, providers face a growing challenge: how to deliver safe, respectful, and compliant content delivery at scale. Viewer expectations have never been higher, like

Subscribe to our newsletter

Get the latest industry trends, exclusive insights, and Gcore updates delivered straight to your inbox.