As 2024 draws to a close, we’re excited to share significant updates to Gcore Edge AI, which includes Inference at the Edge and GPU Cloud. These updates are designed to enhance operational efficiency, improve compliance with data sovereignty requirements, and provide cost-effective solutions as you scale AI workloads in 2025. Let’s dive into the latest enhancements that set the stage for an exciting 2025.
Partnership with Sesterce
We are thrilled to announce our partnership with Sesterce, which combines Gcore’s Inference at the Edge infrastructure with Sesterce’s industry-specific AI models, training frameworks, and deployment solutions. This collaboration provides businesses with an end-to-end AI platform, simplifying the development, training, and deployment of AI models across cloud, on-premises, and edge environments. By leveraging Gcore’s global edge network for low-latency processing and Sesterce’s expertise in AI applications, customers can accelerate AI adoption, reduce operational complexity, and achieve real-time insights. End users benefit from faster, smarter, and more reliable AI-driven services, tailored to their specific needs. Learn more in our dedicated blog.
Native logs for inference deployment tracking
Operational visibility is a cornerstone of effective AI deployment, and our new Logs for Inference Deployments feature delivers precisely that. With this tool, you can track and analyze model logs directly from the Gcore Customer Portal, enabling you to optimize performance, troubleshoot issues, and gain actionable insights. This capability is available for all deployments, providing an intuitive and centralized way to monitor your AI operations. This feature is available now for all Inference at the Edge customers.
New models available in the Gcore model library
We’ve expanded the Gcore model library, which now includes even more cutting-edge models across domains. These additions mean your AI projects can harness the most advanced AI tools available at the click of a button for creativity, automation, and analytical insights.
Image generation models:
- Stable Diffusion XL Base 1.0 delivers remarkable image quality and precision.
- SDXL Lightning is optimized for rapid image generation.
- Stable Cascade is a versatile model for diverse creative applications.
- FLUX.1-schnell and FLUX.1-dev offers high-performance options for demanding image tasks.
- Stable Diffusion 3.5 Large Turbo combines speed and quality.
- Stable Diffusion 3.5 Large focuses on intricate detail and vibrant outputs.
Reasoning AI models:
- Mistral-Nemo-Instruct-2407 is designed for handling complex instructional tasks and nuanced responses.
- Pixtral-12B-2409 acts as a multimodal powerhouse for visual processing.
- Llama-3.2-1B-Instruct is a lightweight yet efficient model tailored for instructional use cases.
- Qwen2.5-7B-Instruct and Qwen2-VL-7B-Instruct delivers next-generation vision-language models with exceptional performance.
- QwQ-32B-Preview offers cutting-edge advancements in large-scale AI modeling, enabling more complex and accurate applications.
Looking Ahead to 2025
These updates represent our commitment to AI innovation and to delivering solutions that empower businesses to thrive in the AI era. From expanding our model library to new partnerships, we’re shaping the future of AI for enterprises worldwide.
Gcore Edge AI has major news coming in early 2025. Be the first to hear about it by subscribing to our newsletter.