Skip to main content

Documentation Index

Fetch the complete documentation index at: https://gcore.com/docs/llms.txt

Use this file to discover all available pages before exploring further.

The Application Catalog is a curated library of pre-built open-source AI models available in Gcore Everywhere Inference. Models deploy directly without building or configuring a custom container image. The catalog groups models into the following categories:
CategoryDescription
Text LLMLarge language models for text generation, summarization, and reasoning
Text + Image LLMMultimodal models that accept both text and image inputs
Text + Audio LLMMultimodal models that accept both text and audio inputs
Text embeddingModels that convert text into vector representations for semantic search and retrieval
EmbeddingMultimodal embedding models
Image generationModels that generate images from text prompts
Video Super-ResolutionModels that upscale and enhance video quality
Speech recognitionModels that transcribe spoken audio to text
Text to speechModels that convert text to spoken audio
Safety modelModels that classify content for safety and policy compliance

Browse the catalog

1. In the Gcore Customer Portal, navigate to Everywhere Inference > Application Catalog.
Application Catalog main page
The catalog page shows all available models grouped by category. Each model card displays the model name, category, and provider. 2. To filter models by category, click the corresponding category button above the model list. 3. To find a specific model, type its name in the Type to search field.
Application Catalog with Text LLM category filter applied
4. Click any model card to open its detail page. It shows the model name, provider, category, description, and application modules with license links.
Application Catalog model detail page

Deploy a model

Request a quota increase if the account quota is insufficient for the selected flavor.
1. Open the model detail page and click Deploy Application. The Create Deployment form opens with the selected application pre-filled.
Create Deployment form — Deployment Configuration and Routing placement sections
2. Under Routing placement, click Select region and select up to six regions where the model will run. 3. Under Application modules, configure the compute resources:
  • Flavor type — select CPU-optimized or GPU-optimized depending on the model requirements.
  • Flavor — select the hardware configuration from the dropdown.
  • Minimum pods — the minimum number of pods to keep running during low-traffic periods.
  • Maximum pods — the maximum number of pods Gcore provisions during peak traffic.
Create Deployment form — Application modules, Deployment details, and Additional options sections
The main application module cannot be removed. The Expose toggle controls whether its endpoint is publicly accessible.
4. Under Deployment details, enter a name for the deployment. 5. (Optional) Under Additional options, enable the Enable API Key authentication toggle to restrict access to the deployment using API keys. 6. Review the estimated cost in the right panel, then click Deploy model. Gcore creates the deployment and opens the Deployments page, where the deployment status is visible.

Available models

ModelProvider
meta-llama/Llama-3.2-1B-InstructMeta
MiniMaxAI/MiniMax-M2.1MiniMaxAI
MiniMaxAI/MiniMax-M2.5MiniMaxAI
mistralai/Devstral-2-123B-Instruct-2512Mistral AI
mistralai/Devstral-Small-2505Mistral
mistralai/Ministral-3-14B-Reasoning-2512Mistral AI
openai/gpt-oss-120bOpenAI
openai/gpt-oss-20bOpenAI
Qwen/Qwen3-14BQwen
Qwen/Qwen3-235B-A22B-Instruct-2507Qwen
Qwen/Qwen3-30B-A3BQwen
Qwen/Qwen3-30B-A3B-Instruct-2507Qwen
Qwen/Qwen3-30B-A3B-Thinking-2507Qwen
Qwen/Qwen3-32BQwen
Qwen/Qwen3-Coder-30B-A3B-InstructQwen
Qwen/Qwen3.5-122B-A10BQwen
Qwen/Qwen3.5-35B-A3BQwen
Qwen/Qwen3.5-397B-A17B-FP8Qwen
xai-org/grok-2xAI
zai-org/GLM-4.7Z.ai
zai-org/GLM-4.7-FlashZ.ai
ModelProvider
google/gemma-3-27b-itGoogle
Qwen/QVQ-72B-PreviewQwen
Qwen/Qwen3-VL-235B-A22B-InstructQwen
ModelProvider
mistralai/Voxtral-Small-24B-2507Mistral
ModelProvider
Alibaba-NLP/gte-Qwen2-7B-instructAlibaba
BAAI/bge-m3BAAI
intfloat/e5-mistral-7b-instructIntfloat
Qwen/Qwen3-Embedding-4BQwen
Qwen/Qwen3-Embedding-8BQwen
ModelProvider
nvidia/nemotron-colembed-vl-8b-v2NVIDIA
vidore/colpali-v1.3Vidore
ModelProvider
ByteDance/SDXL-LightningByteDance
FLUX.1-devBlack Forest Labs
FLUX.1-schnellBlack Forest Labs
stable-cascadeStability AI
stable-diffusion-3.5-largeStability AI
stable-diffusion-3.5-large-turboStability AI
stable-diffusion-xlStability AI
ModelProvider
facebook/seamless-m4t-v2-largeMeta
mistralai/Voxtral-Mini-4B-Realtime-2602Mistral AI
openai/whisper-large-v3OpenAI
openai/whisper-large-v3-turboOpenAI
ModelProvider
microsoft/VibeVoice-1.5BMicrosoft
microsoft/VibeVoice-7BMicrosoft
ResembleAI/chatterboxResembleAI
ModelProvider
ByteDance-Seed/SeedVR2-7BByteDance
ModelProvider
openai/gpt-oss-safeguard-120bOpenAI
openai/gpt-oss-safeguard-20bOpenAI
Qwen/Qwen3Guard-Gen-8BQwen