Advanced reasoning over documents — tenancy-exclusive GPUs in both EU regions
Dedicated · EU-FRA + EU-LON
🔒
Meta Llama 4 Maverick / Scout
Latest Llama 4 multimodal models — EU-LON dedicated only
Dedicated · EU-LON
🔤
Cohere Embed 3 / 4 + Rerank 3.5
Semantic search & reranking with EU data residency
Dedicated · EU-FRA + EU-LON
On-demand only (ext.) — EU-FRA
🌐
Google Gemini 2.5 Pro / Flash / Flash-Lite
External API routed through EU-FRA endpoint — data does not reside on OCI hardware
On-demand (ext.) · EU-FRA
⚠️
xAI Grok — No EU Presence
All xAI Grok models are US-only (Ashburn, Chicago, Phoenix) — not suitable for EU data residency requirements
Not available in EU
🏗️ Dedicated AI Clusters
🔒
Required for Fine-Tuning
Fine-tuning jobs run exclusively on dedicated GPU clusters — Cohere T-Few/Vanilla and Meta LoRA cannot run on-demand
Dedicated onlyCohere Command RLlama 3.3 70B
🏢
Data Residency & Compliance
Tenancy-exclusive GPUs; your data never shares hardware — suited for regulated industries (GDPR, HIPAA, financial)
DedicatedAll CohereAll MetaOpenAI gpt-oss
⚠️
Not Available — Google & xAI
Google Gemini and xAI Grok route through external APIs — dedicated clusters are not supported for these providers
On-demand (ext.) only
⚡ On-demand (No Cluster Needed)
🚀
Google Gemini & xAI Grok — Always On-demand
External API call; no cluster provisioning, instant availability, pay-per-token billing
On-demand (ext.)
🌍
Cohere & OpenAI gpt-oss — On-demand in Select Regions
On-demand access available without a dedicated cluster — ideal for PoC and variable workloads
On-demand + Ded.
📝
Fine-Tuning Requires Dedicated
On-demand mode supports inference only — to fine-tune a model you must provision a dedicated AI cluster first
Dedicated required
LEGEND
★ Flagship / Best-in-class
◉ Balanced / Advanced
▷ Speed / Efficiency tier
○ Lightweight / Budget
◆ Specialized
2M/1M+ Context ≥ 1M tokens
256K Context = 192K–512K tokens
128K Context = 128K tokens
✓ Feature supported
✗ Not supported
MoE Mixture of Experts (sparse activation)
US-CHI Region: on-demand + dedicated
EU-FRA Region: dedicated AI clusters only
US-ASH Region: on-demand / external call only
¹ Parameter counts are shown only when officially disclosed by the provider. Proprietary models (Google Gemini, xAI Grok) do not publish parameter counts and are omitted. "—" means not publicly disclosed.
² Fine-tuning on OCI uses dedicated AI clusters (GPU resources belonging exclusively to your tenancy). Cohere supports T-Few & Vanilla strategies; Meta Llama supports LoRA.
³ Retired/deprecated models (Command R legacy, Llama 3 70B, Llama 3.1 70B) are omitted from the main tables.
⁴ Model Import feature (GA 2025) lets you bring your own LLMs from Hugging Face or OCI Object Storage.
⁶ Grok 4.1 Fast context window (2M tokens) is confirmed by OCI documentation. Grok 4 and Grok 4 Fast context windows are not disclosed in OCI official docs.
⚠ AI-generated content — not an official Oracle document. This page was assembled with AI assistance from OCI public documentation. Data may contain errors or be out of date. Always verify against docs.oracle.com before making production decisions.