Supported Mistral Model
You can import large language models from Hugging Face and OCI Object Storage buckets into OCI Generative AI, create endpoints for those models, and use them in the Generative AI service.
These models have a high-performance, decoder-only Transformer architecture featuring Sliding Window Attention (SWA) for efficient long-context handling and optional Grouped Query Attention (GQA) for improved scalability. For more details, see Mistral in the Hugging Face documentation.
Mistral
| Hugging Face Model ID | Model Capability | Recommended Dedicated AI Cluster Unit Shape |
|---|---|---|
| intfloat/e5-mistral-7b-instruct | EMBEDDING | A10_X1 |
-
While you can import any chat, embedding, (and fine-tuned) model supported through Open Model Engine (with vLLM or SGLang runtime), only models explicitly listed on this page are supported for this model family. Unlisted models might have compatibility issues and we recommend that you test any unlisted model before production use. Learn about OCI Generative AI Imported Model Architecture.
- Imported models support the native context length specified by the model provider. However, the effective maximum context length is also limited by OCI Generative AI's underlying hardware setup. To take full advantage of a model's native context length, you might need to provision more hardware resources.
- Fine-tuned models are supported only if they match the supported base model's transformer version and have a parameter count within ±10% of the original.
- For available hardware and steps on how to deploy the imported models, see Managing Imported Models.
- If the recommended unit shape isn't available in the region, select a higher-tier option. For example, if A100 isn't available, select H100.