Our Large Language Model as a Service (LLMaaS) offering gives you access to cutting-edge language models, inferred using SecNumCloud-qualified infrastructure, HDS-certified for healthcare data hosting, and therefore sovereign, calculated in France. Benefit from high performance and optimal security for your AI applications. Your data remains strictly confidential, and is neither exploited nor stored after processing.
Large models
Our large models offer state-of-the-art performance for the most demanding tasks. They are particularly well-suited to applications requiring a deep understanding of language, complex reasoning or the processing of long documents.
Specialised models
Our specialised models are optimised for specific tasks such as code generation, image analysis or structured data processing. They offer an excellent performance/cost ratio for targeted use cases.
ministral-3:3b
ministral-3:8b
functiongemma:270m
granite3.2-vision:2b
qwen3-embedding:0.6b
granite-embedding:278m
qwen3-embedding:4b
bge-m3:567m
embeddinggemma:300m
gpt-oss:20b
qwen3-2507-think:4b
rnj-1:8b
qwen3-vl:2b
qwen3-vl:4b
qwen3.5:0.8b
qwen3.5:4b
qwen3.5:9b
qwen3:0.6b
qwen3-vl:8b
devstral-small-2:24b
deepseek-ocr
mistral-small3.2:24b
translategemma:12b
translategemma:4b
translategemma:27b
voxtral
z-image:16b
Model comparison
This comparison table will help you choose the model best suited to your needs, based on various criteria such as context size, performance and specific use cases.
| Model | Publisher | Parameters | Context (k tokens) | Vision | Agent | Reasoning | Security | Quick * | Energy efficiency * |
|---|---|---|---|---|---|---|---|---|---|
| Large models | |||||||||
| Specialised models | |||||||||
| ministral-3:3b | Mistral AI | 3B | 250000 | ||||||
| ministral-3:8b | Mistral AI | 8B | 250000 | ||||||
| functiongemma:270m | 270M | 32768 | |||||||
| granite3.2-vision:2b | IBM | 2B | 16384 | ||||||
| qwen3-embedding:0.6b | Qwen Team | 0.6B | 32768 | ||||||
| granite-embedding:278m | IBM | 278M | 512 | ||||||
| qwen3-embedding:4b | Qwen Team | 4B | 40000 | ||||||
| bge-m3:567m | BAAI | 567M | 8192 | ||||||
| embeddinggemma:300m | 300M | 2048 | |||||||
| gpt-oss:20b | OpenAI | 20B | 120000 | ||||||
| qwen3-2507-think:4b | Qwen Team | 4B | 250000 | ||||||
| rnj-1:8b | Essential AI | 8B | 32000 | ||||||
| qwen3-vl:2b | Qwen Team | 2B | 250000 | ||||||
| qwen3-vl:4b | Qwen Team | 4B | 250000 | ||||||
| qwen3.5:0.8b | Qwen Team | 0.8B | 250000 | ||||||
| qwen3.5:4b | Qwen Team | 4B | 250000 | ||||||
| qwen3.5:9b | Qwen Team | 9B | 250000 | ||||||
| qwen3:0.6b | Qwen Team | 0.6B | 40000 | ||||||
| qwen3-vl:8b | Qwen Team | 8B | 250000 | ||||||
| devstral-small-2:24b | Mistral AI & All Hands AI | 24B | 200000 | ||||||
| deepseek-ocr | DeepSeek AI | 3B | 8192 | ||||||
| mistral-small3.2:24b | Mistral AI | 24B | 128000 | ||||||
| translategemma:12b | 12B | 128000 | |||||||
| translategemma:4b | 4B | 128000 | |||||||
| translategemma:27b | 27B | 120000 | |||||||
| voxtral | Mistral AI | 4B | 32768 | N.C. | |||||
| z-image:16b | Community | 16B | N.C. | ||||||
Recommended use cases
Here are some common use cases and the most suitable models for each. These recommendations are based on the specific performance and capabilities of each model.
Multilingual dialogue
- nemotron-3-super:120b
- qwen3.5:27b
- nemotron3-nano:30b
- gpt-oss:120b
Analysis of long documents
- nemotron-3-super:120b
- qwen3.5:27b
- qwen3-2507:235b
Programming and development
- qwen3.5:27b
- qwen3-2507:235b
- qwen-coder-next:80b
- nemotron-3-super:120b
Visual analysis
- qwen3.5:27b
- deepseek-ocr
- qwen3.5:35b
Safety and compliance
- granite3-guardian:8b
- qwen3.5:27b
- granite3-guardian:2b
Light and on-board deployments
- qwen3.5:0.8b
- qwen3-vl:2b
- ministral-3:3b