Our offer Large Language Model as a Service (LLMaaS) gives you access to state-of-the-art language models, inferred using a qualified infrastructure SecNumCloudcertified HDS for hosting health data, and therefore sovereign, calculated in France. Benefit from high performance and optimum security for your AI applications. Your data remains strictly confidential and is not used or stored after processing.
Modèles de grande taille
Our large models offer state-of-the-art performance for the most demanding tasks. They are particularly well-suited to applications requiring a deep understanding of language, complex reasoning or the processing of long documents.
Llama 3.3 70B
Gemma 3 27B
DeepSeek-R1 70B
Qwen3 30B-A3B FP8
Modèles spécialisés
Our specialised models are optimised for specific tasks such as code generation, image analysis or structured data processing. They offer an excellent performance/cost ratio for targeted use cases.
Qwen3 14B
Gemma 3 12B
Gemma 3 4B
Gemma 3 1B
Lucie-7B-Instruct
Mistral Small 3.1
DeepCoder
Granite 3.2 Vision
Granite 3.3 8B
Granite 3.3 2B
Granite 3.1 MoE
Cogito 14B
Cogito 32B
QwQ-32B
DeepSeek-R1 14B
DeepSeek-R1 32B
Cogito 3B
Granite Embedding
Granite 3 Guardian 2B
Granite 3 Guardian 8B
Qwen 2.5 0.5B
Qwen 2.5 1.5B
Qwen 2.5 14B
Qwen 2.5 32B
Qwen 2.5 3B
Qwen3 0.6b
Qwen3 1.7b
Qwen3 4b
Qwen3 8b
Foundation-Sec-8B
Comparaison des modèles
This comparison table will help you choose the model best suited to your needs, based on various criteria such as context size, performance and specific use cases.
Modèle | Editeur | Paramètres | Contexte (k tokens) | Vision | Agent | Raisonnement | Sécurité | Rapide * | Efficience énergétique * |
---|---|---|---|---|---|---|---|---|---|
Modèles de grande taille | |||||||||
Llama 3.3 70B | Meta | 70B | 60000 | ||||||
Gemma 3 27B | 27B | 120000 | |||||||
DeepSeek-R1 70B | DeepSeek AI | 70B | 60000 | ||||||
Qwen3 30B-A3B FP8 | Qwen Team | 30B-A3B | 60000 | ||||||
Modèles spécialisés | |||||||||
Qwen3 14B | Qwen Team | 14B | 60000 | ||||||
Gemma 3 12B | 12B | 120000 | |||||||
Gemma 3 4B | 4B | 120000 | |||||||
Gemma 3 1B | 1B | 32000 | |||||||
Lucie-7B-Instruct | OpenLLM-France | 7B | 32000 | ||||||
Mistral Small 3.1 | Mistral AI | 24B | 60000 | ||||||
DeepCoder | Agentica x Together AI | 14B | 32000 | ||||||
Granite 3.2 Vision | IBM | 2B | 16384 | ||||||
Granite 3.3 8B | IBM | 8B | 60000 | ||||||
Granite 3.3 2B | IBM | 2B | 120000 | ||||||
Granite 3.1 MoE | IBM | 3B | 32000 | ||||||
Cogito 14B | Deep Cogito | 14B | 32000 | ||||||
Cogito 32B | Deep Cogito | 32B | 32000 | ||||||
QwQ-32B | Qwen Team | 32B | 32000 | ||||||
DeepSeek-R1 14B | DeepSeek AI | 14B | 32000 | ||||||
DeepSeek-R1 32B | DeepSeek AI | 32B | 32000 | ||||||
Cogito 3B | Deep Cogito | 3B | 32000 | ||||||
Granite Embedding | IBM | 278M | 32000 | N.C. | |||||
Granite 3 Guardian 2B | IBM | 2B | 8192 | N.C. | |||||
Granite 3 Guardian 8B | IBM | 8B | 32000 | N.C. | |||||
Qwen 2.5 0.5B | Qwen Team | 0.5B | 32000 | ||||||
Qwen 2.5 1.5B | Qwen Team | 1.5B | 32000 | ||||||
Qwen 2.5 14B | Qwen Team | 14B | 32000 | ||||||
Qwen 2.5 32B | Qwen Team | 32B | 32000 | ||||||
Qwen 2.5 3B | Qwen Team | 3B | 32000 | ||||||
Qwen3 0.6b | Qwen Team | 0.6B | 32000 | ||||||
Qwen3 1.7b | Qwen Team | 1.7B | 32000 | ||||||
Qwen3 4b | Qwen Team | 4B | 32000 | ||||||
Qwen3 8b | Qwen Team | 8B | 60000 | ||||||
Foundation-Sec-8B | Foundation AI - Cisco | 8B | 16000 |
Cas d'usage recommandés
Here are some common use cases and the most suitable models for each. These recommendations are based on the specific performance and capabilities of each model.
Multilingual dialogue
- Llama 3.3
- Mistral Small 3.1
- Qwen 2.5
- Granite 3.3
Analysis of long documents
- Gemma 3
- DeepSeek-R1
- Granite 3.3
Programming and development
- DeepCoder
- QwQ
- DeepSeek-R1
- Granite 3.3
Visual analysis
- Granite 3.2 Vision
- Mistral Small 3.1
- Gemma 3
Safety and compliance
- Granite Guardian
- Granite 3.3
- Lucie
- Mistral Small 3.1
Light and on-board deployments
- Gemma 3
- Granite 3.1 MoE
- Granite Guardian
- Granite 3.3