Sie befinden sich im Service: LLM Hosting

Available Models

Available Models

information

This page contains information on models available in the LLM context.

We provide a collection of open-weight large language models (LLMs) which are self-hosted on our HPC infrastructure, giving you flexibility and choice in your workflows to choose the ones that best fit your needs. User prompts and requests are processed in real time only. The content of these prompts is not saved, logged, or stored at any point. Your data is therefore handled with high protection standards to ensure confidentiality and security.


Model updates

Our model list will be updated in regular intervals, taking into account our available hardware resources as well as observed demand, usage patterns and utilization.

Current model list

ModelProviderRelease Datemax. Content LengthCapabilitiesLimitation and Comments
Mixtral-8x22BMistral AI2024-04-1764KExcels in reasoning, mathematics, coding, multilingual benchmarkslarge but a bit older
Mistral-Small-3.2-24BMistral AI2025-06-25128KCompact 24B model optimized for low-latency inference; Good overall performance and quality 

Details: Mixtral-8x22B

Mixtral 8×22B is an open-source large language model (LLM) developed by Mistral AI, released in April 2024 under the permissive Apache 2.0 license. It employs a Sparse Mixture-of-Experts (SMoE) architecture:

  • The model spans 141 billion total parameters, but typically activates only 39 billion parameters per inference, thanks to the MoE design
  • It consists of 8 experts, each with 22B parameters, with 2 experts activated per token

Key Capabilities & Strengths:

  • Efficiency & Cost-effectiveness: The SMoE setup allows Mixtral 8x22B to be faster and more cost-efficient than many dense models of similar or larger size (e.g., LLaMA 2 70B)
  • Large Context Window: It supports a context window of 64,000 tokens, enabling it to process and recall large documents with precision
  • Multilingual Proficiency: Mixtral 8x22B is capable in several languages, i.e. English, French, German, Italian, and Spanish, performing strongly on multilingual benchmarks compared to models
  • Strong Performance in Math & Coding: It excels in reasoning-intensive tasks, coding, and mathematics. On benchmarks like GSM8K, HumanEval, and others, Mixtral 8×22B achieves top-tier scores—e.g., ~90.8% on GSM8K maj@8

Details: Mistral-Small-3.2-24B

Mistral-Small-3.2-24B belongs to the "Small" series from Mistral AI, specifically an enhanced iteration of the earlier Mistral Small 3.1 and Small 3 series. All 24B-parameter models are released under the Apache 2.0 license. It was officially introduced around June 2025, with model card version 2506 representing the "Instruct" variant optimized for instruction following.

Key Capabilities & Strengths:

  • Stronger Instruction Following: Shows major improvements on tough benchmarks (Wildbench, Arena Hard), making it more reliable for guided tasks.
  • Reduced Repetition & Infinite Loops: Generates cleaner outputs with fewer cases of runaway or repetitive text.
  • Enhanced STEM & Coding Abilities: Higher accuracy on programming (HumanEval+, MBPP+) and reasoning-heavy benchmarks.
  • Vision Input Support: Can handle images as input, enabling multimodal tasks like document parsing or visual Q&A.
  • Extended Context Window (128K tokens): Capable of working with very long documents, transcripts, or multi-step workflows.

Deprecated models

None.

zuletzt geändert am 10.09.2025

Wie hat Ihnen dieser Inhalt geholfen?

Creative Commons Lizenzvertrag
Dieses Werk ist lizenziert unter einer Creative Commons Namensnennung - Weitergabe unter gleichen Bedingungen 3.0 Deutschland Lizenz