Browse Models
Mixtral is a family of open-weight large language models developed by Mistral AI that utilizes Sparse Mixture of Experts (SMoE) transformer architectures. The models activate only a subset of experts per token, enabling large parameter counts while maintaining efficient inference. Key versions include Mixtral 8x7B and 8x22B, with various fine-tuned derivatives supporting multilingual tasks, code generation, and instruction following.