Browse Models
Note: Codestral 22B v0.1 weights are released under a Non-Production License, and cannot be utilized for commercial purposes. Please read the license to verify if your use case is permitted.
The simplest way to self-host Codestral 22B v0.1. Launch a dedicated cloud GPU server running Lab Station OS to download and serve the model using any compatible app or framework.
Download model weights for local inference. Must be used with a compatible app, notebook, or codebase. May run slowly, or not work at all, depending on your system resources, particularly GPU(s) and available VRAM.
Codestral 22B is a code-focused model built on MistralForCausalLM architecture, featuring a 32,000-token context window. Trained on 80+ programming languages with emphasis on Python, Java, C++, and JavaScript, it excels at code completion, testing, documentation, and Fill-in-the-Middle prediction tasks.
Codestral 22B v0.1 represents Mistral AI's first specialized code generation model, built on the MistralForCausalLM architecture and utilizing the Mistral tokenizer (v3). The model specializes in code generation and understanding across a diverse range of programming languages, positioning it as a significant advancement in AI-powered coding assistance.
The model was trained on an extensive dataset encompassing more than 80 programming languages, with particular emphasis on widely-used languages such as Python, Java, C, C++, JavaScript, and Bash. A distinguishing feature of Codestral 22B is its 32,000-token context window, significantly surpassing competitors that typically offer 4,000 to 16,000 tokens. This expanded context enables superior performance in long-range code generation tasks, as demonstrated through RepoBench evaluations.
Codestral 22B v0.1 demonstrates robust capabilities across multiple programming tasks:
The model's performance has been extensively benchmarked across various frameworks:
Particularly noteworthy is its performance in Python-specific tasks, where it shows improvements over GPT-4-Turbo and GPT-3.5-Turbo in certain metrics. The model's Fill-in-the-Middle capabilities have been benchmarked against DeepSeek Coder 33B, showing competitive results:
The model can be implemented using either mistral_inference
or the transformers
library. However, users should note that the transformers
tokenizer may require community corrections to match the mistral_common
reference implementation. The model supports integration with various development environments and frameworks, making it versatile for different coding workflows.
A notable limitation is the absence of built-in moderation mechanisms, which should be considered when implementing the model in production environments. The model is available under the MNPL-0.1 license, which allows for research and testing purposes while requiring separate licensing for commercial applications.
The images used above show: