Browse Models
The simplest way to self-host CodeLlama 13B. Launch a dedicated cloud GPU server running Lab Station OS to download and serve the model using any compatible app or framework.
Download model weights for local inference. Must be used with a compatible app, notebook, or codebase. May run slowly, or not work at all, depending on your system resources, particularly GPU(s) and available VRAM.
CodeLlama-13b is Meta's mid-sized code generation model, capable of handling 100k token sequences. Built on Llama 2, it excels at code generation, translation between languages, and code infilling. The 13B parameter size balances efficiency with capability, particularly strong in Python tasks.
CodeLlama-13b is a large language model (LLM) developed by Meta, specifically designed for code generation and understanding. It's part of the Code Llama family of models, which includes 7B, 13B, 34B, and 70B parameter variants. Built upon the Llama 2 architecture, CodeLlama-13b represents a strategic balance between computational efficiency and performance capability.
The model utilizes an optimized transformer architecture, inheriting the foundational strengths of Llama 2 while incorporating specialized modifications for code-related tasks. Trained between January and July 2023, CodeLlama-13b underwent extensive development using Meta's Research Super Cluster, resulting in an estimated 65.3 tCO2eq emissions (fully offset by Meta's sustainability program).
The training process utilized data similar to Llama 2 but with different weights, as detailed in the research paper. While specific details about the training data composition aren't fully disclosed, it's known to be substantially focused on code and related natural language instructions.
CodeLlama-13b demonstrates impressive capabilities across various coding tasks, including:
Within the Code Llama family, the 13B parameter model serves as a middle ground between its siblings. The 7B variant offers faster processing but reduced capabilities, while the 34B model provides superior performance at the cost of increased computational requirements. This positioning makes CodeLlama-13b particularly attractive for applications requiring a balance between performance and resource efficiency.
The model is available in the Hugging Face Transformers format and requires the transformers
and accelerate
libraries for implementation. It accepts text input and generates text output, with generation behavior controllable through parameters such as top_k
, temperature
, and top_p
.
CodeLlama-13b is released under Meta's custom commercial license, which restricts its use in commercial applications. This licensing approach emphasizes Meta's focus on fostering research and development in the field of coding AI. Users interested in deployment should consult the responsible use guide provided by Meta.
The Code Llama family includes specialized variants beyond the base models: