Browse Models
The simplest way to self-host CodeLlama 70B. Launch a dedicated cloud GPU server running Lab Station OS to download and serve the model using any compatible app or framework.
Download model weights for local inference. Must be used with a compatible app, notebook, or codebase. May run slowly, or not work at all, depending on your system resources, particularly GPU(s) and available VRAM.
CodeLlama-70b is Meta AI's largest code generation model with 70B parameters, built on Llama 2. Available in base, Python-specialized, and instruction-following variants, it handles 100k token contexts and excels at code completion, generation, debugging, and infilling across multiple programming languages.
CodeLlama-70b is a large language model specializing in code generation and understanding, representing the largest and most capable variant in the Code Llama family of models. Built upon the Llama 2 architecture, it features 70 billion parameters and utilizes an optimized transformer architecture designed specifically for coding tasks.
The model was trained between January 2023 and January 2024 using Meta's Research Super Cluster and custom training libraries. The training process involved 1400K GPU hours on A100-80GB hardware, resulting in estimated emissions of 228.55 tCO2eq, which Meta fully offset through their sustainability program. The model supports context lengths of up to 16k tokens during training and can handle up to 100k tokens during inference.
The training focused on a massive dataset of code and code-related text, though specific details about the dataset composition remain undisclosed. This specialized training enables the model to excel at various coding tasks, including code generation, completion, debugging, and instruction following.
The Code Llama family includes models of various sizes (7B, 13B, 34B, and 70B parameters) and three distinct variants for each size:
The 70B parameter model represents the most advanced version, demonstrating superior performance across various coding benchmarks compared to its smaller siblings. While larger models generally exhibit better performance on complex tasks, they also require more computational resources.
CodeLlama-70b accepts text input and generates text output, capable of handling various tasks including:
The model is released under Meta's non-commercial license, making it free for research and non-commercial purposes. This licensing approach aims to promote research and development in the field of AI-assisted coding while maintaining certain usage restrictions.
More detailed information about the model's training and evaluation can be found in the research paper "Code Llama: Open Foundation Models for Code".