Browse Models
The simplest way to self-host Llama 2 13B. Launch a dedicated cloud GPU server running Lab Station OS to download and serve the model using any compatible app or framework.
Download model weights for local inference. Must be used with a compatible app, notebook, or codebase. May run slowly, or not work at all, depending on your system resources, particularly GPU(s) and available VRAM.
Llama 2 13B is Meta AI's mid-sized language model with 13 billion parameters, trained on 2 trillion tokens of public data. It balances computational efficiency with capability, showing strength in dialogue tasks through RLHF training. Performs well in code, reasoning, and comprehension while maintaining a 4k token context window.
Llama 2 13B is a large language model developed by Meta AI, released on July 18, 2023, as part of the Llama 2 family of models. Built upon an optimized transformer architecture, it represents a middle ground in the model family, offering a balance between performance and computational requirements with its 13 billion parameters.
The model utilizes an auto-regressive transformer architecture, trained on approximately 2 trillion tokens of publicly available online data between January and July 2023. Meta explicitly excludes user data from their platforms in the training dataset. The model supports a context length of 4k tokens and accepts text input to generate text output. While the larger 70B parameter variant employs Grouped-Query Attention (GQA) for improved inference scalability, this specific optimization is not implemented in the 13B version.
The training process incorporated both supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF) to enhance the model's helpfulness and safety. This approach has resulted in significant improvements over its predecessor, Llama 1, particularly in instruction-following capabilities and general performance across various benchmarks, as detailed in the research publication.
Llama 2 13B demonstrates improved performance compared to the smaller 7B parameter model across multiple benchmarks, including:
While the 13B model shows marked improvements over the 7B variant, it consistently performs below the larger 70B parameter model across all benchmarks. This positioning makes it an attractive middle-ground option for users who require better performance than the 7B model but cannot accommodate the computational requirements of the 70B version.
The model is primarily designed for English language tasks but shows capability in multilingual applications. It comes in both pretrained and fine-tuned versions, with the latter specifically optimized for chat applications through extensive testing and refinement.
Meta offers Llama 2 13B under a custom commercial license, making it accessible for both research and commercial applications. The model's release represents a significant contribution to the field of AI research, providing a powerful open foundation model that researchers can build upon.
The model's significance stems from its improved performance over previous versions and its accessibility to the research community. It serves as a valuable resource for advancing natural language processing research and developing practical applications in areas such as chatbots, code generation, and text analysis.