Browse Models
The simplest way to self-host Dolphin 2.6 Mistral. Launch a dedicated cloud GPU server running Lab Station OS to download and serve the model using any compatible app or framework.
Download model weights for local inference. Must be used with a compatible app, notebook, or codebase. May run slowly, or not work at all, depending on your system resources, particularly GPU(s) and available VRAM.
Dolphin 2.6 Mistral is a fine-tuned version of Mistral-7B with a 16k context window. Trained on 4.5M examples combining GPT-4 and GPT-3.5 completions using the Orca approach, it shows balanced performance across reasoning and knowledge tasks while maintaining coherent long-form responses.
Dolphin 2.6 Mistral is an open-source language model built on the Mistral-7b architecture, featuring a 16k context window. The model is part of the broader Dolphin project, which creates instruct-tuned language models inspired by Microsoft's Orca paper methodology. This particular iteration represents a significant advancement in the project's evolution, combining the capabilities of the Mistral architecture with specialized training approaches.
The model's development, sponsored by Convai, involved an intensive training process spanning three epochs over two days, utilizing four A100 GPUs. The training implemented full-weight finetuning through the Axolotl framework. A notable technical aspect is its ChatML prompt format implementation, where <|im_end|>
maps to token ID 2 (identical to </s>
), ensuring broad compatibility with various applications including koboldAI.
The Dolphin project's training methodology builds upon a substantial dataset containing approximately 1 million FLANv2 examples enhanced with GPT-4 completions and 3.5 million FLANv2 examples augmented with GPT-3.5 completions. For Dolphin 2.6 Mistral specifically, the model underwent DPO tuning using the argilla/ultrafeedback-binarized-preferences-cleaned
dataset.
A distinguishing characteristic of Dolphin 2.6 Mistral is its "uncensored and unbiased" nature, achieved through careful filtering of the training dataset to remove alignment and bias. However, the creators recommend implementing a custom alignment layer before deployment in production services, acknowledging the model's potential to generate responses to unethical prompts.
The model's capabilities have been extensively evaluated on the Open LLM Leaderboard across multiple datasets, demonstrating strong performance across various tasks:
The Dolphin project continues to evolve, with plans for Dolphin 3.0 focusing on expanding capabilities in several key areas:
The broader Dolphin project roadmap includes releases on various base models, including Xgen 7b 8k, LLaMA 13b, MPT 30b 8k, LLaMA 33b, Falcon 40b, and LLaMA 65b, with licensing dependent on the underlying base models.