Launch a dedicated cloud GPU server running Laboratory OS to download and run Dolphin 2.7 Mixtral 8X7B using any compatible app or framework.
Direct Download
Must be used with a compatible app, notebook, or codebase. May run slowly, or not work at all, depending on local system resources, particularly GPU(s) and available VRAM.
Open WebUI is an open-source, self-hosted web interface with a polished, ChatGPT-like user experience for interacting with LLMs. Integrates seamlessly with local Ollama installation.
The most full-featured web interface for experimenting with open source Large Language Models. Featuring a wide range of configurable settings, inference engines, and plugins.
Model Report
cognitivecomputations / Dolphin 2.7 Mixtral 8X7B
Dolphin 2.7 Mixtral 8X7B is an instruction-tuned language model built on the Mixtral-8x7B architecture that follows Orca research methodologies for progressive learning from complex explanation traces. The model features uncensored output with high compliance to user instructions, trained using qLoRA methods on approximately 4.5 million examples that combine FLANv2 instances with GPT-4 and GPT-3.5 generated responses, emphasizing detailed reasoning explanations and code generation capabilities.
Explore the Future of AI
Your server, your data, under your control
Dolphin 2.7 Mixtral 8X7B is an open-source, instruction-tuned language model developed as a retraining of previous Dolphin models, leveraging the Mixtral-8x7B backbone and influenced by the methodologies described in Microsoft's Orca research. This model emphasizes high compliance and uncensored output, aiming to support research applications that require flexible control over alignment and behavior. Dolphin 2.7 Mixtral 8X7B is released under a commercial-friendly license, with its dataset available under Apache-2.0.
Symbolic artwork of Dolphin 2.7 Mixtral 8X7B, visually merging advanced technology with the 'dolphin' theme to represent artificial intelligence and reasoning.
Dolphin 2.7 Mixtral 8X7B is based on the Mixtral-8x7B architecture, a Mixture-of-Experts transformer model supporting efficient scaling and robust performance across tasks. The original context window of Mixtral-8x7B is 32,000 tokens, which is finetuned to 16,000 for Dolphin 2.7, balancing capacity with computational efficiency.
Training utilized the qLoRA method for parameter-efficient finetuning and the Axolotl framework, enabling scalable distributed training. Notably, Dolphin 2.7 unfroze the gate layers of the transformer—distinguishing it from Dolphin-2.5/2.6—to further enhance performance. The model's instruction tuning draws directly on the Orca framework, which advocates for learning from complex explanation traces and detailed demonstrations generated by large teacher models such as GPT-4 and ChatGPT. This progressive learning process leverages intermediate tutors (e.g., GPT-3.5-turbo), paralleling the Orca research design.
Training Data and Methodologies
The dataset underlying Dolphin 2.7 Mixtral 8X7B is constructed to closely follow the principles of the Orca paper, focusing on augmenting open instruction data with high-quality, model-generated explanation traces. This involved assembling approximately 1 million FLANv2 instances augmented with completions from GPT-4, and 3.5 million FLANv2 tasks augmented with GPT-3.5 responses. The dataset further includes 75,000 examples of chain-of-thought explanations concentrated in the FLAN-1m subset.
Throughout training, system prompts were systematically distributed to mirror the Orca submix methodology. Explanation tuning was central: the model learned not only from input-output pairs, but from richer signals explaining the rationale behind completions—elicited via system instructions such as "explain like I’m five" or "think step-by-step and justify your response."
Tokenization is handled via LLaMA BPE, with variable-length sequence packing up to a maximum of 2048 tokens. To optimize learning, loss computation was focused only on teacher-generated tokens. The overall approach is aimed at maximizing sample diversity and instructional clarity, supporting robust reasoning and problem-solving abilities.
Technical Capabilities and Behavior
Dolphin 2.7 Mixtral 8X7B is designed to maximize obedience and compliance to user instructions as a function of its uncensored policy, a distinguishing feature among modern large language models. Dataset filtering intentionally removed alignment layers, requiring users to impose custom constraints where necessary. This design choice is intended for research and development settings, enabling investigation of foundational behaviors and alignment strategies.
A particular area of focus has been on code generation and general-purpose problem solving, with sustained exposure to extensive coding data during finetuning. The model natively supports the ChatML conversational format, improving compatibility with downstream applications that expect richly formatted dialogue.
Example output from Dolphin 2.7 Mixtral 8X7B providing detailed, step-by-step technical responses, illustrating the model’s capacity for compliant and thorough instruction following.
Researchers have emphasized that Dolphin 2.7 is not Direct Preference Optimization (DPO) tuned; therefore, system prompts often play an influential role in shaping output style and content. Enhanced obedience has been reported, but optimal results depend on carefully crafted system-level guidance.
Evaluation, Performance, and Limitations
While Dolphin 2.7 Mixtral 8X7B aspires to replicate and extend the capabilities observed in the Orca model family, a comprehensive benchmark for this specific release is not publicly documented. The Orca paper provides relevant context: Orca-13B demonstrated marked improvements on complex reasoning benchmarks, outperforming models such as Vicuna-13B on Big-Bench Hard and AGIEval, and showing strong performance on professional and academic examination datasets.
Orca's open-ended generation quality approached 95% of ChatGPT and 85% of GPT-4, while demonstrating instruction following and more truthful outputs relative to open-source baselines. Toxic content generation rates were mitigated when compared with Vicuna, though still behind the proprietary reference models. A key remaining limitation cited is a persistent performance gap compared to GPT-4, especially on knowledge-intensive and nuanced tasks.
For Dolphin 2.7 Mixtral 8X7B, its uncensored nature enables compliance with a wide range of requests, but this brings increased responsibility: users are advised to implement independent alignment and moderation mechanisms, as outputs may reflect the biases, knowledge gaps, and ethical limitations of the data sources and teacher models. Hallucination, lack of real-world contextual understanding, and stochasticity in reply consistency are additional limitations shared with other models in this class. The model is primarily intended for research, and any use in downstream or production-facing scenarios should be preceded by rigorous safety evaluation.
Applications and Use Cases
Given its architecture and training, Dolphin 2.7 Mixtral 8X7B is suitable for research in instruction following, coding, and reasoning. Its highly compliant conversational behavior, alongside comprehensive exposure to coding datasets, makes it applicable for technical assistants, code generation research, and as a base for continued alignment experiments. The explicit removal of content alignment enables investigations into novel alignment strategies and adversarial robustness, but requires implementers to take full responsibility for any outputs produced.
The Orca-inspired training regime, with its thoroughly annotated explanations, also suggests research utility in interpretable reasoning, explainable artificial intelligence, and automated grading of complex academic or professional examinations.
Open Licensing and Availability
The Dolphin dataset is released under Apache-2.0, promoting both commercial and non-commercial use. The Dolphin 2.7 Mixtral 8X7B model itself carries a permissive commercial license, though versions based on foundational models such as LLaMA are subject to the respective upstream license restrictions.