Browse Models
The simplest way to self-host OpenHermes 2.5 Mistral 7B. Launch a dedicated cloud GPU server running Lab Station OS to download and serve the model using any compatible app or framework.
Download model weights for local inference. Must be used with a compatible app, notebook, or codebase. May run slowly, or not work at all, depending on your system resources, particularly GPU(s) and available VRAM.
OpenHermes 2.5 Mistral 7B is a fine-tuned version of Mistral 7B trained on GPT-4 outputs and code datasets. It uses ChatML format for structured conversations and shows improved performance on HumanEval (50.7%) and TruthfulQA benchmarks. Available in multiple quantization formats.
OpenHermes-2.5-Mistral-7B represents a significant advancement in open-source language models, building upon the successful OpenHermes series. This fine-tuned version of the Mistral architecture demonstrates substantial improvements across multiple benchmarks while maintaining efficient resource usage.
The model is built on the Mistral 7B architecture and was trained on approximately 1,000,000 entries. The training data primarily consists of GPT-4 generated content, supplemented with high-quality data from open datasets. A notable improvement in this version is the inclusion of additional code datasets, which unexpectedly enhanced performance on non-code benchmarks as well.
The training process involved extensive filtering and format conversion to ShareGPT and then ChatML formats. The model utilizes ChatML as its prompt format, enabling structured multi-turn conversations and maintaining compatibility with OpenAI endpoints. A key feature is the support for system prompts, allowing for more nuanced instructions across multiple conversation turns.
OpenHermes-2.5-Mistral-7B shows significant performance improvements compared to its predecessors and other Mistral fine-tunes. The model achieved notable gains across several key benchmarks:
The model demonstrates particular strength in coding tasks, as shown by its improved HumanEval scores. This improvement came somewhat surprisingly through the addition of code datasets, which enhanced performance across both code and non-code tasks.
OpenHermes-2.5-Mistral-7B exhibits versatile capabilities across different types of tasks. Here are some key application areas:
The model excels at:
The model is available in various quantized formats (GGUF, GPTQ, AWQ, EXL2) to accommodate different deployment scenarios and hardware configurations.
Image Captions: