Browse Models
The simplest way to self-host MythoMax L2. Launch a dedicated cloud GPU server running Lab Station OS to download and serve the model using any compatible app or framework.
Download model weights for local inference. Must be used with a compatible app, notebook, or codebase. May run slowly, or not work at all, depending on your system resources, particularly GPU(s) and available VRAM.
MythoMax-L2-13b is a 13B parameter language model created by merging MythoLogic-L2 and Huginn models using gradient-based tensor manipulation. Built on Llama architecture, it focuses on creative writing and roleplaying tasks, using Alpaca formatting for structured inputs and outputs.
MythoMax-L2-13b represents an innovative advancement in language model development, built upon the foundation of the Llama architecture. This model emerged as an enhanced iteration of MythoMix, created through a sophisticated merger of two predecessor models: MythoLogic-L2-13b and Huginn-13b. What sets this model apart is its novel tensor type merge technique, which specifically focuses on intermingling tensors at the beginning and end of the models to enhance coherency.
The development process involved a meticulous approach to model merging, with unique ratios and gradients applied to each of its 363 tensors. This careful attention to tensor manipulation can be explored in detail through the BlockMerge_Gradient YAML files, which document the scripts and templates used in the model's creation.
MythoMax-L2-13b demonstrates particular excellence in two key areas: roleplaying and storywriting. The model leverages the Alpaca formatting structure, which has proven crucial for optimal performance. This formatting approach requires users to include specific components in their prompts: a system prompt or character card, clear instructions, and a designated section for the model's response.
The model is available in multiple quantized versions through TheBloke's Hugging Face repository, offering flexibility in deployment and usage. These versions include:
Each quantized version maintains the model's core capabilities while optimizing for different deployment scenarios and computational requirements.
MythoMax-L2-13b is part of the broader Mytho Series (Llama 2) Collection, representing an evolution in the series' capabilities. As an improvement over MythoMix, it builds upon the strengths of its predecessors while introducing new architectural innovations through its unique tensor merging approach.
The model operates under an "other" license designation, with the creator promising additional details about the model's architecture and training data in future releases. This transparency approach aligns with the collaborative nature of the AI research community while maintaining appropriate intellectual property protections.