medius-erebus-magnum-14b-GGUF

Maintained By
QuantFactory

Medius-Erebus-Magnum-14B-GGUF

PropertyValue
Parameter Count14.8B
Model TypeTransformers/GGUF
Base ModelQwen/Qwen2.5-14B
Training FrameworkAxolotl 0.4.1

What is medius-erebus-magnum-14b-GGUF?

Medius-Erebus-Magnum is a quantized version of the original model created using llama.cpp, built on the foundation of Qwen2.5-14B. It represents a significant advancement in conversational AI, leveraging advanced training techniques and optimization strategies.

Implementation Details

The model was trained using state-of-the-art techniques including gradient checkpointing with unsloth, flash attention, and specialized plugins like LigerPlugin. Training utilized a sequence length of 32768 with sample packing, and employed the AdamW 8-bit optimizer with a cosine learning rate scheduler.

  • Training conducted over 2 epochs with a learning rate of 0.000008
  • Implements flash attention and RoPE optimizations
  • Uses ChatML template for conversation formatting
  • Trained on diverse datasets including instruction and conversation data

Core Capabilities

  • Advanced conversational abilities with extended context handling
  • Optimized for efficiency through GGUF quantization
  • Enhanced attention mechanisms through Liger optimizations
  • Robust instruction following and response generation

Frequently Asked Questions

Q: What makes this model unique?

The model combines the powerful Qwen2.5-14B architecture with advanced optimization techniques, including GGUF quantization and specialized attention mechanisms, making it both powerful and efficient for deployment.

Q: What are the recommended use cases?

This model is particularly well-suited for conversational applications, instruction-following tasks, and general-purpose language understanding and generation where efficiency and performance are crucial.

The first platform built for prompt engineering