dolphin-2.5-mixtral-8x7b-GPTQ

Maintained By
TheBloke

Dolphin 2.5 Mixtral 8x7B GPTQ

PropertyValue
Model Size6.09B parameters
LicenseApache 2.0
AuthorTheBloke (Quantized) / Eric Hartford (Original)
Tensor TypeI32, BF16, FP16

What is dolphin-2.5-mixtral-8x7b-GPTQ?

This is a GPTQ-quantized version of the Dolphin 2.5 Mixtral model, specifically optimized for efficient deployment while maintaining high performance. The model stands out for its exceptional coding capabilities and general-purpose use, trained on multiple high-quality datasets including OpenHermes, Magicoder, and PureDove.

Implementation Details

The model offers multiple quantization options, from 3-bit to 8-bit versions, with various group sizes to accommodate different hardware configurations. It uses the ChatML prompt format and requires Transformers 4.36.0 or later, along with specific AutoGPTQ requirements.

  • Multiple GPTQ variants available (3-bit, 4-bit, 8-bit)
  • Context length of 8192 tokens
  • Optimized for both coding and general tasks
  • Supports various group sizes for different VRAM requirements

Core Capabilities

  • Advanced coding assistance and generation
  • General conversation and task completion
  • Structured output handling
  • High compliance and instruction-following
  • Multi-turn dialogue support

Frequently Asked Questions

Q: What makes this model unique?

This model combines the power of Mixtral's architecture with specific optimizations for coding tasks, while maintaining strong general-purpose capabilities. Its various quantization options make it highly accessible across different hardware setups.

Q: What are the recommended use cases?

The model excels in coding tasks, general conversation, and structured output generation. It's particularly well-suited for developers needing coding assistance, technical writing, and complex problem-solving tasks.

The first platform built for prompt engineering