Cydonia-v1.3-Magnum-v4-22B-GGUF
Property | Value |
---|---|
Parameter Count | 22.2B |
License | MRL |
Base Model | TheDrummer/Cydonia-22B-v1.3 |
Language | English |
What is Cydonia-v1.3-Magnum-v4-22B-GGUF?
This is a GGUF-quantized version of a sophisticated language model that combines TheDrummer's Cydonia-22B-v1.3 with anthracite-org's magnum-v4-22b using SLERP merge methodology. The model offers various quantization options ranging from 8.27GB to 44.5GB, making it adaptable to different hardware configurations.
Implementation Details
The model was created using mergekit with a specific SLERP merge configuration using varying interpolation weights (t: [0.1, 0.3, 0.6, 0.3, 0.1]). It's available in multiple quantization formats including F16, Q8_0, Q6_K, Q5_K_M, Q4_K_S, and Q3_K variants, allowing users to balance between model size and performance.
- Implements llama.cpp release b4148 for quantization
- Uses bfloat16 dtype in the base merger
- Offers 11 different quantization options
Core Capabilities
- Optimized for conversational AI applications
- Supports efficient inference with reduced memory footprint
- Maintains model quality through careful quantization
- Flexible deployment options through various GGUF formats
Frequently Asked Questions
Q: What makes this model unique?
This model uniquely combines the capabilities of Cydonia-22B-v1.3 and Magnum-v4-22b using carefully tuned SLERP merge parameters, while offering multiple quantization options for various deployment scenarios.
Q: What are the recommended use cases?
The model is well-suited for conversational AI applications, with different quantization options allowing deployment on various hardware configurations, from high-end servers (F16 variant) to more modest systems (Q2_K variant).