Cydonia-v1.3-Magnum-v4-22B-GGUF

Maintained By
knifeayumu

Cydonia-v1.3-Magnum-v4-22B-GGUF

PropertyValue
Parameter Count22.2B
LicenseMRL
Base ModelTheDrummer/Cydonia-22B-v1.3
LanguageEnglish

What is Cydonia-v1.3-Magnum-v4-22B-GGUF?

This is a GGUF-quantized version of a sophisticated language model that combines TheDrummer's Cydonia-22B-v1.3 with anthracite-org's magnum-v4-22b using SLERP merge methodology. The model offers various quantization options ranging from 8.27GB to 44.5GB, making it adaptable to different hardware configurations.

Implementation Details

The model was created using mergekit with a specific SLERP merge configuration using varying interpolation weights (t: [0.1, 0.3, 0.6, 0.3, 0.1]). It's available in multiple quantization formats including F16, Q8_0, Q6_K, Q5_K_M, Q4_K_S, and Q3_K variants, allowing users to balance between model size and performance.

  • Implements llama.cpp release b4148 for quantization
  • Uses bfloat16 dtype in the base merger
  • Offers 11 different quantization options

Core Capabilities

  • Optimized for conversational AI applications
  • Supports efficient inference with reduced memory footprint
  • Maintains model quality through careful quantization
  • Flexible deployment options through various GGUF formats

Frequently Asked Questions

Q: What makes this model unique?

This model uniquely combines the capabilities of Cydonia-22B-v1.3 and Magnum-v4-22b using carefully tuned SLERP merge parameters, while offering multiple quantization options for various deployment scenarios.

Q: What are the recommended use cases?

The model is well-suited for conversational AI applications, with different quantization options allowing deployment on various hardware configurations, from high-end servers (F16 variant) to more modest systems (Q2_K variant).

The first platform built for prompt engineering