NeuralMonarch-7B

Maintained By
mlabonne

NeuralMonarch-7B

PropertyValue
Parameter Count7.24B
LicenseCC-BY-NC-4.0
Context Window8,192 tokens
Model TypeInstruction-following LLM
Base ArchitectureMistral-7B

What is NeuralMonarch-7B?

NeuralMonarch-7B is an advanced language model that represents a significant evolution in the 7B parameter space. It's a DPO (Direct Preference Optimization) fine-tuned model based on Monarch-7B, incorporating training from high-quality preference datasets including truthy-dpo-v0.1 and distilabel-intel-orca-dpo-pairs. The model stands out for its exceptional performance on various benchmarks, notably achieving strong scores on the Nous benchmark suite and outperforming larger 70B and 120B parameter models on EQ-bench.

Implementation Details

The model is implemented using LazyMergekit, combining three foundation models: OmniTruthyBeagle-7B-v0, NeuBeagle-7B, and NeuralOmniBeagle-7B. It utilizes FP16 precision and supports the Mistral Instruct chat template.

  • 8k context window for handling longer sequences
  • Optimized for instruction following and reasoning tasks
  • Implements DPO fine-tuning for improved preference alignment
  • Available in GGUF format for efficient deployment

Core Capabilities

  • Strong performance in reasoning tasks (73.21% on ARC-Challenge)
  • Exceptional accuracy on HellaSwag (89.09%)
  • Robust MMLU performance (64.41% accuracy)
  • High truthfulness scores (77.79% on TruthfulQA)
  • Advanced mathematical reasoning (67.78% on GSM8k)

Frequently Asked Questions

Q: What makes this model unique?

NeuralMonarch-7B stands out for its exceptional balance of performance across various benchmarks, particularly in reasoning and truthfulness tasks. It achieves competitive scores against much larger models while maintaining a relatively small 7B parameter footprint.

Q: What are the recommended use cases?

The model excels in instruction following, reasoning tasks, and general conversational applications. It's particularly well-suited for applications requiring strong reasoning capabilities and truthful responses, making it ideal for educational, research, and general-purpose AI applications.

The first platform built for prompt engineering