una-xaberius-34b-v1beta

Maintained By
fblgit

una-xaberius-34b-v1beta

PropertyValue
Parameter Count34.4B
Base ArchitectureLLaMa Yi-34B
LicenseCC-BY-NC-ND-4.0
FrameworkTransformers 4.35.2-UNA

What is una-xaberius-34b-v1beta?

una-xaberius-34b-v1beta is an advanced language model developed by juanako.ai that implements a novel Uniform Neural Alignment (UNA) technique. The model achieved remarkable performance, scoring 74.18 on average across major benchmarks and setting new records for open-source models in specific tasks like MMLU with a score of 78.15.

Implementation Details

Built on the LLaMa Yi-34B architecture, this model incorporates specialized training using SFT, DPO, and the proprietary UNA technique. It operates in FP16 precision and was trained on multiple high-quality datasets including Tree of Knowledge, Open-Platypus, UltraFeedback, and OpenOrca.

  • Implements novel UNA technique (distinct from layer merging or SLERP)
  • Optimized for both ChatML and Alpaca system prompts
  • Trained on Cybertron's H100 infrastructure
  • Compatible with text-generation-inference systems

Core Capabilities

  • Exceptional performance on MMLU (78.15)
  • Strong reasoning capabilities (ARC Challenge: 70.39)
  • Advanced common sense understanding (HellaSwag: 86.77)
  • Robust mathematical reasoning (GSM8K: 63.38)
  • High truthfulness score (TruthfulQA: 61.45)

Frequently Asked Questions

Q: What makes this model unique?

The model's distinctive feature is its UNA (Uniform Neural Alignment) implementation, which sets it apart from traditional merging techniques while achieving state-of-the-art performance for its parameter size.

Q: What are the recommended use cases?

The model excels in various tasks including reasoning, mathematics, and general knowledge applications. It's particularly well-suited for applications requiring high accuracy in multiple-choice scenarios and complex reasoning tasks.

The first platform built for prompt engineering