Teleut-7b

Maintained By
allura-org

Teleut-7b

PropertyValue
Parameter Count7.62B
Base ModelQwen/Qwen2.5-7B
LicenseApache 2.0
Training Dataallenai/tulu-3-sft-mixture
PrecisionBF16

What is Teleut-7b?

Teleut-7b is an advanced language model that represents a replication attempt of Tulu 3 using the Qwen 2.5 architecture as its foundation. This model demonstrates impressive performance across various benchmarks, particularly excelling in tasks like BBH (64.4%) and MMLU (73.2%) in zero-shot settings.

Implementation Details

The model was trained using the Axolotl framework with specific optimizations including liger rope, rms norm, and GLU activation. Training utilized 8 GPUs with a batch size of 128 and employed the paged_ademamix_8bit optimizer with a cosine learning rate scheduler.

  • Learning rate: 3.5e-06
  • Sequence length: 8192
  • Training epochs: 1
  • Gradient accumulation steps: 2

Core Capabilities

  • Strong performance on Big-Bench Hard (BBH) tasks with 64.4% accuracy
  • Excellent MMLU performance (73.2%) in zero-shot settings
  • Competitive GSM8K performance (78.5%)
  • Robust instruction following capabilities (IFEval: 66.3%)

Frequently Asked Questions

Q: What makes this model unique?

The model combines the architectural advantages of Qwen 2.5 with the training methodology of Tulu 3, resulting in strong performance across various benchmarks while maintaining a relatively compact 7B parameter size.

Q: What are the recommended use cases?

The model is particularly well-suited for tasks requiring strong reasoning capabilities, including mathematical problem-solving, knowledge-intensive tasks, and general instruction following. It performs especially well in zero-shot and few-shot scenarios.

The first platform built for prompt engineering