Llama-2-7b-chat-hf

Maintained By
NousResearch

Llama-2-7b-chat-hf

PropertyValue
Parameter Count6.74B
Model TypeChat-optimized Language Model
Training Data2.0T Tokens
LicenseMeta Custom Commercial License
Context Length4k tokens

What is Llama-2-7b-chat-hf?

Llama-2-7b-chat-hf is Meta's chat-optimized version of their foundational language model, specifically designed for dialogue applications. This model represents the smallest variant in the Llama 2 family, optimized for production deployment while maintaining strong performance across various tasks.

Implementation Details

The model utilizes an optimized transformer architecture, trained on 2 trillion tokens of publicly available data. It implements supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF) to enhance its dialogue capabilities and safety features.

  • FP16 tensor format for efficient deployment
  • 4k token context window
  • Trained with a learning rate of 3.0 x 10-4
  • Requires specific formatting with INST and SYS tags for optimal performance

Core Capabilities

  • Achieves 45.3% accuracy on MMLU benchmarks
  • Demonstrates 33.29% truthfulness in TruthfulQA evaluations
  • Shows strong performance in code generation with 16.8% pass@1 score
  • Exhibits enhanced safety features with only 21.25% toxicity rate in ToxiGen

Frequently Asked Questions

Q: What makes this model unique?

This model stands out for its optimized balance between performance and resource requirements, making it ideal for production deployment. It incorporates specific safety features and dialogue optimization while maintaining commercial usability.

Q: What are the recommended use cases?

The model is specifically designed for English language assistant-like chat applications. It excels in dialogue scenarios and can be used for various natural language generation tasks, though it should be implemented with appropriate safety testing for specific applications.

The first platform built for prompt engineering