Qwen1.5-110B-Chat

Maintained By
Qwen

Qwen1.5-110B-Chat

PropertyValue
Parameter Count111B
Model TypeDecoder-only Language Model
Licensetongyi-qianwen
PaperTechnical Report
Tensor TypeBF16

What is Qwen1.5-110B-Chat?

Qwen1.5-110B-Chat is a state-of-the-art transformer-based language model representing the beta version of Qwen2. As the largest variant in the Qwen1.5 series with 111B parameters, it introduces significant improvements in chat capabilities and multilingual understanding. The model stands out for its stable 32K context length support and simplified implementation that doesn't require trust_remote_code.

Implementation Details

The model architecture incorporates advanced features including SwiGLU activation, attention QKV bias, and group query attention. It utilizes a sophisticated mixture of sliding window attention and full attention mechanisms, complemented by an improved tokenizer designed for multiple natural languages and code processing.

  • Transformer-based decoder-only architecture
  • Advanced attention mechanisms including QKV bias
  • Improved multilingual tokenizer
  • Supports 32K context length across all sizes
  • Requires transformers >= 4.37.0

Core Capabilities

  • Enhanced chat performance with improved human preference alignment
  • Robust multilingual support for both base and chat models
  • Extensive context processing with 32K token support
  • Simplified integration without trust_remote_code requirement
  • Efficient text generation and processing

Frequently Asked Questions

Q: What makes this model unique?

The model's massive scale (111B parameters), combined with its improved tokenizer and attention mechanisms, makes it particularly powerful for complex language understanding tasks. Its ability to handle 32K context length without compromising performance sets it apart from many other language models.

Q: What are the recommended use cases?

The model excels in chat applications, multilingual content generation, and complex language understanding tasks. It's particularly well-suited for applications requiring long context understanding and natural language processing across multiple languages.

The first platform built for prompt engineering