gte-Qwen2-1.5B-instruct

Maintained By
Alibaba-NLP

gte-Qwen2-1.5B-instruct

PropertyValue
Parameter Count1.5B
Embedding Dimension1536
Max Input Length32k tokens
LicenseApache 2.0
PaperResearch Paper

What is gte-Qwen2-1.5B-instruct?

gte-Qwen2-1.5B-instruct is the latest addition to the General Text Embedding (GTE) model family, built on the Qwen2-1.5B architecture. This model represents a significant advancement in multilingual text embedding, achieving impressive performance across English, Chinese, French, and Polish languages while maintaining a relatively compact size compared to its larger 7B parameter counterpart.

Implementation Details

The model incorporates several sophisticated architectural elements including bidirectional attention mechanisms and instruction tuning specifically optimized for the query side. It's trained on a comprehensive multilingual corpus using both weakly supervised and supervised approaches, resulting in robust cross-lingual capabilities.

  • Advanced bidirectional attention architecture
  • Query-specific instruction tuning
  • Comprehensive multilingual training
  • Efficient 1536-dimensional embeddings
  • Support for long sequences up to 32k tokens

Core Capabilities

  • Strong performance on MTEB benchmark (67.16% average score)
  • Excellent multilingual capabilities across C-MTEB (67.65%), MTEB-fr (66.60%), and MTEB-pl (64.04%)
  • Efficient text embedding generation for retrieval and classification tasks
  • Robust performance in semantic similarity tasks

Frequently Asked Questions

Q: What makes this model unique?

The model stands out for its excellent balance between model size and performance, offering near state-of-the-art results while being significantly smaller than 7B parameter models. It's particularly notable for maintaining strong performance across multiple languages and tasks.

Q: What are the recommended use cases?

The model excels in various applications including semantic search, document retrieval, text classification, and cross-lingual information retrieval. It's particularly well-suited for production environments where computational efficiency is important but high-quality multilingual embeddings are required.

🍰 Interesting in building your own agents?
PromptLayer provides Huggingface integration tools to manage and monitor prompts with your whole team. Get started here.