nli-mpnet-base-v2
Property | Value |
---|---|
Parameter Count | 109M |
License | Apache 2.0 |
Paper | Sentence-BERT Paper |
Vector Dimensions | 768 |
What is nli-mpnet-base-v2?
nli-mpnet-base-v2 is a sophisticated sentence transformer model developed by the sentence-transformers team. It's designed to map sentences and paragraphs into a dense 768-dimensional vector space, making it particularly effective for semantic search, clustering, and similarity tasks.
Implementation Details
The model is built on the MPNet architecture and includes a two-component pipeline: a transformer encoder followed by a pooling layer. It processes input text with a maximum sequence length of 75 tokens and employs mean pooling for generating sentence embeddings.
- Built on MPNet architecture with 109M parameters
- Implements mean pooling strategy for sentence embeddings
- Supports both PyTorch and TensorFlow frameworks
- Compatible with ONNX, Safetensors, and OpenVINO
Core Capabilities
- Sentence and paragraph embedding generation
- Semantic similarity computation
- Text clustering and classification
- Cross-lingual understanding
Frequently Asked Questions
Q: What makes this model unique?
This model stands out for its optimized performance on Natural Language Inference (NLI) tasks and its efficient architecture that combines MPNet's power with practical deployment capabilities across multiple frameworks.
Q: What are the recommended use cases?
The model excels in semantic search applications, document similarity comparison, clustering textual data, and creating sentence embeddings for downstream NLP tasks. It's particularly well-suited for production environments requiring robust sentence embeddings.