distiluse-base-multilingual-cased

Maintained By
sentence-transformers

distiluse-base-multilingual-cased

PropertyValue
Parameter Count135M
LicenseApache 2.0
PaperSentence-BERT Paper
Output Dimension512

What is distiluse-base-multilingual-cased?

This is a specialized sentence transformer model designed for generating multilingual sentence embeddings. Built on the DistilBERT architecture, it maps sentences and paragraphs into a 512-dimensional dense vector space, enabling efficient semantic search and clustering across multiple languages.

Implementation Details

The model employs a three-component architecture: a DistilBERT transformer with a maximum sequence length of 128 tokens, a pooling layer that performs mean token pooling, and a dense layer that reduces the 768-dimensional embeddings to 512 dimensions with tanh activation. It's implemented using PyTorch and supports multiple inference frameworks including ONNX and TensorFlow.

  • Multilingual support with maintained case sensitivity
  • Optimized architecture through knowledge distillation
  • Efficient 512-dimensional dense vector output
  • Support for sequences up to 128 tokens

Core Capabilities

  • Semantic sentence similarity computation
  • Cross-lingual text clustering
  • Multilingual document matching
  • Efficient semantic search implementation

Frequently Asked Questions

Q: What makes this model unique?

This model stands out for its multilingual capabilities while maintaining a relatively small size (135M parameters) through distillation. It provides a balance between performance and efficiency, making it suitable for production environments.

Q: What are the recommended use cases?

The model excels in multilingual applications requiring semantic understanding, such as cross-lingual information retrieval, document clustering, and semantic search systems. It's particularly useful when working with multiple languages simultaneously.

The first platform built for prompt engineering