stsb-roberta-large

Maintained By
sentence-transformers

stsb-roberta-large

PropertyValue
Parameter Count355M
LicenseApache 2.0
PaperSentence-BERT Paper
Embedding Dimension1024

What is stsb-roberta-large?

stsb-roberta-large is a deprecated sentence embedding model based on RoBERTa architecture that maps sentences and paragraphs to 1024-dimensional dense vector space. While historically used for semantic tasks like clustering and similarity search, it's now considered outdated due to producing lower quality embeddings compared to modern alternatives.

Implementation Details

The model utilizes a RoBERTa-large backbone with a mean pooling layer. It processes text with a maximum sequence length of 128 tokens and includes both transformer and pooling components in its architecture. The model can be easily implemented using either the sentence-transformers library or HuggingFace Transformers.

  • Built on RoBERTa architecture with 355M parameters
  • Produces 1024-dimensional embeddings
  • Implements mean pooling strategy
  • Supports multiple framework implementations (PyTorch, TensorFlow, JAX)

Core Capabilities

  • Sentence and paragraph embedding generation
  • Semantic similarity computation
  • Text clustering applications
  • Semantic search functionality

Frequently Asked Questions

Q: What makes this model unique?

This model was one of the early implementations of sentence-transformers using RoBERTa architecture. However, it's now deprecated in favor of more modern alternatives that provide better quality embeddings.

Q: What are the recommended use cases?

Given its deprecated status, it's recommended to use newer models from the SBERT.net collection instead. However, if still using this model, it's suitable for basic sentence similarity tasks and semantic search applications where absolute state-of-the-art performance isn't critical.

The first platform built for prompt engineering