pubmedbert-base-embeddings

Maintained By
NeuML

PubMedBERT Base Embeddings

PropertyValue
LicenseApache 2.0
Vector Dimension768
Downloads121,361
FrameworkPyTorch, Transformers
LanguageEnglish

What is pubmedbert-base-embeddings?

PubMedBERT Base Embeddings is a specialized language model fine-tuned using sentence-transformers framework on medical literature. Built upon Microsoft's BiomedNLP-PubMedBERT, it transforms medical text into 768-dimensional dense vectors, specifically optimized for medical domain tasks like semantic search and clustering.

Implementation Details

The model utilizes a sophisticated architecture combining BERT-based transformation with mean pooling. It achieved state-of-the-art performance across multiple medical text evaluation benchmarks, outperforming general-purpose models with averages of 95.64% on key medical datasets including PubMed QA, PubMed Subset, and PubMed Summary.

  • Trained using MultipleNegativesRankingLoss with a scale of 20.0
  • Implements AdamW optimizer with 2e-05 learning rate
  • Uses WarmupLinear scheduler with 10,000 warmup steps
  • Supports max sequence length of 512 tokens

Core Capabilities

  • Generates high-quality medical text embeddings
  • Supports semantic search in medical literature
  • Enables document clustering and similarity analysis
  • Facilitates retrieval augmented generation (RAG)
  • Compatible with multiple frameworks (txtai, sentence-transformers, HuggingFace)

Frequently Asked Questions

Q: What makes this model unique?

This model stands out for its specialized focus on medical literature, achieving superior performance compared to general-purpose models. It consistently outperforms other models in medical text similarity tasks, with an average correlation of 95.64% across standard benchmarks.

Q: What are the recommended use cases?

The model excels in medical literature applications including semantic search, document similarity matching, clustering of medical papers, and as a component in RAG systems for medical AI applications. It's particularly effective when working with PubMed-style medical content.

The first platform built for prompt engineering