bge-reranker-base

Maintained By
BAAI

BGE Reranker Base

PropertyValue
Parameter Count278M
LicenseMIT
LanguagesChinese, English
FrameworkPyTorch, ONNX

What is bge-reranker-base?

BGE Reranker Base is a cross-encoder model designed for efficient reranking of search results and document pairs. Unlike traditional embedding models, it directly computes relevance scores between query-document pairs, offering higher accuracy for ranking tasks. The model contains 278M parameters and supports both Chinese and English languages.

Implementation Details

The model utilizes a cross-encoder architecture based on XLM-RoBERTa, performing full-attention over input pairs to generate accurate relevance scores. It's optimized using cross-entropy loss and can be deployed using PyTorch or ONNX for efficient inference.

  • Achieves MAP scores of 81.27% on CMedQAv1 and 84.10% on CMedQAv2
  • Supports both PyTorch and ONNX runtime for flexible deployment
  • Includes Safetensors implementation for enhanced security

Core Capabilities

  • Bilingual reranking support for Chinese and English content
  • Efficient reranking of top-k search results
  • Strong performance on medical domain tasks
  • Compatible with text-embeddings-inference systems

Frequently Asked Questions

Q: What makes this model unique?

The model uniquely combines cross-encoder architecture with bilingual capabilities, offering superior reranking performance compared to traditional bi-encoder models while maintaining practical inference speeds for production use.

Q: What are the recommended use cases?

The model is ideal for reranking top-k search results from a first-stage retriever, particularly in bilingual contexts or medical information retrieval scenarios. It's best used as a second-stage ranker after initial candidate retrieval.

The first platform built for prompt engineering