bleurt-tiny-512

Maintained By
Elron

BLEURT-tiny-512

PropertyValue
DeveloperGoogle Research (Elron Bandel, Thibault Sellam, et al.)
Model TypeText Classification
Base ArchitectureBERT
PaperBLEURT: Learning Robust Metrics for Text Generation
Downloads61,161

What is bleurt-tiny-512?

BLEURT-tiny-512 is a compact PyTorch implementation of Google's BLEURT (Bilingual Evaluation Understudy with Representations from Transformers) metric. This model is specifically designed for evaluating natural language generation quality, offering a more efficient alternative to the original implementation while maintaining robust performance.

Implementation Details

The model is built on BERT architecture and trained on WMT Metrics Shared Task data from 2017-2019, encompassing thousands of sentence pairs with human ratings from the news domain. The training dataset includes 5,360, 9,492, and 147,691 records for each respective year.

  • PyTorch-based implementation for better compatibility and deployment
  • Optimized for 512 token sequence length
  • Trained on extensive WMT dataset with human annotations
  • Supports direct text comparison and evaluation

Core Capabilities

  • Text quality assessment and comparison
  • Generation evaluation for machine translation
  • Semantic similarity scoring
  • Automated evaluation of natural language generation

Frequently Asked Questions

Q: What makes this model unique?

BLEURT-tiny-512 stands out for its efficient implementation of the BLEURT metric in PyTorch, making it more accessible for production environments while maintaining the robust evaluation capabilities of the original model.

Q: What are the recommended use cases?

The model is ideal for evaluating machine translation output, assessing natural language generation quality, and comparing generated text against reference sentences in production environments.

The first platform built for prompt engineering