roberta-large-openai-detector

Maintained By
openai-community

RoBERTa Large OpenAI Detector

PropertyValue
Parameter Count356M
LicenseMIT
PaperResearch Paper
LanguageEnglish
Accuracy~95% on GPT-2 generated text

What is roberta-large-openai-detector?

The RoBERTa Large OpenAI Detector is a specialized model designed to identify text generated by GPT-2 models. Developed by OpenAI, this model is built on the RoBERTa large architecture and fine-tuned specifically to distinguish between human-written text and machine-generated content from GPT-2 models, particularly the 1.5B parameter version.

Implementation Details

The model leverages the RoBERTa large architecture with 355 million parameters and employs sequence classification to analyze text segments. It was trained on a combination of WebText dataset and GPT-2 generated outputs, making it particularly effective at detecting synthetic text across various sampling methods.

  • Built on RoBERTa large architecture
  • Fine-tuned on GPT-2 1.5B outputs
  • Optimized for 510-token text segments
  • Supports various sampling detection methods

Core Capabilities

  • 95% accuracy in detecting GPT-2 generated text
  • Robust performance across different sampling methods
  • Specialized in long-form text analysis
  • Effective transfer learning capabilities

Frequently Asked Questions

Q: What makes this model unique?

This model is specifically optimized for detecting GPT-2 generated content with high accuracy, making it a valuable tool for synthetic text detection research. Its robustness across different sampling methods sets it apart from other detection models.

Q: What are the recommended use cases?

The model is best suited for research related to synthetic text detection, content authenticity verification, and academic studies on AI-generated text. However, it should be used in conjunction with other detection methods and human judgment for optimal results.

The first platform built for prompt engineering