transformer-turkish-summarization

Maintained By
mukayese

transformer-turkish-summarization

PropertyValue
Parameter Count125M
LicenseMIT
PaperMukayese: Turkish NLP Strikes Back
ROUGE-1 Score43.2049

What is transformer-turkish-summarization?

This is a specialized transformer model designed for Turkish text summarization, developed as part of the Mukayese project. The model was trained from scratch without pre-training, specifically on the mlsum/tu dataset. It represents a significant advancement in Turkish language processing, utilizing modern transformer architecture to generate high-quality summaries of Turkish text.

Implementation Details

The model employs a sophisticated training approach with mixed-precision training and distributed computing across 8 GPUs. It was trained using the Adam optimizer with carefully tuned hyperparameters, including a learning rate of 0.0001 and a batch size of 64.

  • Training duration: 15 epochs
  • Label smoothing factor: 0.1
  • Gradient accumulation steps: 2
  • Native AMP implementation for mixed-precision training

Core Capabilities

  • Achieves 43.2049 ROUGE-1 score on evaluation
  • 30.7082 ROUGE-2 score
  • 38.1981 ROUGE-L score
  • Optimized for Turkish language summarization
  • Supports text-to-text generation tasks

Frequently Asked Questions

Q: What makes this model unique?

This model is specifically designed for Turkish language summarization, trained from scratch without relying on pre-training. It achieves competitive ROUGE scores while being optimized for Turkish text processing.

Q: What are the recommended use cases?

The model is ideal for automatic summarization of Turkish news articles, documents, and other long-form content where concise summaries are needed. It's particularly well-suited for media organizations and content platforms dealing with Turkish language content.

The first platform built for prompt engineering