gpt2-chinese-cluecorpussmall

Maintained By
uer

GPT2 Chinese CLUECorpusSmall

PropertyValue
AuthorUER
FrameworkPyTorch, TensorFlow
Training DataCLUECorpusSmall
Research PaperLink

What is gpt2-chinese-cluecorpussmall?

This is a specialized Chinese language model based on the GPT2 architecture, pre-trained on the CLUECorpusSmall dataset. It comes in multiple variants ranging from a lightweight 6-layer distil version to a massive 48-layer xlarge version, offering flexibility for different computational requirements.

Implementation Details

The model was trained using two frameworks: UER-py for smaller variants and TencentPretrain for the xlarge model. Training occurred in two stages: first with 1,000,000 steps at 128 sequence length, followed by 250,000 steps at 1024 sequence length.

  • Multiple size variants: distil (L=6/H=768), base (L=12/H=768), medium (L=24/H=1024), large (L=36/H=1280), xlarge (L=48/H=1600)
  • Utilizes BertTokenizer for text tokenization
  • Supports text generation with customizable parameters

Core Capabilities

  • Chinese text generation
  • Contextual understanding of Chinese language
  • Flexible deployment options from lightweight to high-capacity models
  • Support for both inference and fine-tuning tasks

Frequently Asked Questions

Q: What makes this model unique?

This model stands out for its specialized Chinese language capabilities and its range of size variants, allowing users to choose between computational efficiency and model capacity based on their needs.

Q: What are the recommended use cases?

The model is ideal for Chinese text generation tasks, creative writing, content completion, and can be fine-tuned for specific domain applications requiring Chinese language understanding.

The first platform built for prompt engineering