EVA-Qwen2.5-32B-v0.2

Maintained By
EVA-UNIT-01

EVA-Qwen2.5-32B-v0.2

PropertyValue
Parameter Count32.8B
Model TypeText Generation/Roleplay
Base ModelQwen2.5-32B
LicenseApache 2.0
Training FormatChatML

What is EVA-Qwen2.5-32B-v0.2?

EVA-Qwen2.5-32B-v0.2 is a specialized language model fine-tuned for roleplay and creative writing applications. It represents a significant improvement over its previous version, featuring cleaned and reprocessed datasets to eliminate non-unicode character issues and improve generation stability.

Implementation Details

The model was trained for 7 hours on 8xH100 SXM hardware, implementing full-parameter fine-tuning on the Qwen2.5-32B base model. It uses advanced training techniques including sample packing and a sequence length of 10240 tokens.

  • Trained on 10 carefully curated datasets including Celeste 70B mixture and specialized roleplay content
  • Implements ChatML format for consistent interaction
  • Uses BF16 precision for optimal performance

Core Capabilities

  • Creative writing and storytelling
  • Character-based roleplay interactions
  • Extended context handling up to 10K tokens
  • Improved stability and coherence in outputs

Frequently Asked Questions

Q: What makes this model unique?

The model combines extensive parameter fine-tuning with a diverse dataset mixture, specifically optimized for creative and roleplay applications. Version 0.2 introduces significant improvements in data cleaning and processing, eliminating previous artifacts and ensuring more stable output.

Q: What are the recommended use cases?

The model excels at creative writing, character roleplay, and story generation. It's optimized with specific sampling parameters (Temperature: 1, Min-P: 0.05, Top-A: 0.2, Repetition Penalty: 1.03) for optimal creative output.

The first platform built for prompt engineering