phi-2-super

Maintained By
abacaj

Phi-2-super

PropertyValue
Parameter Count2.78B
LicenseMIT
Tensor TypeBF16
LanguageEnglish

What is phi-2-super?

Phi-2-super is an enhanced version of Microsoft's Phi-2 model, fine-tuned using SFT (Supervised Fine-Tuning) and cDPO techniques. This 2.78B parameter model is specifically designed for improved instruction following and conversational abilities, achieving a prompt-level loose accuracy of 0.2717 on the Instruction Following Eval dataset.

Implementation Details

The model implements a chat template similar to Mistral instruct models, utilizing the transformers library for inference. It supports efficient text generation with customizable parameters like temperature and top-p sampling.

  • Supports batch processing and GPU acceleration
  • Uses BF16 tensor format for efficient computation
  • Implements sophisticated chat templating for conversation handling
  • Provides built-in safety measures and token management

Core Capabilities

  • Text generation and completion tasks
  • Conversational AI applications
  • Instruction following with high accuracy
  • Code generation and technical writing
  • Custom inference endpoint support

Frequently Asked Questions

Q: What makes this model unique?

The model combines the efficiency of the Phi-2 architecture with enhanced instruction-following capabilities through SFT and cDPO optimization, making it particularly suitable for conversational AI applications while maintaining a relatively small parameter count.

Q: What are the recommended use cases?

The model excels in conversational AI, code generation, and general text generation tasks. It's particularly well-suited for applications requiring both computational efficiency and high-quality instruction following capabilities.

The first platform built for prompt engineering