wizard-mega-13B-GPTQ

Maintained By
TheBloke

Wizard Mega 13B GPTQ

PropertyValue
Parameter Count13B
Model TypeLLaMA-based
Quantization4-bit GPTQ
LicenseOther
Size7.45 GB

What is wizard-mega-13B-GPTQ?

Wizard Mega 13B GPTQ is a quantized version of the original Wizard Mega model, optimized for efficient deployment while maintaining high performance. It's built on the LLaMA 13B architecture and fine-tuned on a carefully curated combination of ShareGPT, WizardLM, and Wizard-Vicuna datasets, with responses filtered to remove typical AI disclaimers and refusals.

Implementation Details

The model utilizes 4-bit quantization with a group size of 128 and no act-order, making it compatible with ExLlama and various GPTQ implementations. It's optimized for a 2048 token context window and uses the Vicuna prompt template for consistent interaction.

  • GPTQ quantization parameters: 4-bit, GS=128
  • Model size reduced to 7.45GB from original
  • Compatible with AutoGPTQ and Transformers libraries
  • Trained using Axolotl on 8xA100 80GB GPUs

Core Capabilities

  • Comprehensive instruction following and chat interactions
  • Code generation and technical explanations
  • Detailed and polite responses to complex queries
  • Efficient deployment on consumer hardware due to quantization

Frequently Asked Questions

Q: What makes this model unique?

This model combines the strengths of multiple high-quality datasets with efficient quantization, making it particularly suitable for deployment on consumer hardware while maintaining strong performance across a wide range of tasks.

Q: What are the recommended use cases?

The model excels at general dialogue, technical writing, code generation, and detailed explanations. It's particularly well-suited for applications requiring balanced performance and efficiency.

The first platform built for prompt engineering