Tiny-Vicuna-1B

Maintained By
Jiayi-Pan

Tiny-Vicuna-1B

PropertyValue
Parameter Count1.1B
LicenseApache 2.0
Tensor TypeF32
Downloads37,461

What is Tiny-Vicuna-1B?

Tiny-Vicuna-1B is a compact language model that represents a fine-tuned version of TinyLlama, specifically optimized using the WizardVicuna Dataset. This model maintains full compatibility with the Vicuna-v1.5 series while offering a more efficient architecture with just 1.1B parameters.

Implementation Details

The model utilizes PyTorch framework and implements the transformer architecture, with support for text generation tasks. It's distributed in Safetensors format and includes text-generation-inference capabilities.

  • Achieves 34.76% average performance across standard benchmarks
  • Implements few-shot learning capabilities (5-25 shots depending on task)
  • Optimized for English language processing

Core Capabilities

  • HellaSwag (10-Shot): 55.92% normalized accuracy
  • Winogrande (5-shot): 58.41% accuracy
  • MMLU (5-Shot): 25.45% accuracy
  • TruthfulQA (0-shot): 33.82% MC2 score
  • AI2 Reasoning Challenge: 33.45% normalized accuracy

Frequently Asked Questions

Q: What makes this model unique?

This model stands out for its efficient architecture while maintaining compatibility with Vicuna-v1.5. It's particularly suitable for early experiments and iterations due to its smaller size and balanced performance across various tasks.

Q: What are the recommended use cases?

The model is well-suited for text generation tasks, particularly in scenarios requiring few-shot learning. It's ideal for development environments where computational resources are limited but reasonable performance is still required.

🍰 Interesting in building your own agents?
PromptLayer provides Huggingface integration tools to manage and monitor prompts with your whole team. Get started here.