Hermes-2-Pro-Llama-3-8B

Maintained By
NousResearch

Hermes-2-Pro-Llama-3-8B

PropertyValue
Parameter Count8.03B
Model TypeLanguage Model (Llama-3 based)
LicenseLlama3
Tensor TypeFP16

What is Hermes-2-Pro-Llama-3-8B?

Hermes-2-Pro-Llama-3-8B is an advanced language model built on Meta's Llama-3 architecture, specifically designed to excel at function calling and structured JSON outputs. This model represents an upgraded version of Nous Hermes 2, incorporating a refined OpenHermes 2.5 Dataset and proprietary function calling capabilities.

Implementation Details

The model utilizes the ChatML format for prompt structuring and introduces special tokens for enhanced agentic capabilities. It achieved impressive benchmarks with 90% accuracy on function calling evaluations and 84% on structured JSON outputs.

  • Built on Llama-3 8B base model architecture
  • Implements ChatML format with specific system prompts
  • Features specialized tokens for streaming: <tools>, <tool_call>, <tool_response>
  • Supports both function calling and JSON mode operations

Core Capabilities

  • Advanced function calling with reliable parsing
  • Structured JSON output generation
  • Multi-turn dialogue support
  • High-performance on general task solving
  • Efficient 4-bit quantization support (requires ~5GB VRAM)
  • Strong performance on standard benchmarks (72.62% on GPT4All)

Frequently Asked Questions

Q: What makes this model unique?

The model's distinctive feature is its dual specialization in function calling and JSON structured outputs, combined with its efficient architecture that maintains high performance while requiring relatively low VRAM usage.

Q: What are the recommended use cases?

This model is particularly well-suited for applications requiring structured data outputs, API integrations through function calling, conversational AI implementations, and general task-solving scenarios where precise, formatted responses are necessary.

The first platform built for prompt engineering