Kunoichi-7B

Maintained By
SanjiWatsuki

Kunoichi-7B

PropertyValue
Parameter Count7.24B
Model TypeText Generation
ArchitectureMistral-based Transformer
LicenseCC-BY-NC-4.0
Tensor TypeBF16

What is Kunoichi-7B?

Kunoichi-7B is an advanced language model created through a SLERP merger between Silicon-Maid-7B and an unreleased "Ninja-7B" model. It represents a significant achievement in balancing roleplay capabilities with enhanced cognitive abilities, demonstrating impressive performance across multiple benchmarks including MT-Bench (8.14) and MMLU (64.9).

Implementation Details

The model is built on the Mistral architecture and supports context windows up to 8k tokens, with experimental support for 16k using NTK RoPE alpha of 2.6. It implements both custom and Alpaca prompt formats, making it versatile for various applications.

  • SLERP merger architecture combining roleplay and cognitive capabilities
  • 8k standard context window (16k experimental)
  • BF16 tensor format for optimal performance
  • Compatible with multiple prompt templates

Core Capabilities

  • Strong performance on MT-Bench (8.14) - comparable to commercial models
  • High EQ-Bench scores (44.32)
  • Excellent MMLU performance (64.9)
  • Enhanced logical reasoning capabilities (Logic Test: 0.58)
  • Optimized for roleplay while maintaining general intelligence

Frequently Asked Questions

Q: What makes this model unique?

Kunoichi-7B stands out for its exceptional balance between roleplay capabilities and cognitive performance, achieving benchmarks that rival commercial models while maintaining strong character adherence in roleplay scenarios.

Q: What are the recommended use cases?

The model excels in general-purpose tasks, roleplay scenarios, and applications requiring both creative and analytical capabilities. It's particularly well-suited for SillyTavern implementations and can be used with various prompt templates.

The first platform built for prompt engineering