ELYZA-japanese-Llama-2-7b

Maintained By
elyza

ELYZA-japanese-Llama-2-7b

PropertyValue
Parameter Count6.27B
Vocabulary Size32,000
LicenseLLAMA 2 Community License
LanguagesJapanese, English
PaperarXiv:2307.09288

What is ELYZA-japanese-Llama-2-7b?

ELYZA-japanese-Llama-2-7b is an advanced language model that extends Meta's LLaMA 2 architecture with enhanced Japanese language capabilities. The model underwent additional pre-training specifically to improve its Japanese language understanding and generation abilities while maintaining its English language capabilities.

Implementation Details

The model architecture is based on the LLaMA 2 framework and comes in several variants, including a base model, an instruct-tuned version, and fast versions with expanded vocabulary. The implementation uses PyTorch and the Transformers library, making it easily accessible for developers.

  • Base model with 6.27B parameters and 32,000 vocab size
  • Fast version with 6.37B parameters and 45,043 vocab size
  • Includes both regular and instruct-tuned variants
  • Supports text generation and inference endpoints

Core Capabilities

  • Bilingual text generation in Japanese and English
  • Instruction-following capabilities in Japanese
  • Enhanced Japanese language understanding
  • Seamless integration with Hugging Face Transformers

Frequently Asked Questions

Q: What makes this model unique?

The model's unique strength lies in its specialized Japanese language capabilities while maintaining the robust architecture of LLaMA 2. It's particularly noteworthy for being one of the few large language models specifically optimized for Japanese language tasks.

Q: What are the recommended use cases?

The model is well-suited for Japanese text generation, bilingual applications, and general language understanding tasks. It's particularly effective when used with the provided system prompts for Japanese language interaction.

The first platform built for prompt engineering