ko-sbert-nli

Maintained By
jhgan

ko-sbert-nli

PropertyValue
Authorjhgan
Downloads28,533
PaperResearch Paper
Embedding Dimension768
Performance82.24% Cosine Pearson on KorSTS

What is ko-sbert-nli?

ko-sbert-nli is a specialized Korean language model based on SBERT (Sentence-BERT) architecture, designed specifically for generating meaningful sentence embeddings. It transforms Korean text into 768-dimensional dense vector representations, making it particularly effective for semantic similarity tasks and natural language understanding applications.

Implementation Details

The model utilizes a BERT-based architecture with mean pooling and is trained using MultipleNegativesRankingLoss with a scale factor of 20.0. It employs the AdamW optimizer with a learning rate of 2e-05 and implements a WarmupLinear scheduler with 889 warmup steps.

  • Maximum sequence length: 128 tokens
  • Pooling strategy: Mean tokens pooling
  • Training dataset: KorNLI
  • Evaluation dataset: KorSTS

Core Capabilities

  • Sentence embedding generation for Korean text
  • Semantic similarity computation
  • Cross-sentence relationship understanding
  • Support for clustering and semantic search applications

Frequently Asked Questions

Q: What makes this model unique?

This model is specifically optimized for Korean language understanding, trained on KorNLI dataset and evaluated on KorSTS, achieving impressive performance metrics including 82.24% Cosine Pearson correlation.

Q: What are the recommended use cases?

The model excels in tasks such as semantic similarity comparison, document clustering, information retrieval, and semantic search applications specifically for Korean language content.

🍰 Interesting in building your own agents?
PromptLayer provides the tools to manage and monitor prompts with your whole team. Get started here.