ko-sbert-nli

Maintained By
jhgan

ko-sbert-nli

PropertyValue
Authorjhgan
Downloads28,533
PaperResearch Paper
Embedding Dimension768
Performance82.24% Cosine Pearson on KorSTS

What is ko-sbert-nli?

ko-sbert-nli is a specialized Korean language model based on SBERT (Sentence-BERT) architecture, designed specifically for generating meaningful sentence embeddings. It transforms Korean text into 768-dimensional dense vector representations, making it particularly effective for semantic similarity tasks and natural language understanding applications.

Implementation Details

The model utilizes a BERT-based architecture with mean pooling and is trained using MultipleNegativesRankingLoss with a scale factor of 20.0. It employs the AdamW optimizer with a learning rate of 2e-05 and implements a WarmupLinear scheduler with 889 warmup steps.

  • Maximum sequence length: 128 tokens
  • Pooling strategy: Mean tokens pooling
  • Training dataset: KorNLI
  • Evaluation dataset: KorSTS

Core Capabilities

  • Sentence embedding generation for Korean text
  • Semantic similarity computation
  • Cross-sentence relationship understanding
  • Support for clustering and semantic search applications

Frequently Asked Questions

Q: What makes this model unique?

This model is specifically optimized for Korean language understanding, trained on KorNLI dataset and evaluated on KorSTS, achieving impressive performance metrics including 82.24% Cosine Pearson correlation.

Q: What are the recommended use cases?

The model excels in tasks such as semantic similarity comparison, document clustering, information retrieval, and semantic search applications specifically for Korean language content.

The first platform built for prompt engineering