geo-bert-multilingual

Maintained By
k4tel

geo-bert-multilingual

PropertyValue
AuthorKateryna Lutsai
Base Modelbert-base-multilingual-cased
PaperResearch Paper
Training DataTwitter dataset with text and metadata

What is geo-bert-multilingual?

geo-bert-multilingual is an advanced natural language processing model designed to predict geographical locations from short text content. Built upon BERT's multilingual architecture, it uniquely outputs predictions as Gaussian Mixture Models (GMM), providing both location estimates and uncertainty measures for text up to 500 words.

Implementation Details

The model implements a custom wrapper layer of linear regression on top of BERT's classification token. It's trained using carefully tuned hyperparameters including a cosine learning rate scheduler starting at 1e-5 and ending at 1e-6, with Adam optimization over 3 epochs.

  • Handles both text-only and metadata-enhanced predictions
  • Outputs 5 prediction points with associated confidence levels
  • Supports multilingual input
  • Achieves up to 82% accuracy within 161km for user location prediction

Core Capabilities

  • Text-based geolocation prediction with mean accuracy of 1588km
  • Enhanced prediction with user metadata achieving 800km mean accuracy
  • Gaussian Mixture Model output for uncertainty estimation
  • Support for both tweet and user home location prediction

Frequently Asked Questions

Q: What makes this model unique?

The model's ability to output geographical predictions as probability distributions (GMMs) sets it apart, providing not just point estimates but also confidence measures about predicted locations.

Q: What are the recommended use cases?

The model is ideal for geo-tagging of big data, particularly social media content. It performs best when user metadata is available alongside text content, making it particularly suitable for Twitter data analysis.

The first platform built for prompt engineering