sentence-bert-base-ja-mean-tokens-v2

sentence-bert-base-ja-mean-tokens-v2

sonoisa

Japanese Sentence-BERT model (v2) with 111M parameters, optimized with MultipleNegativesRankingLoss for better sentence embeddings. Built on BERT-base Japanese.

PropertyValue
Parameter Count111M
LicenseCC-BY-SA-4.0
Authorsonoisa
Base Modelcl-tohoku/bert-base-japanese-whole-word-masking

What is sentence-bert-base-ja-mean-tokens-v2?

This is an improved Japanese Sentence-BERT model that leverages the MultipleNegativesRankingLoss function for better sentence embedding generation. Built upon the successful cl-tohoku/bert-base-japanese-whole-word-masking architecture, this v2 model demonstrates 1.5-2 points higher accuracy compared to its predecessor on private datasets.

Implementation Details

The model utilizes PyTorch framework and requires fugashi and ipadic for inference. It implements mean pooling strategy for generating sentence embeddings and supports batch processing for efficient computation.

  • Improved loss function using MultipleNegativesRankingLoss
  • Built on BERT-base Japanese whole word masking
  • Supports batch processing with customizable batch sizes
  • Implements efficient mean pooling strategy

Core Capabilities

  • Japanese sentence embedding generation
  • Semantic similarity comparison
  • Feature extraction for Japanese text
  • Efficient batch processing of sentences

Frequently Asked Questions

Q: What makes this model unique?

This model stands out due to its improved training approach using MultipleNegativesRankingLoss and its specific optimization for Japanese language processing, showing measurable improvements over the v1 model.

Q: What are the recommended use cases?

The model is ideal for Japanese text processing tasks including semantic similarity comparison, document classification, and feature extraction for downstream NLP tasks.

Socials
PromptLayer
Company
All services online
Location IconPromptLayer is located in the heart of New York City
PromptLayer © 2026