Brief Details: RoBERTa-base model fine-tuned for Chinese sentiment analysis on Dianping restaurant reviews. Part of UER-py toolkit, optimized for text classification tasks.
Brief Details: Chinese ALBERT Large model trained on CLUECorpusSmall dataset. Features 24 layers, 1024 hidden size. Specialized for Chinese masked language modeling and text representation tasks.
Brief Details: Chinese ALBERT base model trained on CLUECorpusSmall dataset. Optimized for masked language modeling with 12 layers, 768 hidden dimensions. Supports both PyTorch and TensorFlow.
Brief-details: A bilingual ELECTRA model trained on Korean-English data (100GB+). Strong performance on both Korean & English NLP tasks with 133M parameters.
Brief Details: XLM-RoBERTa-based multilingual NER model fine-tuned on PAN-X dataset, achieving 84.9% F1 score. Optimized for token classification across languages.
Brief Details: LongLM-large is a 993M parameter Chinese language model specialized in long-text generation and understanding, featuring text infilling and conditional continuation capabilities
BRIEF DETAILS: Lawformer is a specialized pre-trained language model for Chinese legal long documents. Built on Longformer architecture with 19 likes and 729 downloads on Hugging Face.
Brief Details: HuBERT-Large model fine-tuned for emotion recognition, achieving 67.62% accuracy on IEMOCAP dataset. Handles 16kHz speech audio for 4-class emotion classification.
Brief-details: FastSpeech2-based Chinese text-to-speech model trained on Baker dataset, offering high-quality end-to-end speech synthesis with adjustable speed and pitch control
Brief Details: Estonian GPT2-based language model trained on 2.2B words. Features domain-specific prefixes and 118.68M parameters for text generation tasks.
Brief Details: HuBERT-based model for emotion recognition in speech, trained on IEMOCAP dataset. Achieves 63.59% accuracy for 4-class emotion classification.
Brief Details: SqueezeBERT is an efficient BERT variant that replaces fully-connected layers with grouped convolutions, achieving 4.3x faster inference on mobile devices.
BRIEF DETAILS: German GPT-2 language model with 137M parameters, trained on 90GB of clean German web text (GC4 corpus). MIT-licensed, optimized for text generation and fine-tuning tasks.
Brief-details: SqueezeBERT-MNLI: Efficient BERT variant using grouped convolutions, 4.3x faster on mobile devices, pretrained on BookCorpus/Wikipedia and finetuned for MNLI tasks.
Brief Details: SepFormer-WHAM is a state-of-the-art audio source separation model achieving 16.3dB SI-SNRi on WHAM! dataset, ideal for separating mixed speech signals with environmental noise.
Brief Details: SepFormer speech enhancement model trained on WHAM! dataset. Achieves 14.35dB SI-SNR and 3.07 PESQ. Specialized for 8kHz audio denoising using transformer architecture.
Brief-details: Large Polish language model for spaCy with comprehensive NLP capabilities including NER (84.15% F-score), POS tagging (97.81% accuracy), and dependency parsing.
Brief-details: Large Russian language model for NLP tasks with high accuracy (95%+ on NER/POS/Morphology). Includes tok2vec, parser, NER & more components.
Brief-details: State-of-the-art English language transformer model for NLP tasks with 98.13% tagging accuracy and 90.19% NER F-score, built on RoBERTa architecture.
Brief Details: Compact English language model for NLP tasks with 97.25% tagging accuracy. Features tok2vec, tagger, parser & NER components. MIT licensed.
BRIEF-DETAILS: Korean-specific BERT model optimized for Korean language processing with 16,424 vocab size and 99M parameters. Features unique BidirectionalWordPiece tokenization and character/sub-character support.