Brief Details: A merged 70B parameter LLaMA-3.1 model combining Hermes-3, Dracarys, and SauerkrautLM, achieving 7% MMLU-PRO improvement over base LLaMA 3.1
Brief Details: Modified Mistral-Nemo model with ablated safety restrictions, maintaining 128k context window & strong multilingual capabilities, performing similarly to base model.
Brief-details: Flux.1 Dev model optimized with NF4 quantization for efficient inference, requiring bitsandbytes package. Developed by sayakpaul for improved performance.
BRIEF DETAILS: Shuka v1 is an advanced encoder-decoder model combining Saaras v1 audio encoder and Llama3-8B-Instruct for native Indic language audio understanding, supporting 11+ languages.
BRIEF-DETAILS: Qwen2-Math-72B is a specialized 72B parameter math-focused LLM that excels at arithmetic and mathematical reasoning, outperforming both open and closed-source models.
Brief-details: FLUX.1 dev model quantized to FP8 precision, optimized for efficiency while maintaining FLUX capabilities under non-commercial license.
Brief-details: SoteDiffusion V2 is an anime-focused fine-tune of Würstchen V3/Stable Cascade, trained on 12M image-text pairs with full FP32 and MAE Loss on 8xH100 GPUs.
Brief Details: A transformer-based model focused on processing data with complex bandwidth transformations, developed by KimberleyJSN and hosted on HuggingFace.
Brief-details: Ultimate Vocal Remover (UVR) model collection for isolating vocals from audio tracks using deep learning, maintained by Blane187. Comprehensive suite for audio separation.
BRIEF DETAILS: Multilingual phoneme representation model for TTS with 88M params, trained on 330M phoneme-level sentences across ~100 languages using BERT architecture.
Brief Details: A specialized variant of Baichuan2-13B model created by katuni4ka, likely focused on reduced size while maintaining core capabilities of the original architecture.
Brief Details: BERT-base model fine-tuned on SNLI dataset for natural language inference tasks, optimized for entailment classification and semantic relationship analysis
Brief-details: Neural machine translation model for English to Indian languages conversion, featuring 1.1B parameters and flash attention support, by AI4Bharat
BRIEF DETAILS: Advanced speaker diarization model by Revai offering 16.5% WDER improvement over pyannote3.0, proven on 1.25M+ tokens with strong accuracy metrics.
BRIEF-DETAILS: A compact variant of CodeGen2, developed by katuni4ka, designed for code generation tasks with randomized initialization parameters.
Brief-details: Gemma-2b-it is Google's 2B parameter instruction-tuned language model requiring license acceptance, designed for efficient deployment and reliable text generation
BRIEF-DETAILS: BERT base model optimized for Korean language processing, trained on 70GB Korean text with 42K subword vocabulary. Suitable for various Korean NLP tasks.
Brief Details: Optimized 8B parameter Llama 3.1 model using Unsloth's Dynamic 4-bit quantization, offering 2.4x faster training with 58% less memory usage
Brief-details: RVCModels is a curated collection of voice conversion models hosted on HuggingFace, providing organized access to pre-trained models for audio transformation tasks.
BRIEF-DETAILS: Llama-Guard-3-8B is Meta's 3.8B parameter AI safety model designed to detect and filter potentially harmful content, built on Llama architecture.
Brief Details: FLUX.1-Depth-dev is a depth estimation AI model by Black Forest Labs, requiring non-commercial licensing agreement for usage.