Brief Details: A model by unslothai hosted on Hugging Face, designed for collaborative AI development environments. Focus on accessibility and integration.
Brief Details: ResNet TDNN speaker verification model trained on VoxCeleb, achieving 1.05% EER. Handles speaker embedding extraction and verification using SpeechBrain.
Brief Details: Efficient cross-encoder model for MS Marco passage ranking, optimized for speed (9000 docs/sec) with NDCG@10 of 69.84 and MRR@10 of 32.56
Brief Details: GGML-optimized 13B parameter Vicuna language model (v1.1) - Obsolete version of the popular open-source chat assistant based on LLaMA architecture.
BRIEF-DETAILS: 176B parameter multilingual chat model built on BLOOM, instruction-tuned for conversation and QA across languages with advanced capabilities.
Brief-details: A Facebook-developed model focused on multi-token prediction capabilities, emphasizing privacy-compliant processing according to Meta's policies. Hosted on HuggingFace.
Brief-details: Oni_Mitsubishi_12B is a 12 billion parameter language model created by SicariusSicariiStuff, available on HuggingFace for open-source use and research purposes.
Brief-details: Experimental 12B parameter Gemma model with layerwise abliteration technique for uncensored outputs while maintaining coherent responses, featuring 90%+ acceptance rate.
Brief Details: T5XXL-Unchained is a modified T5-XXL encoder featuring an extended tokenizer for Flux uncensoring, with pending compatibility for ComfyUI and Kohya tools.
Brief-details: A 12B parameter fine-tuned version of Google's Gemma model optimized for creative freedom and reduced content restrictions, featuring 128K context window and enhanced character consistency
Brief Details: A state-of-the-art arbitrary-scale super-resolution model using neural heat fields, featuring anti-aliasing capabilities and RDN backbone architecture.
Brief-details: VisualPRM-8B is an 8B-parameter multimodal Process Reward Model that enhances reasoning capabilities of MLLMs through Best-of-N evaluation strategies
BRIEF-DETAILS: Audio Question Answering model achieving SOTA performance through reinforcement learning, based on Qwen2-Audio-7B-Instruct with GRPO optimization
Brief-details: Advanced chest X-ray interpretation model from Stanford AIMI, based on CLIP architecture with ViT-B/16 backbone, trained on LAION-2B dataset
BRIEF-DETAILS: 4-bit quantized version of Midnight-Miqu-70B-v1.5 using AWQ compression, optimized with lmdeploy v0.4.2 for efficient deployment
Brief Details: IndicTrans2 1B parameter model for high-quality Indian language to English translation, supporting all 22 scheduled Indian languages with advanced features like flash attention
Brief-details: Romanian NER model based on XLM-RoBERTa achieving 95.5% F1-macro score on RONEC dataset, optimized for named entity recognition tasks.
Brief-details: Audio processing model by Descript for 44kHz sampling rate, likely used for audio enhancement or conversion tasks. Limited public information available.
Brief-details: Optimized ONNX version of Gemma 3B instruction-tuned model, offering efficient inference with both ONNX Runtime and Transformers.js support
Brief-details: A 24B parameter merged language model combining Arcee-Blitz base with specialized models for roleplay, uncensored content, and enhanced prompt adherence capabilities.
Brief Details: A pruned version of DeepSeek-V3 optimized for code generation, reduced from 256 to 160 experts while maintaining performance, equivalent to 441B parameters