Brief Details: Compact Russian BERT variant (35M params) trained on 2T tokens, supporting 8K context window with strong performance on Russian NLP tasks
BRIEF-DETAILS: 8B parameter language model optimized with float8_e4m3fn quantization, derived from Freepik's flux.1-lite-8B for efficient deployment
BRIEF-DETAILS: Hunyuan-7B-Instruct is a powerful Chinese-focused 7B parameter LLM from Tencent featuring 256K context length, strong benchmarks, and GQA attention mechanism.
Brief Details: Baichuan-Omni-1.5 is a 7B parameter multimodal model supporting text, image, video, and audio I/O with state-of-the-art performance in medical imaging and real-time voice interactions.
BRIEF-DETAILS: A LoRA model trained using Flux, specialized in generating fashion and modeling images of a female character named ForestRoss in various settings and styles.
Brief-details: A specialized GGUF quantized version of the FuseO1-DeepSeekR1 32B model, offering multiple compression variants from 7.4GB to 27GB with varying quality-size tradeoffs.
BRIEF DETAILS: A 7B parameter GGUF-formatted instruction-tuned language model, optimized for research with Q6_K quantization, deployable via llama.cpp
Brief-details: Selene-1-Mini-Llama-3.1-8B is an 8B parameter evaluation model that outperforms larger models in scoring tasks, supporting multiple languages with 128K context.
BRIEF-DETAILS: 32B parameter LLM combining DeepSeek-R1 and Qwen2.5-Coder capabilities, specialized in long-short reasoning fusion for enhanced mathematics and coding tasks
BRIEF DETAILS: Optimized 11B vision-language model using Unsloth's Dynamic 4-bit quantization, offering 2x faster performance and 60% less memory usage while maintaining accuracy.
Brief-details: Layout processing model by vikp for handling document layouts and structure analysis, optimized for the Surya framework.
Brief Details: Qwen2.5-32B optimized for 4-bit quantization (BNB), featuring 32.5B parameters, 128K context window, and multilingual support for 29+ languages
Brief-details: Merlin is a specialized 3D Vision Language Model for computed tomography scans, combining EHR data and radiology reports for enhanced medical image understanding.
BRIEF-DETAILS: Meta's latest 70B parameter LLM, part of the Llama family. Advanced language model with robust capabilities for various NLP tasks.
Brief Details: Norwegian Wav2vec2 model optimized for Bokmål ASR, featuring 1B parameters and built on advanced speech recognition architecture.
Brief-details: LingMess is a specialized coreference resolution model achieving 81.4 F1 on OntoNotes, using linguistic categorization across 6 types of decisions
Brief Details: A specialized MoE (Mixture of Experts) variant of Qwen 1.5, featuring a randomized tiny architecture developed by katuni4ka for experimental purposes.
BRIEF DETAILS: Qwen2-1.5B-Instruct-IMat-GGUF is a quantized version of Qwen's 1.5B parameter instruction model, offering multiple compression variants from 3GB to 436MB using IMatrix optimization.
Brief Details: RWKV's 7B parameter Finch model with improved performance over Eagle-7B. Specializes in both English and Chinese content generation with strong evaluation metrics.
BRIEF DETAILS: CausalLM 7B: A powerful LLaMA 2-compatible model trained on 1.3B tokens of synthetic data, outperforming most models ≤33B in benchmarks like MMLU and CEval.
Brief Details: A 7B parameter LLaMA-based model converted to GGML format. Now obsolete but historically significant as part of the Vicuna model family.