MiniCPM3-RAG-LoRA

Maintained By
openbmb

MiniCPM3-RAG-LoRA

PropertyValue
Parameter Count4B
Max Input Tokens32,768
LanguagesChinese, English
LicenseApache-2.0 (code), Custom MiniCPM License (weights)

What is MiniCPM3-RAG-LoRA?

MiniCPM3-RAG-LoRA is a specialized language model developed through collaboration between ModelBest Inc., NEUIR, and THUNLP, specifically designed for Retrieval-Augmented Generation (RAG) scenarios. Built upon the MiniCPM3 foundation model, it employs Low-Rank Adaptation (LoRA) and Direct Preference Optimization (DPO) techniques, trained on over 20,000 open-domain QA and logical reasoning examples.

Implementation Details

The model leverages advanced fine-tuning techniques to achieve superior performance in RAG applications. It requires transformers>=4.36.0 and supports both Chinese and English languages.

  • Implements LoRA architecture for efficient fine-tuning
  • Supports extensive context window of 32,768 tokens
  • Utilizes DPO for enhanced preference alignment
  • Achieves 13% performance improvement over baseline

Core Capabilities

  • Open-domain question answering (NQ, TQA, MARCO)
  • Multi-hop reasoning (HotpotQA)
  • Dialogue generation (WoW)
  • Fact verification (FEVER)
  • Information filling (T-REx)
  • Outperforms larger models like Llama3-8B and Baichuan2-13B

Frequently Asked Questions

Q: What makes this model unique?

The model's specialized RAG optimization and impressive performance metrics despite its relatively small size (4B parameters) make it stand out. It achieves state-of-the-art results across multiple benchmarks while maintaining efficiency.

Q: What are the recommended use cases?

The model excels in scenarios requiring document retrieval and generation, including question answering, fact checking, and information synthesis. It's particularly suitable for applications needing strong reasoning capabilities with external knowledge integration.

🍰 Interesting in building your own agents?
PromptLayer provides Huggingface integration tools to manage and monitor prompts with your whole team. Get started here.