Pygmalion 7B
Property | Value |
---|---|
Base Model | LLaMA-7B |
Primary Use Case | Conversational AI / Dialogue Generation |
Language | English |
Training Type | Fine-tuned |
What is pygmalion-7b?
Pygmalion-7B is a specialized conversational AI model that builds upon Meta's LLaMA-7B architecture. It's specifically designed for dialogue generation with a unique approach to persona-based interactions. The model represents version 1 of the Pygmalion series at the 7B parameter scale, fine-tuned using a curated subset of data from Pygmalion-6B-v8-pt4.
Implementation Details
The model implements a distinctive prompting format that separates persona information from dialogue using specific tokens. It requires XOR decoding with original LLaMA weights due to licensing requirements, making deployment a multi-step process that requires access to Meta's original LLaMA weights.
- Persona-based dialogue architecture
- Custom prompting format with <START> delimiter
- Automatic end-of-text token generation
- Sliding window dialogue history support
Core Capabilities
- Character-based conversation generation
- Context-aware responses with history tracking
- Flexible persona implementation
- Natural conversation flow management
Frequently Asked Questions
Q: What makes this model unique?
The model's distinct feature is its specialized persona-based conversation system, allowing for character-specific dialogue generation with maintained context and natural flow.
Q: What are the recommended use cases?
The model is specifically designed for fictional conversation and entertainment purposes. It's important to note that it's not fine-tuned for safety or factual accuracy, and should not be used for critical or production applications.