StableLM-Tuned-Alpha-3B
| Property | Value |
|---|---|
| Parameter Count | 3 Billion |
| Architecture | NeoX Transformer |
| Hidden Size | 4096 |
| Layers | 16 |
| Attention Heads | 32 |
| License | CC BY-NC-SA-4.0 |
What is stablelm-tuned-alpha-3b?
StableLM-Tuned-Alpha-3B is a sophisticated decoder-only language model developed by Stability AI, specifically designed for chat and instruction-following tasks. Built upon the StableLM-Base-Alpha architecture, this model has been fine-tuned on a diverse collection of high-quality datasets to enhance its conversational abilities and instruction-following capabilities.
Implementation Details
The model implements a NeoX transformer architecture with a 4096 sequence length capability. It's trained using mixed-precision (FP16) and optimized with AdamW, using a batch size of 256, learning rate of 2e-5, and 50 warm-up steps. The training process incorporated weight decay of 0.01 and beta values of (0.9, 0.99).
- Utilizes a specialized stopping criteria system for generating responses
- Implements a structured prompt format: <|SYSTEM|>...<|USER|>...<|ASSISTANT|>...
- Trained on multiple high-quality datasets including Alpaca, GPT4All, Anthropic HH, DataBricks Dolly, and ShareGPT Vicuna
Core Capabilities
- Enhanced chat and dialogue generation
- Instruction following and task completion
- Poetry and creative writing
- Question answering and information provision
- Safe and ethical response generation
Frequently Asked Questions
Q: What makes this model unique?
The model's distinctive feature is its comprehensive fine-tuning on five diverse datasets, combined with its focus on safe and helpful interactions. It's specifically designed to refuse harmful requests while maintaining creative capabilities.
Q: What are the recommended use cases?
The model is ideal for chat applications, creative writing tasks, and general instruction following. It's particularly suited for non-commercial applications requiring a balance of helpfulness and safety in AI interactions.





