WizardLM-30B-Uncensored-GGUF
Property | Value |
---|---|
Parameter Count | 32.5B |
Model Type | LLaMA |
License | Other |
Author | TheBloke (Quantized) / Eric Hartford (Original) |
What is WizardLM-30B-Uncensored-GGUF?
WizardLM-30B-Uncensored-GGUF is a powerful language model that represents a specialized version of WizardLM, trained specifically without alignment constraints or moralizing responses. This GGUF version offers various quantization options ranging from 2-bit to 8-bit precision, making it adaptable to different hardware configurations and performance requirements.
Implementation Details
The model is available in multiple GGUF quantization formats, each optimized for different use cases. The recommended Q4_K_M version provides a balanced compromise between model size (19.62 GB) and performance. The model supports context lengths up to 2048 tokens and can be run with GPU acceleration using various frameworks.
- Multiple quantization options from Q2_K (13.50 GB) to Q8_0 (34.57 GB)
- Compatible with llama.cpp and various third-party UIs
- Supports GPU layer offloading for improved performance
- Uses standard WizardLM prompt template
Core Capabilities
- Unrestricted text generation without built-in alignment constraints
- Flexible deployment options across different hardware configurations
- High performance text generation with various quality-size tradeoffs
- Compatible with popular frameworks like text-generation-webui and LangChain
Frequently Asked Questions
Q: What makes this model unique?
This model is unique in its uncensored nature and the wide range of quantization options available, allowing users to choose the optimal balance between model size and performance for their specific needs.
Q: What are the recommended use cases?
The model is suited for applications requiring unrestricted text generation where custom alignment can be added separately. It's particularly useful for researchers and developers who want to implement their own safety measures or fine-tuning.