WizardLM-30B-Uncensored-GGML
Property | Value |
---|---|
Model Size | 30B parameters |
License | Other |
Author | TheBloke |
Format | GGML |
What is WizardLM-30B-Uncensored-GGML?
WizardLM-30B-Uncensored-GGML is a specialized variant of the WizardLM language model, optimized for CPU and GPU inference using the GGML format. This model is unique in that it has been trained without typical alignment constraints, making it more flexible for custom implementations while requiring responsible usage.
Implementation Details
The model is available in multiple quantization formats, ranging from 2-bit to 8-bit precision, offering different trade-offs between model size, performance, and accuracy. The implementation includes both traditional llama.cpp quantization methods (q4_0, q4_1, q5_0, q5_1, q8_0) and newer k-quant methods (q2_K through q6_K).
- File sizes range from 13.60GB (q2_K) to 34.56GB (q8_0)
- Supports both CPU and GPU inference
- Compatible with various frameworks including text-generation-webui, KoboldCpp, and llama-cpp-python
Core Capabilities
- Efficient CPU+GPU inference with multiple quantization options
- Flexible deployment across different hardware configurations
- Customizable context window up to 2048 tokens
- Advanced tensor processing with specialized quantization for attention and feed-forward layers
Frequently Asked Questions
Q: What makes this model unique?
This model stands out due to its uncensored training approach and the variety of quantization options available, allowing users to balance between performance and resource requirements. It's specifically optimized for deployment in resource-constrained environments while maintaining model capability.
Q: What are the recommended use cases?
The model is suitable for research and development purposes where custom alignment is desired. Users should carefully consider the responsibility of implementation as the model comes without built-in guardrails. It's particularly useful for scenarios requiring local deployment with varying hardware constraints.