Luna-AI-Llama2-Uncensored

Maintained By
Tap-M

Luna-AI-Llama2-Uncensored

PropertyValue
Licensecc-by-sa-4.0
FrameworkPyTorch
Training Infrastructure8x A100 80GB GPUs
Average Benchmark Score0.5114

What is Luna-AI-Llama2-Uncensored?

Luna-AI-Llama2-Uncensored is a specialized chat model built on the Llama2 architecture, fine-tuned by Tap on over 40,000 long-form chat discussions. The model is designed to provide more direct and unrestricted responses while maintaining coherent conversation abilities.

Implementation Details

The model follows the Vicuna 1.1/OpenChat format for prompts and offers two deployment options: a 4-bit GPTQ version for GPU inference and a GGML version for CPU inference. The training process utilized high-end infrastructure with 8x A100 80GB GPUs, focusing on synthetic outputs that include multiple rounds of human-AI interactions.

  • Benchmark performance: Arc Challenge (0.5512), MMLU (0.46521), TruthfulQA MC (0.4716)
  • Supports both GPU and CPU inference implementations
  • Follows standard chat prompt format: USER: [input] ASSISTANT: [response]

Core Capabilities

  • Long-form conversation handling
  • Unrestricted response generation
  • Multiple deployment options for different hardware configurations
  • Strong performance on various benchmark tasks

Frequently Asked Questions

Q: What makes this model unique?

The model's distinctive feature is its uncensored nature combined with extensive fine-tuning on long-form conversations, making it particularly suitable for open-ended dialogue scenarios.

Q: What are the recommended use cases?

The model is best suited for applications requiring natural conversation flow, particularly where unrestricted responses are desired. It's optimized for both GPU and CPU deployments, making it versatile for different infrastructure requirements.

🍰 Interesting in building your own agents?
PromptLayer provides Huggingface integration tools to manage and monitor prompts with your whole team. Get started here.