miqu-1-70b

Maintained By
miqudev

miqu-1-70b

PropertyValue
Parameter Count69B parameters
Model TypeGGUF Format
Context Length32k tokens
FormatMistral

What is miqu-1-70b?

miqu-1-70b is a large language model featuring 69 billion parameters, optimized for deployment through the GGUF format. It represents the first model in a potential series, implementing the Mistral architecture and instruction format for enhanced performance and versatility.

Implementation Details

The model utilizes a specific prompt format following the Mistral structure: "<s> [INST] QUERY [/INST] ANSWER</s>". It's configured with specialized RoPE (Rotary Position Embedding) settings optimized for a 32k token context window and high-frequency base, which shouldn't be modified for optimal performance.

  • Optimized GGUF format for efficient deployment
  • Specialized RoPE settings for extended context handling
  • Recommended settings: Temperature 1.0, top_p 0.95

Core Capabilities

  • Extended context processing with 32k token window
  • Efficient inference through GGUF optimization
  • Consistent performance with Mistral architecture

Frequently Asked Questions

Q: What makes this model unique?

The model combines a large parameter count (69B) with optimized GGUF format and extended context window, making it suitable for complex tasks while maintaining deployment efficiency.

Q: What are the recommended use cases?

While specific use cases aren't detailed, the model's architecture and size make it suitable for general-purpose language tasks, particularly those requiring extended context understanding and complex reasoning.

🍰 Interesting in building your own agents?
PromptLayer provides Huggingface integration tools to manage and monitor prompts with your whole team. Get started here.