miqu-1-70b

miqu-1-70b

miqudev

A 70B parameter LLM using Mistral architecture, optimized for GGUF format with 32k context window and high-frequency base settings

PropertyValue
Parameter Count69B parameters
Model TypeGGUF Format
Context Length32k tokens
FormatMistral

What is miqu-1-70b?

miqu-1-70b is a large language model featuring 69 billion parameters, optimized for deployment through the GGUF format. It represents the first model in a potential series, implementing the Mistral architecture and instruction format for enhanced performance and versatility.

Implementation Details

The model utilizes a specific prompt format following the Mistral structure: "<s> [INST] QUERY [/INST] ANSWER</s>". It's configured with specialized RoPE (Rotary Position Embedding) settings optimized for a 32k token context window and high-frequency base, which shouldn't be modified for optimal performance.

  • Optimized GGUF format for efficient deployment
  • Specialized RoPE settings for extended context handling
  • Recommended settings: Temperature 1.0, top_p 0.95

Core Capabilities

  • Extended context processing with 32k token window
  • Efficient inference through GGUF optimization
  • Consistent performance with Mistral architecture

Frequently Asked Questions

Q: What makes this model unique?

The model combines a large parameter count (69B) with optimized GGUF format and extended context window, making it suitable for complex tasks while maintaining deployment efficiency.

Q: What are the recommended use cases?

While specific use cases aren't detailed, the model's architecture and size make it suitable for general-purpose language tasks, particularly those requiring extended context understanding and complex reasoning.

Related Models

Socials
PromptLayer
Company
All services online
Location IconPromptLayer is located in the heart of New York City
PromptLayer © 2026