llava-llama-3-8b-v1_1-imat-gguf

Maintained By
city96

LLaVA Llama 3 8B iMatrix GGUF Model

PropertyValue
Authorcity96
Model Size8B parameters
FormatGGUF with iMatrix
SourceHugging Face Repository

What is llava-llama-3-8b-v1_1-imat-gguf?

This model is a specialized conversion of the xtuner/llava-llama-3-8b-v1_1-transformers to GGUF format with iMatrix optimization. It's primarily designed for use as a text encoder in Hunyuan Video applications, while maintaining capability for vision tasks when paired with appropriate mmproj files.

Implementation Details

The model uses the Bartowski calibration_datav3.txt dataset for iMatrix quantization, showing superior performance compared to wikitext and non-iMatrix versions. It maintains a vocabulary size of 128,320 tokens, aligned with the official Hunyuan Video specifications.

  • Optimized quantization using iMatrix technology
  • Compatible with vision tasks through mmproj integration
  • Enhanced performance through calibrated quantization
  • Specialized vocabulary handling for Hunyuan Video compatibility

Core Capabilities

  • Text encoding for Hunyuan Video applications
  • Vision-language tasks with appropriate mmproj files
  • Efficient model compression while maintaining performance
  • Specialized quantization for improved accuracy

Frequently Asked Questions

Q: What makes this model unique?

The model's unique feature is its iMatrix GGUF conversion optimized specifically for Hunyuan Video applications, with carefully calibrated quantization that outperforms standard approaches.

Q: What are the recommended use cases?

Primary use cases include text encoding for Hunyuan Video and vision-language tasks when used with appropriate mmproj files. Note that IQ quantization operations may be slower in ComfyUI due to numpy fallback.

🍰 Interesting in building your own agents?
PromptLayer provides Huggingface integration tools to manage and monitor prompts with your whole team. Get started here.