LLM2CLIP-EVA02-L-14-336
Property | Value |
---|---|
License | Apache 2.0 |
Paper | arXiv:2411.04997 |
Training Data | CC3M, CC12M, YFCC15M, Recap-DataComp-1B(30M subset) |
Primary Task | Zero-Shot Image Classification |
What is LLM2CLIP-EVA02-L-14-336?
LLM2CLIP-EVA02-L-14-336 is a groundbreaking vision-language model that combines the power of Large Language Models (LLMs) with CLIP architecture to enhance visual representation capabilities. Developed by Microsoft, this model introduces an innovative approach where LLMs are fine-tuned in the caption space using contrastive learning, significantly improving textual discriminability.
Implementation Details
The model employs a sophisticated architecture where a fine-tuned LLM acts as a teacher for CLIP's visual encoder. It's built upon the EVA02 architecture and supports processing of longer and more complex captions, overcoming traditional CLIP text encoder limitations. The implementation achieves a remarkable 16.5% performance improvement over the base EVA02 model in both long-text and short-text retrieval tasks.
- Leverages PyTorch framework for implementation
- Supports 336x336 image resolution
- Incorporates contrastive learning techniques
- Features cross-lingual capabilities
Core Capabilities
- Zero-shot image classification
- Cross-modal retrieval tasks
- Enhanced textual discriminability
- Multi-lingual support despite English-only training
- Integration capability with multimodal systems like Llava 1.5
Frequently Asked Questions
Q: What makes this model unique?
This model uniquely leverages LLMs to enhance CLIP's capabilities, allowing for better handling of complex and longer text descriptions while maintaining strong visual understanding. It achieves state-of-the-art performance in cross-lingual tasks despite being trained only on English data.
Q: What are the recommended use cases?
The model is ideal for zero-shot image classification, cross-modal retrieval tasks, and applications requiring sophisticated understanding of image-text relationships. It's particularly valuable in scenarios requiring multilingual support or processing of complex textual descriptions.