tiny-random-Phi3ForCausalLM

Maintained By
Xenova

tiny-random-Phi3ForCausalLM

PropertyValue
Parameter Count2.07M
Model TypeText Generation
ArchitecturePhi-3 (Minimal)
Tensor TypeF32
Downloads325,708

What is tiny-random-Phi3ForCausalLM?

tiny-random-Phi3ForCausalLM is a minimalistic implementation of the Phi-3 architecture, developed by Xenova. This compact model features just 2.07M parameters, making it ideal for experimentation and testing purposes. The model is built with a simplified architecture consisting of 2 hidden layers, 4 attention heads, and maintains a sliding window of 2047 tokens.

Implementation Details

The model implements a streamlined version of the Phi-3 architecture with the following specifications: hidden size of 32, intermediate size of 64, and 4 key-value heads. It utilizes the microsoft/Phi-3-mini-4k-instruct tokenizer and is optimized for F32 tensor operations.

  • Minimal architecture with 2 hidden layers
  • 4 attention heads for efficient processing
  • Sliding window of 2047 tokens
  • 32 hidden size dimensions
  • 64 intermediate size dimensions

Core Capabilities

  • Basic text generation tasks
  • Experimental model testing
  • Lightweight deployment scenarios
  • Educational purposes and architecture study

Frequently Asked Questions

Q: What makes this model unique?

This model stands out for its extremely compact size while maintaining the core Phi-3 architecture components. It's particularly useful for testing and educational purposes where a full-scale model would be unnecessary.

Q: What are the recommended use cases?

The model is best suited for experimental implementations, proof-of-concept testing, and educational scenarios where understanding the basic mechanics of transformer-based models is the primary goal.

🍰 Interesting in building your own agents?
PromptLayer provides Huggingface integration tools to manage and monitor prompts with your whole team. Get started here.