TinyLLama-v0

TinyLLama-v0

Maykeye

A compact Llama-architecture model (4.62M params) trained on TinyStories dataset, featuring BF16 precision and Apache 2.0 license. Ideal for text generation tasks.

PropertyValue
Parameter Count4.62M
Model TypeText Generation
ArchitectureLlama
LicenseApache 2.0
Tensor TypeBF16

What is TinyLLama-v0?

TinyLLama-v0 is an innovative recreation of the TinyStories-1M model using the Llama architecture. Developed by Maykeye, this compact model represents a proof-of-concept approach to implementing lightweight language models for text generation tasks.

Implementation Details

The model was trained on the TinyStoriesV2-GPT4 dataset, utilizing an A100 GPU with 40GB VRAM, completing training in approximately 9 hours across three epochs. It employs the tokenizer from open_llama_3b and is implemented using PyTorch with Safetensors support.

  • Training process fully documented in train.ipynb notebook
  • Utilizes approximately 30GB VRAM during training
  • Implements a basic caching mechanism for story shuffling
  • Context size limitation with truncation for longer stories

Core Capabilities

  • Text generation using Llama architecture
  • Compatible with text-generation-inference endpoints
  • Efficient memory usage with BF16 precision
  • Simple deployment and integration options

Frequently Asked Questions

Q: What makes this model unique?

TinyLLama-v0 stands out for its extremely lightweight implementation of the Llama architecture, making it accessible for experimentation and learning purposes while maintaining functionality for text generation tasks.

Q: What are the recommended use cases?

The model is best suited for educational purposes, prototyping, and scenarios where a lightweight language model is needed. It's particularly useful for understanding Llama architecture implementation and training processes.

Socials
PromptLayer
Company
All services online
Location IconPromptLayer is located in the heart of New York City
PromptLayer © 2026