replit-code-1.5

Maintained By
modularai

Replit Code V-1.5 3B

PropertyValue
Parameter Count3.32B
LicenseApache 2.0
Context Length4096 tokens
Training Data1T tokens
ArchitectureCausal Language Model

What is replit-code-1.5?

Replit Code V-1.5 is a specialized code completion model trained by Replit using MosaicML's platform. This 3.32B parameter model is designed specifically for code generation tasks, trained on a diverse dataset of 30 programming languages from the Stack Dedup dataset and RedPajama's StackExchange data.

Implementation Details

The model utilizes a custom-trained GPTNeoX tokenizer with an optimized vocabulary of 32,768 tokens, achieving improved compression while maintaining high coverage. Training was conducted on 128 H100-80GB GPUs using MosaicML's LLM Foundry and Composer framework.

  • Trained in bfloat16 precision
  • Supports 30 programming languages including Python, JavaScript, Java, C++, and more
  • Implements Triton Flash Attention for improved performance
  • Context window of 4096 tokens

Core Capabilities

  • Advanced code completion across multiple programming languages
  • Natural language processing for development-related content
  • Efficient token compression with optimized vocabulary
  • Support for both CPU and GPU inference

Frequently Asked Questions

Q: What makes this model unique?

The model's specialized training on permissively licensed code and its optimized vocabulary make it particularly effective for code completion tasks while maintaining efficient token compression.

Q: What are the recommended use cases?

The model is best suited for code completion tasks, development assistance, and as a foundation for application-specific fine-tuning in commercial and non-commercial settings. It's particularly effective when used with appropriate temperature and repetition penalty settings for optimal performance.

🍰 Interesting in building your own agents?
PromptLayer provides Huggingface integration tools to manage and monitor prompts with your whole team. Get started here.