snowflake-arctic-base

Maintained By
Snowflake

Snowflake Arctic Base

PropertyValue
Parameter Count482B parameters (17B active)
Model TypeDense-MoE Hybrid Transformer
LicenseApache-2.0
Tensor TypeBF16
Release DateApril 24th, 2024

What is snowflake-arctic-base?

Snowflake Arctic Base is an innovative language model that combines dense transformer architecture with Mixture of Experts (MoE) technology. It features a 10B dense transformer model integrated with a residual 128x3.66B MoE MLP, resulting in 482B total parameters while maintaining efficiency with only 17B active parameters during operation.

Implementation Details

The model leverages advanced features from DeepSpeed and requires version 4.39 or higher of the transformers library. It utilizes BF16 precision and supports both FP8 and FP6 quantization options for optimal performance.

  • Hybrid architecture combining dense and sparse components
  • Top-2 gating mechanism for expert selection
  • Custom code implementation with transformers library
  • DeepSpeed integration for efficient inference

Core Capabilities

  • Text and code generation
  • Enterprise-focused AI applications
  • Efficient parameter utilization through MoE architecture
  • Scalable deployment on high-performance hardware

Frequently Asked Questions

Q: What makes this model unique?

The model's hybrid architecture combining dense transformers with MoE allows it to achieve impressive performance while maintaining efficiency through selective parameter activation. This makes it particularly suitable for enterprise applications requiring both power and efficiency.

Q: What are the recommended use cases?

The model is optimized for enterprise AI applications, particularly text and code generation tasks. It's designed to run on high-performance hardware, with recommended deployment on 8xH100 GPU instances from major cloud providers.

🍰 Interesting in building your own agents?
PromptLayer provides Huggingface integration tools to manage and monitor prompts with your whole team. Get started here.