Llama_3.1_8b_DodoWild_v2.01

Maintained By
Nexesenex

Llama_3.1_8b_DodoWild_v2.01

PropertyValue
Base ModelLLaMA 3.1 8B
Merge MethodModel Stock
Model URLHugging Face
Average Benchmark Score30.31

What is Llama_3.1_8b_DodoWild_v2.01?

Llama_3.1_8b_DodoWild_v2.01 is a sophisticated merge of multiple LLaMA 3.1-based models, created by Nexesenex using mergekit. Built upon the SentientAGI/Dobby-Mini-Unhinged-Llama-3.1-8B base, it combines the strengths of Dolermed and Smarteaz variants to create a more capable language model.

Implementation Details

The model utilizes the Model Stock merge method with equal weights (1.0) assigned to both Dolermed and Smarteaz variants. It employs bfloat16 dtype for efficient computation and includes normalized weights with an auto chat template.

  • Custom tokenizer configuration using union method
  • Normalized weight distribution across merged models
  • Built on LLaMA 3.1 8B architecture

Core Capabilities

  • Strong instruction-following capability (IFEval: 79.78)
  • Decent performance on BBH 3-shot tasks (32.11)
  • Mathematical reasoning capabilities (MATH Lvl 5: 19.86)
  • Professional knowledge evaluation (MMLU-PRO: 30.43)

Frequently Asked Questions

Q: What makes this model unique?

This model's uniqueness lies in its balanced merge of two specialized LLaMA 3.1 variants, optimized for instruction following while maintaining general-purpose capabilities.

Q: What are the recommended use cases?

Given its strong performance in instruction following and reasonable scores across various benchmarks, this model is well-suited for general-purpose tasks, particularly those requiring precise instruction following and professional domain knowledge.

🍰 Interesting in building your own agents?
PromptLayer provides Huggingface integration tools to manage and monitor prompts with your whole team. Get started here.