cabrita-lora-v0-1

Maintained By
22h

Cabrita-LoRA v0.1

PropertyValue
LanguagePortuguese
Base ModelLLaMA 7B
Training MethodLoRA Fine-tuning
LicenseOpenRAIL

What is cabrita-lora-v0-1?

Cabrita-LoRA is a Portuguese language instruction-tuned variant of the LLaMA model, developed using the LoRA (Low-Rank Adaptation) fine-tuning approach. The model was trained on a Portuguese translation of the Stanford Alpaca dataset, making it particularly effective for Portuguese language instruction-following tasks.

Implementation Details

The model leverages the PEFT (Parameter-Efficient Fine-Tuning) framework from Hugging Face and was trained on a single A100 GPU for approximately 4 hours. The training data was created by translating the original Alpaca dataset using ChatGPT, providing a cost-effective solution for Portuguese language model development.

  • Built on LLaMA-7B base model
  • Uses LoRA for efficient fine-tuning
  • Trained on translated Alpaca dataset
  • Implements PEFT methodology

Core Capabilities

  • Portuguese language instruction following
  • Creative text generation
  • Structured response formatting
  • Natural language understanding in Portuguese

Frequently Asked Questions

Q: What makes this model unique?

This model represents one of the first Portuguese-language instruction-tuned variants of LLaMA, offering efficient fine-tuning through LoRA while maintaining high-quality outputs in Portuguese. The cost-effective approach to dataset translation (US$8.00) demonstrates an innovative way to create language-specific models.

Q: What are the recommended use cases?

The model is particularly suitable for Portuguese language applications requiring instruction following, creative text generation, and structured responses. It can be deployed for tasks ranging from content generation to providing detailed explanations in Portuguese.

🍰 Interesting in building your own agents?
PromptLayer provides Huggingface integration tools to manage and monitor prompts with your whole team. Get started here.