--- license: llama3 datasets: - ai-eldorado/Brazilian_CLT_preferences language: - en - pt base_model: - meta-llama/Llama-3.1-8B-Instruct tags: - legal --- **Model Description** This repository provides a **LoRA adapter** trained on the same human-preference dataset used for the DPO model. It enables efficient fine-tuning of **LLaMA-3 8B Instruct (4-bit)** for Brazilian labor law applications without requiring full model retraining. **Intended Use** The LoRA adapter is intended for developers and researchers who want to **adapt LLaMA-3 models for CLT-related tasks** while minimizing computational costs. It is particularly useful for **resource-constrained environments** or for integrating into **multi-agent legal assistant systems**. **Training Details** - **Base Model:** LLaMA-3 8B (4-bit quantized) - **Method:** LoRA fine-tuning with DPO-aligned dataset - **Dataset:** 736 human-preference entries on CLT-related questions - **Hyperparameters:** Same as the DPO model **Performance Summary** When merged with the base model, the LoRA adapter reproduces the improvements observed in the full DPO model: - Increased factual accuracy - Better semantic alignment with CLT regulations #### **Ethical Considerations** - **Legal Disclaimer:** Not a substitute for professional legal advice. - **Risk of Misuse:** Incorrect merging or use outside intended domain may lead to inaccurate outputs. - **Data Privacy:** No personal data was used in training. #### **Bias and Fairness** - Same considerations as the DPO model: - Regional and interpretation biases may exist. - Limited dataset size could affect fairness in edge cases. #### **Limitations** - Requires merging with the base LLaMA-3 model for inference. - Domain-specific; not suitable for general-purpose legal reasoning. #### **Citation** soon