Model Description
This repository provides a LoRA adapter trained on the same human-preference dataset used for the DPO model. It enables efficient fine-tuning of LLaMA-3 8B Instruct (4-bit) for Brazilian labor law applications without requiring full model retraining.
Intended Use
The LoRA adapter is intended for developers and researchers who want to adapt LLaMA-3 models for CLT-related tasks while minimizing computational costs. It is particularly useful for resource-constrained environments or for integrating into multi-agent legal assistant systems.
Training Details
- Base Model: LLaMA-3 8B (4-bit quantized)
- Method: LoRA fine-tuning with DPO-aligned dataset
- Dataset: 736 human-preference entries on CLT-related questions
- Hyperparameters: Same as the DPO model
Performance Summary
When merged with the base model, the LoRA adapter reproduces the improvements observed in the full DPO model:
- Increased factual accuracy
- Better semantic alignment with CLT regulations
Ethical Considerations
- Legal Disclaimer: Not a substitute for professional legal advice.
- Risk of Misuse: Incorrect merging or use outside intended domain may lead to inaccurate outputs.
- Data Privacy: No personal data was used in training.
Bias and Fairness
- Same considerations as the DPO model:
- Regional and interpretation biases may exist.
- Limited dataset size could affect fairness in edge cases.
Limitations
- Requires merging with the base LLaMA-3 model for inference.
- Domain-specific; not suitable for general-purpose legal reasoning.
Citation
soon
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
🙋
Ask for provider support
Model tree for ai-eldorado/Brazilian_CLT_lora
Base model
meta-llama/Llama-3.1-8B
Finetuned
meta-llama/Llama-3.1-8B-Instruct