Update README.md
Browse filesfixing the markdown tags
README.md
CHANGED
|
@@ -9,15 +9,6 @@ tags:
|
|
| 9 |
- merged
|
| 10 |
---
|
| 11 |
|
| 12 |
-
base_model: NousResearch/Meta-Llama-3.1-8B-Instruct
|
| 13 |
-
tags:
|
| 14 |
-
- llama3
|
| 15 |
-
- instruction-tuning
|
| 16 |
-
- summarization
|
| 17 |
-
- fine-tuned
|
| 18 |
-
- merged
|
| 19 |
-
---
|
| 20 |
-
|
| 21 |
# 🧠 FlamingNeuron / llama381binstruct_summarize_short_merged
|
| 22 |
|
| 23 |
This is a **merged model** based on [NousResearch/Meta-Llama-3.1-8B-Instruct](https://huggingface.co/NousResearch/Meta-Llama-3.1-8B-Instruct), fine-tuned using LoRA adapters for legal-domain summarization. The LoRA weights have been merged with the base model for standalone use.
|
|
@@ -44,7 +35,19 @@ inputs = tokenizer(prompt, return_tensors="pt").to("cuda")
|
|
| 44 |
outputs = model.generate(**inputs, max_new_tokens=128)
|
| 45 |
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
|
| 46 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 47 |
## 📚 Citations
|
| 48 |
|
| 49 |
-
```markdown
|
| 50 |
This model was fine-tuned using [TRL](https://github.com/huggingface/trl).
|
|
|
|
| 9 |
- merged
|
| 10 |
---
|
| 11 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 12 |
# 🧠 FlamingNeuron / llama381binstruct_summarize_short_merged
|
| 13 |
|
| 14 |
This is a **merged model** based on [NousResearch/Meta-Llama-3.1-8B-Instruct](https://huggingface.co/NousResearch/Meta-Llama-3.1-8B-Instruct), fine-tuned using LoRA adapters for legal-domain summarization. The LoRA weights have been merged with the base model for standalone use.
|
|
|
|
| 35 |
outputs = model.generate(**inputs, max_new_tokens=128)
|
| 36 |
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
|
| 37 |
|
| 38 |
+
```
|
| 39 |
+
## 🏋️ Training Procedure
|
| 40 |
+
|
| 41 |
+
This model was trained using **Supervised Fine-Tuning (SFT)** on legal document summaries using the [legal_summarization](https://github.com/lauramanor/legal_summarization) dataset. LoRA adapters were applied during training and merged afterward using `merge_and_unload()`.
|
| 42 |
+
|
| 43 |
+
### ⚙️ Framework Versions
|
| 44 |
+
|
| 45 |
+
- TRL: 0.16.1
|
| 46 |
+
- Transformers: 4.51.3
|
| 47 |
+
- PyTorch: 2.6.0+cu124
|
| 48 |
+
- Datasets: 3.5.0
|
| 49 |
+
- Tokenizers: 0.21.1
|
| 50 |
+
|
| 51 |
## 📚 Citations
|
| 52 |
|
|
|
|
| 53 |
This model was fine-tuned using [TRL](https://github.com/huggingface/trl).
|