Update README.md
Browse files
README.md
CHANGED
|
@@ -4,6 +4,10 @@ tags:
|
|
| 4 |
- unsloth
|
| 5 |
- trl
|
| 6 |
- grpo
|
|
|
|
|
|
|
|
|
|
|
|
|
| 7 |
---
|
| 8 |
|
| 9 |
# Model Card for Model ID
|
|
@@ -18,37 +22,43 @@ tags:
|
|
| 18 |
|
| 19 |
<!-- Provide a longer summary of what this model is. -->
|
| 20 |
|
| 21 |
-
|
| 22 |
|
| 23 |
-
- **Developed by:** [
|
| 24 |
-
- **Funded by [optional]:** [
|
| 25 |
-
- **Shared by [optional]:** [
|
| 26 |
-
- **Model type:** [
|
| 27 |
-
- **Language(s) (NLP):** [
|
| 28 |
-
- **License:** [
|
| 29 |
-
- **Finetuned from model [optional]:** [
|
| 30 |
|
| 31 |
### Model Sources [optional]
|
| 32 |
|
| 33 |
<!-- Provide the basic links for the model. -->
|
| 34 |
|
| 35 |
-
- **Repository:** [
|
| 36 |
- **Paper [optional]:** [More Information Needed]
|
| 37 |
- **Demo [optional]:** [More Information Needed]
|
| 38 |
|
| 39 |
## Uses
|
| 40 |
|
| 41 |
-
|
| 42 |
|
| 43 |
### Direct Use
|
| 44 |
|
| 45 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
| 46 |
|
| 47 |
[More Information Needed]
|
| 48 |
|
| 49 |
### Downstream Use [optional]
|
| 50 |
|
| 51 |
-
|
|
|
|
|
|
|
| 52 |
|
| 53 |
[More Information Needed]
|
| 54 |
|
|
@@ -60,7 +70,11 @@ This is the model card of a 🤗 transformers model that has been pushed on the
|
|
| 60 |
|
| 61 |
## Bias, Risks, and Limitations
|
| 62 |
|
| 63 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
| 64 |
|
| 65 |
[More Information Needed]
|
| 66 |
|
|
@@ -95,7 +109,21 @@ Use the code below to get started with the model.
|
|
| 95 |
|
| 96 |
#### Training Hyperparameters
|
| 97 |
|
| 98 |
-
- **Training regime:** [More Information Needed]
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 99 |
|
| 100 |
#### Speeds, Sizes, Times [optional]
|
| 101 |
|
|
@@ -173,7 +201,14 @@ Carbon emissions can be estimated using the [Machine Learning Impact calculator]
|
|
| 173 |
|
| 174 |
## Citation [optional]
|
| 175 |
|
| 176 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 177 |
|
| 178 |
**BibTeX:**
|
| 179 |
|
|
|
|
| 4 |
- unsloth
|
| 5 |
- trl
|
| 6 |
- grpo
|
| 7 |
+
license: apache-2.0
|
| 8 |
+
base_model:
|
| 9 |
+
- Qwen/Qwen2.5-3B
|
| 10 |
+
- Qwen/Qwen2.5-VL-3B-Instruct
|
| 11 |
---
|
| 12 |
|
| 13 |
# Model Card for Model ID
|
|
|
|
| 22 |
|
| 23 |
<!-- Provide a longer summary of what this model is. -->
|
| 24 |
|
| 25 |
+
DeutscheLexAI_BGB is a fine-tuned Qwen2.5-3B model specializing in German legal text processing, trained on the Bürgerliches Gesetzbuch (BGB) dataset. It enhances legal text understanding, summarization, and reasoning for German legal documents.
|
| 26 |
|
| 27 |
+
- **Developed by:** [Ali Asghar (jaffry258@gmail.com)]
|
| 28 |
+
- **Funded by [optional]:** [still under progress ]
|
| 29 |
+
- **Shared by [optional]:** []
|
| 30 |
+
- **Model type:** [Large Language Model (LLM)]
|
| 31 |
+
- **Language(s) (NLP):** [pytorch,transformers,python]
|
| 32 |
+
- **License:** [Appache 2.0]
|
| 33 |
+
- **Finetuned from model [optional]:** [Qwen2.5-3B]
|
| 34 |
|
| 35 |
### Model Sources [optional]
|
| 36 |
|
| 37 |
<!-- Provide the basic links for the model. -->
|
| 38 |
|
| 39 |
+
- **Repository:** [https://huggingface.co/Alijeff1214/DeutscheLexAI_BGB/tree/main]
|
| 40 |
- **Paper [optional]:** [More Information Needed]
|
| 41 |
- **Demo [optional]:** [More Information Needed]
|
| 42 |
|
| 43 |
## Uses
|
| 44 |
|
| 45 |
+
DeutscheLexAI_BGB is a fine-tuned Qwen2.5-3B model specializing in German legal text processing, trained on the Bürgerliches Gesetzbuch (BGB) dataset. It enhances legal text understanding, summarization, and reasoning for German legal documents.
|
| 46 |
|
| 47 |
### Direct Use
|
| 48 |
|
| 49 |
+
Legal research: Extract, summarize, and analyze BGB texts.
|
| 50 |
+
|
| 51 |
+
AI-powered legal assistants: Provide insights into German law.
|
| 52 |
+
|
| 53 |
+
Academic purposes: Assists in legal document structuring.
|
| 54 |
|
| 55 |
[More Information Needed]
|
| 56 |
|
| 57 |
### Downstream Use [optional]
|
| 58 |
|
| 59 |
+
Chatbots for legal guidance.
|
| 60 |
+
|
| 61 |
+
AI-based contract analysis.
|
| 62 |
|
| 63 |
[More Information Needed]
|
| 64 |
|
|
|
|
| 70 |
|
| 71 |
## Bias, Risks, and Limitations
|
| 72 |
|
| 73 |
+
The model may reflect biases in the BGB dataset.
|
| 74 |
+
|
| 75 |
+
Not suitable for real-time legal decision-making.
|
| 76 |
+
|
| 77 |
+
Might struggle with non-German legal texts.
|
| 78 |
|
| 79 |
[More Information Needed]
|
| 80 |
|
|
|
|
| 109 |
|
| 110 |
#### Training Hyperparameters
|
| 111 |
|
| 112 |
+
- **Training regime:** [More Information Needed]
|
| 113 |
+
- trainer = GRPOTrainer(
|
| 114 |
+
model = model,
|
| 115 |
+
processing_class = tokenizer,
|
| 116 |
+
reward_funcs = [
|
| 117 |
+
xmlcount_reward_func,
|
| 118 |
+
soft_format_reward_func,
|
| 119 |
+
strict_format_reward_func,
|
| 120 |
+
int_reward_func,
|
| 121 |
+
correctness_reward_func,
|
| 122 |
+
],
|
| 123 |
+
args = training_args,
|
| 124 |
+
train_dataset = dataset,
|
| 125 |
+
)
|
| 126 |
+
trainer.train()
|
| 127 |
|
| 128 |
#### Speeds, Sizes, Times [optional]
|
| 129 |
|
|
|
|
| 201 |
|
| 202 |
## Citation [optional]
|
| 203 |
|
| 204 |
+
@article{DeutscheLexAI_BGB,
|
| 205 |
+
title={DeutscheLexAI_BGB: A Fine-Tuned Qwen2.5-3B Model for German Legal Texts},
|
| 206 |
+
author={Your Name or Organization},
|
| 207 |
+
journal={Hugging Face Model Hub},
|
| 208 |
+
year={2025},
|
| 209 |
+
url={https://huggingface.co/Alijeff1214/DeutscheLexAI_BGB}
|
| 210 |
+
}
|
| 211 |
+
|
| 212 |
|
| 213 |
**BibTeX:**
|
| 214 |
|