Update README.md
Browse files
README.md
CHANGED
|
@@ -5,8 +5,6 @@ language:
|
|
| 5 |
- fr
|
| 6 |
- en
|
| 7 |
tags:
|
| 8 |
-
- pretrained
|
| 9 |
-
- llama-3
|
| 10 |
- openllm-france
|
| 11 |
datasets:
|
| 12 |
- cmh/alpaca_data_cleaned_fr_52k
|
|
@@ -20,11 +18,11 @@ datasets:
|
|
| 20 |
base_model:
|
| 21 |
- OpenLLM-France/Lucie-7B
|
| 22 |
widget:
|
| 23 |
-
|
| 24 |
-
|
| 25 |
-
|
| 26 |
-
|
| 27 |
-
|
| 28 |
training_progress:
|
| 29 |
context_length: 32000
|
| 30 |
---
|
|
@@ -57,7 +55,7 @@ Note that this instruction training is light and is meant to allow Lucie to prod
|
|
| 57 |
|
| 58 |
Due to its size, Lucie-7B is limited in the information that it can memorize; its ability to produce correct answers could be improved by implementing the model in a retrieval augmented generation pipeline.
|
| 59 |
|
| 60 |
-
While Lucie-7B-Instruct is trained on sequences of 4096 tokens, its base model, Lucie-7B has a context size of 32K tokens. Based on Needle-in-a-haystack evaluations, Lucie-7B-Instruct
|
| 61 |
|
| 62 |
|
| 63 |
## Training details
|
|
@@ -232,4 +230,4 @@ Finally, we thank the entire OpenLLM-France community, whose members have helped
|
|
| 232 |
|
| 233 |
## Contact
|
| 234 |
|
| 235 |
-
contact@openllm-france.fr
|
|
|
|
| 5 |
- fr
|
| 6 |
- en
|
| 7 |
tags:
|
|
|
|
|
|
|
| 8 |
- openllm-france
|
| 9 |
datasets:
|
| 10 |
- cmh/alpaca_data_cleaned_fr_52k
|
|
|
|
| 18 |
base_model:
|
| 19 |
- OpenLLM-France/Lucie-7B
|
| 20 |
widget:
|
| 21 |
+
- text: |-
|
| 22 |
+
Quelle est la capitale de l'Espagne ? Madrid.
|
| 23 |
+
Quelle est la capitale de la France ?
|
| 24 |
+
example_title: Capital cities in French
|
| 25 |
+
group: 1-shot Question Answering
|
| 26 |
training_progress:
|
| 27 |
context_length: 32000
|
| 28 |
---
|
|
|
|
| 55 |
|
| 56 |
Due to its size, Lucie-7B is limited in the information that it can memorize; its ability to produce correct answers could be improved by implementing the model in a retrieval augmented generation pipeline.
|
| 57 |
|
| 58 |
+
While Lucie-7B-Instruct is trained on sequences of 4096 tokens, its base model, Lucie-7B has a context size of 32K tokens. Based on Needle-in-a-haystack evaluations, Lucie-7B-Instruct-v1.1 has a context window size of 22K tokens. This window could be increasd by fine-tuning on longer data samples.
|
| 59 |
|
| 60 |
|
| 61 |
## Training details
|
|
|
|
| 230 |
|
| 231 |
## Contact
|
| 232 |
|
| 233 |
+
contact@openllm-france.fr
|