Update README.md
Browse files
README.md
CHANGED
|
@@ -97,10 +97,14 @@ It's important to note that AI is not a single entity, but rather a rapidly evol
|
|
| 97 |
|
| 98 |
#### Conclusion
|
| 99 |
|
| 100 |
-
LLama 2 can understand the question and gives the user a very specific and overall a better answer compared to the one given by the fine-tuned model. However the fine-tuned model answers with a sentence written in a perfect italian.
|
| 101 |
|
| 102 |
### Training Data and Details
|
| 103 |
|
| 104 |
The dataset used is [seeweb/Seeweb-it-292-forLLM](https://huggingface.co/datasets/seeweb/Seeweb-it-292-forLLM), a dataset containing approx. 300 italian prompt-answer conversations.
|
| 105 |
|
| 106 |
-
The training has been made on RTX A6000, inside [Seeweb's Cloud Server GPU](https://www.seeweb.it/prodotti/cloud-server-gpu)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 97 |
|
| 98 |
#### Conclusion
|
| 99 |
|
| 100 |
+
LLama 2 can understand the question and gives the user a very specific and overall a better answer compared to the one given by the fine-tuned model. However the fine-tuned model answers with a sentence written in a perfect italian, which is what we were trying to achieve with this fine-tuning process.
|
| 101 |
|
| 102 |
### Training Data and Details
|
| 103 |
|
| 104 |
The dataset used is [seeweb/Seeweb-it-292-forLLM](https://huggingface.co/datasets/seeweb/Seeweb-it-292-forLLM), a dataset containing approx. 300 italian prompt-answer conversations.
|
| 105 |
|
| 106 |
+
The training has been made on RTX A6000, inside [Seeweb's Cloud Server GPU](https://www.seeweb.it/prodotti/cloud-server-gpu)
|
| 107 |
+
|
| 108 |
+
### What next?
|
| 109 |
+
|
| 110 |
+
The model must be improved: a much bigger dataset needs to be created so that the model can learn many more ways to answer.
|