itsrocchi commited on
Commit
01fc5b7
·
1 Parent(s): a6d1714

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -2
README.md CHANGED
@@ -97,10 +97,14 @@ It's important to note that AI is not a single entity, but rather a rapidly evol
97
 
98
  #### Conclusion
99
 
100
- LLama 2 can understand the question and gives the user a very specific and overall a better answer compared to the one given by the fine-tuned model. However the fine-tuned model answers with a sentence written in a perfect italian.
101
 
102
  ### Training Data and Details
103
 
104
  The dataset used is [seeweb/Seeweb-it-292-forLLM](https://huggingface.co/datasets/seeweb/Seeweb-it-292-forLLM), a dataset containing approx. 300 italian prompt-answer conversations.
105
 
106
- The training has been made on RTX A6000, inside [Seeweb's Cloud Server GPU](https://www.seeweb.it/prodotti/cloud-server-gpu)
 
 
 
 
 
97
 
98
  #### Conclusion
99
 
100
+ LLama 2 can understand the question and gives the user a very specific and overall a better answer compared to the one given by the fine-tuned model. However the fine-tuned model answers with a sentence written in a perfect italian, which is what we were trying to achieve with this fine-tuning process.
101
 
102
  ### Training Data and Details
103
 
104
  The dataset used is [seeweb/Seeweb-it-292-forLLM](https://huggingface.co/datasets/seeweb/Seeweb-it-292-forLLM), a dataset containing approx. 300 italian prompt-answer conversations.
105
 
106
+ The training has been made on RTX A6000, inside [Seeweb's Cloud Server GPU](https://www.seeweb.it/prodotti/cloud-server-gpu)
107
+
108
+ ### What next?
109
+
110
+ The model must be improved: a much bigger dataset needs to be created so that the model can learn many more ways to answer.