Update README.md
Browse files
README.md
CHANGED
|
@@ -14,7 +14,7 @@ This AQ-model is useful in coversations with another LLM-QA-chatbot, so that the
|
|
| 14 |
If you have an automatic conversation between two LLMs, one QA-LLM and one AQ-LLM the conversation will not get stuck and repetitive but continue forever :-)
|
| 15 |
|
| 16 |
The model was finetuned starting from t5-small on a NVidia RTX 3090 in about 1 1/2h with a batch size of 8, using 4 GB of RAM on the GPU.
|
| 17 |
-
The same model trained with a batch size of 32 (14.3 RAM GB on the GPU in 1 hour)
|
| 18 |
|
| 19 |
Test with
|
| 20 |
|
|
|
|
| 14 |
If you have an automatic conversation between two LLMs, one QA-LLM and one AQ-LLM the conversation will not get stuck and repetitive but continue forever :-)
|
| 15 |
|
| 16 |
The model was finetuned starting from t5-small on a NVidia RTX 3090 in about 1 1/2h with a batch size of 8, using 4 GB of RAM on the GPU.
|
| 17 |
+
The same model trained with a batch size of 32 gave sligthly worse results (14.3 RAM GB on the GPU in 1 hour).
|
| 18 |
|
| 19 |
Test with
|
| 20 |
|