A finetuned model based on t5-small (~60M parameters) that given an answer it responds with a question. I call it AQ model because it does the opposite of the usual question answering LLM model.

This AQ-model is useful in coversations with another LLM-QA-chatbot, so that the conversation does not get stuck but moves continously to new topics. If you have an automatic conversation between two LLMs, one QA-LLM and one AQ-LLM the conversation will not get stuck and repetitive but continue forever :-)

The model was finetuned starting from t5-small on a NVidia RTX 3090 in about 1 1/2h with a batch size of 8, using 4 GB of RAM on the GPU. As the GPU was running at 320W, the energy to train this model was 480Wh.
The same model trained with a batch size of 32 gave sligthly worse results (14.3 RAM GB on the GPU in 1 hour).

Test with

./test_aqmodel.py "The hypothesis fails because of the decay with radius to the power of 3"

Output: What is the reason the hypothesis is a faulty hypothesis?

Last but not least: this model was finetuned with help of python scripts suggested by ChatGPT-4o 8-) Using vibe programming as Karpathy names it ...

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for dangermouse77/FromAnswerToQuestion-T5-small-60M

Base model

google-t5/t5-small
Finetuned
(2226)
this model

Dataset used to train dangermouse77/FromAnswerToQuestion-T5-small-60M