Revert "Replace use of transformers library with vLLM for high-speed inference" 74aae71 timadair commited on Sep 23, 2025
Replace use of transformers library with vLLM for high-speed inference 56c4132 timadair commited on Sep 23, 2025
Change duration, reasoning level, max_new_tokens to ensure a smooth demo 7cea9da timadair commited on Sep 10, 2025
Split out system prompt to use gpt-oss-20b's low reasoning mode, max_new_tokens->1500, comment cleanup 593fd81 timadair commited on Sep 6, 2025
Update readme with awareness of ZeroGPU limitations for visitors 5e8e9f3 timadair commited on Sep 5, 2025
One last time. Duration of 45s, which should allow almost all requests to succeed. Visitors can't generate if the duration exceeds their remaining daily allotment. aca3670 timadair commited on Sep 5, 2025
Populate the README, fix 'inference not run' formatting error c4da447 timadair commited on Sep 5, 2025
Try to have the quiz_json box visible while the quiz is being generated. 246c399 timadair commited on Sep 5, 2025
That works functionally but the layout is bad. Have questions be inline with answer radio button groups. 463477c timadair commited on Sep 5, 2025
Another UI rebuild. Have a fixed number of questions slots that are populated instead of trying to dynamically create UI element. 1142597 timadair commited on Sep 5, 2025
Change to grabbing only text after the final instance of the start of the JSON. da4ab71 timadair commited on Sep 5, 2025
Be more explicit with instructions to remove code-generating behavior c131aa6 timadair commited on Sep 4, 2025
Use roleplaying to try to influence the LLM away from writing code that generates a quiz eb22e5c timadair commited on Sep 4, 2025