Increase token limits for SmolLM2-360M: better quality answers d27c8a4 Aditya-1301 commited on Nov 12, 2025
Complete Ollama removal: Switch to local transformers with SmolLM2-135M for CPU inference 9498975 Aditya-1301 commited on Nov 12, 2025
Switch to meta-llama/Llama-3.2-1B-Instruct - available via HF Inference API 204b852 Aditya-1301 commited on Nov 12, 2025
Fix HuggingFace model name format: use HuggingFaceTB/SmolLM2-360M instead of Ollama format 95b613c Aditya-1301 commited on Nov 12, 2025
Fix HuggingFace Spaces compatibility: standardize InferenceClient API and improve backend auto-detection f7e1108 Aditya-1301 commited on Nov 12, 2025
Make web search truly optional with ENABLE_WEB_SEARCH flag and graceful fallback 70821a6 Aditya-1301 commited on Nov 9, 2025
Switch to SmolLM2-1.7B for CPU-friendly inference on HF Spaces a68b4d6 Aditya-1301 commited on Nov 9, 2025
Add HuggingFace Inference API as LLM backend with auto-detection 5170f73 Aditya-1301 commited on Nov 9, 2025
Implement code structure updates and remove redundant sections b1703c9 Aditya-1301 commited on Nov 2, 2025