Spaces:
Paused
Paused
Commit History
halucination fixes af19ca9
prompt 6b9e619
prompt change b97884f
prompt 3466be8
prompt a313e26
prompt optimization acad387
user friendly prompt change 24a085f
fixes for formatted output cd9dd67
prompt update 67d7777
update prompt 2f0674a
Modified Dockerfile to include build-essential and cmake. 032e872
docker and requirements.txt b0bb219
pinned torch==2.3.1, torchvision==0.18.1, and torchaudio==2.3.1 in requirements.txt to resolve the version conflict. b1d7235
removed openai-whisper from requirements.txt. 35ce356
RUN pip install --no-cache-dir --no-build-isolation openai-whisper==20231117 ba7396d
fixed the build error related to openai-whisper. It failed because it relies on pkg_resources, which was removed in recent versions of setuptools. I've updated the Dockerfile to use a compatible version (setuptools<70.0.0). 569dec6
preload disabled c7103dc
entrypoint.sh !models_config.json in docker ignore 10cb238
fix for the "file not found" error ef49b1e
docker files fix ec7a30f
Merge branch 'FT-DEV-20/01/2026' 3e3e71c
rag 79f3ec0
Revert "refactor(ai-service): optimize prompts, token counting, and benchmarking - Externalize system prompts to text files for better maintainability. - Integrate tiktoken for faster and more accurate token counting. - Refactor BenchmarkLogger to use asynchronous logging for zero latency impact. - Improve prompt echo removal logic with more robust markers. - Add specialized system instruction for medical document processing. - Update test expectations to reflect precise token counting." 8c76d6f
Revert "update path for verify_cache.py" 9a2be98
Revert "disable preload models" 7c67d4d
Revert "minimal docker" a032d5a
Revert "use existing entrypoint.sh instead of inline generation"" 4612788
Revert "entry point docker ignore" b5ed223
entry point docker ignore 2f7babe
use existing entrypoint.sh instead of inline generation" 2fb56a3
minimal docker b6b8cb1
disable preload models 9cd4351
update path for verify_cache.py e23bd85
refactor(ai-service): optimize prompts, token counting, and benchmarking - Externalize system prompts to text files for better maintainability. - Integrate tiktoken for faster and more accurate token counting. - Refactor BenchmarkLogger to use asynchronous logging for zero latency impact. - Improve prompt echo removal logic with more robust markers. - Add specialized system instruction for medical document processing. - Update test expectations to reflect precise token counting. 4f8dcad
class 'generator' error. e00d3b6
traceback in benchmark api ed63cbf
error success benchmark 144f3be
exceptions and error logs added 7f8ace3
bench mark 188ec8d
update unified manager.py 2ee5ce0
resolve streaming generation errors and improve robustness 7fdf84c
dynamic cache fix 16d588e
dynamic cache patch e23b3f0
openvino fix b13236b
changes for publishing the latest including generate_generic api 4156c57
Optimize performance and fix torchvision error 3600c13
Adhil Krishna G commited on
Local running issue fix e913b78
Adhil Krishna G commited on
Update requirements.txt 9ea5231
Adhil Krishna G commited on
Environment variables issue fix 91bf800
Adhil Krishna G commited on