ZeroGPU-LLM-Inference / requirements.txt

Commit History

Clarify LLM Compressor optional status - vLLM has native AWQ support
b2bf767

Alikestocode commited on

Implement vLLM with LLM Compressor and performance optimizations
a79facb

Alikestocode commited on

Migrate to AWQ quantization with FlashAttention-2
06b4cf5

Alikestocode commited on

Update app.py and requirements.txt for CourseGPT-Pro router models
4c3d05b

Alikestocode commited on

Initial commit: ZeroGPU LLM Inference Space
f91e906

Alikestocode commited on