Hugging Face
Models
Datasets
Spaces
Buckets
new
Docs
Enterprise
Pricing
Log In
Sign Up
Spaces:
Alovestocode
/
ZeroGPU-LLM-Inference
like
0
Sleeping
App
Files
Files
Community
Fetching metadata from the HF Docker repository...
03689e3
ZeroGPU-LLM-Inference
68.7 kB
Ctrl+K
Ctrl+K
1 contributor
History:
24 commits
Alikestocode
Fix Gradio UI structure and add comprehensive fallback logging
03689e3
5 months ago
.gitattributes
Safe
1.52 kB
Initial commit: ZeroGPU LLM Inference Space
5 months ago
.gitignore
Safe
27 Bytes
Add .gitignore and remove cache files
5 months ago
README.md
Safe
4.23 kB
Implement vLLM with LLM Compressor and performance optimizations
5 months ago
UI_UX_IMPROVEMENTS.md
Safe
6.81 kB
Initial commit: ZeroGPU LLM Inference Space
5 months ago
USER_GUIDE.md
Safe
7.9 kB
Initial commit: ZeroGPU LLM Inference Space
5 months ago
app.py
Safe
34.5 kB
Fix Gradio UI structure and add comprehensive fallback logging
5 months ago
apt.txt
Safe
11 Bytes
Initial commit: ZeroGPU LLM Inference Space
5 months ago
requirements.txt
Safe
197 Bytes
Implement vLLM with LLM Compressor and performance optimizations
5 months ago
style.css
Safe
2.84 kB
Initial commit: ZeroGPU LLM Inference Space
5 months ago
test_api.py
Safe
3.43 kB
Migrate to AWQ quantization with FlashAttention-2
5 months ago
test_api_gradio_client.py
Safe
7.2 kB
Implement vLLM with LLM Compressor and performance optimizations
5 months ago