Hugging Face's logo Hugging Face
  • Models
  • Datasets
  • Spaces
  • Buckets new
  • Docs
  • Enterprise
  • Pricing

  • Log In
  • Sign Up

Spaces:
Alovestocode
/
ZeroGPU-LLM-Inference
Sleeping

App Files Files Community
Fetching metadata from the HF Docker repository...
ZeroGPU-LLM-Inference
68.7 kB
Ctrl+K
Ctrl+K
  • 1 contributor
History: 24 commits
Alikestocode's picture
Alikestocode
Fix Gradio UI structure and add comprehensive fallback logging
03689e3 5 months ago
  • .gitattributes
    1.52 kB
    Initial commit: ZeroGPU LLM Inference Space 5 months ago
  • .gitignore
    27 Bytes
    Add .gitignore and remove cache files 5 months ago
  • README.md
    4.23 kB
    Implement vLLM with LLM Compressor and performance optimizations 5 months ago
  • UI_UX_IMPROVEMENTS.md
    6.81 kB
    Initial commit: ZeroGPU LLM Inference Space 5 months ago
  • USER_GUIDE.md
    7.9 kB
    Initial commit: ZeroGPU LLM Inference Space 5 months ago
  • app.py
    34.5 kB
    Fix Gradio UI structure and add comprehensive fallback logging 5 months ago
  • apt.txt
    11 Bytes
    Initial commit: ZeroGPU LLM Inference Space 5 months ago
  • requirements.txt
    197 Bytes
    Implement vLLM with LLM Compressor and performance optimizations 5 months ago
  • style.css
    2.84 kB
    Initial commit: ZeroGPU LLM Inference Space 5 months ago
  • test_api.py
    3.43 kB
    Migrate to AWQ quantization with FlashAttention-2 5 months ago
  • test_api_gradio_client.py
    7.2 kB
    Implement vLLM with LLM Compressor and performance optimizations 5 months ago