FROM pytorch/pytorch:2.2.1-cuda12.1-cudnn8-runtime # Instalujemy vLLM jednym poleceniem. RUN pip install vllm # Uruchamiamy serwer vLLM z Pana modelem. CMD ["vllm", "serve", "adowu/falcon-h1-3b-instruct-pl-bfloat16", "--host", "0.0.0.0", "--trust-remote-code"]