| # Use the Hugging Face text-generation inference image as a base image | |
| FROM ghcr.io/huggingface/text-generation-inference:latest | |
| # Install Python 3.11 and necessary dependencies | |
| RUN apt-get update && \ | |
| apt-get install -y python3.11 python3.11-dev libpython3.11-dev && \ | |
| apt-get clean && \ | |
| ldconfig | |
| # Set environment variables | |
| ENV PYTHONIOENCODING=UTF-8 | |
| # Copy the model from Hugging Face repository (using GitHub URL for reference) | |
| # If your model is uploaded in the model directory within Hugging Face, it will be automatically accessible | |
| # Command to run the model inference server | |
| CMD ["--model-id", "Sidreds06/MHCV1", "--max-batch-prefill-tokens", "4096"] | |