# Use an official Python runtime as a parent image FROM python:3.9-slim # Set the working directory in the container WORKDIR /app # Install wget and other necessary tools RUN apt-get update && apt-get install -y wget && rm -rf /var/lib/apt/lists/* # Copy the current directory contents into the container at /app COPY . /app # Download the Llama model RUN wget https://huggingface.co/Mozilla/Meta-Llama-3.1-8B-Instruct-llamafile/resolve/main/Meta-Llama-3.1-8B-Instruct.Q6_K.llamafile && \ chmod +x Meta-Llama-3.1-8B-Instruct.Q6_K.llamafile # Make port 8080 available to the world outside this container EXPOSE 8080 # Run the Llama model when the container launches CMD ["./Meta-Llama-3.1-8B-Instruct.Q6_K.llamafile"]