|
|
FROM ubuntu:22.04 |
|
|
|
|
|
ENV DEBIAN_FRONTEND=noninteractive |
|
|
WORKDIR /app |
|
|
|
|
|
RUN apt update && apt install -y \ |
|
|
git \ |
|
|
cmake \ |
|
|
build-essential \ |
|
|
curl \ |
|
|
ca-certificates \ |
|
|
&& rm -rf /var/lib/apt/lists/* |
|
|
|
|
|
RUN git clone https://github.com/ggerganov/llama.cpp |
|
|
|
|
|
RUN cmake -S llama.cpp -B llama.cpp/build \ |
|
|
-DLLAMA_BUILD_SERVER=ON \ |
|
|
-DCMAKE_BUILD_TYPE=Release \ |
|
|
&& cmake --build llama.cpp/build --target llama-server |
|
|
|
|
|
RUN mkdir -p /models |
|
|
|
|
|
|
|
|
RUN curl -L --fail -o /models/model.gguf \ |
|
|
https://huggingface.co/LiquidAI/LFM2.5-1.2B-Instruct-GGUF/resolve/main/LFM2.5-1.2B-Instruct-Q8_0.gguf |
|
|
|
|
|
|
|
|
RUN head -c 4 /models/model.gguf |
|
|
|
|
|
COPY start.sh /start.sh |
|
|
RUN chmod +x /start.sh |
|
|
|
|
|
EXPOSE 7860 |
|
|
CMD ["/start.sh"] |