| | FROM ubuntu:22.04 |
| |
|
| | ENV DEBIAN_FRONTEND=noninteractive |
| | WORKDIR /app |
| |
|
| | RUN apt update && apt install -y \ |
| | git \ |
| | cmake \ |
| | build-essential \ |
| | curl \ |
| | ca-certificates \ |
| | && rm -rf /var/lib/apt/lists/* |
| |
|
| | RUN git clone https://github.com/ggerganov/llama.cpp |
| |
|
| | RUN cmake -S llama.cpp -B llama.cpp/build \ |
| | -DLLAMA_BUILD_SERVER=ON \ |
| | -DCMAKE_BUILD_TYPE=Release \ |
| | && cmake --build llama.cpp/build --target llama-server |
| |
|
| | RUN mkdir -p /models |
| |
|
| | |
| | RUN curl -L --fail -o /models/model.gguf \ |
| | https://huggingface.co/LiquidAI/LFM2-350M-GGUF/resolve/main/LFM2-350M-Q8_0.gguf |
| |
|
| | |
| | RUN head -c 4 /models/model.gguf |
| |
|
| | COPY start.sh /start.sh |
| | RUN chmod +x /start.sh |
| |
|
| | EXPOSE 7860 |
| | CMD ["/start.sh"] |