Javedalam commited on
Commit
fad24d6
·
1 Parent(s): bc59401

Switch to TinyLlama 1.1B Chat (llama.cpp-supported)

Browse files
Files changed (1) hide show
  1. Dockerfile +2 -2
Dockerfile CHANGED
@@ -7,10 +7,10 @@ RUN set -eux; \
7
  elif command -v apt-get >/dev/null 2>&1; then apt-get update && apt-get install -y --no-install-recommends curl ca-certificates && rm -rf /var/lib/apt/lists/*; \
8
  else echo "no supported pkg manager" && exit 1; fi
9
 
10
- # bake the Liquid AI GGUF at build time (fail build if download fails)
11
  RUN mkdir -p /models && \
12
  curl -fL --retry 5 --retry-delay 2 -o /models/model.gguf \
13
- "https://huggingface.co/LiquidAI/LFM2-350M-GGUF/resolve/main/LFM2-350M-Q4_K_M.gguf?download=true"
14
 
15
  EXPOSE 7860
16
  # base image ENTRYPOINT is ["llama-server"]; pass only args:
 
7
  elif command -v apt-get >/dev/null 2>&1; then apt-get update && apt-get install -y --no-install-recommends curl ca-certificates && rm -rf /var/lib/apt/lists/*; \
8
  else echo "no supported pkg manager" && exit 1; fi
9
 
10
+ # bake tiny model into the image (fail build if download fails)
11
  RUN mkdir -p /models && \
12
  curl -fL --retry 5 --retry-delay 2 -o /models/model.gguf \
13
+ "https://huggingface.co/TheBloke/TinyLlama-1.1B-Chat-v1.0-GGUF/resolve/main/tinyllama-1.1b-chat-v1.0.Q4_K_M.gguf?download=true"
14
 
15
  EXPOSE 7860
16
  # base image ENTRYPOINT is ["llama-server"]; pass only args: