build error

Job failed with exit code: 1. Reason: cache miss: [builder 14/18] RUN npm install cache miss: [runtime 4/9] RUN mkdir -p /data && chmod 777 /data cache miss: [builder 4/18] COPY --chown=user . /home/user/app cache miss: [builder 13/18] WORKDIR /home/user/llama.cpp/examples/server/webui cache miss: [builder 9/18] RUN git clone https://github.com/ggml-org/llama.cpp.git /home/user/llama.cpp cache miss: [builder 17/18] RUN cmake -B build -DBUILD_SHARED_LIBS=OFF -DCMAKE_BUILD_TYPE=Release cache miss: [builder 7/18] RUN mkdir -p /data && chmod 777 /data cache miss: [runtime 2/9] RUN useradd -m -u 1000 user cache miss: [builder 2/18] RUN useradd -m -u 1000 user cache miss: [runtime 9/9] WORKDIR /home/user/llama.cpp cache miss: resolve image config for docker-image://docker.io/docker/dockerfile:1.6 cache miss: docker-image://docker.io/docker/dockerfile:1.6@sha256:ac85f380a63b13dfcefa89046420e1781752bab202122f8f50032edf31be0021 cache miss: [builder 15/18] RUN npm run build cache miss: [builder 12/18] RUN git apply /home/user/app/helloworld.patch cache miss: [builder 5/18] RUN apt-get update && apt-get install -y --no-install-recommends git cmake build-essential g++ wget curl ca-certificates python3 patch pkg-config libcurl4-openssl-dev && rm -rf /var/lib/apt/lists/* cache miss: [builder 11/18] RUN git checkout b34443923 cache miss: [builder 10/18] WORKDIR /home/user/llama.cpp cache miss: [builder 6/18] RUN curl -fsSL https://deb.nodesource.com/setup_20.x | bash - && apt-get update && apt-get install -y --no-install-recommends nodejs && rm -rf /var/lib/apt/lists/* cache miss: [builder 18/18] RUN cmake --build build --config Release -j $(nproc) -t llama-server cache miss: [builder 16/18] WORKDIR /home/user/llama.cpp cache miss: [runtime 5/9] RUN mkdir -p /home/user/llama.cpp cache miss: [runtime 8/9] RUN cat > /usr/local/bin/start.sh << 'EOF'\nset -euo pipefail\n\nmkdir -p \"${MODEL_DIR}\"\n\nURL_NO_QUERY=\"${MODEL_DOWNLOAD_LINK%%\\?*}\"\nFNAME=\"$(basename \"${URL_NO_QUERY}\")\"\nMODEL_PATH=\"${MODEL_DIR}/${FNAME}\"\n\nif [[ ! -f \"${MODEL_PATH}\" ]]; then\n echo \"Model not found at ${MODEL_PATH}\"\n echo \"Downloading: ${MODEL_DOWNLOAD_LINK}\"\n tmp=\"${MODEL_PATH}.tmp\"\n rm -f \"${tmp}\"\n curl -L --fail --retry 5 --retry-all-errors --connect-timeout 30 \\\n -o \"${tmp}\" \"${MODEL_DOWNLOAD_LINK}\"\n mv \"${tmp}\" \"${MODEL_PATH}\"\n echo \"Downloaded model to ${MODEL_PATH}\"\nelse\n echo \"Using cached model: ${MODEL_PATH}\"\nfi\n\ncd /home/user/llama.cpp\n\nexec /usr/local/bin/llama-server \\\n --host \"${LLAMA_HOST}\" \\\n --port \"${LLAMA_PORT}\" \\\n --path \"/home/user/llama.cpp/examples/server\" \\\n -m \"${MODEL_PATH}\" \\\n ${LLAMA_EXTRA_ARGS}\nEOF\n && chmod +x /usr/local/bin/start.sh cache miss: [runtime 3/9] RUN apt-get update && apt-get install -y --no-install-recommends ca-certificates curl libstdc++6 libgomp1 && rm -rf /var/lib/apt/lists/* cache miss: [builder 8/18] RUN python3 replace_hw.py cache miss: [runtime 7/9] COPY --from=builder /home/user/llama.cpp/examples/server /home/user/llama.cpp/examples/server cache miss: [runtime 6/9] COPY --from=builder /home/user/llama.cpp/build/bin/llama-server /usr/local/bin/llama-server cache miss: [builder 3/18] WORKDIR /home/user/app {"total":33,"completed":32,"user_total":28,"user_cached":0,"user_completed":27,"user_cacheable":27,"from":1,"miss":27,"client_duration_ms":240301}

Build logs:

Fetching error logs...