jaothan's picture
Upload 13 files
9c1ceb7 verified
raw
history blame
964 Bytes
APP := llamacpp_python
PORT ?= 8001
CHAT_FORMAT ?=
include ../common/Makefile.common
IMAGE_NAME ?= $(REGISTRY_ORG)/$(COMPONENT)/$(APP):latest
IMAGE := $(REGISTRY)/$(IMAGE_NAME)
CUDA_IMAGE := $(REGISTRY)/$(REGISTRY_ORG)/$(COMPONENT)/$(APP)_cuda:latest
VULKAN_IMAGE := $(REGISTRY)/$(REGISTRY_ORG)/$(COMPONENT)/$(APP)_vulkan:latest
MODELS_PATH := /locallm/models
MODEL_NAME ?= granite-7b-lab-Q4_K_M.gguf
.Phony: all
all: build download-model-granite run
.PHONY: build-cuda
build-cuda:
"${CONTAINER_TOOL}" build --squash-all -t $(CUDA_IMAGE) . -f cuda/Containerfile
.PHONY: build-vulkan-amd64 build-vulkan-arm64
build-vulkan-amd64:
"${CONTAINER_TOOL}" build --squash-all -t $(VULKAN_IMAGE) . -f vulkan/amd64/Containerfile
build-vulkan-arm64:
"${CONTAINER_TOOL}" build --squash-all -t $(VULKAN_IMAGE) . -f vulkan/arm64/Containerfile
.PHONY: download-model-granite # default model
download-model-granite:
cd ../../models/ && \
make download-model-granite