|
|
APP := llamacpp_python |
|
|
PORT ?= 8001 |
|
|
CHAT_FORMAT ?= |
|
|
|
|
|
include ../common/Makefile.common |
|
|
|
|
|
IMAGE_NAME ?= $(REGISTRY_ORG)/$(COMPONENT)/$(APP):latest |
|
|
IMAGE := $(REGISTRY)/$(IMAGE_NAME) |
|
|
CUDA_IMAGE := $(REGISTRY)/$(REGISTRY_ORG)/$(COMPONENT)/$(APP)_cuda:latest |
|
|
VULKAN_IMAGE := $(REGISTRY)/$(REGISTRY_ORG)/$(COMPONENT)/$(APP)_vulkan:latest |
|
|
|
|
|
MODELS_PATH := /locallm/models |
|
|
MODEL_NAME ?= granite-7b-lab-Q4_K_M.gguf |
|
|
|
|
|
.Phony: all |
|
|
all: build download-model-granite run |
|
|
|
|
|
.PHONY: build-cuda |
|
|
build-cuda: |
|
|
"${CONTAINER_TOOL}" build --squash-all -t $(CUDA_IMAGE) . -f cuda/Containerfile |
|
|
|
|
|
.PHONY: build-vulkan-amd64 build-vulkan-arm64 |
|
|
build-vulkan-amd64: |
|
|
"${CONTAINER_TOOL}" build --squash-all -t $(VULKAN_IMAGE) . -f vulkan/amd64/Containerfile |
|
|
build-vulkan-arm64: |
|
|
"${CONTAINER_TOOL}" build --squash-all -t $(VULKAN_IMAGE) . -f vulkan/arm64/Containerfile |
|
|
|
|
|
.PHONY: download-model-granite # default model |
|
|
download-model-granite: |
|
|
cd ../../models/ && \ |
|
|
make download-model-granite |
|
|
|