| APP := llamacpp_python |
| PORT ?= 8001 |
| CHAT_FORMAT ?= |
|
|
| include ../common/Makefile.common |
|
|
| IMAGE_NAME ?= $(REGISTRY_ORG)/$(COMPONENT)/$(APP):latest |
| IMAGE := $(REGISTRY)/$(IMAGE_NAME) |
| CUDA_IMAGE := $(REGISTRY)/$(REGISTRY_ORG)/$(COMPONENT)/$(APP)_cuda:latest |
| VULKAN_IMAGE := $(REGISTRY)/$(REGISTRY_ORG)/$(COMPONENT)/$(APP)_vulkan:latest |
|
|
| MODELS_PATH := /locallm/models |
| MODEL_NAME ?= granite-7b-lab-Q4_K_M.gguf |
|
|
| .Phony: all |
| all: build download-model-granite run |
|
|
| .PHONY: build-cuda |
| build-cuda: |
| "${CONTAINER_TOOL}" build --squash-all -t $(CUDA_IMAGE) . -f cuda/Containerfile |
|
|
| .PHONY: build-vulkan-amd64 build-vulkan-arm64 |
| build-vulkan-amd64: |
| "${CONTAINER_TOOL}" build --squash-all -t $(VULKAN_IMAGE) . -f vulkan/amd64/Containerfile |
| build-vulkan-arm64: |
| "${CONTAINER_TOOL}" build --squash-all -t $(VULKAN_IMAGE) . -f vulkan/arm64/Containerfile |
|
|
| .PHONY: download-model-granite # default model |
| download-model-granite: |
| cd ../../models/ && \ |
| make download-model-granite |
|
|