| | APP := llamacpp_python |
| | PORT ?= 8001 |
| | CHAT_FORMAT ?= |
| |
|
| | include ../common/Makefile.common |
| |
|
| | IMAGE_NAME ?= $(REGISTRY_ORG)/$(COMPONENT)/$(APP):latest |
| | IMAGE := $(REGISTRY)/$(IMAGE_NAME) |
| | CUDA_IMAGE := $(REGISTRY)/$(REGISTRY_ORG)/$(COMPONENT)/$(APP)_cuda:latest |
| | VULKAN_IMAGE := $(REGISTRY)/$(REGISTRY_ORG)/$(COMPONENT)/$(APP)_vulkan:latest |
| |
|
| | MODELS_PATH := /locallm/models |
| | MODEL_NAME ?= granite-7b-lab-Q4_K_M.gguf |
| |
|
| | .Phony: all |
| | all: build download-model-granite run |
| |
|
| | .PHONY: build-cuda |
| | build-cuda: |
| | "${CONTAINER_TOOL}" build --squash-all -t $(CUDA_IMAGE) . -f cuda/Containerfile |
| |
|
| | .PHONY: build-vulkan-amd64 build-vulkan-arm64 |
| | build-vulkan-amd64: |
| | "${CONTAINER_TOOL}" build --squash-all -t $(VULKAN_IMAGE) . -f vulkan/amd64/Containerfile |
| | build-vulkan-arm64: |
| | "${CONTAINER_TOOL}" build --squash-all -t $(VULKAN_IMAGE) . -f vulkan/arm64/Containerfile |
| |
|
| | .PHONY: download-model-granite # default model |
| | download-model-granite: |
| | cd ../../models/ && \ |
| | make download-model-granite |
| |
|