diff --git a/.gitignore b/.gitignore index b537ac6b..40fef6b7 100644 --- a/.gitignore +++ b/.gitignore @@ -2,8 +2,6 @@ model-runner model-runner.sock docker-model -# Default MODELS_PATH in Makefile -models-store/ # Directory where we store the updated llama.cpp updated-inference/ vendor/ diff --git a/Makefile b/Makefile index 4d9435cc..a5dcd1e0 100644 --- a/Makefile +++ b/Makefile @@ -11,7 +11,6 @@ DOCKER_IMAGE_SGLANG := docker/model-runner:latest-sglang DOCKER_IMAGE_DIFFUSERS := docker/model-runner:latest-diffusers DOCKER_TARGET ?= final-llamacpp PORT := 8080 -MODELS_PATH := $(shell pwd)/models-store LLAMA_ARGS ?= DOCKER_BUILD_ARGS := \ --load \ @@ -63,7 +62,6 @@ run: build clean: rm -f $(APP_NAME) rm -f model-runner.sock - rm -rf $(MODELS_PATH) # Run tests test: @@ -164,12 +162,11 @@ docker-run-diffusers: docker-build-diffusers # Common implementation for running Docker container docker-run-impl: @echo "" - @echo "Starting service on port $(PORT) with model storage at $(MODELS_PATH)..." + @echo "Starting service on port $(PORT)..." @echo "Service will be available at: http://localhost:$(PORT)" @echo "Example usage: curl http://localhost:$(PORT)/models" @echo "" PORT="$(PORT)" \ - MODELS_PATH="$(MODELS_PATH)" \ DOCKER_IMAGE="$(DOCKER_IMAGE)" \ LLAMA_ARGS="$(LLAMA_ARGS)" \ DMR_ORIGINS="$(DMR_ORIGINS)" \ diff --git a/cmd/cli/Dockerfile b/cmd/cli/Dockerfile index 1509b74c..be4eca65 100644 --- a/cmd/cli/Dockerfile +++ b/cmd/cli/Dockerfile @@ -22,7 +22,7 @@ ARG DOCS_FORMATS RUN --mount=target=/context \ --mount=target=.,type=tmpfs <&2