@@ -156,6 +156,7 @@ ALL_GRPC_BACKENDS=backend-assets/grpc/langchain-huggingface
156
156
ALL_GRPC_BACKENDS+ =backend-assets/grpc/bert-embeddings
157
157
ALL_GRPC_BACKENDS+ =backend-assets/grpc/llama-cpp
158
158
ALL_GRPC_BACKENDS+ =backend-assets/grpc/llama-cpp-noavx
159
+ ALL_GRPC_BACKENDS+ =backend-assets/grpc/llama-cpp-fallback
159
160
ALL_GRPC_BACKENDS+ =backend-assets/grpc/llama-ggml
160
161
ALL_GRPC_BACKENDS+ =backend-assets/grpc/gpt4all
161
162
ALL_GRPC_BACKENDS+ =backend-assets/grpc/rwkv
@@ -294,6 +295,7 @@ clean: ## Remove build related file
294
295
rm -rf backend-assets/*
295
296
$(MAKE ) -C backend/cpp/grpc clean
296
297
$(MAKE ) -C backend/cpp/llama clean
298
+ rm -rf backend/cpp/llama-* || true
297
299
$(MAKE ) dropreplace
298
300
$(MAKE ) protogen-clean
299
301
rmdir pkg/grpc/proto || true
@@ -312,7 +314,7 @@ build: prepare backend-assets grpcs ## Build the project
312
314
CGO_LDFLAGS=" $( CGO_LDFLAGS) " $(GOCMD ) build -ldflags " $( LD_FLAGS) " -tags " $( GO_TAGS) " -o $(BINARY_NAME ) ./
313
315
314
316
build-minimal :
315
- BUILD_GRPC_FOR_BACKEND_LLAMA=true GRPC_BACKENDS=" backend-assets/grpc/llama-cpp backend-assets/grpc/llama-cpp-noavx " GO_TAGS=none $(MAKE ) build
317
+ BUILD_GRPC_FOR_BACKEND_LLAMA=true GRPC_BACKENDS=" backend-assets/grpc/llama-cpp" GO_TAGS=none $(MAKE ) build
316
318
317
319
build-api :
318
320
BUILD_GRPC_FOR_BACKEND_LLAMA=true BUILD_API_ONLY=true GO_TAGS=none $(MAKE ) build
@@ -657,22 +659,17 @@ ifeq ($(BUILD_TYPE),metal)
657
659
endif
658
660
659
661
backend-assets/grpc/llama-cpp-noavx : backend-assets/grpc
660
- $(info ${GREEN}I llama-cpp build info:noavx${RESET})
661
662
cp -rf backend/cpp/llama backend/cpp/llama-noavx
662
663
$(MAKE ) -C backend/cpp/llama-noavx purge
663
- CMAKE_ARGS+=-DLLAMA_AVX2=OFF
664
- $(MAKE ) VARIANT=" llama-noavx" build-llama-cpp-grpc-server
664
+ $( info ${GREEN}I llama-cpp build info:noavx${RESET})
665
+ CMAKE_ARGS= " $( CMAKE_ARGS ) -DLLAMA_AVX512=OFF -DLLAMA_AVX2=OFF " $(MAKE ) VARIANT=" llama-noavx" build-llama-cpp-grpc-server
665
666
cp -rfv backend/cpp/llama-noavx/grpc-server backend-assets/grpc/llama-cpp-noavx
666
667
667
668
backend-assets/grpc/llama-cpp-fallback : backend-assets/grpc
668
- $(info ${GREEN}I llama-cpp build info:fallback${RESET})
669
669
cp -rf backend/cpp/llama backend/cpp/llama-fallback
670
670
$(MAKE ) -C backend/cpp/llama-fallback purge
671
- CMAKE_ARGS+=-DLLAMA_F16C=OFF
672
- CMAKE_ARGS+=-DLLAMA_AVX512=OFF
673
- CMAKE_ARGS+=-DLLAMA_AVX2=OFF
674
- CMAKE_ARGS+=-DLLAMA_FMA=OFF
675
- $(MAKE ) VARIANT=" llama-fallback" build-llama-cpp-grpc-server
671
+ $(info ${GREEN}I llama-cpp build info:fallback${RESET})
672
+ CMAKE_ARGS=" $( CMAKE_ARGS) -DLLAMA_F16C=OFF -DLLAMA_AVX512=OFF -DLLAMA_AVX2=OFF -DLLAMA_FMA=OFF" $(MAKE ) VARIANT=" llama-fallback" build-llama-cpp-grpc-server
676
673
cp -rfv backend/cpp/llama-fallback/grpc-server backend-assets/grpc/llama-cpp-fallback
677
674
678
675
backend-assets/grpc/llama-ggml : sources/go-llama.cpp sources/go-llama.cpp/libbinding.a backend-assets/grpc
0 commit comments