mirror of
https://github.com/mudler/LocalAI.git
synced 2025-06-29 14:14:59 +00:00
Bump llama.cpp to support full CUDA offload
This commit is contained in:
parent
9b498dc76a
commit
a9559fa8ce
2 changed files with 4 additions and 7 deletions
|
@ -41,10 +41,6 @@ ENV NVIDIA_VISIBLE_DEVICES=all
|
|||
|
||||
WORKDIR /build
|
||||
|
||||
COPY Makefile .
|
||||
RUN make get-sources
|
||||
COPY go.mod .
|
||||
RUN make prepare
|
||||
COPY . .
|
||||
RUN make build
|
||||
|
||||
|
@ -62,8 +58,9 @@ RUN if [ "${FFMPEG}" = "true" ]; then \
|
|||
|
||||
WORKDIR /build
|
||||
|
||||
COPY . .
|
||||
RUN make prepare-sources
|
||||
COPY --from=builder /build/local-ai ./
|
||||
COPY entrypoint.sh .
|
||||
|
||||
# Define the health check command
|
||||
HEALTHCHECK --interval=1m --timeout=10m --retries=10 \
|
||||
|
|
2
Makefile
2
Makefile
|
@ -3,7 +3,7 @@ GOTEST=$(GOCMD) test
|
|||
GOVET=$(GOCMD) vet
|
||||
BINARY_NAME=local-ai
|
||||
|
||||
GOLLAMA_VERSION?=5f1620443a59c5531b5a15a16cd68f600a8437e9
|
||||
GOLLAMA_VERSION?=7ad833b67070fd3ec46d838f5e38d21111013f98
|
||||
GPT4ALL_REPO?=https://github.com/nomic-ai/gpt4all
|
||||
GPT4ALL_VERSION?=b004c53a7bba182cd4483d95ba9e1f68d8e56da3
|
||||
GOGGMLTRANSFORMERS_VERSION?=01b8436f44294d0e1267430f9eda4460458cec54
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue