You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
feat(llama.cpp): do not specify backends to autoload and add llama.cpp variants (#2232)
* feat(initializer): do not specify backends to autoload
We can simply try to autoload the backends extracted in the asset dir.
This will allow to build variants of the same backend (for e.g. with different instructions sets),
so to have a single binary for all the variants.
Signed-off-by: mudler <[email protected]>
* refactor(prepare): refactor out llama.cpp prepare steps
Make it so are idempotent and that we can re-build
Signed-off-by: mudler <[email protected]>
* [TEST] feat(build): build noavx version along
Signed-off-by: mudler <[email protected]>
* build: make build parallel
Signed-off-by: Ettore Di Giacinto <[email protected]>
* build: do not override CMAKE_ARGS
Signed-off-by: Ettore Di Giacinto <[email protected]>
* build: add fallback variant
Signed-off-by: Ettore Di Giacinto <[email protected]>
* Fixups
Signed-off-by: Ettore Di Giacinto <[email protected]>
* fix(huggingface-langchain): fail if no token is set
Signed-off-by: Ettore Di Giacinto <[email protected]>
* fix(huggingface-langchain): rename
Signed-off-by: Ettore Di Giacinto <[email protected]>
* fix: do not autoload local-store
Signed-off-by: Ettore Di Giacinto <[email protected]>
* fix: give priority between the listed backends
Signed-off-by: Ettore Di Giacinto <[email protected]>
---------
Signed-off-by: mudler <[email protected]>
Signed-off-by: Ettore Di Giacinto <[email protected]>
Expect(err.Error()).To(ContainSubstring(fmt.Sprintf("error, status code: 500, message: could not load model - all backends returned error: %d errors occurred:", backends)))
792
+
Expect(err.Error()).To(ContainSubstring("error, status code: 500, message: could not load model - all backends returned error:"))
0 commit comments