Closed
Description
The main llama.cpp has been updated to support GPUs on Mac's with the following flag (tested on my system):
LLAMA_METAL=1 make -j && ./main -m /Downloads/guanaco-65B.ggmlv3.q4_0.bin -p "I believe the meaning of life is" --ignore-eos -n 64 -ngl 1
It look like the following flag needs to be added to CMake options:
CMAKE_ARGS="LLAMA_METAL=1" FORCE_CMAKE=1 pip install -e .
While it appears that it installs successfully, the library cannot be loaded.
>>> from llama_cpp import Llama, LlamaCache
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "/opt/homebrew/Caskroom/mambaforge/base/lib/python3.10/site-packages/llama_cpp/__init__.py", line 1, in <module>
from .llama_cpp import *
File "/opt/homebrew/Caskroom/mambaforge/base/lib/python3.10/site-packages/llama_cpp/llama_cpp.py", line 73, in <module>
_lib = _load_shared_library(_lib_base_name)
File "/opt/homebrew/Caskroom/mambaforge/base/lib/python3.10/site-packages/llama_cpp/llama_cpp.py", line 64, in _load_shared_library
raise FileNotFoundError(
FileNotFoundError: Shared library with base name 'llama' not found
This happens regardless of whether the GitHub repo or PyPy are used.