You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+8-2Lines changed: 8 additions & 2 deletions
Original file line number
Diff line number
Diff line change
@@ -21,7 +21,7 @@ Documentation is available at [https://llama-cpp-python.readthedocs.io/en/latest
21
21
> Starting with version 0.1.79 the model format has changed from `ggmlv3` to `gguf`. Old model files can be converted using the `convert-llama-ggmlv3-to-gguf.py` script in [`llama.cpp`](https://github.com/ggerganov/llama.cpp)
22
22
23
23
24
-
## Installation from PyPI (recommended)
24
+
## Installation from PyPI
25
25
26
26
Install from PyPI (requires a c compiler):
27
27
@@ -45,7 +45,7 @@ bash Miniforge3-MacOSX-arm64.sh
45
45
```
46
46
Otherwise, while installing it will build the llama.ccp x86 version which will be 10x slower on Apple Silicon (M1) Mac.
47
47
48
-
### Installation with OpenBLAS / cuBLAS / CLBlast / Metal
48
+
### Installation with Hardware Acceleration
49
49
50
50
`llama.cpp` supports multiple BLAS backends for faster processing.
51
51
Use the `FORCE_CMAKE=1` environment variable to force the use of `cmake` and install the pip package for the desired BLAS backend.
@@ -74,6 +74,12 @@ To install with Metal (MPS), set the `LLAMA_METAL=on` environment variable befor
0 commit comments