File tree Expand file tree Collapse file tree 1 file changed +6
-6
lines changed Expand file tree Collapse file tree 1 file changed +6
-6
lines changed Original file line number Diff line number Diff line change @@ -38,19 +38,19 @@ llama-cpp-python 0.1.68
38
38
39
39
```
40
40
41
- ** (5) Download a v3 ggml model**
42
- - ** ggmlv3 **
43
- - file name ends with ** q4_0.bin ** - indicating it is 4bit quantized, with quantisation method 0
41
+ ** (5) Download a v3 gguf v2 model**
42
+ - ** ggufv2 **
43
+ - file name ends with ** Q4_0.gguf ** - indicating it is 4bit quantized, with quantisation method 0
44
44
45
- https://huggingface.co/TheBloke/open-llama-7b-open-instruct-GGML
45
+ https://huggingface.co/TheBloke/CodeLlama-7B-GGUF
46
46
47
47
48
48
** (6) run the llama-cpp-python API server with MacOS Metal GPU support**
49
49
```
50
50
# config your ggml model path
51
- # make sure it is ggml v3
51
+ # make sure it is gguf v2
52
52
# make sure it is q4_0
53
- export MODEL=[path to your llama.cpp ggml models]]/[ggml-model-name]]q4_0.bin
53
+ export MODEL=[path to your llama.cpp ggml models]]/[ggml-model-name]]Q4_0.gguf
54
54
python3 -m llama_cpp.server --model $MODEL --n_gpu_layers 1
55
55
```
56
56
You can’t perform that action at this time.
0 commit comments