8000 Update llama.cpp · coderonion/llama-cpp-python@011cd84 · GitHub
[go: up one dir, main page]

Skip to content

Commit 011cd84

Browse files
committed
Update llama.cpp
1 parent da003d8 commit 011cd84

File tree

3 files changed

+9
-1
lines changed

3 files changed

+9
-1
lines changed

Makefile

Lines changed: 6 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -30,6 +30,12 @@ build.metal:
3030
build.vulkan:
3131
CMAKE_ARGS="-DLLAMA_VULKAN=on" python3 -m pip install --verbose -e .
3232

33+
build.kompute:
34+
CMAKE_ARGS="-DLLAMA_KOMPUTE=on" python3 -m pip install --verbose -e .
35+
36+
build.sycl:
37+
CMAKE_ARGS="-DLLAMA_SYCL=on" python3 -m pip install --verbose -e .
38+
3339
build.sdist:
3440
python3 -m build --sdist
3541

llama_cpp/llama_cpp.py

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -187,6 +187,7 @@ def _load_shared_library(lib_base_name: str):
187187
# LLAMA_FTYPE_MOSTLY_IQ2_XS = 20, // except 1d tensors
188188
# LLAMA_FTYPE_MOSTLY_Q2_K_S = 21, // except 1d tensors
189189
# LLAMA_FTYPE_MOSTLY_Q3_K_XS = 22, // except 1d tensors
190+
# LLAMA_FTYPE_MOSTLY_IQ3_XXS = 23, // except 1d tensors
190191

191192
# LLAMA_FTYPE_GUESSED = 1024, // not specified in the model file
192193
# };
@@ -211,6 +212,7 @@ def _load_shared_library(lib_base_name: str):
211212
LLAMA_FTYPE_MOSTLY_IQ2_XS = 20
212213
LLAMA_FTYPE_MOSTLY_Q2_K_S = 21
213214
LLAMA_FTYPE_MOSTLY_Q3_K_XS = 22
215+
LLAMA_FTYPE_MOSTLY_IQ3_XXS = 23
214216
LLAMA_FTYPE_GUESSED = 1024
215217

216218
# enum llama_rope_scaling_type {

vendor/llama.cpp

0 commit comments

Comments
 (0)
0