8000 feat: Update llama.cpp · coderonion/llama-cpp-python@5585f8a · GitHub
[go: up one dir, main page]

Skip to content

Commit 5585f8a

Browse files
committed
feat: Update llama.cpp
1 parent b9b50e5 commit 5585f8a

File tree

2 files changed

+7
-7
lines changed

2 files changed

+7
-7
lines changed

llama_cpp/llama_cpp.py

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -344,9 +344,9 @@
344344
# LLAMA_FTYPE_MOSTLY_IQ4_XS = 30, // except 1d tensors
345345
# LLAMA_FTYPE_MOSTLY_IQ1_M = 31, // except 1d tensors
346346
# LLAMA_FTYPE_MOSTLY_BF16 = 32, // except 1d tensors
347-
# LLAMA_FTYPE_MOSTLY_Q4_0_4_4 = 33, // except 1d tensors
348-
# LLAMA_FTYPE_MOSTLY_Q4_0_4_8 = 34, // except 1d tensors
349-
# LLAMA_FTYPE_MOSTLY_Q4_0_8_8 = 35, // except 1d tensors
347+
# //LLAMA_FTYPE_MOSTLY_Q4_0_4_4 = 33, // removed from gguf files, use Q4_0 and runtime repack
348+
# //LLAMA_FTYPE_MOSTLY_Q4_0_4_8 = 34, // removed from gguf files, use Q4_0 and runtime repack
349+
# //LLAMA_FTYPE_MOSTLY_Q4_0_8_8 = 35, // removed from gguf files, use Q4_0 and runtime repack
350350
# LLAMA_FTYPE_MOSTLY_TQ1_0 = 36, // except 1d tensors
351351
# LLAMA_FTYPE_MOSTLY_TQ2_0 = 37, // except 1d tensors
352352
#
@@ -382,9 +382,9 @@
382382
LLAMA_FTYPE_MOSTLY_IQ4_XS = 30
383383
LLAMA_FTYPE_MOSTLY_IQ1_M = 31
384384
LLAMA_FTYPE_MOSTLY_BF16 = 32
385-
LLAMA_FTYPE_MOSTLY_Q4_0_4_4 = 33
386-
LLAMA_FTYPE_MOSTLY_Q4_0_4_8 = 34
387-
LLAMA_FTYPE_MOSTLY_Q4_0_8_8 = 35
385+
# LLAMA_FTYPE_MOSTLY_Q4_0_4_4 = 33
386+
# LLAMA_FTYPE_MOSTLY_Q4_0_4_8 = 34
387+
# LLAMA_FTYPE_MOSTLY_Q4_0_8_8 = 35
388388
LLAMA_FTYPE_MOSTLY_TQ1_0 = 36
389389
LLAMA_FTYPE_MOSTLY_TQ2_0 = 37
390390
LLAMA_FTYPE_GUESSED = 1024

vendor/llama.cpp

0 commit comments

Comments
 (0)
0