|
344 | 344 | # LLAMA_FTYPE_MOSTLY_IQ4_XS = 30, // except 1d tensors
|
345 | 345 | # LLAMA_FTYPE_MOSTLY_IQ1_M = 31, // except 1d tensors
|
346 | 346 | # LLAMA_FTYPE_MOSTLY_BF16 = 32, // except 1d tensors
|
347 |
| -# LLAMA_FTYPE_MOSTLY_Q4_0_4_4 = 33, // except 1d tensors |
348 |
| -# LLAMA_FTYPE_MOSTLY_Q4_0_4_8 = 34, // except 1d tensors |
349 |
| -# LLAMA_FTYPE_MOSTLY_Q4_0_8_8 = 35, // except 1d tensors |
| 347 | +# //LLAMA_FTYPE_MOSTLY_Q4_0_4_4 = 33, // removed from gguf files, use Q4_0 and runtime repack |
| 348 | +# //LLAMA_FTYPE_MOSTLY_Q4_0_4_8 = 34, // removed from gguf files, use Q4_0 and runtime repack |
| 349 | +# //LLAMA_FTYPE_MOSTLY_Q4_0_8_8 = 35, // removed from gguf files, use Q4_0 and runtime repack |
350 | 350 | # LLAMA_FTYPE_MOSTLY_TQ1_0 = 36, // except 1d tensors
|
351 | 351 | # LLAMA_FTYPE_MOSTLY_TQ2_0 = 37, // except 1d tensors
|
352 | 352 | #
|
|
382 | 382 | LLAMA_FTYPE_MOSTLY_IQ4_XS = 30
|
383 | 383 | LLAMA_FTYPE_MOSTLY_IQ1_M = 31
|
384 | 384 | LLAMA_FTYPE_MOSTLY_BF16 = 32
|
385 |
| -LLAMA_FTYPE_MOSTLY_Q4_0_4_4 = 33 |
386 |
| -LLAMA_FTYPE_MOSTLY_Q4_0_4_8 = 34 |
387 |
| -LLAMA_FTYPE_MOSTLY_Q4_0_8_8 = 35 |
| 385 | +# LLAMA_FTYPE_MOSTLY_Q4_0_4_4 = 33 |
| 386 | +# LLAMA_FTYPE_MOSTLY_Q4_0_4_8 = 34 |
| 387 | +# LLAMA_FTYPE_MOSTLY_Q4_0_8_8 = 35 |
388 | 388 | LLAMA_FTYPE_MOSTLY_TQ1_0 = 36
|
389 | 389 | LLAMA_FTYPE_MOSTLY_TQ2_0 = 37
|
390 | 390 | LLAMA_FTYPE_GUESSED = 1024
|
|
0 commit comments