8000
We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 8d9560e commit c928e0aCopy full SHA for c928e0a
llama_cpp/llama_cpp.py
@@ -47,7 +47,6 @@ class llama_context_params(Structure):
47
("n_ctx", c_int), # text context
48
("n_parts", c_int), # -1 for default
49
("seed", c_int), # RNG seed, 0 for random
50
-
51
("f16_kv", c_bool), # use fp16 for KV cache
52
(
53
"logits_all",
@@ -56,7 +55,6 @@ class llama_context_params(Structure):
56
55
("vocab_only", c_bool), # only load the vocabulary, no weights
57
("use_mlock", c_bool), # force system to keep model in RAM
58
("embedding", c_bool), # embedding mode only
59
60
# called with a progress value between 0 and 1, pass NULL to disable
61
("progress_callback", llama_progress_callback),
62
# context pointer passed to the progress callback
0 commit comments