8000
We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent b5fca91 commit 85d3374Copy full SHA for 85d3374
llama_cpp/server/model.py
@@ -6,6 +6,7 @@
6
7
import llama_cpp
8
import llama_cpp.llama_speculative as llama_speculative
9
+import llama_cpp.llama_tokenizer as llama_tokenizer
10
11
from llama_cpp.server.settings import ModelSettings
12
@@ -95,7 +96,7 @@ def load_llama_from_model_settings(settings: ModelSettings) -> llama_cpp.Llama:
95
96
97
tokenizer: Optional[llama_cpp.BaseLlamaTokenizer] = None
98
if settings.hf_pretrained_model_name_or_path is not None:
- tokenizer = llama_cpp.LlamaHFTokenizer.from_pretrained(settings.hf_pretrained_model_name_or_path)
99
+ tokenizer = llama_tokenizer.LlamaHFTokenizer.from_pretrained(settings.hf_pretrained_model_name_or_path)
100
101
draft_model = None
102
if settings.draft_model is not None:
0 commit comments