8000 llama_cpp server: add missing top_k param to CreateChatCompletionRequest · Stonelinks/llama-cpp-python@a5aa6c1 · GitHub
[go: up one dir, main page]

Skip to content

Commit a5aa6c1

Browse files
committed
llama_cpp server: add missing top_k param to CreateChatCompletionRequest
`llama.create_chat_completion` definitely has a `top_k` argument, but its missing from `CreateChatCompletionRequest`. decision: add it
1 parent 1e42913 commit a5aa6c1

File tree

1 file changed

+1
-0
lines changed

1 file changed

+1
-0
lines changed

llama_cpp/server/app.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -169,6 +169,7 @@ class CreateChatCompletionRequest(BaseModel):
169169
model: str = model_field
170170

171171
# llama.cpp specific parameters
172+
top_k: int = 40,
172173
repeat_penalty: float = 1.1
173174

174175
class Config:

0 commit comments

Comments
 (0)
0