8000 Update n_batch for server · MohamedEmad300/llama-cpp-python@3cab3ef · GitHub
[go: up one dir, main page]

Skip to content

Commit 3cab3ef

Browse files
committed
Update n_batch for server
1 parent cc706fb commit 3cab3ef

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

llama_cpp/server/__main__.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -28,7 +28,7 @@
2828
class Settings(BaseSettings):
2929
model: str
3030
n_ctx: int = 2048
31-
n_batch: int = 8
31+
n_batch: int = 512
3232
n_threads: int = max((os.cpu_count() or 2) // 2, 1)
3333
f16_kv: bool = True
3434
use_mlock: bool = False # This causes a silent failure on platforms that don't support mlock (e.g. Windows) took forever to figure out...

0 commit comments

Comments
 (0)
0