8000 Bugfix: enable embeddings for fastapi server · coderonion/llama-cpp-python@f28bf3f · GitHub
[go: up one dir, main page]

Skip to content
< 8000 script crossorigin="anonymous" type="application/javascript" src="https://github.githubassets.com/assets/sessions-eed3aa0554dd.js" defer="defer">

Commit f28bf3f

Browse files
committed
Bugfix: enable embeddings for fastapi server
1 parent c25b7df commit f28bf3f

File tree

1 file changed

+4
-1
lines changed

1 file changed

+4
-1
lines changed

examples/fastapi_server.py

Lines changed: 4 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -31,6 +31,7 @@ class Settings(BaseSettings):
3131
settings.model,
3232
f16_kv=True,
3333
use_mlock=True,
34+
embedding=True,
3435
n_threads=6,
3536
n_batch=2048,
3637
)
@@ -93,4 +94,6 @@ class Config:
9394
response_model=CreateEmbeddingResponse,
9495
)
9596
def create_embedding(request: CreateEmbeddingRequest):
96-
return llama.create_embedding(**request.dict())
97+
# print(request)
98+
# return llama.create_embedding(**request.dict(exclude={"model", "user"}))
99+
return llama.create_embedding(request.input)

0 commit comments

Comments
 (0)
0