8000 clarify supported parameters better in api server · Issue #3 · Stonelinks/llama-cpp-python · GitHub
[go: up one dir, main page]

Skip to content

clarify supported parameters better in api server #3

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Open
Stonelinks opened this issue Apr 29, 2023 · 1 comment
Open

clarify supported parameters better in api server #3

Stonelinks opened this issue Apr 29, 2023 · 1 comment

Comments

@Stonelinks
Copy link
Owner

There's a comment for "ignored or currently unsupported" parameters but I'm pretty sure some of them are supported (for instance, I think logprobs is supported?).

Also when I generate an openapi client off the openapi.json there are errors because of these.

So just delete the things that are unsupported or clarify that they're actually supported / provide more info on the existing support.

@Stonelinks
Copy link
Owner Author
Stonelinks commented Apr 29, 2023

Just leaving some notes here for parameters i see in app.py

  • model: ignored, but currently marked "optional"... on the one hand could mark "required" to make it explicit in case the server supports multiple llama's at the same time, but also could delete it since its ignored. decision: mark it required for the sake of openai api compatibility
  • n, presence_penalty, frequency_penalty, best_of, logit_bias, user: not supported, excluded from the calls into llama. decision: delete it
  • logprobs (CreateCompletionRequest only): i think this is actually supported (its in the arguments of LLama.__call__, which is how the completion is invoked). decision: mark as supported
  • top_k (CreateChatCompletionRequest only): llama.create_chat_completion definitely has a top_k argument, but its missing from CreateChatCompletionRequest. decision: add it

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels 3A76
None yet
Projects
None yet
Development

No branches or pull requests

1 participant
0