-
Notifications
You must be signed in to change notification settings - Fork 1.1k
Insights: abetlen/llama-cpp-python
Overview
-
0 Active pull requests
-
- 0 Merged pull requests
- 0 Open pull requests
- 0 Closed issues
- 1 New issue
There hasn’t been any commit activity on abetlen/llama-cpp-python in the last week.
Want to help out?
1 Issue opened by 1 person
-
Support for jinja for custom chat templates
#2023 opened
May 22, 2025
11 Unresolved conversations
Sometimes conversations happen on old items that aren’t yet closed. Here is a list of all the Issues and Pull Requests with unresolved conversations.
-
Failed building wheel for llama-cpp-python
#1932 commented on
May 21, 2025 • 0 new comments -
llama-server not using GPU
#1826 commented on
May 21, 2025 • 0 new comments -
How to install the latest version with GPU support
#2012 commented on
May 21, 2025 • 0 new comments -
Include usage key in create_completion when streaming
#1498 commented on
May 23, 2025 • 0 new comments -
Feature request: add support for streaming tool use
#1883 commented on
May 25, 2025 • 0 new comments -
Can't install with GPU support with Cuda toolkit 12.9 and Cuda 12.9
#2013 commented on
May 27, 2025 • 0 new comments -
pyinstaller hook script
#709 commented on
May 23, 2025 • 0 new comments -
Feat: Support Ranking Method
#1820 commented on
May 27, 2025 • 0 new comments -
feat: Add Gemma3 chat handler (#1976)
#1989 commented on
May 23, 2025 • 0 new comments -
Added support for overriding tensor buffer types
#2007 commented on
May 22, 2025 • 0 new comments -
Add support for Cohere Command models
#2018 commented on
May 23, 2025 • 0 new comments