8000 feat: Generic Chat Formats, Tool Calling, and Huggingface Pull Suppor… · xhedit/llama-cpp-conv@425673c · GitHub
[go: up one dir, main page]

Skip to content

Commit 425673c

Browse files
abetlenxhedit
authored andcommitted
feat: Generic Chat Formats, Tool Calling, and Huggingface Pull Support for Multimodal Models (Obsidian, LLaVA1.6, Moondream) (abetlen#1147)
* Test dummy image tags in chat templates * Format and improve types for llava_cpp.py * Add from_pretrained support to llava chat format. * Refactor llava chat format to use a jinja2 * Revert chat format test * Add moondream support (wip) * Update moondream chat format * Update moondream chat format * Update moondream prompt * Add function calling support * Cache last image embed * Add Llava1.6 support * Add nanollava support * Add obisidian support * Remove unnecessary import * Re-order multimodal chat formats * Logits all no longer required for multi-modal models * Update README.md * Update docs * Update README * Fix typo * Update README * Fix typo
1 parent bae86ef commit 425673c

File tree

4 files changed

+675
-140
lines changed

4 files changed

+675
-140
lines changed

docs/server.md

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -98,6 +98,8 @@ You'll first need to download one of the available multi-modal models in GGUF fo
9898
- [llava-v1.5-7b](https://huggingface.co/mys/ggml_llava-v1.5-7b)
9999
- [llava-v1.5-13b](https://huggingface.co/mys/ggml_llava-v1.5-13b)
100100
- [bakllava-1-7b](https://huggingface.co/mys/ggml_bakllava-1)
101+
- [llava-v1.6-34b](https://huggingface.co/cjpais/llava-v1.6-34B-gguf)
102+
- [moondream2](https://huggingface.co/vikhyatk/moondream2)
101103

102104
Then when you run the server you'll need to also specify the path to the clip model used for image embedding and the `llava-1-5` chat_format
103105

0 commit comments

Comments
 (0)
0