@@ -61,8 +61,8 @@ The llama.cpp CANN backend is designed to support Ascend NPU. It utilize the abi
61
61
| Llama-2 | √ | √ | √ |
62
62
| Llama-3 | √ | √ | √ |
63
63
| Mistral-7B | √ | √ | √ |
64
- | Mistral MOE | x | x | x |
65
- | DBRX | x | x | x |
64
+ | Mistral MOE | √ | √ | √ |
65
+ | DBRX | ? | ? | ? |
66
66
| Falcon | √ | √ | √ |
67
67
| Chinese LLaMA/Alpaca | √ | √ | √ |
68
68
| Vigogne(French) | √ | √ | √ |
@@ -80,7 +80,7 @@ The llama.cpp CANN backend is designed to support Ascend NPU. It utilize the abi
80
80
| Qwen models | √ | √ | √ |
81
81
| PLaMo-13B | √ | √ | √ |
82
82
| Phi models | √ | √ | √ |
83
- | PhiMoE | x | x | x |
83
+ | PhiMoE | ? | ?
10000
| ? |
84
84
| GPT-2 | √ | √ | √ |
85
85
| Orion | √ | √ | √ |
86
86
| InternlLM2 | √ | √ | √ |
@@ -89,45 +89,46 @@ The llama.cpp CANN backend is designed to support Ascend NPU. It utilize the abi
89
89
| Mamba | √ | √ | √ |
90
90
| Xverse | √ | √ | √ |
91
91
| command-r models | √ | √ | √ |
92
- | Grok-1 | x | x | x |
92
+ | Grok-1 | ? | ? | ? |
93
93
| SEA-LION | √ | √ | √ |
94
94
| GritLM-7B | √ | √ | √ |
95
95
| OLMo | √ | √ | √ |
96
96
| OLMo 2 | √ | √ | √ |
97
- | OLMoE | x | x | x |
97
+ | OLMoE | ? | ? | ? |
98
98
| Granite models | √ | √ | √ |
99
- | GPT-NeoX + Pythia | x | x | x |
100
- | Snowflake-Arctic MoE | x | × | × |
99
+ | GPT-NeoX | ? | ? | ? |
100
+ | Pythia | √ | √ | √ |
101
+ | Snowflake-Arctic MoE | ? | ? | ? |
101
102
| Smaug | √ | √ | √ |
102
103
| Poro 34B | √ | √ | √ |
103
- | Bitnet b1.58 models | √ | √ | √ |
104
+ | Bitnet b1.58 models | √ | x | x |
104
105
| Flan-T5 | √ | √ | √ |
105
- | Open Elm models | √ | √ | √ |
106
+ | Open Elm models | x | x | x |
106
107
| chatGLM3-6B + ChatGLM4-9b + GLMEdge-1.5b + GLMEdge-4b | √ | √ | √ |
107
108
| GLM-4-0414 | √ | √ | √ |
108
109
| SmolLM | √ | √ | √ |
109
110
| EXAONE-3.0-7.8B-Instruct | √ | √ | √ |
110
111
| FalconMamba Models | √ | √ | √ |
111
- | Jais Models | x | x | x |
112
+ | Jais Models | ? | ? | ? |
112
113
| Bielik-11B-v2.3 | √ | √ | √ |
113
- | RWKV-6 | x | x | x |
114
- | QRWKV-6 | x | x | x |
114
+ | RWKV-6 | √ | √ | √ |
115
+ | QRWKV-6 | √ | √ | √ |
115
116
| GigaChat-20B-A3B | x | x | x |
116
117
| Trillion-7B-preview | √ | √ | √ |
117
118
<
CEC3
/td>| Ling models | √ | √ | √ |
118
119
119
120
120
121
** Multimodal**
121
- | LLaVA 1.5 models, LLaVA 1.6 models | √ | √ | √ |
122
- | BakLLaVA | x | x | x |
123
- | Obsidian | x | x | x |
124
- | ShareGPT4V | x | x | x |
125
- | MobileVLM 1.7B/3B models | x | x | x |
126
- | Yi-VL | x | x | x |
122
+ | LLaVA 1.5 models, LLaVA 1.6 models | ? | ? | ? |
123
+ | BakLLaVA | ? | ? | ? |
124
+ | Obsidian | ? | ? | ? |
125
+ | ShareGPT4V | ? | ? | ? |
126
+ | MobileVLM 1.7B/3B models | ? | ? | ? |
127
+ | Yi-VL | ? | ? | ? |
127
128
| Mini CPM | √ | √ | √ |
128
129
| Moondream | √ | √ | √ |
129
- | Bunny | x | x | x |
130
- | GLM-EDGE | x | x | x |
130
+ | Bunny | ? | ? | ? |
131
+ | GLM-EDGE | √ | √ | √ |
131
132
| Qwen2-VL | √ | √ | √ |
132
133
133
134
0 commit comments