8000 [CANN]: add the basic supports of Flash Attention kernel by shibizhao · Pull Request #13627 · ggml-org/llama.cpp · GitHub
[go: up one dir, main page]

Skip to content

[CANN]: add the basic supports of Flash Attention kernel#13627

Merged
hipudding merged 18 commits intoggml-org:masterfrom
shibizhao:flash-attn-cann
May 26, 2025

Commits

Commits on May 14, 2025

Commits on May 15, 2025

Commits on May 18, 2025

Commits on May 19, 2025

Commits on May 21, 2025

Commits on May 22, 2025

Commits on May 23, 2025

0