#machine-learning #tensor

candle-flash-attn-v1

Flash attention V1 layer for the candle ML framework

1 unstable release

0.0.1 Apr 15, 2025

#211 in #tensor

Download history 47/week @ 2025-07-15 68/week @ 2025-07-22 36/week @ 2025-07-29 54/week @ 2025-08-05 37/week @ 2025-08-12 60/week @ 2025-08-19 41/week @ 2025-08-26 58/week @ 2025-09-02 51/week @ 2025-09-09 36/week @ 2025-09-16 69/week @ 2025-09-23 52/week @ 2025-09-30 24/week @ 2025-10-07 34/week @ 2025-10-14 43/week @ 2025-10-21 38/week @ 2025-10-28

149 downloads per month

MIT/Apache

2MB
849 lines

Candle Flash Attention v1 Layer

Flash Attention v2 does not support Turing GPUs (T4, RTX 2080). This layer can be used in replacement of the official flash attention Candle layer in the meantime.

Dependencies

~21MB
~469K SLoC