v0.3.2
📣 Highlights
Starting with this release, every time we ship a new version of flash-linear-attention, we will simultaneously publish fla-core: a minimal-dependency subset of the main repo that contains only the essentials.
🧠 New Models
What's Changed
- [Conv] Provide fn interface for
causal_conv1dby @yzhangcs in #578 - [Log Linear Attention] add backward pass by @2022tgoel in #577
- [PaTH] Fix q init & dq masking by @yzhangcs in #581
- [TokenShift] Fix a bug in decoding by @zhiyuan1i in #583
- [Deps] Lock
transformers<4.56.0by @zhiyuan1i in #582 - [Log-Linear Attention] add models by @2022tgoel in #579
- [Deps] Upgrade to transformers 4.56.x by @zhiyuan1i in #587
- [Build] Split package distribution[skip test] by @zhiyuan1i in #588
Full Changelog: v0.3.1...v0.3.2