Skip to content

Releases: vllm-project/flash-attention

v2.6.2

05 Sep 16:09
Compare
Choose a tag to compare
v2.6.2 Pre-release
Pre-release

What's Changed

  • Fix a wrong reference to seqlen_k variable in the varlen kernel by @cakeng in #18
  • Fix ima for split-kv kernel by @bfontain in #20

New Contributors

Full Changelog: v2.6.1...v2.6.2

v2.6.1

01 Aug 04:09
f9d2c10
Compare
Choose a tag to compare
v2.6.1 Pre-release
Pre-release

What's Changed

New Contributors

  • @mgoin made their first contribution in #10

Full Changelog: v2.6.0...v2.6.1

v2.6.0

29 Jul 19:09
Compare
Choose a tag to compare
v2.6.0 Pre-release
Pre-release

What's Changed

New Contributors

Full Changelog: v2.5.9...v2.6.0

v2.5.9.post1

07 Jun 01:54
537f75e
Compare
Choose a tag to compare
v2.5.9.post1 Pre-release
Pre-release

What's Changed

Full Changelog: v2.5.9...v2.5.9.post1

v2.5.9

31 May 17:10
Compare
Choose a tag to compare
v2.5.9 Pre-release
Pre-release

What's Changed

  • Fix out kwarg shape check with ngroups swapped by @Yard1 in #4

Full Changelog: v2.5.8.post3...v2.5.9

v2.5.8.post3

27 May 23:57
03bf1f8
Compare
Choose a tag to compare
v2.5.8.post3 Pre-release
Pre-release

What's Changed

  • Expose out in python API by @Yard1 in #2
  • Don't use kwargs in autograd functions by @Yard1 in #3

New Contributors

  • @Yard1 made their first contribution in #2

Full Changelog: v2.5.8.post2...v2.5.8.post3

v2.5.8.post2

19 May 09:57
Compare
Choose a tag to compare
v2.5.8.post2 Pre-release
Pre-release

v2.5.8.post1

07 May 01:48
Compare
Choose a tag to compare
v2.5.8.post1 Pre-release
Pre-release

What's Changed

New Contributors

Full Changelog: https://github.com/vllm-project/flash-attention/commits/v2.5.8.post1