Skip to content

Releases: nshepperd/flash_attn_jax

v0.2.2

24 May 10:57
Compare
Choose a tag to compare
Build cuda12 version with Hopper support.

v0.2.1

21 May 12:29
Compare
Choose a tag to compare
Bump minor version to 0.2.1.

v0.2.0

01 May 23:46
Compare
Choose a tag to compare
Expanded vmap support for flash_mha. Vmapping q but not k,v reduces t…

…o a grouped-query attention, which we now support.

v0.1.0a3

18 Mar 14:24
Compare
Choose a tag to compare
Try cibuildwheel.

v0.1.0a2

18 Mar 14:22
Compare
Choose a tag to compare
Try cibuildwheel.

v0.1.0a1

17 Mar 14:13
Compare
Choose a tag to compare
Set CUDA_HOME for the sdist

v0.1.0

17 Mar 12:16
Compare
Choose a tag to compare
Try release with github actions and rebased repo.

v2.5.5a2

28 Feb 16:52
Compare
Choose a tag to compare
Implement ring attention backward pass. More tests.

v2.5.5a1

27 Feb 07:09
Compare
Choose a tag to compare
Merge up to v2.5.5

v2.5.0a4

21 Feb 00:12
Compare
Choose a tag to compare
Implement custom sharding for flash_mha, to allow efficient multi-gpu…

… computation when sharded across batch or head dimensions.