Skip to content

Conversation

@cpersson-amd
Copy link

This PR implements the following:

  • TransformerEngine flash attention for WAN training and inference.
  • A new fsdp sharding parallelism optimized for use on GPUs.
  • Some minor changes to allow for training on flax version 0.11.2.

The code has been tested on WAN 2.1 (training and inference) and flux (only training) using GPUs.

@google-cla
Copy link

google-cla bot commented Dec 16, 2025

Thanks for your pull request! It looks like this may be your first contribution to a Google open source project. Before we can look at your pull request, you'll need to sign a Contributor License Agreement (CLA).

View this failed invocation of the CLA check for more information.

For the most up to date status, view the checks section at the bottom of the pull request.

@cpersson-amd cpersson-amd marked this pull request as draft December 17, 2025 00:18
@cpersson-amd cpersson-amd marked this pull request as ready for review December 17, 2025 10:21
@cpersson-amd cpersson-amd reopened this Dec 17, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant