No module named 'dropout_layer_norm'.

#10
by hassanjbara - opened

When loading the model as recommended, I'm getting the following error message: Flash Attention or Flash Attention Submodules not found, consider installing for better performance: No module named 'dropout_layer_norm'.
I do have flash-attn 2.5.7 and transformers 4.41-dev. By the way, transformers 4.40 is out, so I guess I don't need to be on the dev branch, right? Despite what is suggested in the model card.

You are missing some of the flash submodules: https://github.com/Dao-AILab/flash-attention/tree/main/csrc/layer_norm
The minimum version for the example code to work is version 4.38.

Microsoft org

If you clone the flash attention repo, navigate to csrc/layer_norm (https://github.com/Dao-AILab/flash-attention/tree/main/csrc/layer_norm), then do pip install ., it will install the dropout_layer_norm module and you should see better performance.

Thanks for the help! It would be nice if this is mentioned in the model card.

hassanjbara changed discussion status to closed

Sign up or log in to comment