8000 [ROCM] Properly disable Flash Attention/Efficient Attention with environment variables by xinyazhang · Pull Request #1541 · ROCm/pytorch · GitHub
[go: up one dir, main page]

Skip to content

[ROCM] Properly disable Flash Attention/Efficient Attention with environment variables #1541

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Closed
wants to merge 3 commits into from

Conversation

xinyazhang
Copy link
@xinyazhang xinyazhang commented Aug 19, 2024

Now USE_FLASH_ATTENTION=0 USE_MEM_EFF_ATTENTION=0 python setup.py can compile correctly

This is a backported version from pytorch#133866

Tested with USE_FLASH_ATTENTION=0 USE_MEM_EFF_ATTENTION=0 python setup.py develop --user and python -c 'import torch'

@pruthvistony
Copy link
Collaborator

The PR #1536 was merged. MEM_EFF_ATTENTION is always turned off, when it will be enabled?

@pruthvistony
Copy link
Collaborator

Not required.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants
0