Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How the performance of Sage Attention compares to that of FA3 on Hopper GPUs? #59

Open
alexngng opened this issue Dec 3, 2024 · 2 comments
Labels
enhancement New feature or request

Comments

@alexngng
Copy link

alexngng commented Dec 3, 2024

Did not see that comparison in the paper.

@jt-zhang
Copy link
Member

jt-zhang commented Dec 3, 2024

SageAttention is currently designed to accelerate attention mechanisms across various GPUs, unlike FA3, which is optimized exclusively for and can only run on Hopper GPUs. We plan to optimize for Hopper in the future.

@alexngng
Copy link
Author

alexngng commented Dec 6, 2024

SageAttention is currently designed to accelerate attention mechanisms across various GPUs, unlike FA3, which is optimized exclusively for and can only run on Hopper GPUs. We plan to optimize for Hopper in the future.

Thanks for reply.

@jason-huang03 jason-huang03 added the enhancement New feature or request label Dec 11, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

3 participants