Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Do you plan to integrate this algorithm into the vllm project? #15

Open
Alienfeel opened this issue Oct 21, 2024 · 2 comments
Open

Do you plan to integrate this algorithm into the vllm project? #15

Alienfeel opened this issue Oct 21, 2024 · 2 comments

Comments

@Alienfeel
Copy link

RT

@RyeYuan
Copy link

RyeYuan commented Nov 19, 2024

I had tried to use the sageAttention in prefilling phase of vllm based on latest 2.0.0 branch, however, after replacing fa2, it doesn't seem to have much effect, and the end-to-end throughput performance (tokens/s) remains almost the same

@jason-huang03
Copy link
Member

We are discussing with sglang team, and there is possibility that sageattention will be used in sglang in the future.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants