-
-
Notifications
You must be signed in to change notification settings - Fork 6.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[ROCm] include gfx908 as supported #2792
Conversation
@zhuohan123 would you have time to review this? |
@WoosukKwon is there anything I should be doing differently to get a review here? |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM! Thanks for the fix!
get this error in MI100: |
ROCm/flash-attention supports the
gfx908
architecture.Without this change, vLLM appears to build successfully for me, but serving an LLM on an MI100 results in gibberish output.
With this change, everything works as expected.