Skip to content

Conversation

@Isotr0py
Copy link
Member

@Isotr0py Isotr0py commented Mar 22, 2025

Signed-off-by: Isotr0py <2037008807@qq.com>
@github-actions
Copy link

👋 Hi! Thank you for contributing to the vLLM project.

💬 Join our developer Slack at https://slack.vllm.ai to discuss your PR in #pr-reviews, coordinate on features in #feat- channels, or join special interest groups in #sig- channels.

Just a reminder: PRs would not trigger full CI run by default. Instead, it would only run fastcheck CI which starts running only a small and essential subset of CI tests to quickly catch errors. You can run other CI tests on top of those by going to your fastcheck build on Buildkite UI (linked in the PR checks section) and unblock them. If you do not have permission to unblock, ping simon-mo or khluu to add you in our Buildkite org.

Once the PR is approved and ready to go, your PR reviewer(s) can run CI to test the changes comprehensively before merging.

To run CI, PR reviewers can either: Add ready label to the PR or enable auto-merge.

🚀

Signed-off-by: Isotr0py <2037008807@qq.com>
@Isotr0py
Copy link
Member Author

@gshtras Can you check if this PR can fix the Mllama FP8 regression issue? I can generate reasonable outputs with neuralmagic/Llama-3.2-11B-Vision-Instruct-FP8-dynamic, but not amd/Llama-3.2-11B-Vision-Instruct-FP8-KV.

I'm not sure if it's due to hardware difference or something other, because I can't generate reasonable outputs from this checkpoint with an earlier commit (ec79b67) before QKVCrossParallelLinear introduction as well.

Signed-off-by: Isotr0py <2037008807@qq.com>
@Isotr0py Isotr0py marked this pull request as ready for review March 24, 2025 05:57
@NickLucche
Copy link
Collaborator

Thanks for the fix!

@gshtras
Copy link
Collaborator

gshtras commented Mar 24, 2025

@gshtras Can you check if this PR can fix the Mllama FP8 regression issue? I can generate reasonable outputs with neuralmagic/Llama-3.2-11B-Vision-Instruct-FP8-dynamic, but not amd/Llama-3.2-11B-Vision-Instruct-FP8-KV.

I'm not sure if it's due to hardware difference or something other, because I can't generate reasonable outputs from this checkpoint with an earlier commit (ec79b67) before QKVCrossParallelLinear introduction as well.

This fixes the weight loading, but there are also runtime issues in attention now, since 77a318b
Not yet sure if it is related, or a new issue

@Isotr0py
Copy link
Member Author

there are also runtime issues in attention now, since 77a318b

Hmmm, that's weird. I can run this model without any errors, but the problem is that the output hidden states will contain NaN, causing model outputs all "!".

@gshtras
Copy link
Collaborator

gshtras commented Mar 24, 2025

there are also runtime issues in attention now, since 77a318b

Hmmm, that's weird. I can run this model without any errors, but the problem is that the output hidden states will contain NaN, causing model outputs all "!".

Possibly hardware related, to clarify, I'm checking with V0 on a MI300X machine.
Applying this fix to the ROCm fork solves the issue, so there appears to be an unrelated regression, I think this PR is GTG, thanks

@gshtras
Copy link
Collaborator

gshtras commented Mar 31, 2025

there are also runtime issues in attention now, since 77a318b

Hmmm, that's weird. I can run this model without any errors, but the problem is that the output hidden states will contain NaN, causing model outputs all "!".

So that is due to an issue with the graph mode, which shouldn't be used on ROCm and is addressed in #15413

Copy link
Contributor

@SageMoore SageMoore left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I generally think this looks fine, but I'd like someone with more experience in this part of the codebase to take a look. @robertgshaw2-redhat @tlrmchlsmth @mgoin

@mgoin mgoin added ready ONLY add when PR is ready to merge/full CI is needed bug Something isn't working labels Apr 8, 2025
@vllm-bot vllm-bot merged commit 40b4284 into vllm-project:main Apr 8, 2025
43 of 45 checks passed
@Isotr0py Isotr0py deleted the fix-xqkv-quant branch April 8, 2025 17:03
yangw-dev pushed a commit to yangw-dev/vllm that referenced this pull request Apr 21, 2025
…Linear` (vllm-project#15328) Signed-off-by: Isotr0py <2037008807@qq.com> Signed-off-by: Yang Wang <elainewy@meta.com>
jikunshang pushed a commit to jikunshang/vllm that referenced this pull request Apr 29, 2025
lk-chen pushed a commit to lk-chen/vllm that referenced this pull request Apr 29, 2025
RichardoMrMu pushed a commit to RichardoMrMu/vllm that referenced this pull request May 12, 2025
…Linear` (vllm-project#15328) Signed-off-by: Isotr0py <2037008807@qq.com> Signed-off-by: Mu Huai <tianbowen.tbw@antgroup.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

bug Something isn't working force-merge ready ONLY add when PR is ready to merge/full CI is needed

6 participants