Skip to content

Conversation

Aidyn-A
Copy link
Contributor

@Aidyn-A Aidyn-A commented Sep 10, 2025

Purpose

Fixes #24464 "Building vLLM with CUDA 13.0".

Test Plan

No plans. Just build.

Test Result

Test results must be the same as on CUDA 12.x.


Essential Elements of an Effective PR Description Checklist
  • The purpose of the PR, such as "Fix some issue (link existing issues this PR will resolve)".
  • The test plan, such as providing test command.
  • The test results, such as pasting the results comparison before and after, or e2e results
  • (Optional) The necessary documentation update, such as updating supported_models.md and examples for a new model.
  • (Optional) Release notes update. If your change is user facing, please update the release notes draft in the Google Doc.

Copy link
Contributor

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request introduces fixes for building with CUDA 13 by abstracting CUB API differences into a new cub_helpers.h header. The changes are well-implemented and consistently applied across all relevant CUDA kernel files. This is a good approach to handle version-specific API changes. I have one suggestion for CMakeLists.txt to improve the robustness of how compiler flags are added.

Copy link
Contributor

@atalman atalman left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

lgtm

Copy link
Collaborator

@ProExpertProg ProExpertProg left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Looks good apart from 1 nit, also Gemini suggestion is good. Did you try to build vLLM from source locally (pip install -e .) with cuda 13.0?

@Aidyn-A
Copy link
Contributor Author

Aidyn-A commented Sep 11, 2025

Did you try to build vLLM from source locally (pip install -e .) with cuda 13.0?

Yes, I did. I ran VLLM_TARGET_DEVICE=cuda pip install --verbose --no-deps --no-build-isolation -e . to confirm it:

$ pip show vllm
Name: vllm
Version: 0.10.2rc2.dev89+gf94619747.d20250911.cu130
...

Copy link

pytorch-bot bot commented Sep 15, 2025

No ciflow labels are configured for this repo.
For information on how to enable CIFlow bot see this wiki

@zou3519 zou3519 added the ready ONLY add when PR is ready to merge/full CI is needed label Sep 15, 2025
@Aidyn-A
Copy link
Contributor Author

Aidyn-A commented Sep 17, 2025

@huydhn, @atalman are you able to merge it?

@zou3519 zou3519 merged commit bfe9380 into vllm-project:main Sep 17, 2025
92 of 93 checks passed
FeiDaLI pushed a commit to FeiDaLI/vllm that referenced this pull request Sep 25, 2025
charlifu pushed a commit to ROCm/vllm that referenced this pull request Sep 25, 2025
xuebwang-amd pushed a commit to xuebwang-amd/vllm that referenced this pull request Oct 10, 2025
Signed-off-by: Aidyn-A <[email protected]>
Signed-off-by: xuebwang-amd <[email protected]>
choprahetarth pushed a commit to Tandemn-Labs/vllm that referenced this pull request Oct 11, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

ci/build ready ONLY add when PR is ready to merge/full CI is needed

Projects

None yet

Development

Successfully merging this pull request may close these issues.

[Bug]: Building vLLM with CUDA 13.0

4 participants