Skip to content

Conversation

Isotr0py
Copy link
Member

@Isotr0py Isotr0py commented Sep 22, 2025

Purpose

  • Remove _set_default_args_v0 since v0 engine has been removed

Test Plan

Test Result


Essential Elements of an Effective PR Description Checklist
  • The purpose of the PR, such as "Fix some issue (link existing issues this PR will resolve)".
  • The test plan, such as providing test command.
  • The test results, such as pasting the results comparison before and after, or e2e results
  • (Optional) The necessary documentation update, such as updating supported_models.md and examples for a new model.
  • (Optional) Release notes update. If your change is user facing, please update the release notes draft in the Google Doc.

Copy link
Contributor

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request removes the _set_default_args_v0 function and related logic as part of the V0 engine deprecation. The changes are straightforward, renaming _set_default_args_v1 to _set_default_args and simplifying the call site. However, the removal of V0 components appears to be incomplete in other parts of the codebase. I've pointed out a critical issue where setting VLLM_USE_V1=0 could lead to a runtime crash due to lingering references to V0 components in platform-specific files. This should be addressed to ensure the V0 removal is clean and does not introduce new bugs.

Comment on lines +1150 to +1158
# Set default arguments for V1 Engine.
self._set_default_args(usage_context, model_config)
# Disable chunked prefill for POWER (ppc64le)/ARM/s390x CPUs in V1
if current_platform.is_cpu() and current_platform.get_cpu_architecture(
) in (CpuArchEnum.POWERPC, CpuArchEnum.S390X, CpuArchEnum.ARM):
logger.info("Chunked prefill is not supported for ARM and POWER "
"and S390X CPUs; "
"disabling it for V1 backend.")
self.enable_chunked_prefill = False
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

critical

While this change correctly removes the V0-specific default arguments, the removal of V0 components seems incomplete across the codebase. Several platform-specific files still contain logic that depends on VLLM_USE_V1 and might attempt to use V0 components, which are presumably removed. This could lead to runtime errors if a user sets VLLM_USE_V1=0.

For example, in vllm/platforms/cuda.py and vllm/platforms/rocm.py, the code still selects vllm.worker.worker.Worker if VLLM_USE_V1 is false. If this V0 worker has been removed, this will cause a crash.

It is recommended to update these files to either always use the V1 worker or raise an error if VLLM_USE_V1=0, similar to the approach in cpu.py, xpu.py, and tpu.py which explicitly disallow VLLM_USE_V1=0.

@WoosukKwon WoosukKwon added the ready ONLY add when PR is ready to merge/full CI is needed label Sep 23, 2025
@WoosukKwon WoosukKwon enabled auto-merge (squash) September 23, 2025 00:04
@WoosukKwon WoosukKwon merged commit c625f90 into vllm-project:main Sep 23, 2025
56 checks passed
linfeng-yuan pushed a commit to linfeng-yuan/vllm that referenced this pull request Sep 23, 2025
@Isotr0py Isotr0py deleted the rm-v0-args branch September 23, 2025 05:05
FeiDaLI pushed a commit to FeiDaLI/vllm that referenced this pull request Sep 25, 2025
charlifu pushed a commit to ROCm/vllm that referenced this pull request Sep 25, 2025
yewentao256 pushed a commit that referenced this pull request Oct 3, 2025
gjc0824 pushed a commit to gjc0824/vllm that referenced this pull request Oct 10, 2025
xuebwang-amd pushed a commit to xuebwang-amd/vllm that referenced this pull request Oct 10, 2025
choprahetarth pushed a commit to Tandemn-Labs/vllm that referenced this pull request Oct 11, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

ready ONLY add when PR is ready to merge/full CI is needed

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants