[Bugfix][Hardware][Gaudi] Bump vllm_hpu_extension version #11028
+12
−1
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
vllm_hpu_extension had
vllm.utils.get_vllm_instance_id
dependency, which was removed in #10976, causing HPU backend to crash on vllm_hpu_extension import. Extension was updated to remove that dependency (HabanaAI/vllm-hpu-extension#52), and this PR contains that fix.Small changes were needed to hpu_attn.py, as vllm_hpu_extension had a minor paged attention API change, and the existing code was not compatible with the newer revision. With these changes, stays functionally the same as before.