Skip to content

Conversation

kerthcet
Copy link
Contributor

@kerthcet kerthcet commented Feb 13, 2025

xref: #6234

With #6234, we can donwload the loras dynamically from HF, however, once people set the environment VLLM_USE_MODELSCOPE=true, we should download the loras from MS as well.

It's counterintuitive to download the model from one hub, but lora from another hub.

If this is acceptable, I'll continue the work, like adding tests. cc @Jeffwan @Yard1 as the original author and reviewer.

Copy link

👋 Hi! Thank you for contributing to the vLLM project.

💬 Join our developer Slack at https://slack.vllm.ai to discuss your PR in #pr-reviews, coordinate on features in #feat- channels, or join special interest groups in #sig- channels.

Just a reminder: PRs would not trigger full CI run by default. Instead, it would only run fastcheck CI which starts running only a small and essential subset of CI tests to quickly catch errors. You can run other CI tests on top of those by going to your fastcheck build on Buildkite UI (linked in the PR checks section) and unblock them. If you do not have permission to unblock, ping simon-mo or khluu to add you in our Buildkite org.

Once the PR is approved and ready to go, your PR reviewer(s) can run CI to test the changes comprehensively before merging.

To run CI, PR reviewers can either: Add ready label to the PR or enable auto-merge.

🚀

@jeejeelee
Copy link
Collaborator

See: #10762

@kerthcet
Copy link
Contributor Author

kerthcet commented Feb 14, 2025

See: #10762

Thanks for the response, I read the comments but seems no conclusion yet. From my point of view, if we set the VLLM_USE_MODELSCOPE=true, we're switching from HF to MS, then both the models and loras should from the same source, I believe this is the default semantic of the env variable VLLM_USE_MODELSCOPE, or we need to update the document which noted as:

By default, vLLM downloads models from HuggingFace. 
If you would like to use models from ModelScope, 
set the environment variable VLLM_USE_MODELSCOPE before initializing the engine.

Regarding to the download problem, HF is under the same situation I think. My two cents. wdyt?

@jeejeelee
Copy link
Collaborator

jeejeelee commented Feb 19, 2025

HF is under the same situation I think

We have tested locally, and this situation doesn't exist on HF.

Previously, we were concerned about the risk of lora service crashes during loading, but after merging #12102 and #11727 , I think we can consider supporting downloads from MODELSCOPE

Copy link

This pull request has been automatically marked as stale because it has not had any activity within 90 days. It will be automatically closed if no further activity occurs within 30 days. Leave a comment if you feel this pull request should remain open. Thank you!

@github-actions github-actions bot added the stale Over 90 days of inactivity label May 21, 2025
Copy link

This pull request has been automatically closed due to inactivity. Please feel free to reopen if you intend to continue working on it. Thank you!

@github-actions github-actions bot closed this Jun 20, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
stale Over 90 days of inactivity
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants