-
-
Notifications
You must be signed in to change notification settings - Fork 10.5k
Closed
Labels
usageHow to use vllmHow to use vllm
Description
Your current environment
The output of `python collect_env.py`
How would you like to use vllm
Hi
I want to attach lora using docker command
docker run --runtime nvidia --gpus all
-v ~/.cache/huggingface:/root/.cache/huggingface
-v /datadrive/finetune_model/infosys:/app/lora/xyz
-v /datadrive/finetune_model/dummy:/app/lora/abc
-p 8000:8000
--env "HUGGING_FACE_HUB_TOKEN="
vllm/vllm-openai --enable-lora
--model meta-llama/Meta-Llama-3-8B-Instruct
--lora-modules xyz-lora=/datadrive/finetune_model/xyz
--lora-modules abc-lora=/datadrive/finetune_model/abc
However, I am getting below error
{'object': 'error', 'message': 'The model `xyz-lora` does not exist.', 'type': 'NotFoundError', 'param': None, 'code': 404}
Can anyone help here? Did I use wrong command?
Thanks you
Metadata
Metadata
Assignees
Labels
usageHow to use vllmHow to use vllm