You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
This issue has been automatically marked as stale because it has not had any activity within 90 days. It will be automatically closed if no further activity occurs within 30 days. Leave a comment if you feel this issue should remain open. Thank you!
Uh oh!
There was an error while loading. Please reload this page.
Hi, i am doing a load test on vllm server. below is the way to reproduce:
instance: 1xRTX 3090
load test tool: k6
server command:
python -m vllm.entrypoints.api_server --model mistralai/Mistral-7B-v0.1 --disable-log-requests --port 9009 --max-num-seqs 500
then run k6 with 100 VU:
i tried to adjust the --max-num-seqs and --max-num-batched-tokens but still cant pass 100 VU. is there any best config for the server?
any help is appreciate, thank you.
The text was updated successfully, but these errors were encountered: