-
-
Notifications
You must be signed in to change notification settings - Fork 10.5k
[Feature] Add vllm bench
CLI
#13993
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Feature] Add vllm bench
CLI
#13993
Conversation
👋 Hi! Thank you for contributing to the vLLM project. 💬 Join our developer Slack at https://slack.vllm.ai to discuss your PR in #pr-reviews, coordinate on features in #feat- channels, or join special interest groups in #sig- channels. Just a reminder: PRs would not trigger full CI run by default. Instead, it would only run Once the PR is approved and ready to go, your PR reviewer(s) can run CI to test the changes comprehensively before merging. To run CI, PR reviewers can either: Add 🚀 |
Thanks for the PR! My two cents:
|
Thanks for the suggestions! I’ll keep working on improving points 1 and 3. |
6e27b04
to
49267b5
Compare
The new commits introduce the nested version for benchmark CLI.
vllm bench [throughput|latency|serving] --opts
vllm bench --help
vllm bench [throughput|latency|serving] --help |
This pull request has merge conflicts that must be resolved before it can be |
@khluu do you know why the changes made by local pre-commit is different from the CI and how should we fix it? |
A bug fix related to precommit was merged not too long ago.. maybe merge this branch with main and try again? |
80b1aed
to
872db2b
Compare
Signed-off-by: Randy Chen <[email protected]>
Reframing the PR
Follow-up Items
|
Signed-off-by: Randy Chen <[email protected]>
Signed-off-by: Randy Chen <[email protected]>
Signed-off-by: Randy Chen <[email protected]>
Signed-off-by: Randy Chen <[email protected]>
This reverts commit d5bc88b. Signed-off-by: Randy Chen <[email protected]>
Signed-off-by: Randy Chen <[email protected]>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
It's in a good shape in general. Will take a deep look tomorrow.
Meanwhile please update the PR description with the latest scope of this PR and the remaining follow-up tasks.
…project#13840) Signed-off-by: Randy Chen <[email protected]>
Signed-off-by: Randy Chen <[email protected]>
Signed-off-by: Cody Yu <[email protected]>
Signed-off-by: Cody Yu <[email protected]>
vllm bench
CLI
Signed-off-by: Cody Yu <[email protected]>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
We have to options to move forward:
I actually prefer (1) because it should take less time to merge this PR, but if you feel supporting datasets won't postponing the review process (because the logic is mostly copied), I'm also ok with it. |
yea I'm also okay with (1) just to get this PR in! |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM!
Signed-off-by: Randy Chen <[email protected]>
Signed-off-by: Randy Chen <[email protected]> Signed-off-by: Cody Yu <[email protected]> Co-authored-by: Cody Yu <[email protected]> Signed-off-by: Richard Liu <[email protected]>
Signed-off-by: Randy Chen <[email protected]> Signed-off-by: Cody Yu <[email protected]> Co-authored-by: Cody Yu <[email protected]> Signed-off-by: Louis Ulmer <[email protected]>
Signed-off-by: Randy Chen <[email protected]> Signed-off-by: Cody Yu <[email protected]> Co-authored-by: Cody Yu <[email protected]>
Signed-off-by: Randy Chen <[email protected]> Signed-off-by: Cody Yu <[email protected]> Co-authored-by: Cody Yu <[email protected]> Signed-off-by: Mu Huai <[email protected]>
Part of #13840
This PR adds CLI commands of
vllm bench
. We only supportvllm bench serve
in this PR to align the interface, the following benchmark modes will be added in follow-up PRs (contributions are welcome).What has been covered in this PR:
vllm bench serve
.Future work:
vllm bench latency
.vllm bench throughput
.benchmarks/
to use this CLI.