-
-
Notifications
You must be signed in to change notification settings - Fork 10.5k
[Feature] Add visionarena offline support for benchmark_throughput #14654
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
Signed-off-by: Jennifer Zhao <[email protected]>
👋 Hi! Thank you for contributing to the vLLM project. 💬 Join our developer Slack at https://slack.vllm.ai to discuss your PR in #pr-reviews, coordinate on features in #feat- channels, or join special interest groups in #sig- channels. Just a reminder: PRs would not trigger full CI run by default. Instead, it would only run Once the PR is approved and ready to go, your PR reviewer(s) can run CI to test the changes comprehensively before merging. To run CI, PR reviewers can either: Add 🚀 |
Signed-off-by: Jennifer Zhao <[email protected]>
Signed-off-by: Jennifer Zhao <[email protected]>
Signed-off-by: Jennifer Zhao <[email protected]>
Signed-off-by: Jennifer Zhao <[email protected]>
Signed-off-by: Jennifer Zhao <[email protected]>
Signed-off-by: Jennifer Zhao <[email protected]>
Signed-off-by: Jennifer Zhao <[email protected]>
[Feature] Huggingface Dataset support for benchmark_throughput
Signed-off-by: Jennifer Zhao <[email protected]>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I left a few comments - thank you for your work!
Could you please also do a comparison run between main and this PR when running vLLM with ShareGPT? I'd like to see the differences between the two if there's any.
--------- Signed-off-by: Jennifer Zhao <[email protected]>
test results here https://github.com/JenZhao/vllm/pull/5#issue-2918317590 |
Signed-off-by: Jennifer Zhao <[email protected]>
Signed-off-by: Jennifer Zhao <[email protected]>
Signed-off-by: Jennifer Zhao <[email protected]>
Signed-off-by: Jennifer Zhao <[email protected]>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM! I left some final comments.
Co-authored-by: Roger Wang <[email protected]>
Signed-off-by: Jennifer Zhao <[email protected]>
…llm-project#14654) Signed-off-by: Jennifer Zhao <[email protected]> Signed-off-by: Jennifer Zhao <[email protected]> Co-authored-by: Jennifer Zhao <[email protected]> Co-authored-by: Jennifer Zhao <[email protected]> Co-authored-by: Roger Wang <[email protected]> Signed-off-by: Richard Liu <[email protected]>
…llm-project#14654) Signed-off-by: Jennifer Zhao <[email protected]> Signed-off-by: Jennifer Zhao <[email protected]> Co-authored-by: Jennifer Zhao <[email protected]> Co-authored-by: Jennifer Zhao <[email protected]> Co-authored-by: Roger Wang <[email protected]> Signed-off-by: Louis Ulmer <[email protected]>
…llm-project#14654) Signed-off-by: Jennifer Zhao <[email protected]> Signed-off-by: Jennifer Zhao <[email protected]> Co-authored-by: Jennifer Zhao <[email protected]> Co-authored-by: Jennifer Zhao <[email protected]> Co-authored-by: Roger Wang <[email protected]>
…llm-project#14654) Signed-off-by: Jennifer Zhao <[email protected]> Signed-off-by: Jennifer Zhao <[email protected]> Co-authored-by: Jennifer Zhao <[email protected]> Co-authored-by: Jennifer Zhao <[email protected]> Co-authored-by: Roger Wang <[email protected]> Signed-off-by: Mu Huai <[email protected]>
This is a follow-up PR for #14036 to support vision arena in the
benchmark_throughput
script.With the
vllm-chat
backend, the prompt token count will now include image token counts, as we will be using the request output to perform the calculation.CLOSES #14600
hf:
lmarena-ai/vision-arena-bench-v0.1
Try another model
hf:
lmms-lab/LLaVA-OneVision-Data
since we changed the token calculation for vllm and vllm-chat, here we are using sharegpt dataset to verify the change
This Branch
Main Branch
vllm-chat is not support on the main branch
python3 vllm/benchmarks/benchmark_throughput.py --model NousResearch/Hermes-3-Llama-3.1-8B --backend vllm --dataset-name sharegpt --dataset-path ~/data/vllm_benchmark_datasets/ShareGPT_V3_unfiltered_cleaned_split.json --num-prompts 1000