Skip to content

Conversation

@vMaroon
Copy link
Member

@vMaroon vMaroon commented Sep 25, 2025

Summary

Upgrade the llm-d-kv-cache-manager dependency to v0.3.1, which adds support for the updated vLLM v0.11.0 KVEvents format while maintaining backwards compatibility. This was tested and benchmarked with vLLM v0.10.2 and v0.11.0, built into the image: quay.io/vmaroon/llm-d-inference-scheduler:v0.3.1.

This is required for the llm-d v0.3.0 release.

Context: since the llm-d release took (much) longer than planned, and the vLLM native CPU offloading work landed, this llm-d v0.4.0 change is required now.

@kfswain
Copy link
Collaborator

kfswain commented Sep 25, 2025

/lgtm
/approve

@github-actions github-actions bot added the lgtm "Looks good to me", indicates that a PR is ready to be merged. label Sep 25, 2025
@kfswain kfswain mentioned this pull request Sep 25, 2025
@github-actions github-actions bot merged commit 4829137 into llm-d:main Sep 25, 2025
6 checks passed
kfswain pushed a commit that referenced this pull request Sep 26, 2025
usize pushed a commit to usize/llm-d-inference-scheduler that referenced this pull request Oct 6, 2025
usize pushed a commit to usize/llm-d-inference-scheduler that referenced this pull request Oct 8, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

lgtm "Looks good to me", indicates that a PR is ready to be merged.

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants