-
-
Notifications
You must be signed in to change notification settings - Fork 11.7k
Install pre-built xformers-0.0.32.post2 built with pt-2.9.0 #27598
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
Signed-off-by: Huy Do <[email protected]>
Signed-off-by: Huy Do <[email protected]>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Code Review
This pull request aims to install a prebuilt version of xformers-0.0.32.post2 built with PyTorch 2.9.0, instead of waiting for an official release. The changes involve removing the temporary installation of xformers from the Dockerfile and updating the requirements file to include the prebuilt wheel URL. I have identified a critical issue related to the hardcoding of the xformers version in the requirements file.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
💡 Codex Review
Here are some automated review suggestions for this pull request.
ℹ️ About Codex in GitHub
Codex has been enabled to automatically review pull requests in this repo. Reviews are triggered when you
- Open a pull request for review
- Mark a draft as ready
- Comment "@codex review".
If Codex has suggestions, it will comment; otherwise it will react with 👍.
When you sign up for Codex through ChatGPT, Codex can also answer questions or update the PR, like "@codex address that feedback".
| && uv pip install --system dist/*.whl --verbose \ | ||
| --extra-index-url ${PYTORCH_CUDA_INDEX_BASE_URL}/cu$(echo $CUDA_VERSION | cut -d. -f1,2 | tr -d '.') | ||
|
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Ensure runtime image still installs xformers
The Docker runtime stage no longer installs xformers after the manual uv pip install git+…[email protected] block was removed. The new wheel URL was added to requirements/cuda.txt, but that requirements file is consumed only in the earlier build stage (COPY requirements/cuda.txt … followed by uv pip install --python /opt/venv/bin/python3 -r requirements/cuda.txt). The vllm-base stage (lines shown) now installs only the built vLLM wheel and FlashInfer, so the final images vllm-base, vllm-openai, etc. ship without xformers. Any runtime paths that import xformers (memory-efficient attention, sliding window, etc.) will fail with ModuleNotFoundError. A separate uv pip install for the new wheel needs to run in the runtime stage as before.
Useful? React with 👍 / 👎.
|
AFIAK this doesn't work when we distribute this as wheel on PyPI and users run |
|
Hmm I also actually run into this error when installing this branch |
|
Let's wait for CI and and I can check what is the |
I was trying to install editable - but yea I turned on ready label so you can inspect |
Signed-off-by: Huy Do <[email protected]>
I had a mental lapse lol, |
|
LoRA failed today on nightly https://buildkite.com/organizations/vllm/analytics/suites/ci-1/tests/f36a301a-3b8a-8a11-a011-b58b871880bf?branch=main&period=1day&execution_id=019a2dd1-0f20-7305-b8e6-e4e27a676f44 Quantization is also known failure https://buildkite.com/vllm/ci/builds/36507/steps/canvas?sid=019a28fa-289d-4cb9-ab28-f16f76f5ba29#019a28fa-2967-4891-b480-35d514706161/156-3583 I'm forcing merging this so we can cut a new rc and lower the build time to save CI cost. |
|
I encountered the following error while attempting to install the latest main branch code. |
|
True, I just realize that there are only:
Let me find out where that cu130 wheel is, get that one ready, and we can retry cu130 build. |
Could you try it with |
|
@huydhn I'm going to revert this to unbreak main 😢 |
…ject#27598) Signed-off-by: Huy Do <[email protected]> Co-authored-by: Roger Wang <[email protected]> Signed-off-by: Bhagyashri <[email protected]>
…ject#27598) Signed-off-by: Huy Do <[email protected]> Co-authored-by: Roger Wang <[email protected]>
…ject#27598) Signed-off-by: Huy Do <[email protected]> Co-authored-by: Roger Wang <[email protected]>
…ject#27598) Signed-off-by: Huy Do <[email protected]> Co-authored-by: Roger Wang <[email protected]>
…ject#27598) Signed-off-by: Huy Do <[email protected]> Co-authored-by: Roger Wang <[email protected]>
Purpose
Instead of waiting for xformers to release a new version for PyTorch 2.9.0, I have built
0.0.32.post2locally and made the wheel available.For more context, we don’t want to wait for
xformerspackage for 2.9 to become available. So, I opt to build it from source. This works for CI, but has several issues like (1) increasing build time and (2) not listed as a dependency incuda.txt. So, installing a pre-built wheel would help in the meantime until there is a new xformers version@ywang96