Skip to content

Conversation

ji-huazhong
Copy link
Collaborator

@ji-huazhong ji-huazhong commented Feb 12, 2025

What this PR does / why we need it?

In open-r1, the rank 0 process will create an LLM instance and load the model to npu:7. We need to force the output tensor to be created on the same device as the query tensor.

Does this PR introduce any user-facing change?

No

How was this patch tested?

Test by main branch

@wangxiyuan wangxiyuan merged commit 49f3cb3 into vllm-project:v0.7.1-release Feb 13, 2025
1 check passed
@Yikun
Copy link
Collaborator

Yikun commented Feb 13, 2025

Looks like we should also fix this on main, would you mind cherry pick this? @ji-huazhong

@ji-huazhong
Copy link
Collaborator Author

ji-huazhong commented Feb 13, 2025

This issue has been addressed on main, see #25
@Yikun

@ji-huazhong ji-huazhong deleted the fix branch February 17, 2025 07:14
Angazenn pushed a commit to Angazenn/vllm-ascend that referenced this pull request Feb 21, 2025
vllm-project#54)

### What this PR does / why we need it?
In open-r1, the rank 0 process will create an LLM instance and load the model to `npu:7`. We need to force the output tensor to be created on the same device as the query tensor.

### Does this PR introduce _any_ user-facing change?
No

### How was this patch tested?
Test by main branch

Signed-off-by: angazenn <[email protected]>
ZhengWG pushed a commit to ZhengWG/vllm-ascend that referenced this pull request Jun 18, 2025
move get_init_expert_map to forward_before
offline893 pushed a commit to offline893/vllm-ascend that referenced this pull request Sep 9, 2025
move get_init_expert_map to forward_before
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants