Skip to content

Commit 8630432

Browse files
committed
fix: graniteMoE does support PP (but requires enforce-eager)
Signed-off-by: Travis Johnson <[email protected]>
1 parent 4467c5d commit 8630432

File tree

2 files changed

+4
-4
lines changed

2 files changed

+4
-4
lines changed

docs/source/models/supported_models.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -297,12 +297,12 @@ See [this page](#generative-models) for more information on how to use generativ
297297
* Granite 3.0 MoE, PowerMoE
298298
* `ibm-granite/granite-3.0-1b-a400m-base`, `ibm-granite/granite-3.0-3b-a800m-instruct`, `ibm/PowerMoE-3b`, etc.
299299
* ✅︎
300-
*
300+
* ✅︎
301301
- * `GraniteMoeSharedForCausalLM`
302302
* Granite MoE Shared
303303
* `ibm-research/moe-7b-1b-active-shared-experts` (test model)
304304
* ✅︎
305-
*
305+
* ✅︎
306306
- * `GritLM`
307307
* GritLM
308308
* `parasail-ai/GritLM-7B-vllm`.

vllm/model_executor/models/granitemoeshared.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -28,7 +28,7 @@
2828

2929
from . import mixtral
3030
from .granitemoe import GraniteMoeAttention, GraniteMoeMoE
31-
from .interfaces import SupportsLoRA
31+
from .interfaces import SupportsLoRA, SupportsPP
3232
from .utils import make_layers, maybe_prefix
3333

3434

@@ -208,7 +208,7 @@ def forward(
208208
return hidden_states
209209

210210

211-
class GraniteMoeSharedForCausalLM(nn.Module, SupportsLoRA):
211+
class GraniteMoeSharedForCausalLM(nn.Module, SupportsLoRA, SupportsPP):
212212
fall_back_to_pt_during_load = False
213213

214214
packed_modules_mapping = {

0 commit comments

Comments
 (0)