Skip to content

Commit b5f882c

Browse files
Mixtral 8x7B support (#2011)
Co-authored-by: Pierre Stock <[email protected]> Co-authored-by: Zhuohan Li <[email protected]>
1 parent 2e8fc0d commit b5f882c

File tree

4 files changed

+538
-0
lines changed

4 files changed

+538
-0
lines changed

README.md

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -60,6 +60,7 @@ vLLM seamlessly supports many Hugging Face models, including the following archi
6060
- InternLM (`internlm/internlm-7b`, `internlm/internlm-chat-7b`, etc.)
6161
- LLaMA & LLaMA-2 (`meta-llama/Llama-2-70b-hf`, `lmsys/vicuna-13b-v1.3`, `young-geng/koala`, `openlm-research/open_llama_13b`, etc.)
6262
- Mistral (`mistralai/Mistral-7B-v0.1`, `mistralai/Mistral-7B-Instruct-v0.1`, etc.)
63+
- Mixtral (`mistralai/Mixtral-8x7B-v0.1`, `mistralai/Mixtral-8x7B-Instruct-v0.1`, etc.)
6364
- MPT (`mosaicml/mpt-7b`, `mosaicml/mpt-30b`, etc.)
6465
- OPT (`facebook/opt-66b`, `facebook/opt-iml-max-30b`, etc.)
6566
- Phi-1.5 (`microsoft/phi-1_5`, etc.)

vllm/model_executor/model_loader.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -33,6 +33,7 @@
3333
"LlamaForCausalLM": LlamaForCausalLM,
3434
"LLaMAForCausalLM": LlamaForCausalLM, # For decapoda-research/llama-*
3535
"MistralForCausalLM": MistralForCausalLM,
36+
"MixtralForCausalLM": MixtralForCausalLM,
3637
# transformers's mpt class has lower case
3738
"MptForCausalLM": MPTForCausalLM,
3839
"MPTForCausalLM": MPTForCausalLM,

vllm/model_executor/models/__init__.py

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -10,6 +10,7 @@
1010
from vllm.model_executor.models.internlm import InternLMForCausalLM
1111
from vllm.model_executor.models.llama import LlamaForCausalLM
1212
from vllm.model_executor.models.mistral import MistralForCausalLM
13+
from vllm.model_executor.models.mixtral import MixtralForCausalLM
1314
from vllm.model_executor.models.mpt import MPTForCausalLM
1415
from vllm.model_executor.models.opt import OPTForCausalLM
1516
from vllm.model_executor.models.phi_1_5 import PhiForCausalLM
@@ -35,5 +36,6 @@
3536
"PhiForCausalLM",
3637
"QWenLMHeadModel",
3738
"MistralForCausalLM",
39+
"MixtralForCausalLM",
3840
"YiForCausalLM",
3941
]

0 commit comments

Comments
 (0)