Unverified Commit b5f882cc authored by Pierre Stock's avatar Pierre Stock Committed by GitHub
Browse files

Mixtral 8x7B support (#2011)


Co-authored-by: default avatarPierre Stock <p@mistral.ai>
Co-authored-by: default avatarZhuohan Li <zhuohan123@gmail.com>
parent 2e8fc0d4
...@@ -60,6 +60,7 @@ vLLM seamlessly supports many Hugging Face models, including the following archi ...@@ -60,6 +60,7 @@ vLLM seamlessly supports many Hugging Face models, including the following archi
- InternLM (`internlm/internlm-7b`, `internlm/internlm-chat-7b`, etc.) - InternLM (`internlm/internlm-7b`, `internlm/internlm-chat-7b`, etc.)
- LLaMA & LLaMA-2 (`meta-llama/Llama-2-70b-hf`, `lmsys/vicuna-13b-v1.3`, `young-geng/koala`, `openlm-research/open_llama_13b`, etc.) - LLaMA & LLaMA-2 (`meta-llama/Llama-2-70b-hf`, `lmsys/vicuna-13b-v1.3`, `young-geng/koala`, `openlm-research/open_llama_13b`, etc.)
- Mistral (`mistralai/Mistral-7B-v0.1`, `mistralai/Mistral-7B-Instruct-v0.1`, etc.) - Mistral (`mistralai/Mistral-7B-v0.1`, `mistralai/Mistral-7B-Instruct-v0.1`, etc.)
- Mixtral (`mistralai/Mixtral-8x7B-v0.1`, `mistralai/Mixtral-8x7B-Instruct-v0.1`, etc.)
- MPT (`mosaicml/mpt-7b`, `mosaicml/mpt-30b`, etc.) - MPT (`mosaicml/mpt-7b`, `mosaicml/mpt-30b`, etc.)
- OPT (`facebook/opt-66b`, `facebook/opt-iml-max-30b`, etc.) - OPT (`facebook/opt-66b`, `facebook/opt-iml-max-30b`, etc.)
- Phi-1.5 (`microsoft/phi-1_5`, etc.) - Phi-1.5 (`microsoft/phi-1_5`, etc.)
......
...@@ -33,6 +33,7 @@ _MODEL_REGISTRY = { ...@@ -33,6 +33,7 @@ _MODEL_REGISTRY = {
"LlamaForCausalLM": LlamaForCausalLM, "LlamaForCausalLM": LlamaForCausalLM,
"LLaMAForCausalLM": LlamaForCausalLM, # For decapoda-research/llama-* "LLaMAForCausalLM": LlamaForCausalLM, # For decapoda-research/llama-*
"MistralForCausalLM": MistralForCausalLM, "MistralForCausalLM": MistralForCausalLM,
"MixtralForCausalLM": MixtralForCausalLM,
# transformers's mpt class has lower case # transformers's mpt class has lower case
"MptForCausalLM": MPTForCausalLM, "MptForCausalLM": MPTForCausalLM,
"MPTForCausalLM": MPTForCausalLM, "MPTForCausalLM": MPTForCausalLM,
......
...@@ -10,6 +10,7 @@ from vllm.model_executor.models.gpt_neox import GPTNeoXForCausalLM ...@@ -10,6 +10,7 @@ from vllm.model_executor.models.gpt_neox import GPTNeoXForCausalLM
from vllm.model_executor.models.internlm import InternLMForCausalLM from vllm.model_executor.models.internlm import InternLMForCausalLM
from vllm.model_executor.models.llama import LlamaForCausalLM from vllm.model_executor.models.llama import LlamaForCausalLM
from vllm.model_executor.models.mistral import MistralForCausalLM from vllm.model_executor.models.mistral import MistralForCausalLM
from vllm.model_executor.models.mixtral import MixtralForCausalLM
from vllm.model_executor.models.mpt import MPTForCausalLM from vllm.model_executor.models.mpt import MPTForCausalLM
from vllm.model_executor.models.opt import OPTForCausalLM from vllm.model_executor.models.opt import OPTForCausalLM
from vllm.model_executor.models.phi_1_5 import PhiForCausalLM from vllm.model_executor.models.phi_1_5 import PhiForCausalLM
...@@ -35,5 +36,6 @@ __all__ = [ ...@@ -35,5 +36,6 @@ __all__ = [
"PhiForCausalLM", "PhiForCausalLM",
"QWenLMHeadModel", "QWenLMHeadModel",
"MistralForCausalLM", "MistralForCausalLM",
"MixtralForCausalLM",
"YiForCausalLM", "YiForCausalLM",
] ]
This diff is collapsed.
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment