Unverified Commit b4ec427a authored by Daniël de Kok's avatar Daniël de Kok Committed by GitHub
Browse files

Simplify two ipex conditions (#2755)

parent 38cff84a
...@@ -8,10 +8,10 @@ from text_generation_server.utils.weights import UnquantizedWeight, Weights ...@@ -8,10 +8,10 @@ from text_generation_server.utils.weights import UnquantizedWeight, Weights
if SYSTEM == "rocm": if SYSTEM == "rocm":
from vllm.model_executor.layers.fused_moe import fused_moe from vllm.model_executor.layers.fused_moe import fused_moe
elif SYSTEM != "ipex": elif SYSTEM == "ipex":
from moe_kernels.fused_moe import fused_moe
else:
from intel_extension_for_pytorch.llm.modules import GatedMLPMOE from intel_extension_for_pytorch.llm.modules import GatedMLPMOE
else:
from moe_kernels.fused_moe import fused_moe
class UnquantizedSparseMoELayer(nn.Module): class UnquantizedSparseMoELayer(nn.Module):
......
...@@ -25,10 +25,10 @@ from text_generation_server.utils.import_utils import SYSTEM ...@@ -25,10 +25,10 @@ from text_generation_server.utils.import_utils import SYSTEM
if SYSTEM == "rocm": if SYSTEM == "rocm":
from vllm.model_executor.layers.fused_moe import fused_moe from vllm.model_executor.layers.fused_moe import fused_moe
elif SYSTEM != "ipex": elif SYSTEM == "ipex":
from moe_kernels.fused_moe import fused_moe
else:
from intel_extension_for_pytorch.llm.modules import GatedMLPMOE from intel_extension_for_pytorch.llm.modules import GatedMLPMOE
else:
from moe_kernels.fused_moe import fused_moe
from text_generation_server.layers.attention import ( from text_generation_server.layers.attention import (
paged_attention, paged_attention,
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment