Unverified Commit be0124bd authored by Lianmin Zheng's avatar Lianmin Zheng Committed by GitHub
Browse files

Rename triton_fused_moe -> fused_moe_triton (#2163)

parent fe5d3e81
from sglang.srt.layers.fused_moe.layer import FusedMoE, FusedMoEMethodBase
from sglang.srt.layers.fused_moe_grok.layer import FusedMoE, FusedMoEMethodBase
...@@ -20,7 +20,7 @@ from vllm.model_executor.layers.quantization.base_config import ( ...@@ -20,7 +20,7 @@ from vllm.model_executor.layers.quantization.base_config import (
from vllm.model_executor.layers.quantization.fp8 import Fp8Config from vllm.model_executor.layers.quantization.fp8 import Fp8Config
from vllm.model_executor.utils import set_weight_attrs from vllm.model_executor.utils import set_weight_attrs
from sglang.srt.layers.fused_moe.fused_moe import padding_size from sglang.srt.layers.fused_moe_grok.fused_moe import padding_size
from sglang.srt.utils import is_hip from sglang.srt.utils import is_hip
logger = init_logger(__name__) logger = init_logger(__name__)
...@@ -123,7 +123,7 @@ class UnquantizedFusedMoEMethod(FusedMoEMethodBase, CustomOp): ...@@ -123,7 +123,7 @@ class UnquantizedFusedMoEMethod(FusedMoEMethodBase, CustomOp):
num_expert_group: Optional[int], num_expert_group: Optional[int],
topk_group: Optional[int], topk_group: Optional[int],
) -> torch.Tensor: ) -> torch.Tensor:
from sglang.srt.layers.fused_moe.fused_moe import fused_moe from sglang.srt.layers.fused_moe_grok.fused_moe import fused_moe
return fused_moe( return fused_moe(
x, x,
...@@ -609,7 +609,7 @@ class Fp8MoEMethod(FusedMoEMethodBase): ...@@ -609,7 +609,7 @@ class Fp8MoEMethod(FusedMoEMethodBase):
topk_group: Optional[int] = None, topk_group: Optional[int] = None,
) -> torch.Tensor: ) -> torch.Tensor:
from sglang.srt.layers.fused_moe.fused_moe import fused_moe from sglang.srt.layers.fused_moe_grok.fused_moe import fused_moe
return fused_moe( return fused_moe(
x, x,
......
from contextlib import contextmanager from contextlib import contextmanager
from typing import Any, Dict, Optional from typing import Any, Dict, Optional
import sglang.srt.layers.triton_fused_moe.fused_moe # noqa import sglang.srt.layers.fused_moe_triton.fused_moe # noqa
from sglang.srt.layers.triton_fused_moe.fused_moe import ( from sglang.srt.layers.fused_moe_triton.fused_moe import (
fused_experts, fused_experts,
fused_topk, fused_topk,
get_config_file_name, get_config_file_name,
grouped_topk, grouped_topk,
) )
from sglang.srt.layers.triton_fused_moe.layer import ( from sglang.srt.layers.fused_moe_triton.layer import (
FusedMoE, FusedMoE,
FusedMoEMethodBase, FusedMoEMethodBase,
FusedMoeWeightScaleSupported, FusedMoeWeightScaleSupported,
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment