There was an error while loading. Please reload this page.
1 parent dd2a945 commit 70f8b96Copy full SHA for 70f8b96
vllm/model_executor/layers/fused_moe/layer.py
@@ -43,6 +43,7 @@
43
from .pplx_prepare_finalize import PplxPrepareAndFinalize
44
else:
45
fused_experts = None # type: ignore
46
+ FusedMoEPrepareAndFinalize = None # type: ignore
47
if is_rocm_aiter_moe_enabled():
48
from vllm.model_executor.layers.fused_moe.rocm_aiter_fused_moe import ( # noqa: E501
49
rocm_aiter_biased_group_topk as grouped_topk)
0 commit comments