Upgrade to vllm 0.17.0 corex v4.1 overlay
This commit is contained in:
@@ -6,8 +6,7 @@ from enum import Enum
|
||||
|
||||
import torch
|
||||
import torch.nn.functional as F
|
||||
|
||||
from vllm._custom_ops import silu_and_mul, gelu_and_mul, swigluoai_and_mul
|
||||
from vllm import _custom_ops as ops
|
||||
|
||||
|
||||
class MoEActivation(Enum):
|
||||
@@ -114,14 +113,11 @@ def apply_moe_activation(
|
||||
|
||||
# Activations with gated multiplication (gate × activation(up))
|
||||
if activation == MoEActivation.SILU:
|
||||
# torch.ops._C.silu_and_mul(output, input)
|
||||
silu_and_mul(output, input)
|
||||
ops.silu_and_mul(output, input)
|
||||
elif activation == MoEActivation.GELU:
|
||||
# torch.ops._C.gelu_and_mul(output, input)
|
||||
gelu_and_mul(output, input)
|
||||
ops.gelu_and_mul(output, input)
|
||||
elif activation == MoEActivation.SWIGLUOAI:
|
||||
# torch.ops._C.swigluoai_and_mul(output, input)
|
||||
swigluoai_and_mul(output, input)
|
||||
ops.swigluoai_and_mul(output, input)
|
||||
elif activation == MoEActivation.SWIGLUSTEP:
|
||||
from vllm.model_executor.layers.activation import swiglustep_and_mul_triton
|
||||
|
||||
|
||||
Reference in New Issue
Block a user