flashinfer.activation.gelu_and_mul#
- flashinfer.activation.gelu_and_mul(input: torch.Tensor, out: torch.Tensor | None = None) torch.Tensor #
Fused GeLU and Mul operation.
gelu(input[..., :hidden_size]) * input[..., hidden_size:]
- Parameters:
input (torch.Tensor) – Input tensor, shape (…, 2 * hidden_size).
out (Optional[torch.Tensor]) – The the output tensor, if specified, the kernel will update this tensor inplace.
- Returns:
output – Output tensor, shape (…, hidden_size).
- Return type:
torch.Tensor