[qwen3 next ]add ascend c casual_conv1d_fn (#6661)
### What this PR does / why we need it?
add ascend c casual_conv1d_fn
- vLLM version: v0.15.0
- vLLM main:
13397841ab
---------
Signed-off-by: ZT-AIA <1028681969@qq.com>
Signed-off-by: ZT-AIA <63220130+ZT-AIA@users.noreply.github.com>
Co-authored-by: gemini-code-assist[bot] <176961590+gemini-code-assist[bot]@users.noreply.github.com>
This commit is contained in:
@@ -458,6 +458,22 @@ void transpose_kv_cache_by_block_meta(
|
||||
return;
|
||||
}
|
||||
|
||||
at::Tensor causal_conv1d_fn_meta(
|
||||
const at::Tensor& mixed_qkv_non_spec_T,
|
||||
const at::Tensor& conv_weights,
|
||||
const c10::optional<at::Tensor>& bias_opt,
|
||||
c10::string_view activation,
|
||||
const at::Tensor& conv_state,
|
||||
const at::Tensor& has_initial_state,
|
||||
const at::Tensor& non_spec_state_indices_tensor,
|
||||
const at::Tensor& non_spec_query_start_loc,
|
||||
int64_t pad_slot_id)
|
||||
{
|
||||
|
||||
at::Tensor output = at::empty_symint(mixed_qkv_non_spec_T.sym_sizes(), mixed_qkv_non_spec_T.options());
|
||||
return output;
|
||||
}
|
||||
|
||||
std::vector<at::Tensor> moe_grouped_matmul_meta(
|
||||
at::Tensor x,
|
||||
at::Tensor weight,
|
||||
@@ -527,6 +543,8 @@ TORCH_LIBRARY_IMPL_EXPAND(CONCAT(_C, _ascend), Meta, ops) {
|
||||
ops.impl("npu_add_rms_norm_bias", &vllm_ascend::meta::npu_add_rms_norm_bias_meta);
|
||||
// transpose_kv_cache_by_block
|
||||
ops.impl("transpose_kv_cache_by_block", &vllm_ascend::meta::transpose_kv_cache_by_block_meta);
|
||||
// causal_conv1d_fn
|
||||
ops.impl("causal_conv1d_fn", &vllm_ascend::meta::causal_conv1d_fn_meta);
|
||||
// moe_grouped_matmul
|
||||
ops.impl("moe_grouped_matmul", &vllm_ascend::meta::moe_grouped_matmul_meta);
|
||||
}
|
||||
|
||||
Reference in New Issue
Block a user