Logo
Explore Help
Register Sign In
EngineX/xc-llm-ascend
3
0
Fork 0
You've already forked xc-llm-ascend
Code Issues Pull Requests Projects Releases Wiki Activity
Files
84d4f474c03a4c3387a9fe2cdc849f9a15947f8a
xc-llm-ascend/tests/ut/attention
History
gh924 6880c1b383 [Feature] Support for cross-attention and whisper model (#5592)
### What this PR does / why we need it?
To solve the problem of the
issue:https://github.com/vllm-project/vllm-ascend/issues/2262

- support for cross-attention when the model is encoder-decoder
- support for whisper model

- vLLM version: v0.13.0
- vLLM main:
7157596103

Signed-off-by: gh924 <guihao2@huawei.com>
Co-authored-by: Aoxuan Chen <43376869+chenaoxuan@users.noreply.github.com>
2026-01-11 11:38:45 +08:00
..
test_attention_cp.py
[Refactor]7/N Extract common code to common_cp (#5490)
2026-01-05 17:41:12 +08:00
test_attention_mask.py
[Refactor] 2/N Unify all mask generation methods and cache mask (#4779)
2025-12-09 18:51:00 +08:00
test_attention_v1.py
[Feature] Support for cross-attention and whisper model (#5592)
2026-01-11 11:38:45 +08:00
test_mla_cp.py
[Refactor] Fix AttentionMaskBuilder singleton and remove redundant pcp_prefill_mask (#4870)
2026-01-07 17:09:52 +08:00
test_mla_v1.py
[Refactor] Fix AttentionMaskBuilder singleton and remove redundant pcp_prefill_mask (#4870)
2026-01-07 17:09:52 +08:00
test_sfa_v1.py
[Refactor] Replace the implementations of o_proj, q_b_proj, and kv_b_proj with custom_op for sharded CP (#5698)
2026-01-09 15:58:40 +08:00
utils.py
[refactor](UT,PCP,DCP) refactor pcp&dcp patches in UTs (#5505)
2026-01-05 09:05:45 +08:00
Powered by Gitea Version: 1.24.3 Page: 189ms Template: 66ms
English
Bahasa Indonesia Deutsch English Español Français Gaeilge Italiano Latviešu Magyar nyelv Nederlands Polski Português de Portugal Português do Brasil Suomi Svenska Türkçe Čeština Ελληνικά Български Русский Українська فارسی മലയാളം 日本語 简体中文 繁體中文(台灣) 繁體中文(香港) 한국어
Licenses API