Logo
Explore Help
Register Sign In
EngineX/xc-llm-ascend
3
0
Fork 0
You've already forked xc-llm-ascend
Code Issues Pull Requests Actions Projects Releases Wiki Activity
Files
761bd3d9d72e72e5558cf9dd7343a34ea8b9c10b
xc-llm-ascend/tests
History
wangxiyuan b350edae9a [UT] refactor test_expert_load_balancer and fix broken CI (#1293)
refactor test_expert_load_balancer to keep the ut code style

This PR also fixed the break change from
https://github.com/vllm-project/vllm/pull/16188/files#diff-e2942ece30a5c580437694ffb964bfc664b510c59244c08e5921b8f5cefb4280

This is just a quick fix. We'll support embedding on V1 later

Closes: https://github.com/vllm-project/vllm-ascend/issues/1299

Signed-off-by: wangxiyuan <wangxiyuan1007@gmail.com>
2025-06-20 01:02:52 +08:00
..
e2e
[UT] refactor test_expert_load_balancer and fix broken CI (#1293)
2025-06-20 01:02:52 +08:00
multicard
[DP][V1] Fix rank set in DP scenario & Bump torch-npu version to 2.5.1.post1.dev20250528 (#1235)
2025-06-16 23:09:53 +08:00
ut
[UT] refactor test_expert_load_balancer and fix broken CI (#1293)
2025-06-20 01:02:52 +08:00
__init__.py
[SpecDecode] Add spec decode support (#500)
2025-04-17 20:16:32 +08:00
conftest.py
[CI/UT][Graph] Add ut for torchair graph mode (#1103)
2025-06-14 16:59:00 +08:00
model_utils.py
[CI] Refactor CI (#952)
2025-05-28 06:31:35 +08:00
utils.py
[ModelRunner] Support embedding inputs (#916)
2025-06-06 20:21:13 +08:00
Powered by Gitea Version: 1.24.3 Page: 78ms Template: 6ms
English
Bahasa Indonesia Deutsch English Español Français Gaeilge Italiano Latviešu Magyar nyelv Nederlands Polski Português de Portugal Português do Brasil Suomi Svenska Türkçe Čeština Ελληνικά Български Русский Українська فارسی മലയാളം 日本語 简体中文 繁體中文(台灣) 繁體中文(香港) 한국어
Licenses API