This website requires JavaScript.
Explore
Help
Register
Sign In
EngineX
/
xc-llm-ascend
Watch
3
Star
0
Fork
0
You've already forked xc-llm-ascend
Code
Issues
Pull Requests
Actions
Projects
Releases
Wiki
Activity
Files
4a2505f81f5b5f3eedeb58f08ab522e2c69aa4a5
xc-llm-ascend
/
vllm_ascend
/
worker
History
yiz-liu
5305a2ccf9
[Bugfix] Tweak distributed process group initialization and add dummy… (
#816
)
...
fix batch execution method to enable DP in V1 Signed-off-by: Yizhou Liu <
liu_yizhou@outlook.com
>
2025-05-12 17:31:29 +08:00
..
__init__.py
port deepseekv2 and mtp to main branch (
#429
)
2025-04-19 17:38:18 +08:00
cache_engine.py
support deepseek quant & mix-parallel with graphmode (
#585
)
2025-04-23 16:23:25 +08:00
draft_model_runner.py
[CI] upgrade vllm to 0.8.5 (
#715
)
2025-04-30 09:15:50 +08:00
model_runner_v1.py
[Bugfix] Add max_num_batched_tokens to InputBatch to make main CI pass (
#806
)
2025-05-12 00:36:56 +08:00
model_runner.py
[Core] Support the features of prefix cache and chunked prefill in v0/v1 (
#782
)
2025-05-09 16:39:28 +08:00
multi_step_runner.py
[CI] upgrade vllm to 0.8.5 (
#715
)
2025-04-30 09:15:50 +08:00
multi_step_worker.py
support multistep decode (
#299
)
2025-03-11 19:20:06 +08:00
pooling_model_runner.py
[MISC] Clean up torch_npu (
#688
)
2025-04-29 18:03:38 +08:00
worker_v1.py
[Bugfix] Tweak distributed process group initialization and add dummy… (
#816
)
2025-05-12 17:31:29 +08:00
worker.py
[Disaggregated Prefill] P2P Disaggregated Prefill based on llm_datadist (
#694
)
2025-05-01 22:31:36 +08:00