Files
xc-llm-ascend/docs/source/community/user_stories/index.md
Yikun Jiang 05dec7eda9 [Doc] Refactor and init user story page (#1224)
### What this PR does / why we need it?
This PR refactor the user stories page:
- Move it to community
- Add initial info of LLaMA-Factory, Huggingface/trl, MindIE Turbo,
GPUStack, verl
- Add a new page for LLaMA-Factory

### Does this PR introduce _any_ user-facing change?
No

### How was this patch tested?
Preview locally

Signed-off-by: Yikun Jiang <yikunkero@gmail.com>
2025-06-17 09:36:35 +08:00

1.9 KiB

User Stories

Read case studies on how users and developers solves real, everyday problems with vLLM Ascend

  • LLaMA-Factory is an easy-to-use and efficient platform for training and fine-tuning large language models, it supports vLLM Ascend to speed up inference since LLaMA-Factory#7739, gain 2x performance enhancement of inference.

  • Huggingface/trl is a cutting-edge library designed for post-training foundation models using advanced techniques like SFT, PPO and DPO, it uses vLLM Ascend since v0.17.0 to support RLHF on Ascend NPU.

  • MindIE Turbo is an LLM inference engine acceleration plug-in library developed by Huawei on Ascend hardware, which includes self-developed large language model optimization algorithms and optimizations related to the inference engine framework. It supports vLLM Ascend since 2.0rc1.

  • GPUStack is an open-source GPU cluster manager for running AI models. It supports vLLM Ascend since v0.6.2, see more GPUStack performance evaluation info on link.

  • verl is a flexible, efficient and production-ready RL training library for large language models (LLMs), uses vLLM Ascend since v0.4.0, see more info on verl x Ascend Quickstart.

:::{toctree} :caption: More details :maxdepth: 1 llamafactory :::