Files
xc-llm-ascend/docs/source/community/user_stories/llamafactory.md
zhangxinyuehfad 75de3fa172 [v0.11.0][Doc] Update doc (#3852)
### What this PR does / why we need it?
Update doc


Signed-off-by: hfadzxy <starmoon_zhang@163.com>
2025-10-29 11:32:12 +08:00

1.1 KiB
Raw Blame History

LLaMA-Factory

Introduction

LLaMA-Factory is an easy-to-use and efficient platform for training and fine-tuning large language models. With LLaMA-Factory, you can fine-tune hundreds of pre-trained models locally without writing any code.

LLaMA-Facotory users need to evaluate and inference the model after fine-tuning.

Business challenge

LLaMA-Factory uses Transformers to perform inference on Ascend NPUs, but the speed is slow.

Benefits with vLLM Ascend

With the joint efforts of LLaMA-Factory and vLLM Ascend (LLaMA-Factory#7739), LLaMA-Factory has achieved significant performance gains during model inference. Benchmark results show that its inference speed is now up to 2× faster compared to the Transformers implementation.

Learn more

See more details about LLaMA-Factory and how it uses vLLM Ascend for inference on Ascend NPUs in LLaMA-Factory Ascend NPU Inference.