v0.10.1rc1

This commit is contained in:
2025-09-09 09:40:35 +08:00
parent d6f6ef41fe
commit 9149384e03
432 changed files with 84698 additions and 1 deletions

View File

@@ -0,0 +1,290 @@
# SOME DESCRIPTIVE TITLE.
# Copyright (C) 2025, vllm-ascend team
# This file is distributed under the same license as the vllm-ascend
# package.
# FIRST AUTHOR <EMAIL@ADDRESS>, 2025.
#
#, fuzzy
msgid ""
msgstr ""
"Project-Id-Version: vllm-ascend\n"
"Report-Msgid-Bugs-To: \n"
"POT-Creation-Date: 2025-07-18 09:01+0800\n"
"PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n"
"Last-Translator: FULL NAME <EMAIL@ADDRESS>\n"
"Language-Team: zh_CN <LL@li.org>\n"
"Language: zh_CN\n"
"MIME-Version: 1.0\n"
"Content-Type: text/plain; charset=utf-8\n"
"Content-Transfer-Encoding: 8bit\n"
"Plural-Forms: nplurals=1; plural=0;\n"
"Generated-By: Babel 2.17.0\n"
#: ../../user_guide/configuration/additional_config.md:1
msgid "Additional Configuration"
msgstr "附加配置"
#: ../../user_guide/configuration/additional_config.md:3
msgid ""
"additional configuration is a mechanism provided by vLLM to allow plugins to"
" control inner behavior by their own. vLLM Ascend uses this mechanism to "
"make the project more flexible."
msgstr "额外配置是 vLLM 提供的一种机制允许插件自行控制内部行为。vLLM Ascend 利用这种机制使项目更加灵活。"
#: ../../user_guide/configuration/additional_config.md:5
msgid "How to use"
msgstr "如何使用"
#: ../../user_guide/configuration/additional_config.md:7
msgid ""
"With either online mode or offline mode, users can use additional "
"configuration. Take Qwen3 as an example:"
msgstr "无论是在线模式还是离线模式,用户都可以使用额外的配置。以 Qwen3 为例:"
#: ../../user_guide/configuration/additional_config.md:9
msgid "**Online mode**:"
msgstr "**在线模式**"
#: ../../user_guide/configuration/additional_config.md:15
msgid "**Offline mode**:"
msgstr "**离线模式**"
#: ../../user_guide/configuration/additional_config.md:23
msgid "Configuration options"
msgstr "配置选项"
#: ../../user_guide/configuration/additional_config.md:25
msgid ""
"The following table lists the additional configuration options available in "
"vLLM Ascend:"
msgstr "下表列出了 vLLM Ascend 中可用的其他配置选项:"
#: ../../user_guide/configuration/additional_config.md
msgid "Name"
msgstr "名称"
#: ../../user_guide/configuration/additional_config.md
msgid "Type"
msgstr "类型"
#: ../../user_guide/configuration/additional_config.md
msgid "Default"
msgstr "默认"
#: ../../user_guide/configuration/additional_config.md
msgid "Description"
msgstr "描述"
#: ../../user_guide/configuration/additional_config.md
msgid "`torchair_graph_config`"
msgstr "`torchair_graph_config`"
#: ../../user_guide/configuration/additional_config.md
msgid "dict"
msgstr "dict"
#: ../../user_guide/configuration/additional_config.md
#, python-brace-format
msgid "`{}`"
msgstr "`{}`"
#: ../../user_guide/configuration/additional_config.md
msgid "The config options for torchair graph mode"
msgstr "torchair 图模式的配置选项"
#: ../../user_guide/configuration/additional_config.md
msgid "`ascend_scheduler_config`"
msgstr "`ascend_scheduler_config`"
#: ../../user_guide/configuration/additional_config.md
msgid "The config options for ascend scheduler"
msgstr "ascend 调度器的配置选项"
#: ../../user_guide/configuration/additional_config.md
msgid "`expert_tensor_parallel_size`"
msgstr "`expert_tensor_parallel_size`"
#: ../../user_guide/configuration/additional_config.md
msgid "str"
msgstr "str"
#: ../../user_guide/configuration/additional_config.md
msgid "`0`"
msgstr "`0`"
#: ../../user_guide/configuration/additional_config.md
msgid "Expert tensor parallel size the model to use."
msgstr "专家张量并行的模型大小设置。"
#: ../../user_guide/configuration/additional_config.md
msgid "`refresh`"
msgstr "`刷新`"
#: ../../user_guide/configuration/additional_config.md
msgid "bool"
msgstr "bool"
#: ../../user_guide/configuration/additional_config.md
msgid "`false`"
msgstr "`false`"
#: ../../user_guide/configuration/additional_config.md
msgid ""
"Whether to refresh global ascend config content. This value is usually used "
"by rlhf or ut/e2e test case."
msgstr "是否刷新全局 ascend 配置信息。此值通常由 rlhf 或 ut/e2e 测试用例使用。"
#: ../../user_guide/configuration/additional_config.md
msgid "`expert_map_path`"
msgstr "`expert_map_path`"
#: ../../user_guide/configuration/additional_config.md
msgid "`None`"
msgstr "`None`"
#: ../../user_guide/configuration/additional_config.md
msgid ""
"When using expert load balancing for the MOE model, an expert map path needs"
" to be passed in."
msgstr "在为MOE模型使用专家负载均衡时需要传入专家映射路径。"
#: ../../user_guide/configuration/additional_config.md
msgid "`chunked_prefill_for_mla`"
msgstr "`chunked_prefill_for_mla`"
#: ../../user_guide/configuration/additional_config.md
msgid "`False`"
msgstr "`False`"
#: ../../user_guide/configuration/additional_config.md
msgid "Whether to enable the fused operator-like chunked_prefill."
msgstr "是否启用类似算子融合的 chunked_prefill 功能。"
#: ../../user_guide/configuration/additional_config.md
msgid "`kv_cache_dtype`"
msgstr "`kv_cache_dtype`"
#: ../../user_guide/configuration/additional_config.md
msgid ""
"When using the kv cache quantization method, kv cache dtype needs to be set,"
" currently only int8 is supported."
msgstr "当使用kv缓存量化方法时需要设置kv缓存的数据类型目前仅支持int8。"
#: ../../user_guide/configuration/additional_config.md:37
msgid "The details of each config option are as follows:"
msgstr "每个配置选项的详细信息如下:"
#: ../../user_guide/configuration/additional_config.md:39
msgid "**torchair_graph_config**"
msgstr "**torchair_graph_config**"
#: ../../user_guide/configuration/additional_config.md
msgid "`enabled`"
msgstr "`启用`"
#: ../../user_guide/configuration/additional_config.md
msgid ""
"Whether to enable torchair graph mode. Currently only DeepSeek series models"
" and PanguProMoE are supported to use torchair graph mode"
msgstr "是否启用 torchair 图模式。目前仅支持 DeepSeek 系列模型和 PanguProMoE 使用 torchair 图模式。"
#: ../../user_guide/configuration/additional_config.md
msgid "`enable_multistream_mla`"
msgstr "`enable_multistream_mla`"
#: ../../user_guide/configuration/additional_config.md
msgid ""
"Whether to put vector ops of MLA to another stream. This option only takes "
"effects on models using MLA (e.g., DeepSeek)."
msgstr "是否将MLA的向量操作放到另一个流中。此选项仅对使用MLA的模型例如DeepSeek有效。"
#: ../../user_guide/configuration/additional_config.md
msgid "`enable_multistream_moe`"
msgstr "`enable_multistream_moe`"
#: ../../user_guide/configuration/additional_config.md
msgid ""
"Whether to enable multistream shared expert. This option only takes effects "
"on DeepSeek moe models."
msgstr "是否启用多流共享专家功能。此选项仅对 DeepSeek MoE 模型生效。"
#: ../../user_guide/configuration/additional_config.md
msgid "`enable_view_optimize`"
msgstr "`enable_view_optimize` (启用视图优化)"
#: ../../user_guide/configuration/additional_config.md
msgid "`True`"
msgstr "`True`"
#: ../../user_guide/configuration/additional_config.md
msgid "Whether to enable torchair view optimization"
msgstr "是否启用torchair视图优化"
#: ../../user_guide/configuration/additional_config.md
msgid "`use_cached_graph`"
msgstr "`use_cached_graph`"
#: ../../user_guide/configuration/additional_config.md
msgid "Whether to use cached graph"
msgstr "是否使用缓存的图"
#: ../../user_guide/configuration/additional_config.md
msgid "`graph_batch_sizes`"
msgstr "`graph_batch_sizes`"
#: ../../user_guide/configuration/additional_config.md
msgid "list[int]"
msgstr "list[int]"
#: ../../user_guide/configuration/additional_config.md
msgid "`[]`"
msgstr "`[]`"
#: ../../user_guide/configuration/additional_config.md
msgid "The batch size for torchair graph cache"
msgstr "torchair 图缓存的批量大小"
#: ../../user_guide/configuration/additional_config.md
msgid "`graph_batch_sizes_init`"
msgstr "`graph_batch_sizes_init`"
#: ../../user_guide/configuration/additional_config.md
msgid "Init graph batch size dynamically if `graph_batch_sizes` is empty"
msgstr "如果 `graph_batch_sizes` 为空,则动态初始化图批大小"
#: ../../user_guide/configuration/additional_config.md
msgid "`enable_kv_nz`"
msgstr "`enable_kv_nz`"
#: ../../user_guide/configuration/additional_config.md
msgid ""
"Whether to enable kvcache NZ layout. This option only takes effects on "
"models using MLA (e.g., DeepSeek)."
msgstr "是否启用 kvcache NZ 布局。此选项仅对使用 MLA 的模型(例如 DeepSeek生效。"
#: ../../user_guide/configuration/additional_config.md:52
msgid "**ascend_scheduler_config**"
msgstr "**ascend_scheduler_config**"
#: ../../user_guide/configuration/additional_config.md
msgid "Whether to enable ascend scheduler for V1 engine"
msgstr "是否为 V1 引擎启用 ascend 调度器"
#: ../../user_guide/configuration/additional_config.md:58
msgid ""
"ascend_scheduler_config also support the options from [vllm scheduler "
"config](https://docs.vllm.ai/en/stable/api/vllm/config.html#vllm.config.SchedulerConfig)."
" For example, you can add `enable_chunked_prefill: True` to "
"ascend_scheduler_config as well."
msgstr ""
"ascend_scheduler_config 也支持来自 [vllm scheduler "
"config](https://docs.vllm.ai/en/stable/api/vllm/config.html#vllm.config.SchedulerConfig)"
" 的选项。例如,你也可以在 ascend_scheduler_config 中添加 `enable_chunked_prefill: True`。"
#: ../../user_guide/configuration/additional_config.md:60
msgid "Example"
msgstr "示例"
#: ../../user_guide/configuration/additional_config.md:62
msgid "An example of additional configuration is as follows:"
msgstr "以下是额外配置的一个示例:"

View File

@@ -0,0 +1,28 @@
# Translations template for PROJECT.
# Copyright (C) 2025 ORGANIZATION
# This file is distributed under the same license as the PROJECT project.
# FIRST AUTHOR <EMAIL@ADDRESS>, 2025.
#
#, fuzzy
msgid ""
msgstr ""
"Project-Id-Version: PROJECT VERSION\n"
"Report-Msgid-Bugs-To: EMAIL@ADDRESS\n"
"POT-Creation-Date: 2025-07-18 09:01+0800\n"
"PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n"
"Last-Translator: FULL NAME <EMAIL@ADDRESS>\n"
"Language-Team: LANGUAGE <LL@li.org>\n"
"MIME-Version: 1.0\n"
"Content-Type: text/plain; charset=utf-8\n"
"Content-Transfer-Encoding: 8bit\n"
"Generated-By: Babel 2.17.0\n"
#: ../../user_guide/configuration/env_vars.md:1
msgid "Environment Variables"
msgstr "环境变量"
#: ../../user_guide/configuration/env_vars.md:3
msgid ""
"vllm-ascend uses the following environment variables to configure the "
"system:"
msgstr "vllm-ascend 使用以下环境变量来配置系统:"

View File

@@ -0,0 +1,30 @@
# SOME DESCRIPTIVE TITLE.
# Copyright (C) 2025, vllm-ascend team
# This file is distributed under the same license as the vllm-ascend
# package.
# FIRST AUTHOR <EMAIL@ADDRESS>, 2025.
#
#, fuzzy
msgid ""
msgstr ""
"Project-Id-Version: vllm-ascend\n"
"Report-Msgid-Bugs-To: \n"
"POT-Creation-Date: 2025-07-18 09:01+0800\n"
"PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n"
"Last-Translator: FULL NAME <EMAIL@ADDRESS>\n"
"Language-Team: zh_CN <LL@li.org>\n"
"Language: zh_CN\n"
"MIME-Version: 1.0\n"
"Content-Type: text/plain; charset=utf-8\n"
"Content-Transfer-Encoding: 8bit\n"
"Plural-Forms: nplurals=1; plural=0;\n"
"Generated-By: Babel 2.17.0\n"
#: ../../user_guide/configuration/index.md:1
#: ../../user_guide/configuration/index.md:5
msgid "Configuration Guide"
msgstr "配置指南"
#: ../../user_guide/configuration/index.md:3
msgid "This section provides a detailed configuration guide of vLLM Ascend."
msgstr "本节提供了 vLLM Ascend 的详细配置指南。"

View File

@@ -0,0 +1,121 @@
# Translations template for PROJECT.
# Copyright (C) 2025 ORGANIZATION
# This file is distributed under the same license as the PROJECT project.
# FIRST AUTHOR <EMAIL@ADDRESS>, 2025.
#
#, fuzzy
msgid ""
msgstr ""
"Project-Id-Version: PROJECT VERSION\n"
"Report-Msgid-Bugs-To: EMAIL@ADDRESS\n"
"POT-Creation-Date: 2025-07-18 09:01+0800\n"
"PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n"
"Last-Translator: FULL NAME <EMAIL@ADDRESS>\n"
"Language-Team: LANGUAGE <LL@li.org>\n"
"MIME-Version: 1.0\n"
"Content-Type: text/plain; charset=utf-8\n"
"Content-Transfer-Encoding: 8bit\n"
"Generated-By: Babel 2.17.0\n"
#: ../../user_guide/feature_guide/graph_mode.md:1
msgid "Graph Mode Guide"
msgstr "图模式指南"
#: ../../user_guide/feature_guide/graph_mode.md:4
msgid ""
"This feature is currently experimental. In future versions, there may be "
"behavioral changes around configuration, coverage, performance improvement."
msgstr "此功能目前为实验性功能。在未来的版本中,配置、覆盖率和性能改进等方面的行为可能会有变化。"
#: ../../user_guide/feature_guide/graph_mode.md:7
msgid ""
"This guide provides instructions for using Ascend Graph Mode with vLLM "
"Ascend. Please note that graph mode is only available on V1 Engine. And only"
" Qwen, DeepSeek series models are well tested from 0.9.0rc1. We'll make it "
"stable and generalize in the next release."
msgstr ""
"本指南提供了在 vLLM Ascend 上使用 Ascend 图模式的操作说明。请注意,图模式仅在 V1 引擎上可用,并且从 0.9.0rc1 起,仅对"
" Qwen、DeepSeek 系列模型进行了充分测试。我们将在下一个版本中使其更加稳定和通用。"
#: ../../user_guide/feature_guide/graph_mode.md:9
msgid "Getting Started"
msgstr "快速入门"
#: ../../user_guide/feature_guide/graph_mode.md:11
msgid ""
"From v0.9.1rc1 with V1 Engine, vLLM Ascend will run models in graph mode by "
"default to keep the same behavior with vLLM. If you hit any issues, please "
"feel free to open an issue on GitHub and fallback to eager mode temporarily "
"by set `enforce_eager=True` when initializing the model."
msgstr ""
"从 v0.9.1rc1 版本起,使用 V1 引擎时vLLM Ascend 默认将在图模式下运行模型,以保持与 vLLM "
"同样的行为。如果遇到任何问题,欢迎在 GitHub 上提交 issue并在初始化模型时通过设置 `enforce_eager=True` 临时切换回 "
"eager 模式。"
#: ../../user_guide/feature_guide/graph_mode.md:13
msgid "There are two kinds for graph mode supported by vLLM Ascend:"
msgstr "vLLM Ascend 支持两种图模式:"
#: ../../user_guide/feature_guide/graph_mode.md:14
msgid ""
"**ACLGraph**: This is the default graph mode supported by vLLM Ascend. In "
"v0.9.1rc1, only Qwen series models are well tested."
msgstr ""
"**ACLGraph**:这是 vLLM Ascend 支持的默认图模式。在 v0.9.1rc1 版本中,只有 Qwen 系列模型得到了充分测试。"
#: ../../user_guide/feature_guide/graph_mode.md:15
msgid ""
"**TorchAirGraph**: This is the GE graph mode. In v0.9.1rc1, only DeepSeek "
"series models are supported."
msgstr "**TorchAirGraph**这是GE图模式。在v0.9.1rc1版本中仅支持DeepSeek系列模型。"
#: ../../user_guide/feature_guide/graph_mode.md:17
msgid "Using ACLGraph"
msgstr "使用 ACLGraph"
#: ../../user_guide/feature_guide/graph_mode.md:18
msgid ""
"ACLGraph is enabled by default. Take Qwen series models as an example, just "
"set to use V1 Engine is enough."
msgstr "ACLGraph 默认启用。以 Qwen 系列模型为例,只需设置为使用 V1 引擎即可。"
#: ../../user_guide/feature_guide/graph_mode.md:20
#: ../../user_guide/feature_guide/graph_mode.md:41
#: ../../user_guide/feature_guide/graph_mode.md:64
msgid "offline example:"
msgstr "离线示例:"
#: ../../user_guide/feature_guide/graph_mode.md:31
#: ../../user_guide/feature_guide/graph_mode.md:52
#: ../../user_guide/feature_guide/graph_mode.md:74
msgid "online example:"
msgstr "在线示例:"
#: ../../user_guide/feature_guide/graph_mode.md:37
msgid "Using TorchAirGraph"
msgstr "使用 TorchAirGraph"
#: ../../user_guide/feature_guide/graph_mode.md:39
msgid ""
"If you want to run DeepSeek series models with graph mode, you should use "
"[TorchAirGraph](https://www.hiascend.com/document/detail/zh/Pytorch/700/modthirdparty/torchairuseguide/torchair_0002.html)."
" In this case, additional config is required."
msgstr ""
"如果你想通过图模式运行 DeepSeek 系列模型,你应该使用 "
"[TorchAirGraph](https://www.hiascend.com/document/detail/zh/Pytorch/700/modthirdparty/torchairuseguide/torchair_0002.html)。在这种情况下,需要额外的配置。"
#: ../../user_guide/feature_guide/graph_mode.md:58
msgid ""
"You can find more detail about additional config "
"[here](../configuration/additional_config.md)."
msgstr "你可以在[这里](../configuration/additional_config.md)找到关于附加配置的更多详细信息。"
#: ../../user_guide/feature_guide/graph_mode.md:60
msgid "Fallback to Eager Mode"
msgstr "回退到 Eager 模式"
#: ../../user_guide/feature_guide/graph_mode.md:62
msgid ""
"If both `ACLGraph` and `TorchAirGraph` fail to run, you should fallback to "
"eager mode."
msgstr "如果 `ACLGraph` 和 `TorchAirGraph` 都无法运行,你应该退回到 eager 模式。"

View File

@@ -0,0 +1,30 @@
# SOME DESCRIPTIVE TITLE.
# Copyright (C) 2025, vllm-ascend team
# This file is distributed under the same license as the vllm-ascend
# package.
# FIRST AUTHOR <EMAIL@ADDRESS>, 2025.
#
#, fuzzy
msgid ""
msgstr ""
"Project-Id-Version: vllm-ascend\n"
"Report-Msgid-Bugs-To: \n"
"POT-Creation-Date: 2025-07-18 09:01+0800\n"
"PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n"
"Last-Translator: FULL NAME <EMAIL@ADDRESS>\n"
"Language-Team: zh_CN <LL@li.org>\n"
"Language: zh_CN\n"
"MIME-Version: 1.0\n"
"Content-Type: text/plain; charset=utf-8\n"
"Content-Transfer-Encoding: 8bit\n"
"Plural-Forms: nplurals=1; plural=0;\n"
"Generated-By: Babel 2.17.0\n"
#: ../../user_guide/feature_guide/index.md:1
#: ../../user_guide/feature_guide/index.md:5
msgid "Feature Guide"
msgstr "功能指南"
#: ../../user_guide/feature_guide/index.md:3
msgid "This section provides a detailed usage guide of vLLM Ascend features."
msgstr "本节提供了 vLLM Ascend 功能的详细使用指南。"

View File

@@ -0,0 +1,58 @@
# SOME DESCRIPTIVE TITLE.
# Copyright (C) 2025, vllm-ascend team
# This file is distributed under the same license as the vllm-ascend
# package.
# FIRST AUTHOR <EMAIL@ADDRESS>, 2025.
#
#, fuzzy
msgid ""
msgstr ""
"Project-Id-Version: vllm-ascend\n"
"Report-Msgid-Bugs-To: \n"
"POT-Creation-Date: 2025-07-18 09:01+0800\n"
"PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n"
"Last-Translator: FULL NAME <EMAIL@ADDRESS>\n"
"Language-Team: zh_CN <LL@li.org>\n"
"Language: zh_CN\n"
"MIME-Version: 1.0\n"
"Content-Type: text/plain; charset=utf-8\n"
"Content-Transfer-Encoding: 8bit\n"
"Plural-Forms: nplurals=1; plural=0;\n"
"Generated-By: Babel 2.17.0\n"
#: ../../user_guide/feature_guide/lora.md:1
msgid "LoRA Adapters Guide"
msgstr "LoRA 适配器指南"
#: ../../user_guide/feature_guide/lora.md:3
msgid ""
"Like vLLM, vllm-ascend supports LoRA as well. The usage and more details can"
" be found in [vLLM official "
"document](https://docs.vllm.ai/en/latest/features/lora.html)."
msgstr ""
"与 vLLM 类似vllm-ascend 也支持 LoRA。用法及更多详情可参见 [vLLM "
"官方文档](https://docs.vllm.ai/en/latest/features/lora.html)。"
#: ../../user_guide/feature_guide/lora.md:5
msgid ""
"You can also refer to "
"[this](https://docs.vllm.ai/en/latest/models/supported_models.html#list-of-"
"text-only-language-models) to find which models support LoRA in vLLM."
msgstr ""
"你也可以参考[这个链接](https://docs.vllm.ai/en/latest/models/supported_models.html#list-"
"of-text-only-language-models)来查找哪些模型在 vLLM 中支持 LoRA。"
#: ../../user_guide/feature_guide/lora.md:7
msgid "Tips"
msgstr "提示"
#: ../../user_guide/feature_guide/lora.md:8
msgid ""
"If you fail to run vllm-ascend with LoRA, you may follow [this "
"instruction](https://vllm-"
"ascend.readthedocs.io/en/latest/user_guide/feature_guide/graph_mode.html#fallback-"
"to-eager-mode) to disable graph mode and try again."
msgstr ""
"如果你在使用 LoRA 运行 vllm-ascend 时失败,可以按照[此说明](https://vllm-"
"ascend.readthedocs.io/en/latest/user_guide/feature_guide/graph_mode.html#fallback-"
"to-eager-mode)禁用图模式后再重试。"

View File

@@ -0,0 +1,183 @@
# SOME DESCRIPTIVE TITLE.
# Copyright (C) 2025, vllm-ascend team
# This file is distributed under the same license as the vllm-ascend
# package.
# FIRST AUTHOR <EMAIL@ADDRESS>, 2025.
#
#, fuzzy
msgid ""
msgstr ""
"Project-Id-Version: vllm-ascend\n"
"Report-Msgid-Bugs-To: \n"
"POT-Creation-Date: 2025-07-18 09:01+0800\n"
"PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n"
"Last-Translator: FULL NAME <EMAIL@ADDRESS>\n"
"Language-Team: zh_CN <LL@li.org>\n"
"Language: zh_CN\n"
"MIME-Version: 1.0\n"
"Content-Type: text/plain; charset=utf-8\n"
"Content-Transfer-Encoding: 8bit\n"
"Plural-Forms: nplurals=1; plural=0;\n"
"Generated-By: Babel 2.17.0\n"
#: ../../user_guide/feature_guide/quantization.md:1
msgid "Quantization Guide"
msgstr "量化指南"
#: ../../user_guide/feature_guide/quantization.md:3
msgid ""
"Model quantization is a technique that reduces the size and computational "
"requirements of a model by lowering the data precision of the weights and "
"activation values in the model, thereby saving the memory and improving the "
"inference speed."
msgstr "模型量化是一种通过降低模型中权重和激活值的数据精度,从而减少模型大小和计算需求的技术,这样可以节省内存并提高推理速度。"
#: ../../user_guide/feature_guide/quantization.md:5
msgid ""
"Since 0.9.0rc2 version, quantization feature is experimentally supported in "
"vLLM Ascend. Users can enable quantization feature by specifying "
"`--quantization ascend`. Currently, only Qwen, DeepSeek series models are "
"well tested. Well support more quantization algorithm and models in the "
"future."
msgstr ""
"自 0.9.0rc2 版本起vLLM Ascend 实验性地支持量化特性。用户可以通过指定 `--quantization ascend` "
"启用量化功能。目前,只有 Qwen、DeepSeek 系列模型经过了充分测试。未来我们将支持更多的量化算法和模型。"
#: ../../user_guide/feature_guide/quantization.md:7
msgid "Install modelslim"
msgstr "安装 modelslim"
#: ../../user_guide/feature_guide/quantization.md:9
msgid ""
"To quantize a model, users should install "
"[ModelSlim](https://gitee.com/ascend/msit/blob/master/msmodelslim/README.md)"
" which is the Ascend compression and acceleration tool. It is an affinity-"
"based compression tool designed for acceleration, using compression as its "
"core technology and built upon the Ascend platform."
msgstr ""
"要对模型进行量化,用户应安装[ModelSlim](https://gitee.com/ascend/msit/blob/master/msmodelslim/README.md),这是昇腾的压缩与加速工具。它是一种基于亲和性的压缩工具,专为加速设计,以压缩为核心技术,并基于昇腾平台构建。"
#: ../../user_guide/feature_guide/quantization.md:11
msgid ""
"Currently, only the specific tag [modelslim-"
"VLLM-8.1.RC1.b020_001](https://gitee.com/ascend/msit/blob/modelslim-"
"VLLM-8.1.RC1.b020_001/msmodelslim/README.md) of modelslim works with vLLM "
"Ascend. Please do not install other version until modelslim master version "
"is available for vLLM Ascend in the future."
msgstr ""
"目前,只有 modelslim 的特定标签 [modelslim-"
"VLLM-8.1.RC1.b020_001](https://gitee.com/ascend/msit/blob/modelslim-"
"VLLM-8.1.RC1.b020_001/msmodelslim/README.md) 支持 vLLM Ascend。在未来 modelslim "
"的主版本支持 vLLM Ascend 之前,请不要安装其他版本。"
#: ../../user_guide/feature_guide/quantization.md:13
msgid "Install modelslim:"
msgstr "安装 modelslim"
#: ../../user_guide/feature_guide/quantization.md:21
msgid "Quantize model"
msgstr "量化模型"
#: ../../user_guide/feature_guide/quantization.md:23
#, python-format
msgid ""
"Take [DeepSeek-V2-Lite](https://modelscope.cn/models/deepseek-"
"ai/DeepSeek-V2-Lite) as an example, you just need to download the model, and"
" then execute the convert command. The command is shown below. More info can"
" be found in modelslim doc [deepseek w8a8 dynamic quantization "
"docs](https://gitee.com/ascend/msit/blob/modelslim-"
"VLLM-8.1.RC1.b020_001/msmodelslim/example/DeepSeek/README.md#deepseek-v2-w8a8-dynamic%E9%87%8F%E5%8C%96)."
msgstr ""
"以 [DeepSeek-V2-Lite](https://modelscope.cn/models/deepseek-"
"ai/DeepSeek-V2-Lite) 为例,你只需要下载模型,然后执行转换命令。命令如下所示。更多信息可参考 modelslim 文档 "
"[deepseek w8a8 动态量化文档](https://gitee.com/ascend/msit/blob/modelslim-"
"VLLM-8.1.RC1.b020_001/msmodelslim/example/DeepSeek/README.md#deepseek-v2-w8a8-dynamic%E9%87%8F%E5%8C%96)。"
#: ../../user_guide/feature_guide/quantization.md:32
msgid ""
"You can also download the quantized model that we uploaded. Please note that"
" these weights should be used for test only. For example, "
"https://www.modelscope.cn/models/vllm-ascend/DeepSeek-V2-Lite-W8A8"
msgstr ""
"你也可以下载我们上传的量化模型。请注意这些权重仅应用于测试。例如https://www.modelscope.cn/models/vllm-"
"ascend/DeepSeek-V2-Lite-W8A8"
#: ../../user_guide/feature_guide/quantization.md:35
msgid "Once convert action is done, there are two important files generated."
msgstr "转换操作完成后,会生成两个重要的文件。"
#: ../../user_guide/feature_guide/quantization.md:37
msgid ""
"[config.json](https://www.modelscope.cn/models/vllm-"
"ascend/DeepSeek-V2-Lite-W8A8/file/view/master/config.json?status=1). Please "
"make sure that there is no `quantization_config` field in it."
msgstr ""
"[config.json](https://www.modelscope.cn/models/vllm-"
"ascend/DeepSeek-V2-Lite-W8A8/file/view/master/config.json?status=1)。请确保其中没有 "
"`quantization_config` 字段。"
#: ../../user_guide/feature_guide/quantization.md:39
msgid ""
"[quant_model_description.json](https://www.modelscope.cn/models/vllm-"
"ascend/DeepSeek-V2-Lite-W8A8/file/view/master/quant_model_description.json?status=1)."
" All the converted weights info are recorded in this file."
msgstr ""
"[quant_model_description.json](https://www.modelscope.cn/models/vllm-"
"ascend/DeepSeek-V2-Lite-W8A8/file/view/master/quant_model_description.json?status=1)。所有被转换的权重信息都记录在该文件中。"
#: ../../user_guide/feature_guide/quantization.md:41
msgid "Here is the full converted model files:"
msgstr "以下是完整转换后的模型文件:"
#: ../../user_guide/feature_guide/quantization.md:60
msgid "Run the model"
msgstr "运行模型"
#: ../../user_guide/feature_guide/quantization.md:62
msgid ""
"Now, you can run the quantized models with vLLM Ascend. Here is the example "
"for online and offline inference."
msgstr "现在,你可以使用 vLLM Ascend 运行量化模型。下面是在线和离线推理的示例。"
#: ../../user_guide/feature_guide/quantization.md:64
msgid "Offline inference"
msgstr "离线推理"
#: ../../user_guide/feature_guide/quantization.md:90
msgid "Online inference"
msgstr "在线推理"
#: ../../user_guide/feature_guide/quantization.md:97
msgid "FAQs"
msgstr "常见问题解答"
#: ../../user_guide/feature_guide/quantization.md:99
msgid ""
"1. How to solve the KeyError: 'xxx.layers.0.self_attn.q_proj.weight' "
"problem?"
msgstr "1. 如何解决 KeyError: 'xxx.layers.0.self_attn.q_proj.weight' 问题?"
#: ../../user_guide/feature_guide/quantization.md:101
msgid ""
"First, make sure you specify `ascend` quantization method. Second, check if "
"your model is converted by this `modelslim-VLLM-8.1.RC1.b020_001` modelslim "
"version. Finally, if it still doesn't work, please submit a issue, maybe "
"some new models need to be adapted."
msgstr ""
"首先,请确保你指定了 `ascend` 量化方法。其次,检查你的模型是否由 `modelslim-VLLM-8.1.RC1.b020_001` 这个 "
"modelslim 版本转换。如果仍然无法使用,请提交一个 issue可能有一些新模型需要适配。"
#: ../../user_guide/feature_guide/quantization.md:104
msgid ""
"2. How to solve the error \"Could not locate the "
"configuration_deepseek.py\"?"
msgstr "2. 如何解决“无法找到 configuration_deepseek.py”错误"
#: ../../user_guide/feature_guide/quantization.md:106
msgid ""
"Please convert DeepSeek series models using `modelslim-"
"VLLM-8.1.RC1.b020_001` modelslim, this version has fixed the missing "
"configuration_deepseek.py error."
msgstr ""
"请使用 `modelslim-VLLM-8.1.RC1.b020_001` 的 modelslim 转换 DeepSeek 系列模型,该版本已修复缺少 "
"configuration_deepseek.py 的错误。"

View File

@@ -0,0 +1,156 @@
# SOME DESCRIPTIVE TITLE.
# Copyright (C) 2025, vllm-ascend team
# This file is distributed under the same license as the vllm-ascend
# package.
# FIRST AUTHOR <EMAIL@ADDRESS>, 2025.
#
#, fuzzy
msgid ""
msgstr ""
"Project-Id-Version: vllm-ascend\n"
"Report-Msgid-Bugs-To: \n"
"POT-Creation-Date: 2025-07-18 09:01+0800\n"
"PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n"
"Last-Translator: FULL NAME <EMAIL@ADDRESS>\n"
"Language-Team: zh_CN <LL@li.org>\n"
"Language: zh_CN\n"
"MIME-Version: 1.0\n"
"Content-Type: text/plain; charset=utf-8\n"
"Content-Transfer-Encoding: 8bit\n"
"Plural-Forms: nplurals=1; plural=0;\n"
"Generated-By: Babel 2.17.0\n"
#: ../../user_guide/feature_guide/sleep_mode.md:1
msgid "Sleep Mode Guide"
msgstr "睡眠模式指南"
#: ../../user_guide/feature_guide/sleep_mode.md:3
msgid "Overview"
msgstr "概述"
#: ../../user_guide/feature_guide/sleep_mode.md:5
msgid ""
"Sleep Mode is an API designed to offload model weights and discard KV cache "
"from NPU memory. This functionality is essential for reinforcement learning "
"(RL) post-training workloads, particularly in online algorithms such as PPO,"
" GRPO, or DPO. During training, the policy model typically performs auto-"
"regressive generation using inference engines like vLLM, followed by forward"
" and backward passes for optimization."
msgstr ""
"Sleep Mode 是一个用于卸载模型权重并清除 NPU 内存中 KV 缓存的 API。此功能对于强化学习RL后训练任务尤其重要特别是在 "
"PPO、GRPO 或 DPO 等在线算法中。在训练过程中,策略模型通常会使用像 vLLM "
"这样的推理引擎进行自回归生成,然后进行前向和反向传播以进行优化。"
#: ../../user_guide/feature_guide/sleep_mode.md:7
msgid ""
"Since the generation and training phases may employ different model "
"parallelism strategies, it becomes crucial to free KV cache and even offload"
" model parameters stored within vLLM during training. This ensures efficient"
" memory utilization and avoids resource contention on the NPU."
msgstr ""
"由于生成和训练阶段可能采用不同的模型并行策略,因此在训练过程中及时释放 KV 缓存,甚至卸载存储在 vLLM "
"内的模型参数变得至关重要。这可以确保内存的高效利用,并避免 NPU 上的资源争用。"
#: ../../user_guide/feature_guide/sleep_mode.md:10
msgid "Getting started"
msgstr "快速上手"
#: ../../user_guide/feature_guide/sleep_mode.md:12
#, python-brace-format
msgid ""
"With `enable_sleep_mode=True`, the way we manage memory(malloc, free) in "
"vllm will under a specific memory pool, during loading model and initialize "
"kv_caches, we tag the memory as a map: `{\"weight\": data, \"kv_cache\": "
"data}`."
msgstr ""
"当 `enable_sleep_mode=True` 时,我们在 vllm 中管理内存malloc, "
"free的方式会在一个特定的内存池下进行在加载模型和初始化 kv_caches "
"期间,我们会将内存打上标签,组织成一个映射:`{\"weight\": data, \"kv_cache\": data}`。"
#: ../../user_guide/feature_guide/sleep_mode.md:14
msgid ""
"The engine(v0/v1) supports two sleep levels to manage memory during idle "
"periods:"
msgstr "该引擎v0/v1支持两种睡眠等级以在空闲期间管理内存"
#: ../../user_guide/feature_guide/sleep_mode.md:16
msgid "Level 1 Sleep"
msgstr "一级睡眠"
#: ../../user_guide/feature_guide/sleep_mode.md:17
msgid "Action: Offloads model weights and discards the KV cache."
msgstr "操作卸载模型权重并清除KV缓存。"
#: ../../user_guide/feature_guide/sleep_mode.md:18
msgid "Memory: Model weights are moved to CPU memory; KV cache is forgotten."
msgstr "内存模型权重被移动到CPU内存KV缓存被清除。"
#: ../../user_guide/feature_guide/sleep_mode.md:19
msgid "Use Case: Suitable when reusing the same model later."
msgstr "用例:适用于之后需要重复使用同一个模型的情况。"
#: ../../user_guide/feature_guide/sleep_mode.md:20
msgid ""
"Note: Ensure sufficient CPU memory is available to hold the model weights."
msgstr "注意请确保有足够的CPU内存来存储模型权重。"
#: ../../user_guide/feature_guide/sleep_mode.md:22
msgid "Level 2 Sleep"
msgstr "二级睡眠"
#: ../../user_guide/feature_guide/sleep_mode.md:23
msgid "Action: Discards both model weights and KV cache."
msgstr "操作同时丢弃模型权重和KV缓存。"
#: ../../user_guide/feature_guide/sleep_mode.md:24
msgid ""
"Memory: The content of both the model weights and kv cache is forgotten."
msgstr "内存模型权重和kv缓存的内容都会被遗忘。"
#: ../../user_guide/feature_guide/sleep_mode.md:25
msgid ""
"Use Case: Ideal when switching to a different model or updating the current "
"one."
msgstr "用例:当切换到不同的模型或更新当前模型时非常理想。"
#: ../../user_guide/feature_guide/sleep_mode.md:27
msgid ""
"Since this feature uses the low-level API "
"[AscendCL](https://www.hiascend.com/document/detail/zh/CANNCommunityEdition/82RC1alpha002/API/appdevgapi/appdevgapi_07_0000.html),"
" in order to use sleep mode, you should follow the [installation "
"guide](https://vllm-ascend.readthedocs.io/en/latest/installation.html) and "
"building from source, if you are using v0.7.3, remember to set `export "
"COMPILE_CUSTOM_KERNELS=1`, for the latest version(v0.9.x+), the environment "
"variable `COMPILE_CUSTOM_KERNELS` will be set 1 by default while building "
"from source."
msgstr ""
"由于此功能使用了底层 API "
"[AscendCL](https://www.hiascend.com/document/detail/zh/CANNCommunityEdition/82RC1alpha002/API/appdevgapi/appdevgapi_07_0000.html),为了使用休眠模式,你应按照[安装指南](https://vllm-"
"ascend.readthedocs.io/en/latest/installation.html)进行操作,并从源码编译。如果你使用的是 "
"v0.7.3,请记得设置 `export COMPILE_CUSTOM_KERNELS=1` 对于最新版本v0.9.x+),在从源码编译时环境变量 "
"`COMPILE_CUSTOM_KERNELS` 默认会被设置为 1。"
#: ../../user_guide/feature_guide/sleep_mode.md:29
msgid "Usage"
msgstr "用法"
#: ../../user_guide/feature_guide/sleep_mode.md:31
msgid "The following is a simple example of how to use sleep mode."
msgstr "以下是如何使用睡眠模式的一个简单示例。"
#: ../../user_guide/feature_guide/sleep_mode.md:33
msgid "offline inference:"
msgstr "离线推理:"
#: ../../user_guide/feature_guide/sleep_mode.md:72
msgid "online serving:"
msgstr "在线服务:"
#: ../../user_guide/feature_guide/sleep_mode.md:74
msgid ""
"Considering there may be a risk of malicious access, please make sure you "
"are under a dev-mode, and explicit specify the develop env: "
"`VLLM_SERVER_DEV_MODE` to expose these endpoints(sleep/wake up)."
msgstr ""
"鉴于可能存在恶意访问的风险,请确保您处于开发模式,并明确指定开发环境:`VLLM_SERVER_DEV_MODE`以便开放这些端点sleep/wake"
" up。"

View File

@@ -0,0 +1,220 @@
# SOME DESCRIPTIVE TITLE.
# Copyright (C) 2025, vllm-ascend team
# This file is distributed under the same license as the vllm-ascend
# package.
# FIRST AUTHOR <EMAIL@ADDRESS>, 2025.
#
#, fuzzy
msgid ""
msgstr ""
"Project-Id-Version: vllm-ascend\n"
"Report-Msgid-Bugs-To: \n"
"POT-Creation-Date: 2025-07-18 09:01+0800\n"
"PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n"
"Last-Translator: FULL NAME <EMAIL@ADDRESS>\n"
"Language-Team: zh_CN <LL@li.org>\n"
"Language: zh_CN\n"
"MIME-Version: 1.0\n"
"Content-Type: text/plain; charset=utf-8\n"
"Content-Transfer-Encoding: 8bit\n"
"Plural-Forms: nplurals=1; plural=0;\n"
"Generated-By: Babel 2.17.0\n"
#: ../../user_guide/feature_guide/structured_output.md:1
msgid "Structured Output Guide"
msgstr "结构化输出指南"
#: ../../user_guide/feature_guide/structured_output.md:3
msgid "Overview"
msgstr "概述"
#: ../../user_guide/feature_guide/structured_output.md:5
msgid "What is Structured Output?"
msgstr "什么是结构化输出?"
#: ../../user_guide/feature_guide/structured_output.md:7
msgid ""
"LLMs can be unpredictable when you need output in specific formats. Think of"
" asking a model to generate JSON - without guidance, it might produce valid "
"text that breaks JSON specification. **Structured Output (also called Guided"
" Decoding)** enables LLMs to generate outputs that follow a desired "
"structure while preserving the non-deterministic nature of the system."
msgstr ""
"当你需要特定格式输出时大型语言模型LLMs可能表现出不可预测性。比如让模型生成 "
"JSON如果没有指导模型可能会生成有效的文本但这些文本却不符合 JSON 规范。**结构化输出(也称为引导解码)** "
"能让大型语言模型生成符合预期结构的输出,同时保留系统的非确定性特性。"
#: ../../user_guide/feature_guide/structured_output.md:9
msgid ""
"In simple terms, structured decoding gives LLMs a “template” to follow. "
"Users provide a schema that “influences” the models output, ensuring "
"compliance with the desired structure."
msgstr "简单来说结构化解码为LLM提供了一个“模板”来遵循。用户提供一个模式来“影响”模型的输出从而确保输出符合期望的结构。"
#: ../../user_guide/feature_guide/structured_output.md:11
msgid "![structured decoding](./images/structured_output_1.png)"
msgstr "![结构化解码](./images/structured_output_1.png)"
#: ../../user_guide/feature_guide/structured_output.md:11
msgid "structured decoding"
msgstr "结构化解码"
#: ../../user_guide/feature_guide/structured_output.md:13
msgid "Structured Output in vllm-ascend"
msgstr "vllm-ascend 中的结构化输出"
#: ../../user_guide/feature_guide/structured_output.md:15
msgid ""
"Currently, vllm-ascend supports **xgrammar** and **guidance** backend for "
"structured output with vllm v1 engine."
msgstr "目前vllm-ascend 支持 vllm v1 引擎的结构化输出,后端包括 **xgrammar** 和 **guidance**。"
#: ../../user_guide/feature_guide/structured_output.md:17
msgid ""
"XGrammar introduces a new technique that batch constrained decoding via "
"pushdown automaton (PDA). You can think of a PDA as a “collection of FSMs, "
"and each FSM represents a context-free grammar (CFG).” One significant "
"advantage of PDA is its recursive nature, allowing us to execute multiple "
"state transitions. They also include additional optimisation (for those who "
"are interested) to reduce grammar compilation overhead. Besides, you can "
"also find more details about guidance by yourself."
msgstr ""
"XGrammar 引入了一种通过下推自动机PDA进行批量约束解码的新技术。你可以把 PDA 理解为“有限状态机FSM的集合每个 FSM "
"代表一个上下文无关文法CFG。” PDA 的一个重要优点是其递归特性使我们能够执行多次状态转移。此外PDA "
"还包含了额外的优化(供感兴趣的用户参考),以减少语法编译的开销。除此之外,你还可以自己找到更多关于指导的信息。"
#: ../../user_guide/feature_guide/structured_output.md:19
msgid "How to Use Structured Output?"
msgstr "如何使用结构化输出?"
#: ../../user_guide/feature_guide/structured_output.md:21
msgid "Online Inference"
msgstr "在线推理"
#: ../../user_guide/feature_guide/structured_output.md:23
msgid ""
"You can also generate structured outputs using the OpenAI's Completions and "
"Chat API. The following parameters are supported, which must be added as "
"extra parameters:"
msgstr "你也可以使用 OpenAI 的 Completions 和 Chat API 生成结构化输出。支持以下参数,这些参数必须作为额外参数添加:"
#: ../../user_guide/feature_guide/structured_output.md:25
msgid "`guided_choice`: the output will be exactly one of the choices."
msgstr "`guided_choice`:输出将会是其中一个选项。"
#: ../../user_guide/feature_guide/structured_output.md:26
msgid "`guided_regex`: the output will follow the regex pattern."
msgstr "`guided_regex`:输出将遵循正则表达式模式。"
#: ../../user_guide/feature_guide/structured_output.md:27
msgid "`guided_json`: the output will follow the JSON schema."
msgstr "`guided_json`:输出将遵循 JSON 架构。"
#: ../../user_guide/feature_guide/structured_output.md:28
msgid "`guided_grammar`: the output will follow the context free grammar."
msgstr "`guided_grammar`:输出将遵循上下文无关文法。"
#: ../../user_guide/feature_guide/structured_output.md:30
msgid ""
"Structured outputs are supported by default in the OpenAI-Compatible Server."
" You can choose to specify the backend to use by setting the `--guided-"
"decoding-backend` flag to vllm serve. The default backend is `auto`, which "
"will try to choose an appropriate backend based on the details of the "
"request. You may also choose a specific backend, along with some options."
msgstr ""
"OpenAI 兼容服务器默认支持结构化输出。你可以通过设置 `--guided-decoding-backend` 标志为 vllm serve "
"来指定要使用的后端。默认后端为 `auto`,它会根据请求的详细信息尝试选择合适的后端。你也可以选择特定的后端,并设置一些选项。"
#: ../../user_guide/feature_guide/structured_output.md:32
msgid ""
"Now let´s see an example for each of the cases, starting with the "
"guided_choice, as it´s the easiest one:"
msgstr "现在让我们来看每种情况的示例,首先是 guided_choice因为它是最简单的"
#: ../../user_guide/feature_guide/structured_output.md:51
msgid ""
"The next example shows how to use the guided_regex. The idea is to generate "
"an email address, given a simple regex template:"
msgstr "下一个例子展示了如何使用 guided_regex。其思路是基于一个简单的正则表达式模板生成一个电子邮件地址"
#: ../../user_guide/feature_guide/structured_output.md:67
msgid ""
"One of the most relevant features in structured text generation is the "
"option to generate a valid JSON with pre-defined fields and formats. For "
"this we can use the guided_json parameter in two different ways:"
msgstr ""
"在结构化文本生成中,最相关的特性之一是能够生成具有预定义字段和格式的有效 JSON。为此我们可以通过两种不同的方式使用 guided_json 参数:"
#: ../../user_guide/feature_guide/structured_output.md:69
msgid "Using a JSON Schema."
msgstr "使用 JSON 架构。"
#: ../../user_guide/feature_guide/structured_output.md:70
msgid "Defining a Pydantic model and then extracting the JSON Schema from it."
msgstr "定义一个 Pydantic 模型,然后从中提取 JSON Schema。"
#: ../../user_guide/feature_guide/structured_output.md:72
msgid ""
"The next example shows how to use the guided_json parameter with a Pydantic "
"model:"
msgstr "下一个示例展示了如何将 guided_json 参数与 Pydantic 模型一起使用:"
#: ../../user_guide/feature_guide/structured_output.md:104
msgid ""
"Finally we have the guided_grammar option, which is probably the most "
"difficult to use, but it´s really powerful. It allows us to define complete "
"languages like SQL queries. It works by using a context free EBNF grammar. "
"As an example, we can use to define a specific format of simplified SQL "
"queries:"
msgstr ""
"最后,我们有 guided_grammar 选项,这可能是最难使用的,但它非常强大。它允许我们定义完整的语言,比如 SQL 查询。它通过使用上下文无关的"
" EBNF 语法来实现。例如,我们可以用它来定义一种简化 SQL 查询的特定格式:"
#: ../../user_guide/feature_guide/structured_output.md:134
msgid ""
"Find more examples [here](https://github.com/vllm-"
"project/vllm/blob/main/examples/offline_inference/structured_outputs.py)."
msgstr ""
"在[这里](https://github.com/vllm-"
"project/vllm/blob/main/examples/offline_inference/structured_outputs.py)可以找到更多示例。"
#: ../../user_guide/feature_guide/structured_output.md:136
msgid "Offline Inference"
msgstr "离线推理"
#: ../../user_guide/feature_guide/structured_output.md:138
msgid ""
"To use Structured Output, we'll need to configure the guided decoding using "
"the class `GuidedDecodingParams` inside `SamplingParams`. The main available"
" options inside `GuidedDecodingParams` are:"
msgstr ""
"要使用结构化输出,我们需要在 `SamplingParams` 内通过 `GuidedDecodingParams` "
"类配置引导解码。`GuidedDecodingParams` 中主要可用的选项有:"
#: ../../user_guide/feature_guide/structured_output.md:140
msgid "json"
msgstr "json"
#: ../../user_guide/feature_guide/structured_output.md:141
msgid "regex"
msgstr "正则表达式"
#: ../../user_guide/feature_guide/structured_output.md:142
msgid "choice"
msgstr "选择"
#: ../../user_guide/feature_guide/structured_output.md:143
msgid "grammar"
msgstr "语法"
#: ../../user_guide/feature_guide/structured_output.md:145
msgid "One example for the usage of the choice parameter is shown below:"
msgstr "choice 参数用法的一个示例如下:"
#: ../../user_guide/feature_guide/structured_output.md:163
msgid ""
"Find more examples of other usages [here](https://github.com/vllm-"
"project/vllm/blob/main/examples/offline_inference/structured_outputs.py)."
msgstr ""
"查看更多其他用法的示例 [在这里](https://github.com/vllm-"
"project/vllm/blob/main/examples/offline_inference/structured_outputs.py)。"

File diff suppressed because it is too large Load Diff

View File

@@ -0,0 +1,30 @@
# Translations template for PROJECT.
# Copyright (C) 2025 ORGANIZATION
# This file is distributed under the same license as the PROJECT project.
# FIRST AUTHOR <EMAIL@ADDRESS>, 2025.
#
#, fuzzy
msgid ""
msgstr ""
"Project-Id-Version: PROJECT VERSION\n"
"Report-Msgid-Bugs-To: EMAIL@ADDRESS\n"
"POT-Creation-Date: 2025-07-18 09:01+0800\n"
"PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n"
"Last-Translator: FULL NAME <EMAIL@ADDRESS>\n"
"Language-Team: LANGUAGE <LL@li.org>\n"
"MIME-Version: 1.0\n"
"Content-Type: text/plain; charset=utf-8\n"
"Content-Transfer-Encoding: 8bit\n"
"Generated-By: Babel 2.17.0\n"
#: ../../user_guide/support_matrix/index.md:5
msgid "Support Matrix"
msgstr "支持矩阵"
#: ../../user_guide/support_matrix/index.md:1
msgid "Features and models"
msgstr "特性与模型"
#: ../../user_guide/support_matrix/index.md:3
msgid "This section provides a detailed supported matrix by vLLM Ascend."
msgstr "本节提供了 vLLM Ascend 的详细支持矩阵。"

View File

@@ -0,0 +1,264 @@
# SOME DESCRIPTIVE TITLE.
# Copyright (C) 2025, vllm-ascend team
# This file is distributed under the same license as the vllm-ascend
# package.
# FIRST AUTHOR <EMAIL@ADDRESS>, 2025.
#
#, fuzzy
msgid ""
msgstr ""
"Project-Id-Version: vllm-ascend\n"
"Report-Msgid-Bugs-To: \n"
"POT-Creation-Date: 2025-07-18 09:01+0800\n"
"PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n"
"Last-Translator: FULL NAME <EMAIL@ADDRESS>\n"
"Language-Team: zh_CN <LL@li.org>\n"
"Language: zh_CN\n"
"MIME-Version: 1.0\n"
"Content-Type: text/plain; charset=utf-8\n"
"Content-Transfer-Encoding: 8bit\n"
"Plural-Forms: nplurals=1; plural=0;\n"
"Generated-By: Babel 2.17.0\n"
#: ../../user_guide/support_matrix/supported_features.md:1
msgid "Feature Support"
msgstr "功能支持"
#: ../../user_guide/support_matrix/supported_features.md:3
msgid ""
"The feature support principle of vLLM Ascend is: **aligned with the vLLM**. "
"We are also actively collaborating with the community to accelerate support."
msgstr "vLLM Ascend 的特性支持原则是:**与 vLLM 保持一致**。我们也在积极与社区合作,加快支持进度。"
#: ../../user_guide/support_matrix/supported_features.md:5
msgid ""
"You can check the [support status of vLLM V1 Engine][v1_user_guide]. Below "
"is the feature support status of vLLM Ascend:"
msgstr "你可以查看 [vLLM V1 引擎的支持状态][v1_user_guide]。下面是 vLLM Ascend 的功能支持情况:"
#: ../../user_guide/support_matrix/supported_features.md
msgid "Feature"
msgstr "特性"
#: ../../user_guide/support_matrix/supported_features.md
msgid "vLLM V0 Engine"
msgstr "vLLM V0 引擎"
#: ../../user_guide/support_matrix/supported_features.md
msgid "vLLM V1 Engine"
msgstr "vLLM V1 引擎"
#: ../../user_guide/support_matrix/supported_features.md
msgid "Next Step"
msgstr "下一步"
#: ../../user_guide/support_matrix/supported_features.md
msgid "Chunked Prefill"
msgstr "分块预填充"
#: ../../user_guide/support_matrix/supported_features.md
msgid "🟢 Functional"
msgstr "🟢 功能性"
#: ../../user_guide/support_matrix/supported_features.md
msgid "Functional, see detail note: [Chunked Prefill][cp]"
msgstr "功能性,详见说明:[分块预填充][cp]"
#: ../../user_guide/support_matrix/supported_features.md
msgid "Automatic Prefix Caching"
msgstr "自动前缀缓存"
#: ../../user_guide/support_matrix/supported_features.md
msgid "Functional, see detail note: [vllm-ascend#732][apc]"
msgstr "可用,请参见详细说明:[vllm-ascend#732][apc]"
#: ../../user_guide/support_matrix/supported_features.md
msgid "LoRA"
msgstr "LoRA"
#: ../../user_guide/support_matrix/supported_features.md
msgid "[vllm-ascend#396][multilora], [vllm-ascend#893][v1 multilora]"
msgstr "[vllm-ascend#396][multilora][vllm-ascend#893][v1 multilora]"
#: ../../user_guide/support_matrix/supported_features.md
msgid "Prompt adapter"
msgstr "提示适配器"
#: ../../user_guide/support_matrix/supported_features.md
msgid "🔴 No plan"
msgstr "🔴 无计划"
#: ../../user_guide/support_matrix/supported_features.md
msgid "This feature has been deprecated by vllm."
msgstr "此功能已被 vllm 弃用。"
#: ../../user_guide/support_matrix/supported_features.md
msgid "Speculative decoding"
msgstr "猜测式解码"
#: ../../user_guide/support_matrix/supported_features.md
msgid "Basic support"
msgstr "基础支持"
#: ../../user_guide/support_matrix/supported_features.md
msgid "Pooling"
msgstr "池化"
#: ../../user_guide/support_matrix/supported_features.md
msgid "🟡 Planned"
msgstr "🟡 计划中"
#: ../../user_guide/support_matrix/supported_features.md
msgid "CI needed and adapting more models; V1 support rely on vLLM support."
msgstr "需要持续集成CI并适配更多模型V1 的支持依赖于 vLLM 的支持。"
#: ../../user_guide/support_matrix/supported_features.md
msgid "Enc-dec"
msgstr "Enc-dec编码-解码)"
#: ../../user_guide/support_matrix/supported_features.md
msgid "🔴 NO plan"
msgstr "🔴 没有计划"
#: ../../user_guide/support_matrix/supported_features.md
msgid "Plan in 2025.06.30"
msgstr "2025.06.30 的计划"
#: ../../user_guide/support_matrix/supported_features.md
msgid "Multi Modality"
msgstr "多模态"
#: ../../user_guide/support_matrix/supported_features.md
msgid "[Tutorial][multimodal], optimizing and adapting more models"
msgstr "[教程][multimodal],优化和适配更多模型"
#: ../../user_guide/support_matrix/supported_features.md
msgid "LogProbs"
msgstr "LogProbs"
#: ../../user_guide/support_matrix/supported_features.md
msgid "CI needed"
msgstr "需要持续集成CI"
#: ../../user_guide/support_matrix/supported_features.md
msgid "Prompt logProbs"
msgstr "提示 logProbs"
#: ../../user_guide/support_matrix/supported_features.md
msgid "Async output"
msgstr "异步输出"
#: ../../user_guide/support_matrix/supported_features.md
msgid "Multi step scheduler"
msgstr "多步调度器"
#: ../../user_guide/support_matrix/supported_features.md
msgid "🔴 Deprecated"
msgstr "🔴 已弃用"
#: ../../user_guide/support_matrix/supported_features.md
msgid "[vllm#8779][v1_rfc], replaced by [vLLM V1 Scheduler][v1_scheduler]"
msgstr "[vllm#8779][v1_rfc],已被 [vLLM V1 调度器][v1_scheduler] 替代"
#: ../../user_guide/support_matrix/supported_features.md
msgid "Best of"
msgstr "精选"
#: ../../user_guide/support_matrix/supported_features.md
msgid "[vllm#13361][best_of], CI needed"
msgstr "[vllm#13361][best_of]需要持续集成CI"
#: ../../user_guide/support_matrix/supported_features.md
msgid "Beam search"
msgstr "束搜索"
#: ../../user_guide/support_matrix/supported_features.md
msgid "Guided Decoding"
msgstr "引导解码"
#: ../../user_guide/support_matrix/supported_features.md
msgid "[vllm-ascend#177][guided_decoding]"
msgstr "[vllm-ascend#177][guided_decoding]"
#: ../../user_guide/support_matrix/supported_features.md
msgid "Tensor Parallel"
msgstr "张量并行"
#: ../../user_guide/support_matrix/supported_features.md
msgid "Pipeline Parallel"
msgstr "流水线并行"
#: ../../user_guide/support_matrix/supported_features.md
msgid "Expert Parallel"
msgstr "专家并行"
#: ../../user_guide/support_matrix/supported_features.md
msgid "CI needed; No plan on V0 support"
msgstr "需要持续集成没有支持V0的计划"
#: ../../user_guide/support_matrix/supported_features.md
msgid "Data Parallel"
msgstr "数据并行"
#: ../../user_guide/support_matrix/supported_features.md
msgid "CI needed; No plan on V0 support"
msgstr "需要 CI暂无 V0 支持计划"
#: ../../user_guide/support_matrix/supported_features.md
msgid "Prefill Decode Disaggregation"
msgstr "预填充 解码 拆分"
#: ../../user_guide/support_matrix/supported_features.md
msgid "1P1D available, working on xPyD and V1 support."
msgstr "1P1D 已可用,正在开发 xPyD 和 V1 支持。"
#: ../../user_guide/support_matrix/supported_features.md
msgid "Quantization"
msgstr "量化"
#: ../../user_guide/support_matrix/supported_features.md
msgid "W8A8 available, CI needed; working on more quantization method support"
msgstr "W8A8 已可用需要持续集成CI正在开发对更多量化方法的支持。"
#: ../../user_guide/support_matrix/supported_features.md
msgid "Graph Mode"
msgstr "图模式"
#: ../../user_guide/support_matrix/supported_features.md
msgid "🔵 Experimental"
msgstr "🔵 实验性"
#: ../../user_guide/support_matrix/supported_features.md
msgid "Experimental, see detail note: [vllm-ascend#767][graph_mode]"
msgstr "实验性功能,详见说明:[vllm-ascend#767][graph_mode]"
#: ../../user_guide/support_matrix/supported_features.md
msgid "Sleep Mode"
msgstr "睡眠模式"
#: ../../user_guide/support_matrix/supported_features.md
msgid "level=1 available, CI needed, working on V1 support"
msgstr "level=1 可用需要CI正在开发 V1 支持"
#: ../../user_guide/support_matrix/supported_features.md:33
msgid "🟢 Functional: Fully operational, with ongoing optimizations."
msgstr "🟢 功能性:完全可用,正在持续优化中。"
#: ../../user_guide/support_matrix/supported_features.md:34
msgid ""
"🔵 Experimental: Experimental support, interfaces and functions may change."
msgstr "🔵 实验性:实验性支持,接口和功能可能会发生变化。"
#: ../../user_guide/support_matrix/supported_features.md:35
msgid "🚧 WIP: Under active development, will be supported soon."
msgstr "🚧 WIP正在积极开发中很快将会支持。"
#: ../../user_guide/support_matrix/supported_features.md:36
msgid ""
"🟡 Planned: Scheduled for future implementation (some may have open "
"PRs/RFCs)."
msgstr "🟡 计划中已安排将来实现其中一些可能已有开放的PR/RFC。"
#: ../../user_guide/support_matrix/supported_features.md:37
msgid "🔴 NO plan / Deprecated: No plan for V0 or deprecated by vLLM v1."
msgstr "🔴 没有计划 / 已弃用V0 没有计划或已被 vLLM v1 弃用。"

View File

@@ -0,0 +1,214 @@
# SOME DESCRIPTIVE TITLE.
# Copyright (C) 2025, vllm-ascend team
# This file is distributed under the same license as the vllm-ascend
# package.
# FIRST AUTHOR <EMAIL@ADDRESS>, 2025.
#
#, fuzzy
msgid ""
msgstr ""
"Project-Id-Version: vllm-ascend\n"
"Report-Msgid-Bugs-To: \n"
"POT-Creation-Date: 2025-07-18 09:01+0800\n"
"PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n"
"Last-Translator: FULL NAME <EMAIL@ADDRESS>\n"
"Language-Team: zh_CN <LL@li.org>\n"
"Language: zh_CN\n"
"MIME-Version: 1.0\n"
"Content-Type: text/plain; charset=utf-8\n"
"Content-Transfer-Encoding: 8bit\n"
"Plural-Forms: nplurals=1; plural=0;\n"
"Generated-By: Babel 2.17.0\n"
#: ../../user_guide/support_matrix/supported_models.md:1
msgid "Model Support"
msgstr "模型支持"
#: ../../user_guide/support_matrix/supported_models.md:3
msgid "Text-only Language Models"
msgstr "纯文本语言模型"
#: ../../user_guide/support_matrix/supported_models.md:5
#: ../../user_guide/support_matrix/supported_models.md:38
msgid "Generative Models"
msgstr "生成模型"
#: ../../user_guide/support_matrix/supported_models.md
msgid "Model"
msgstr "模型"
#: ../../user_guide/support_matrix/supported_models.md
msgid "Supported"
msgstr "支持"
#: ../../user_guide/support_matrix/supported_models.md
msgid "Note"
msgstr "注释"
#: ../../user_guide/support_matrix/supported_models.md
msgid "DeepSeek v3"
msgstr "DeepSeek v3"
#: ../../user_guide/support_matrix/supported_models.md
msgid "✅"
msgstr "✅"
#: ../../user_guide/support_matrix/supported_models.md
msgid "DeepSeek R1"
msgstr "DeepSeek R1"
#: ../../user_guide/support_matrix/supported_models.md
msgid "DeepSeek Distill (Qwen/LLama)"
msgstr "DeepSeek 精炼Qwen/LLama"
#: ../../user_guide/support_matrix/supported_models.md
msgid "Qwen3"
msgstr "Qwen3"
#: ../../user_guide/support_matrix/supported_models.md
msgid "Qwen3-Moe"
msgstr "Qwen3-Moe"
#: ../../user_guide/support_matrix/supported_models.md
msgid "Qwen2.5"
msgstr "Qwen2.5"
#: ../../user_guide/support_matrix/supported_models.md
msgid "QwQ-32B"
msgstr "QwQ-32B"
#: ../../user_guide/support_matrix/supported_models.md
msgid "LLama3.1/3.2"
msgstr "LLama3.1/3.2"
#: ../../user_guide/support_matrix/supported_models.md
msgid "Internlm"
msgstr "Internlm"
#: ../../user_guide/support_matrix/supported_models.md
msgid "Baichuan"
msgstr "百川"
#: ../../user_guide/support_matrix/supported_models.md
msgid "Phi-4-mini"
msgstr "Phi-4-mini"
#: ../../user_guide/support_matrix/supported_models.md
msgid "MiniCPM"
msgstr "MiniCPM"
#: ../../user_guide/support_matrix/supported_models.md
msgid "MiniCPM3"
msgstr "MiniCPM3"
#: ../../user_guide/support_matrix/supported_models.md
msgid "LLama4"
msgstr "LLama4"
#: ../../user_guide/support_matrix/supported_models.md
msgid "Mistral"
msgstr "Mistral"
#: ../../user_guide/support_matrix/supported_models.md
msgid "Need test"
msgstr "需要测试"
#: ../../user_guide/support_matrix/supported_models.md
msgid "DeepSeek v2.5"
msgstr "DeepSeek v2.5"
#: ../../user_guide/support_matrix/supported_models.md
msgid "Gemma-2"
msgstr "Gemma-2"
#: ../../user_guide/support_matrix/supported_models.md
msgid "Mllama"
msgstr "Mllama"
#: ../../user_guide/support_matrix/supported_models.md
msgid "Gemma-3"
msgstr "Gemma-3"
#: ../../user_guide/support_matrix/supported_models.md
msgid "❌"
msgstr "❌"
#: ../../user_guide/support_matrix/supported_models.md
msgid "[#496](https://github.com/vllm-project/vllm-ascend/issues/496)"
msgstr "[#496](https://github.com/vllm-project/vllm-ascend/issues/496)"
#: ../../user_guide/support_matrix/supported_models.md
msgid "ChatGLM"
msgstr "ChatGLM"
#: ../../user_guide/support_matrix/supported_models.md
msgid "[#554](https://github.com/vllm-project/vllm-ascend/issues/554)"
msgstr "[#554](https://github.com/vllm-project/vllm-ascend/issues/554)"
#: ../../user_guide/support_matrix/supported_models.md:29
msgid "Pooling Models"
msgstr "池化模型"
#: ../../user_guide/support_matrix/supported_models.md
msgid "XLM-RoBERTa-based"
msgstr "基于XLM-RoBERTa"
#: ../../user_guide/support_matrix/supported_models.md
msgid "Molmo"
msgstr "Molmo"
#: ../../user_guide/support_matrix/supported_models.md:36
msgid "Multimodal Language Models"
msgstr "多模态语言模型"
#: ../../user_guide/support_matrix/supported_models.md
msgid "Qwen2-VL"
msgstr "Qwen2-VL"
#: ../../user_guide/support_matrix/supported_models.md
msgid "Qwen2.5-VL"
msgstr "Qwen2.5-VL"
#: ../../user_guide/support_matrix/supported_models.md
msgid "LLaVA 1.5"
msgstr "LLaVA 1.5"
#: ../../user_guide/support_matrix/supported_models.md
msgid "LLaVA 1.6"
msgstr "LLaVA 1.6"
#: ../../user_guide/support_matrix/supported_models.md
msgid "[#553](https://github.com/vllm-project/vllm-ascend/issues/553)"
msgstr "[#553](https://github.com/vllm-project/vllm-ascend/issues/553)"
#: ../../user_guide/support_matrix/supported_models.md
msgid "InternVL2"
msgstr "InternVL2"
#: ../../user_guide/support_matrix/supported_models.md
msgid "InternVL2.5"
msgstr "InternVL2.5"
#: ../../user_guide/support_matrix/supported_models.md
msgid "Qwen2-Audio"
msgstr "Qwen2-Audio"
#: ../../user_guide/support_matrix/supported_models.md
msgid "LLaVA-Next"
msgstr "LLaVA-Next"
#: ../../user_guide/support_matrix/supported_models.md
msgid "LLaVA-Next-Video"
msgstr "LLaVA-Next-Video"
#: ../../user_guide/support_matrix/supported_models.md
msgid "Phi-3-Vison/Phi-3.5-Vison"
msgstr "Phi-3-Vison/Phi-3.5-Vison"
#: ../../user_guide/support_matrix/supported_models.md
msgid "GLM-4v"
msgstr "GLM-4v"
#: ../../user_guide/support_matrix/supported_models.md
msgid "Ultravox"
msgstr "Ultravox"