v0.10.1rc1
This commit is contained in:
@@ -0,0 +1,290 @@
|
||||
# SOME DESCRIPTIVE TITLE.
|
||||
# Copyright (C) 2025, vllm-ascend team
|
||||
# This file is distributed under the same license as the vllm-ascend
|
||||
# package.
|
||||
# FIRST AUTHOR <EMAIL@ADDRESS>, 2025.
|
||||
#
|
||||
#, fuzzy
|
||||
msgid ""
|
||||
msgstr ""
|
||||
"Project-Id-Version: vllm-ascend\n"
|
||||
"Report-Msgid-Bugs-To: \n"
|
||||
"POT-Creation-Date: 2025-07-18 09:01+0800\n"
|
||||
"PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n"
|
||||
"Last-Translator: FULL NAME <EMAIL@ADDRESS>\n"
|
||||
"Language-Team: zh_CN <LL@li.org>\n"
|
||||
"Language: zh_CN\n"
|
||||
"MIME-Version: 1.0\n"
|
||||
"Content-Type: text/plain; charset=utf-8\n"
|
||||
"Content-Transfer-Encoding: 8bit\n"
|
||||
"Plural-Forms: nplurals=1; plural=0;\n"
|
||||
"Generated-By: Babel 2.17.0\n"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md:1
|
||||
msgid "Additional Configuration"
|
||||
msgstr "附加配置"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md:3
|
||||
msgid ""
|
||||
"additional configuration is a mechanism provided by vLLM to allow plugins to"
|
||||
" control inner behavior by their own. vLLM Ascend uses this mechanism to "
|
||||
"make the project more flexible."
|
||||
msgstr "额外配置是 vLLM 提供的一种机制,允许插件自行控制内部行为。vLLM Ascend 利用这种机制使项目更加灵活。"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md:5
|
||||
msgid "How to use"
|
||||
msgstr "如何使用"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md:7
|
||||
msgid ""
|
||||
"With either online mode or offline mode, users can use additional "
|
||||
"configuration. Take Qwen3 as an example:"
|
||||
msgstr "无论是在线模式还是离线模式,用户都可以使用额外的配置。以 Qwen3 为例:"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md:9
|
||||
msgid "**Online mode**:"
|
||||
msgstr "**在线模式**:"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md:15
|
||||
msgid "**Offline mode**:"
|
||||
msgstr "**离线模式**:"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md:23
|
||||
msgid "Configuration options"
|
||||
msgstr "配置选项"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md:25
|
||||
msgid ""
|
||||
"The following table lists the additional configuration options available in "
|
||||
"vLLM Ascend:"
|
||||
msgstr "下表列出了 vLLM Ascend 中可用的其他配置选项:"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md
|
||||
msgid "Name"
|
||||
msgstr "名称"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md
|
||||
msgid "Type"
|
||||
msgstr "类型"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md
|
||||
msgid "Default"
|
||||
msgstr "默认"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md
|
||||
msgid "Description"
|
||||
msgstr "描述"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md
|
||||
msgid "`torchair_graph_config`"
|
||||
msgstr "`torchair_graph_config`"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md
|
||||
msgid "dict"
|
||||
msgstr "dict"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md
|
||||
#, python-brace-format
|
||||
msgid "`{}`"
|
||||
msgstr "`{}`"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md
|
||||
msgid "The config options for torchair graph mode"
|
||||
msgstr "torchair 图模式的配置选项"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md
|
||||
msgid "`ascend_scheduler_config`"
|
||||
msgstr "`ascend_scheduler_config`"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md
|
||||
msgid "The config options for ascend scheduler"
|
||||
msgstr "ascend 调度器的配置选项"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md
|
||||
msgid "`expert_tensor_parallel_size`"
|
||||
msgstr "`expert_tensor_parallel_size`"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md
|
||||
msgid "str"
|
||||
msgstr "str"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md
|
||||
msgid "`0`"
|
||||
msgstr "`0`"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md
|
||||
msgid "Expert tensor parallel size the model to use."
|
||||
msgstr "专家张量并行的模型大小设置。"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md
|
||||
msgid "`refresh`"
|
||||
msgstr "`刷新`"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md
|
||||
msgid "bool"
|
||||
msgstr "bool"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md
|
||||
msgid "`false`"
|
||||
msgstr "`false`"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md
|
||||
msgid ""
|
||||
"Whether to refresh global ascend config content. This value is usually used "
|
||||
"by rlhf or ut/e2e test case."
|
||||
msgstr "是否刷新全局 ascend 配置信息。此值通常由 rlhf 或 ut/e2e 测试用例使用。"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md
|
||||
msgid "`expert_map_path`"
|
||||
msgstr "`expert_map_path`"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md
|
||||
msgid "`None`"
|
||||
msgstr "`None`"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md
|
||||
msgid ""
|
||||
"When using expert load balancing for the MOE model, an expert map path needs"
|
||||
" to be passed in."
|
||||
msgstr "在为MOE模型使用专家负载均衡时,需要传入专家映射路径。"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md
|
||||
msgid "`chunked_prefill_for_mla`"
|
||||
msgstr "`chunked_prefill_for_mla`"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md
|
||||
msgid "`False`"
|
||||
msgstr "`False`"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md
|
||||
msgid "Whether to enable the fused operator-like chunked_prefill."
|
||||
msgstr "是否启用类似算子融合的 chunked_prefill 功能。"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md
|
||||
msgid "`kv_cache_dtype`"
|
||||
msgstr "`kv_cache_dtype`"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md
|
||||
msgid ""
|
||||
"When using the kv cache quantization method, kv cache dtype needs to be set,"
|
||||
" currently only int8 is supported."
|
||||
msgstr "当使用kv缓存量化方法时,需要设置kv缓存的数据类型,目前仅支持int8。"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md:37
|
||||
msgid "The details of each config option are as follows:"
|
||||
msgstr "每个配置选项的详细信息如下:"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md:39
|
||||
msgid "**torchair_graph_config**"
|
||||
msgstr "**torchair_graph_config**"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md
|
||||
msgid "`enabled`"
|
||||
msgstr "`启用`"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md
|
||||
msgid ""
|
||||
"Whether to enable torchair graph mode. Currently only DeepSeek series models"
|
||||
" and PanguProMoE are supported to use torchair graph mode"
|
||||
msgstr "是否启用 torchair 图模式。目前仅支持 DeepSeek 系列模型和 PanguProMoE 使用 torchair 图模式。"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md
|
||||
msgid "`enable_multistream_mla`"
|
||||
msgstr "`enable_multistream_mla`"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md
|
||||
msgid ""
|
||||
"Whether to put vector ops of MLA to another stream. This option only takes "
|
||||
"effects on models using MLA (e.g., DeepSeek)."
|
||||
msgstr "是否将MLA的向量操作放到另一个流中。此选项仅对使用MLA的模型(例如,DeepSeek)有效。"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md
|
||||
msgid "`enable_multistream_moe`"
|
||||
msgstr "`enable_multistream_moe`"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md
|
||||
msgid ""
|
||||
"Whether to enable multistream shared expert. This option only takes effects "
|
||||
"on DeepSeek moe models."
|
||||
msgstr "是否启用多流共享专家功能。此选项仅对 DeepSeek MoE 模型生效。"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md
|
||||
msgid "`enable_view_optimize`"
|
||||
msgstr "`enable_view_optimize` (启用视图优化)"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md
|
||||
msgid "`True`"
|
||||
msgstr "`True`"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md
|
||||
msgid "Whether to enable torchair view optimization"
|
||||
msgstr "是否启用torchair视图优化"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md
|
||||
msgid "`use_cached_graph`"
|
||||
msgstr "`use_cached_graph`"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md
|
||||
msgid "Whether to use cached graph"
|
||||
msgstr "是否使用缓存的图"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md
|
||||
msgid "`graph_batch_sizes`"
|
||||
msgstr "`graph_batch_sizes`"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md
|
||||
msgid "list[int]"
|
||||
msgstr "list[int]"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md
|
||||
msgid "`[]`"
|
||||
msgstr "`[]`"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md
|
||||
msgid "The batch size for torchair graph cache"
|
||||
msgstr "torchair 图缓存的批量大小"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md
|
||||
msgid "`graph_batch_sizes_init`"
|
||||
msgstr "`graph_batch_sizes_init`"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md
|
||||
msgid "Init graph batch size dynamically if `graph_batch_sizes` is empty"
|
||||
msgstr "如果 `graph_batch_sizes` 为空,则动态初始化图批大小"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md
|
||||
msgid "`enable_kv_nz`"
|
||||
msgstr "`enable_kv_nz`"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md
|
||||
msgid ""
|
||||
"Whether to enable kvcache NZ layout. This option only takes effects on "
|
||||
"models using MLA (e.g., DeepSeek)."
|
||||
msgstr "是否启用 kvcache NZ 布局。此选项仅对使用 MLA 的模型(例如 DeepSeek)生效。"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md:52
|
||||
msgid "**ascend_scheduler_config**"
|
||||
msgstr "**ascend_scheduler_config**"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md
|
||||
msgid "Whether to enable ascend scheduler for V1 engine"
|
||||
msgstr "是否为 V1 引擎启用 ascend 调度器"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md:58
|
||||
msgid ""
|
||||
"ascend_scheduler_config also support the options from [vllm scheduler "
|
||||
"config](https://docs.vllm.ai/en/stable/api/vllm/config.html#vllm.config.SchedulerConfig)."
|
||||
" For example, you can add `enable_chunked_prefill: True` to "
|
||||
"ascend_scheduler_config as well."
|
||||
msgstr ""
|
||||
"ascend_scheduler_config 也支持来自 [vllm scheduler "
|
||||
"config](https://docs.vllm.ai/en/stable/api/vllm/config.html#vllm.config.SchedulerConfig)"
|
||||
" 的选项。例如,你也可以在 ascend_scheduler_config 中添加 `enable_chunked_prefill: True`。"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md:60
|
||||
msgid "Example"
|
||||
msgstr "示例"
|
||||
|
||||
#: ../../user_guide/configuration/additional_config.md:62
|
||||
msgid "An example of additional configuration is as follows:"
|
||||
msgstr "以下是额外配置的一个示例:"
|
||||
@@ -0,0 +1,28 @@
|
||||
# Translations template for PROJECT.
|
||||
# Copyright (C) 2025 ORGANIZATION
|
||||
# This file is distributed under the same license as the PROJECT project.
|
||||
# FIRST AUTHOR <EMAIL@ADDRESS>, 2025.
|
||||
#
|
||||
#, fuzzy
|
||||
msgid ""
|
||||
msgstr ""
|
||||
"Project-Id-Version: PROJECT VERSION\n"
|
||||
"Report-Msgid-Bugs-To: EMAIL@ADDRESS\n"
|
||||
"POT-Creation-Date: 2025-07-18 09:01+0800\n"
|
||||
"PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n"
|
||||
"Last-Translator: FULL NAME <EMAIL@ADDRESS>\n"
|
||||
"Language-Team: LANGUAGE <LL@li.org>\n"
|
||||
"MIME-Version: 1.0\n"
|
||||
"Content-Type: text/plain; charset=utf-8\n"
|
||||
"Content-Transfer-Encoding: 8bit\n"
|
||||
"Generated-By: Babel 2.17.0\n"
|
||||
|
||||
#: ../../user_guide/configuration/env_vars.md:1
|
||||
msgid "Environment Variables"
|
||||
msgstr "环境变量"
|
||||
|
||||
#: ../../user_guide/configuration/env_vars.md:3
|
||||
msgid ""
|
||||
"vllm-ascend uses the following environment variables to configure the "
|
||||
"system:"
|
||||
msgstr "vllm-ascend 使用以下环境变量来配置系统:"
|
||||
@@ -0,0 +1,30 @@
|
||||
# SOME DESCRIPTIVE TITLE.
|
||||
# Copyright (C) 2025, vllm-ascend team
|
||||
# This file is distributed under the same license as the vllm-ascend
|
||||
# package.
|
||||
# FIRST AUTHOR <EMAIL@ADDRESS>, 2025.
|
||||
#
|
||||
#, fuzzy
|
||||
msgid ""
|
||||
msgstr ""
|
||||
"Project-Id-Version: vllm-ascend\n"
|
||||
"Report-Msgid-Bugs-To: \n"
|
||||
"POT-Creation-Date: 2025-07-18 09:01+0800\n"
|
||||
"PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n"
|
||||
"Last-Translator: FULL NAME <EMAIL@ADDRESS>\n"
|
||||
"Language-Team: zh_CN <LL@li.org>\n"
|
||||
"Language: zh_CN\n"
|
||||
"MIME-Version: 1.0\n"
|
||||
"Content-Type: text/plain; charset=utf-8\n"
|
||||
"Content-Transfer-Encoding: 8bit\n"
|
||||
"Plural-Forms: nplurals=1; plural=0;\n"
|
||||
"Generated-By: Babel 2.17.0\n"
|
||||
|
||||
#: ../../user_guide/configuration/index.md:1
|
||||
#: ../../user_guide/configuration/index.md:5
|
||||
msgid "Configuration Guide"
|
||||
msgstr "配置指南"
|
||||
|
||||
#: ../../user_guide/configuration/index.md:3
|
||||
msgid "This section provides a detailed configuration guide of vLLM Ascend."
|
||||
msgstr "本节提供了 vLLM Ascend 的详细配置指南。"
|
||||
@@ -0,0 +1,121 @@
|
||||
# Translations template for PROJECT.
|
||||
# Copyright (C) 2025 ORGANIZATION
|
||||
# This file is distributed under the same license as the PROJECT project.
|
||||
# FIRST AUTHOR <EMAIL@ADDRESS>, 2025.
|
||||
#
|
||||
#, fuzzy
|
||||
msgid ""
|
||||
msgstr ""
|
||||
"Project-Id-Version: PROJECT VERSION\n"
|
||||
"Report-Msgid-Bugs-To: EMAIL@ADDRESS\n"
|
||||
"POT-Creation-Date: 2025-07-18 09:01+0800\n"
|
||||
"PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n"
|
||||
"Last-Translator: FULL NAME <EMAIL@ADDRESS>\n"
|
||||
"Language-Team: LANGUAGE <LL@li.org>\n"
|
||||
"MIME-Version: 1.0\n"
|
||||
"Content-Type: text/plain; charset=utf-8\n"
|
||||
"Content-Transfer-Encoding: 8bit\n"
|
||||
"Generated-By: Babel 2.17.0\n"
|
||||
|
||||
#: ../../user_guide/feature_guide/graph_mode.md:1
|
||||
msgid "Graph Mode Guide"
|
||||
msgstr "图模式指南"
|
||||
|
||||
#: ../../user_guide/feature_guide/graph_mode.md:4
|
||||
msgid ""
|
||||
"This feature is currently experimental. In future versions, there may be "
|
||||
"behavioral changes around configuration, coverage, performance improvement."
|
||||
msgstr "此功能目前为实验性功能。在未来的版本中,配置、覆盖率和性能改进等方面的行为可能会有变化。"
|
||||
|
||||
#: ../../user_guide/feature_guide/graph_mode.md:7
|
||||
msgid ""
|
||||
"This guide provides instructions for using Ascend Graph Mode with vLLM "
|
||||
"Ascend. Please note that graph mode is only available on V1 Engine. And only"
|
||||
" Qwen, DeepSeek series models are well tested from 0.9.0rc1. We'll make it "
|
||||
"stable and generalize in the next release."
|
||||
msgstr ""
|
||||
"本指南提供了在 vLLM Ascend 上使用 Ascend 图模式的操作说明。请注意,图模式仅在 V1 引擎上可用,并且从 0.9.0rc1 起,仅对"
|
||||
" Qwen、DeepSeek 系列模型进行了充分测试。我们将在下一个版本中使其更加稳定和通用。"
|
||||
|
||||
#: ../../user_guide/feature_guide/graph_mode.md:9
|
||||
msgid "Getting Started"
|
||||
msgstr "快速入门"
|
||||
|
||||
#: ../../user_guide/feature_guide/graph_mode.md:11
|
||||
msgid ""
|
||||
"From v0.9.1rc1 with V1 Engine, vLLM Ascend will run models in graph mode by "
|
||||
"default to keep the same behavior with vLLM. If you hit any issues, please "
|
||||
"feel free to open an issue on GitHub and fallback to eager mode temporarily "
|
||||
"by set `enforce_eager=True` when initializing the model."
|
||||
msgstr ""
|
||||
"从 v0.9.1rc1 版本起,使用 V1 引擎时,vLLM Ascend 默认将在图模式下运行模型,以保持与 vLLM "
|
||||
"同样的行为。如果遇到任何问题,欢迎在 GitHub 上提交 issue,并在初始化模型时通过设置 `enforce_eager=True` 临时切换回 "
|
||||
"eager 模式。"
|
||||
|
||||
#: ../../user_guide/feature_guide/graph_mode.md:13
|
||||
msgid "There are two kinds for graph mode supported by vLLM Ascend:"
|
||||
msgstr "vLLM Ascend 支持两种图模式:"
|
||||
|
||||
#: ../../user_guide/feature_guide/graph_mode.md:14
|
||||
msgid ""
|
||||
"**ACLGraph**: This is the default graph mode supported by vLLM Ascend. In "
|
||||
"v0.9.1rc1, only Qwen series models are well tested."
|
||||
msgstr ""
|
||||
"**ACLGraph**:这是 vLLM Ascend 支持的默认图模式。在 v0.9.1rc1 版本中,只有 Qwen 系列模型得到了充分测试。"
|
||||
|
||||
#: ../../user_guide/feature_guide/graph_mode.md:15
|
||||
msgid ""
|
||||
"**TorchAirGraph**: This is the GE graph mode. In v0.9.1rc1, only DeepSeek "
|
||||
"series models are supported."
|
||||
msgstr "**TorchAirGraph**:这是GE图模式。在v0.9.1rc1版本中,仅支持DeepSeek系列模型。"
|
||||
|
||||
#: ../../user_guide/feature_guide/graph_mode.md:17
|
||||
msgid "Using ACLGraph"
|
||||
msgstr "使用 ACLGraph"
|
||||
|
||||
#: ../../user_guide/feature_guide/graph_mode.md:18
|
||||
msgid ""
|
||||
"ACLGraph is enabled by default. Take Qwen series models as an example, just "
|
||||
"set to use V1 Engine is enough."
|
||||
msgstr "ACLGraph 默认启用。以 Qwen 系列模型为例,只需设置为使用 V1 引擎即可。"
|
||||
|
||||
#: ../../user_guide/feature_guide/graph_mode.md:20
|
||||
#: ../../user_guide/feature_guide/graph_mode.md:41
|
||||
#: ../../user_guide/feature_guide/graph_mode.md:64
|
||||
msgid "offline example:"
|
||||
msgstr "离线示例:"
|
||||
|
||||
#: ../../user_guide/feature_guide/graph_mode.md:31
|
||||
#: ../../user_guide/feature_guide/graph_mode.md:52
|
||||
#: ../../user_guide/feature_guide/graph_mode.md:74
|
||||
msgid "online example:"
|
||||
msgstr "在线示例:"
|
||||
|
||||
#: ../../user_guide/feature_guide/graph_mode.md:37
|
||||
msgid "Using TorchAirGraph"
|
||||
msgstr "使用 TorchAirGraph"
|
||||
|
||||
#: ../../user_guide/feature_guide/graph_mode.md:39
|
||||
msgid ""
|
||||
"If you want to run DeepSeek series models with graph mode, you should use "
|
||||
"[TorchAirGraph](https://www.hiascend.com/document/detail/zh/Pytorch/700/modthirdparty/torchairuseguide/torchair_0002.html)."
|
||||
" In this case, additional config is required."
|
||||
msgstr ""
|
||||
"如果你想通过图模式运行 DeepSeek 系列模型,你应该使用 "
|
||||
"[TorchAirGraph](https://www.hiascend.com/document/detail/zh/Pytorch/700/modthirdparty/torchairuseguide/torchair_0002.html)。在这种情况下,需要额外的配置。"
|
||||
|
||||
#: ../../user_guide/feature_guide/graph_mode.md:58
|
||||
msgid ""
|
||||
"You can find more detail about additional config "
|
||||
"[here](../configuration/additional_config.md)."
|
||||
msgstr "你可以在[这里](../configuration/additional_config.md)找到关于附加配置的更多详细信息。"
|
||||
|
||||
#: ../../user_guide/feature_guide/graph_mode.md:60
|
||||
msgid "Fallback to Eager Mode"
|
||||
msgstr "回退到 Eager 模式"
|
||||
|
||||
#: ../../user_guide/feature_guide/graph_mode.md:62
|
||||
msgid ""
|
||||
"If both `ACLGraph` and `TorchAirGraph` fail to run, you should fallback to "
|
||||
"eager mode."
|
||||
msgstr "如果 `ACLGraph` 和 `TorchAirGraph` 都无法运行,你应该退回到 eager 模式。"
|
||||
@@ -0,0 +1,30 @@
|
||||
# SOME DESCRIPTIVE TITLE.
|
||||
# Copyright (C) 2025, vllm-ascend team
|
||||
# This file is distributed under the same license as the vllm-ascend
|
||||
# package.
|
||||
# FIRST AUTHOR <EMAIL@ADDRESS>, 2025.
|
||||
#
|
||||
#, fuzzy
|
||||
msgid ""
|
||||
msgstr ""
|
||||
"Project-Id-Version: vllm-ascend\n"
|
||||
"Report-Msgid-Bugs-To: \n"
|
||||
"POT-Creation-Date: 2025-07-18 09:01+0800\n"
|
||||
"PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n"
|
||||
"Last-Translator: FULL NAME <EMAIL@ADDRESS>\n"
|
||||
"Language-Team: zh_CN <LL@li.org>\n"
|
||||
"Language: zh_CN\n"
|
||||
"MIME-Version: 1.0\n"
|
||||
"Content-Type: text/plain; charset=utf-8\n"
|
||||
"Content-Transfer-Encoding: 8bit\n"
|
||||
"Plural-Forms: nplurals=1; plural=0;\n"
|
||||
"Generated-By: Babel 2.17.0\n"
|
||||
|
||||
#: ../../user_guide/feature_guide/index.md:1
|
||||
#: ../../user_guide/feature_guide/index.md:5
|
||||
msgid "Feature Guide"
|
||||
msgstr "功能指南"
|
||||
|
||||
#: ../../user_guide/feature_guide/index.md:3
|
||||
msgid "This section provides a detailed usage guide of vLLM Ascend features."
|
||||
msgstr "本节提供了 vLLM Ascend 功能的详细使用指南。"
|
||||
@@ -0,0 +1,58 @@
|
||||
# SOME DESCRIPTIVE TITLE.
|
||||
# Copyright (C) 2025, vllm-ascend team
|
||||
# This file is distributed under the same license as the vllm-ascend
|
||||
# package.
|
||||
# FIRST AUTHOR <EMAIL@ADDRESS>, 2025.
|
||||
#
|
||||
#, fuzzy
|
||||
msgid ""
|
||||
msgstr ""
|
||||
"Project-Id-Version: vllm-ascend\n"
|
||||
"Report-Msgid-Bugs-To: \n"
|
||||
"POT-Creation-Date: 2025-07-18 09:01+0800\n"
|
||||
"PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n"
|
||||
"Last-Translator: FULL NAME <EMAIL@ADDRESS>\n"
|
||||
"Language-Team: zh_CN <LL@li.org>\n"
|
||||
"Language: zh_CN\n"
|
||||
"MIME-Version: 1.0\n"
|
||||
"Content-Type: text/plain; charset=utf-8\n"
|
||||
"Content-Transfer-Encoding: 8bit\n"
|
||||
"Plural-Forms: nplurals=1; plural=0;\n"
|
||||
"Generated-By: Babel 2.17.0\n"
|
||||
|
||||
#: ../../user_guide/feature_guide/lora.md:1
|
||||
msgid "LoRA Adapters Guide"
|
||||
msgstr "LoRA 适配器指南"
|
||||
|
||||
#: ../../user_guide/feature_guide/lora.md:3
|
||||
msgid ""
|
||||
"Like vLLM, vllm-ascend supports LoRA as well. The usage and more details can"
|
||||
" be found in [vLLM official "
|
||||
"document](https://docs.vllm.ai/en/latest/features/lora.html)."
|
||||
msgstr ""
|
||||
"与 vLLM 类似,vllm-ascend 也支持 LoRA。用法及更多详情可参见 [vLLM "
|
||||
"官方文档](https://docs.vllm.ai/en/latest/features/lora.html)。"
|
||||
|
||||
#: ../../user_guide/feature_guide/lora.md:5
|
||||
msgid ""
|
||||
"You can also refer to "
|
||||
"[this](https://docs.vllm.ai/en/latest/models/supported_models.html#list-of-"
|
||||
"text-only-language-models) to find which models support LoRA in vLLM."
|
||||
msgstr ""
|
||||
"你也可以参考[这个链接](https://docs.vllm.ai/en/latest/models/supported_models.html#list-"
|
||||
"of-text-only-language-models)来查找哪些模型在 vLLM 中支持 LoRA。"
|
||||
|
||||
#: ../../user_guide/feature_guide/lora.md:7
|
||||
msgid "Tips"
|
||||
msgstr "提示"
|
||||
|
||||
#: ../../user_guide/feature_guide/lora.md:8
|
||||
msgid ""
|
||||
"If you fail to run vllm-ascend with LoRA, you may follow [this "
|
||||
"instruction](https://vllm-"
|
||||
"ascend.readthedocs.io/en/latest/user_guide/feature_guide/graph_mode.html#fallback-"
|
||||
"to-eager-mode) to disable graph mode and try again."
|
||||
msgstr ""
|
||||
"如果你在使用 LoRA 运行 vllm-ascend 时失败,可以按照[此说明](https://vllm-"
|
||||
"ascend.readthedocs.io/en/latest/user_guide/feature_guide/graph_mode.html#fallback-"
|
||||
"to-eager-mode)禁用图模式后再重试。"
|
||||
@@ -0,0 +1,183 @@
|
||||
# SOME DESCRIPTIVE TITLE.
|
||||
# Copyright (C) 2025, vllm-ascend team
|
||||
# This file is distributed under the same license as the vllm-ascend
|
||||
# package.
|
||||
# FIRST AUTHOR <EMAIL@ADDRESS>, 2025.
|
||||
#
|
||||
#, fuzzy
|
||||
msgid ""
|
||||
msgstr ""
|
||||
"Project-Id-Version: vllm-ascend\n"
|
||||
"Report-Msgid-Bugs-To: \n"
|
||||
"POT-Creation-Date: 2025-07-18 09:01+0800\n"
|
||||
"PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n"
|
||||
"Last-Translator: FULL NAME <EMAIL@ADDRESS>\n"
|
||||
"Language-Team: zh_CN <LL@li.org>\n"
|
||||
"Language: zh_CN\n"
|
||||
"MIME-Version: 1.0\n"
|
||||
"Content-Type: text/plain; charset=utf-8\n"
|
||||
"Content-Transfer-Encoding: 8bit\n"
|
||||
"Plural-Forms: nplurals=1; plural=0;\n"
|
||||
"Generated-By: Babel 2.17.0\n"
|
||||
|
||||
#: ../../user_guide/feature_guide/quantization.md:1
|
||||
msgid "Quantization Guide"
|
||||
msgstr "量化指南"
|
||||
|
||||
#: ../../user_guide/feature_guide/quantization.md:3
|
||||
msgid ""
|
||||
"Model quantization is a technique that reduces the size and computational "
|
||||
"requirements of a model by lowering the data precision of the weights and "
|
||||
"activation values in the model, thereby saving the memory and improving the "
|
||||
"inference speed."
|
||||
msgstr "模型量化是一种通过降低模型中权重和激活值的数据精度,从而减少模型大小和计算需求的技术,这样可以节省内存并提高推理速度。"
|
||||
|
||||
#: ../../user_guide/feature_guide/quantization.md:5
|
||||
msgid ""
|
||||
"Since 0.9.0rc2 version, quantization feature is experimentally supported in "
|
||||
"vLLM Ascend. Users can enable quantization feature by specifying "
|
||||
"`--quantization ascend`. Currently, only Qwen, DeepSeek series models are "
|
||||
"well tested. We’ll support more quantization algorithm and models in the "
|
||||
"future."
|
||||
msgstr ""
|
||||
"自 0.9.0rc2 版本起,vLLM Ascend 实验性地支持量化特性。用户可以通过指定 `--quantization ascend` "
|
||||
"启用量化功能。目前,只有 Qwen、DeepSeek 系列模型经过了充分测试。未来我们将支持更多的量化算法和模型。"
|
||||
|
||||
#: ../../user_guide/feature_guide/quantization.md:7
|
||||
msgid "Install modelslim"
|
||||
msgstr "安装 modelslim"
|
||||
|
||||
#: ../../user_guide/feature_guide/quantization.md:9
|
||||
msgid ""
|
||||
"To quantize a model, users should install "
|
||||
"[ModelSlim](https://gitee.com/ascend/msit/blob/master/msmodelslim/README.md)"
|
||||
" which is the Ascend compression and acceleration tool. It is an affinity-"
|
||||
"based compression tool designed for acceleration, using compression as its "
|
||||
"core technology and built upon the Ascend platform."
|
||||
msgstr ""
|
||||
"要对模型进行量化,用户应安装[ModelSlim](https://gitee.com/ascend/msit/blob/master/msmodelslim/README.md),这是昇腾的压缩与加速工具。它是一种基于亲和性的压缩工具,专为加速设计,以压缩为核心技术,并基于昇腾平台构建。"
|
||||
|
||||
#: ../../user_guide/feature_guide/quantization.md:11
|
||||
msgid ""
|
||||
"Currently, only the specific tag [modelslim-"
|
||||
"VLLM-8.1.RC1.b020_001](https://gitee.com/ascend/msit/blob/modelslim-"
|
||||
"VLLM-8.1.RC1.b020_001/msmodelslim/README.md) of modelslim works with vLLM "
|
||||
"Ascend. Please do not install other version until modelslim master version "
|
||||
"is available for vLLM Ascend in the future."
|
||||
msgstr ""
|
||||
"目前,只有 modelslim 的特定标签 [modelslim-"
|
||||
"VLLM-8.1.RC1.b020_001](https://gitee.com/ascend/msit/blob/modelslim-"
|
||||
"VLLM-8.1.RC1.b020_001/msmodelslim/README.md) 支持 vLLM Ascend。在未来 modelslim "
|
||||
"的主版本支持 vLLM Ascend 之前,请不要安装其他版本。"
|
||||
|
||||
#: ../../user_guide/feature_guide/quantization.md:13
|
||||
msgid "Install modelslim:"
|
||||
msgstr "安装 modelslim:"
|
||||
|
||||
#: ../../user_guide/feature_guide/quantization.md:21
|
||||
msgid "Quantize model"
|
||||
msgstr "量化模型"
|
||||
|
||||
#: ../../user_guide/feature_guide/quantization.md:23
|
||||
#, python-format
|
||||
msgid ""
|
||||
"Take [DeepSeek-V2-Lite](https://modelscope.cn/models/deepseek-"
|
||||
"ai/DeepSeek-V2-Lite) as an example, you just need to download the model, and"
|
||||
" then execute the convert command. The command is shown below. More info can"
|
||||
" be found in modelslim doc [deepseek w8a8 dynamic quantization "
|
||||
"docs](https://gitee.com/ascend/msit/blob/modelslim-"
|
||||
"VLLM-8.1.RC1.b020_001/msmodelslim/example/DeepSeek/README.md#deepseek-v2-w8a8-dynamic%E9%87%8F%E5%8C%96)."
|
||||
msgstr ""
|
||||
"以 [DeepSeek-V2-Lite](https://modelscope.cn/models/deepseek-"
|
||||
"ai/DeepSeek-V2-Lite) 为例,你只需要下载模型,然后执行转换命令。命令如下所示。更多信息可参考 modelslim 文档 "
|
||||
"[deepseek w8a8 动态量化文档](https://gitee.com/ascend/msit/blob/modelslim-"
|
||||
"VLLM-8.1.RC1.b020_001/msmodelslim/example/DeepSeek/README.md#deepseek-v2-w8a8-dynamic%E9%87%8F%E5%8C%96)。"
|
||||
|
||||
#: ../../user_guide/feature_guide/quantization.md:32
|
||||
msgid ""
|
||||
"You can also download the quantized model that we uploaded. Please note that"
|
||||
" these weights should be used for test only. For example, "
|
||||
"https://www.modelscope.cn/models/vllm-ascend/DeepSeek-V2-Lite-W8A8"
|
||||
msgstr ""
|
||||
"你也可以下载我们上传的量化模型。请注意,这些权重仅应用于测试。例如:https://www.modelscope.cn/models/vllm-"
|
||||
"ascend/DeepSeek-V2-Lite-W8A8"
|
||||
|
||||
#: ../../user_guide/feature_guide/quantization.md:35
|
||||
msgid "Once convert action is done, there are two important files generated."
|
||||
msgstr "转换操作完成后,会生成两个重要的文件。"
|
||||
|
||||
#: ../../user_guide/feature_guide/quantization.md:37
|
||||
msgid ""
|
||||
"[config.json](https://www.modelscope.cn/models/vllm-"
|
||||
"ascend/DeepSeek-V2-Lite-W8A8/file/view/master/config.json?status=1). Please "
|
||||
"make sure that there is no `quantization_config` field in it."
|
||||
msgstr ""
|
||||
"[config.json](https://www.modelscope.cn/models/vllm-"
|
||||
"ascend/DeepSeek-V2-Lite-W8A8/file/view/master/config.json?status=1)。请确保其中没有 "
|
||||
"`quantization_config` 字段。"
|
||||
|
||||
#: ../../user_guide/feature_guide/quantization.md:39
|
||||
msgid ""
|
||||
"[quant_model_description.json](https://www.modelscope.cn/models/vllm-"
|
||||
"ascend/DeepSeek-V2-Lite-W8A8/file/view/master/quant_model_description.json?status=1)."
|
||||
" All the converted weights info are recorded in this file."
|
||||
msgstr ""
|
||||
"[quant_model_description.json](https://www.modelscope.cn/models/vllm-"
|
||||
"ascend/DeepSeek-V2-Lite-W8A8/file/view/master/quant_model_description.json?status=1)。所有被转换的权重信息都记录在该文件中。"
|
||||
|
||||
#: ../../user_guide/feature_guide/quantization.md:41
|
||||
msgid "Here is the full converted model files:"
|
||||
msgstr "以下是完整转换后的模型文件:"
|
||||
|
||||
#: ../../user_guide/feature_guide/quantization.md:60
|
||||
msgid "Run the model"
|
||||
msgstr "运行模型"
|
||||
|
||||
#: ../../user_guide/feature_guide/quantization.md:62
|
||||
msgid ""
|
||||
"Now, you can run the quantized models with vLLM Ascend. Here is the example "
|
||||
"for online and offline inference."
|
||||
msgstr "现在,你可以使用 vLLM Ascend 运行量化模型。下面是在线和离线推理的示例。"
|
||||
|
||||
#: ../../user_guide/feature_guide/quantization.md:64
|
||||
msgid "Offline inference"
|
||||
msgstr "离线推理"
|
||||
|
||||
#: ../../user_guide/feature_guide/quantization.md:90
|
||||
msgid "Online inference"
|
||||
msgstr "在线推理"
|
||||
|
||||
#: ../../user_guide/feature_guide/quantization.md:97
|
||||
msgid "FAQs"
|
||||
msgstr "常见问题解答"
|
||||
|
||||
#: ../../user_guide/feature_guide/quantization.md:99
|
||||
msgid ""
|
||||
"1. How to solve the KeyError: 'xxx.layers.0.self_attn.q_proj.weight' "
|
||||
"problem?"
|
||||
msgstr "1. 如何解决 KeyError: 'xxx.layers.0.self_attn.q_proj.weight' 问题?"
|
||||
|
||||
#: ../../user_guide/feature_guide/quantization.md:101
|
||||
msgid ""
|
||||
"First, make sure you specify `ascend` quantization method. Second, check if "
|
||||
"your model is converted by this `modelslim-VLLM-8.1.RC1.b020_001` modelslim "
|
||||
"version. Finally, if it still doesn't work, please submit a issue, maybe "
|
||||
"some new models need to be adapted."
|
||||
msgstr ""
|
||||
"首先,请确保你指定了 `ascend` 量化方法。其次,检查你的模型是否由 `modelslim-VLLM-8.1.RC1.b020_001` 这个 "
|
||||
"modelslim 版本转换。如果仍然无法使用,请提交一个 issue,可能有一些新模型需要适配。"
|
||||
|
||||
#: ../../user_guide/feature_guide/quantization.md:104
|
||||
msgid ""
|
||||
"2. How to solve the error \"Could not locate the "
|
||||
"configuration_deepseek.py\"?"
|
||||
msgstr "2. 如何解决“无法找到 configuration_deepseek.py”错误?"
|
||||
|
||||
#: ../../user_guide/feature_guide/quantization.md:106
|
||||
msgid ""
|
||||
"Please convert DeepSeek series models using `modelslim-"
|
||||
"VLLM-8.1.RC1.b020_001` modelslim, this version has fixed the missing "
|
||||
"configuration_deepseek.py error."
|
||||
msgstr ""
|
||||
"请使用 `modelslim-VLLM-8.1.RC1.b020_001` 的 modelslim 转换 DeepSeek 系列模型,该版本已修复缺少 "
|
||||
"configuration_deepseek.py 的错误。"
|
||||
@@ -0,0 +1,156 @@
|
||||
# SOME DESCRIPTIVE TITLE.
|
||||
# Copyright (C) 2025, vllm-ascend team
|
||||
# This file is distributed under the same license as the vllm-ascend
|
||||
# package.
|
||||
# FIRST AUTHOR <EMAIL@ADDRESS>, 2025.
|
||||
#
|
||||
#, fuzzy
|
||||
msgid ""
|
||||
msgstr ""
|
||||
"Project-Id-Version: vllm-ascend\n"
|
||||
"Report-Msgid-Bugs-To: \n"
|
||||
"POT-Creation-Date: 2025-07-18 09:01+0800\n"
|
||||
"PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n"
|
||||
"Last-Translator: FULL NAME <EMAIL@ADDRESS>\n"
|
||||
"Language-Team: zh_CN <LL@li.org>\n"
|
||||
"Language: zh_CN\n"
|
||||
"MIME-Version: 1.0\n"
|
||||
"Content-Type: text/plain; charset=utf-8\n"
|
||||
"Content-Transfer-Encoding: 8bit\n"
|
||||
"Plural-Forms: nplurals=1; plural=0;\n"
|
||||
"Generated-By: Babel 2.17.0\n"
|
||||
|
||||
#: ../../user_guide/feature_guide/sleep_mode.md:1
|
||||
msgid "Sleep Mode Guide"
|
||||
msgstr "睡眠模式指南"
|
||||
|
||||
#: ../../user_guide/feature_guide/sleep_mode.md:3
|
||||
msgid "Overview"
|
||||
msgstr "概述"
|
||||
|
||||
#: ../../user_guide/feature_guide/sleep_mode.md:5
|
||||
msgid ""
|
||||
"Sleep Mode is an API designed to offload model weights and discard KV cache "
|
||||
"from NPU memory. This functionality is essential for reinforcement learning "
|
||||
"(RL) post-training workloads, particularly in online algorithms such as PPO,"
|
||||
" GRPO, or DPO. During training, the policy model typically performs auto-"
|
||||
"regressive generation using inference engines like vLLM, followed by forward"
|
||||
" and backward passes for optimization."
|
||||
msgstr ""
|
||||
"Sleep Mode 是一个用于卸载模型权重并清除 NPU 内存中 KV 缓存的 API。此功能对于强化学习(RL)后训练任务尤其重要,特别是在 "
|
||||
"PPO、GRPO 或 DPO 等在线算法中。在训练过程中,策略模型通常会使用像 vLLM "
|
||||
"这样的推理引擎进行自回归生成,然后进行前向和反向传播以进行优化。"
|
||||
|
||||
#: ../../user_guide/feature_guide/sleep_mode.md:7
|
||||
msgid ""
|
||||
"Since the generation and training phases may employ different model "
|
||||
"parallelism strategies, it becomes crucial to free KV cache and even offload"
|
||||
" model parameters stored within vLLM during training. This ensures efficient"
|
||||
" memory utilization and avoids resource contention on the NPU."
|
||||
msgstr ""
|
||||
"由于生成和训练阶段可能采用不同的模型并行策略,因此在训练过程中及时释放 KV 缓存,甚至卸载存储在 vLLM "
|
||||
"内的模型参数变得至关重要。这可以确保内存的高效利用,并避免 NPU 上的资源争用。"
|
||||
|
||||
#: ../../user_guide/feature_guide/sleep_mode.md:10
|
||||
msgid "Getting started"
|
||||
msgstr "快速上手"
|
||||
|
||||
#: ../../user_guide/feature_guide/sleep_mode.md:12
|
||||
#, python-brace-format
|
||||
msgid ""
|
||||
"With `enable_sleep_mode=True`, the way we manage memory(malloc, free) in "
|
||||
"vllm will under a specific memory pool, during loading model and initialize "
|
||||
"kv_caches, we tag the memory as a map: `{\"weight\": data, \"kv_cache\": "
|
||||
"data}`."
|
||||
msgstr ""
|
||||
"当 `enable_sleep_mode=True` 时,我们在 vllm 中管理内存(malloc, "
|
||||
"free)的方式会在一个特定的内存池下进行,在加载模型和初始化 kv_caches "
|
||||
"期间,我们会将内存打上标签,组织成一个映射:`{\"weight\": data, \"kv_cache\": data}`。"
|
||||
|
||||
#: ../../user_guide/feature_guide/sleep_mode.md:14
|
||||
msgid ""
|
||||
"The engine(v0/v1) supports two sleep levels to manage memory during idle "
|
||||
"periods:"
|
||||
msgstr "该引擎(v0/v1)支持两种睡眠等级,以在空闲期间管理内存:"
|
||||
|
||||
#: ../../user_guide/feature_guide/sleep_mode.md:16
|
||||
msgid "Level 1 Sleep"
|
||||
msgstr "一级睡眠"
|
||||
|
||||
#: ../../user_guide/feature_guide/sleep_mode.md:17
|
||||
msgid "Action: Offloads model weights and discards the KV cache."
|
||||
msgstr "操作:卸载模型权重并清除KV缓存。"
|
||||
|
||||
#: ../../user_guide/feature_guide/sleep_mode.md:18
|
||||
msgid "Memory: Model weights are moved to CPU memory; KV cache is forgotten."
|
||||
msgstr "内存:模型权重被移动到CPU内存;KV缓存被清除。"
|
||||
|
||||
#: ../../user_guide/feature_guide/sleep_mode.md:19
|
||||
msgid "Use Case: Suitable when reusing the same model later."
|
||||
msgstr "用例:适用于之后需要重复使用同一个模型的情况。"
|
||||
|
||||
#: ../../user_guide/feature_guide/sleep_mode.md:20
|
||||
msgid ""
|
||||
"Note: Ensure sufficient CPU memory is available to hold the model weights."
|
||||
msgstr "注意:请确保有足够的CPU内存来存储模型权重。"
|
||||
|
||||
#: ../../user_guide/feature_guide/sleep_mode.md:22
|
||||
msgid "Level 2 Sleep"
|
||||
msgstr "二级睡眠"
|
||||
|
||||
#: ../../user_guide/feature_guide/sleep_mode.md:23
|
||||
msgid "Action: Discards both model weights and KV cache."
|
||||
msgstr "操作:同时丢弃模型权重和KV缓存。"
|
||||
|
||||
#: ../../user_guide/feature_guide/sleep_mode.md:24
|
||||
msgid ""
|
||||
"Memory: The content of both the model weights and kv cache is forgotten."
|
||||
msgstr "内存:模型权重和kv缓存的内容都会被遗忘。"
|
||||
|
||||
#: ../../user_guide/feature_guide/sleep_mode.md:25
|
||||
msgid ""
|
||||
"Use Case: Ideal when switching to a different model or updating the current "
|
||||
"one."
|
||||
msgstr "用例:当切换到不同的模型或更新当前模型时非常理想。"
|
||||
|
||||
#: ../../user_guide/feature_guide/sleep_mode.md:27
|
||||
msgid ""
|
||||
"Since this feature uses the low-level API "
|
||||
"[AscendCL](https://www.hiascend.com/document/detail/zh/CANNCommunityEdition/82RC1alpha002/API/appdevgapi/appdevgapi_07_0000.html),"
|
||||
" in order to use sleep mode, you should follow the [installation "
|
||||
"guide](https://vllm-ascend.readthedocs.io/en/latest/installation.html) and "
|
||||
"building from source, if you are using v0.7.3, remember to set `export "
|
||||
"COMPILE_CUSTOM_KERNELS=1`, for the latest version(v0.9.x+), the environment "
|
||||
"variable `COMPILE_CUSTOM_KERNELS` will be set 1 by default while building "
|
||||
"from source."
|
||||
msgstr ""
|
||||
"由于此功能使用了底层 API "
|
||||
"[AscendCL](https://www.hiascend.com/document/detail/zh/CANNCommunityEdition/82RC1alpha002/API/appdevgapi/appdevgapi_07_0000.html),为了使用休眠模式,你应按照[安装指南](https://vllm-"
|
||||
"ascend.readthedocs.io/en/latest/installation.html)进行操作,并从源码编译。如果你使用的是 "
|
||||
"v0.7.3,请记得设置 `export COMPILE_CUSTOM_KERNELS=1` ;对于最新版本(v0.9.x+),在从源码编译时环境变量 "
|
||||
"`COMPILE_CUSTOM_KERNELS` 默认会被设置为 1。"
|
||||
|
||||
#: ../../user_guide/feature_guide/sleep_mode.md:29
|
||||
msgid "Usage"
|
||||
msgstr "用法"
|
||||
|
||||
#: ../../user_guide/feature_guide/sleep_mode.md:31
|
||||
msgid "The following is a simple example of how to use sleep mode."
|
||||
msgstr "以下是如何使用睡眠模式的一个简单示例。"
|
||||
|
||||
#: ../../user_guide/feature_guide/sleep_mode.md:33
|
||||
msgid "offline inference:"
|
||||
msgstr "离线推理:"
|
||||
|
||||
#: ../../user_guide/feature_guide/sleep_mode.md:72
|
||||
msgid "online serving:"
|
||||
msgstr "在线服务:"
|
||||
|
||||
#: ../../user_guide/feature_guide/sleep_mode.md:74
|
||||
msgid ""
|
||||
"Considering there may be a risk of malicious access, please make sure you "
|
||||
"are under a dev-mode, and explicit specify the develop env: "
|
||||
"`VLLM_SERVER_DEV_MODE` to expose these endpoints(sleep/wake up)."
|
||||
msgstr ""
|
||||
"鉴于可能存在恶意访问的风险,请确保您处于开发模式,并明确指定开发环境:`VLLM_SERVER_DEV_MODE`,以便开放这些端点(sleep/wake"
|
||||
" up)。"
|
||||
@@ -0,0 +1,220 @@
|
||||
# SOME DESCRIPTIVE TITLE.
|
||||
# Copyright (C) 2025, vllm-ascend team
|
||||
# This file is distributed under the same license as the vllm-ascend
|
||||
# package.
|
||||
# FIRST AUTHOR <EMAIL@ADDRESS>, 2025.
|
||||
#
|
||||
#, fuzzy
|
||||
msgid ""
|
||||
msgstr ""
|
||||
"Project-Id-Version: vllm-ascend\n"
|
||||
"Report-Msgid-Bugs-To: \n"
|
||||
"POT-Creation-Date: 2025-07-18 09:01+0800\n"
|
||||
"PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n"
|
||||
"Last-Translator: FULL NAME <EMAIL@ADDRESS>\n"
|
||||
"Language-Team: zh_CN <LL@li.org>\n"
|
||||
"Language: zh_CN\n"
|
||||
"MIME-Version: 1.0\n"
|
||||
"Content-Type: text/plain; charset=utf-8\n"
|
||||
"Content-Transfer-Encoding: 8bit\n"
|
||||
"Plural-Forms: nplurals=1; plural=0;\n"
|
||||
"Generated-By: Babel 2.17.0\n"
|
||||
|
||||
#: ../../user_guide/feature_guide/structured_output.md:1
|
||||
msgid "Structured Output Guide"
|
||||
msgstr "结构化输出指南"
|
||||
|
||||
#: ../../user_guide/feature_guide/structured_output.md:3
|
||||
msgid "Overview"
|
||||
msgstr "概述"
|
||||
|
||||
#: ../../user_guide/feature_guide/structured_output.md:5
|
||||
msgid "What is Structured Output?"
|
||||
msgstr "什么是结构化输出?"
|
||||
|
||||
#: ../../user_guide/feature_guide/structured_output.md:7
|
||||
msgid ""
|
||||
"LLMs can be unpredictable when you need output in specific formats. Think of"
|
||||
" asking a model to generate JSON - without guidance, it might produce valid "
|
||||
"text that breaks JSON specification. **Structured Output (also called Guided"
|
||||
" Decoding)** enables LLMs to generate outputs that follow a desired "
|
||||
"structure while preserving the non-deterministic nature of the system."
|
||||
msgstr ""
|
||||
"当你需要特定格式输出时,大型语言模型(LLMs)可能表现出不可预测性。比如让模型生成 "
|
||||
"JSON,如果没有指导,模型可能会生成有效的文本,但这些文本却不符合 JSON 规范。**结构化输出(也称为引导解码)** "
|
||||
"能让大型语言模型生成符合预期结构的输出,同时保留系统的非确定性特性。"
|
||||
|
||||
#: ../../user_guide/feature_guide/structured_output.md:9
|
||||
msgid ""
|
||||
"In simple terms, structured decoding gives LLMs a “template” to follow. "
|
||||
"Users provide a schema that “influences” the model’s output, ensuring "
|
||||
"compliance with the desired structure."
|
||||
msgstr "简单来说,结构化解码为LLM提供了一个“模板”来遵循。用户提供一个模式来“影响”模型的输出,从而确保输出符合期望的结构。"
|
||||
|
||||
#: ../../user_guide/feature_guide/structured_output.md:11
|
||||
msgid ""
|
||||
msgstr ""
|
||||
|
||||
#: ../../user_guide/feature_guide/structured_output.md:11
|
||||
msgid "structured decoding"
|
||||
msgstr "结构化解码"
|
||||
|
||||
#: ../../user_guide/feature_guide/structured_output.md:13
|
||||
msgid "Structured Output in vllm-ascend"
|
||||
msgstr "vllm-ascend 中的结构化输出"
|
||||
|
||||
#: ../../user_guide/feature_guide/structured_output.md:15
|
||||
msgid ""
|
||||
"Currently, vllm-ascend supports **xgrammar** and **guidance** backend for "
|
||||
"structured output with vllm v1 engine."
|
||||
msgstr "目前,vllm-ascend 支持 vllm v1 引擎的结构化输出,后端包括 **xgrammar** 和 **guidance**。"
|
||||
|
||||
#: ../../user_guide/feature_guide/structured_output.md:17
|
||||
msgid ""
|
||||
"XGrammar introduces a new technique that batch constrained decoding via "
|
||||
"pushdown automaton (PDA). You can think of a PDA as a “collection of FSMs, "
|
||||
"and each FSM represents a context-free grammar (CFG).” One significant "
|
||||
"advantage of PDA is its recursive nature, allowing us to execute multiple "
|
||||
"state transitions. They also include additional optimisation (for those who "
|
||||
"are interested) to reduce grammar compilation overhead. Besides, you can "
|
||||
"also find more details about guidance by yourself."
|
||||
msgstr ""
|
||||
"XGrammar 引入了一种通过下推自动机(PDA)进行批量约束解码的新技术。你可以把 PDA 理解为“有限状态机(FSM)的集合,每个 FSM "
|
||||
"代表一个上下文无关文法(CFG)。” PDA 的一个重要优点是其递归特性,使我们能够执行多次状态转移。此外,PDA "
|
||||
"还包含了额外的优化(供感兴趣的用户参考),以减少语法编译的开销。除此之外,你还可以自己找到更多关于指导的信息。"
|
||||
|
||||
#: ../../user_guide/feature_guide/structured_output.md:19
|
||||
msgid "How to Use Structured Output?"
|
||||
msgstr "如何使用结构化输出?"
|
||||
|
||||
#: ../../user_guide/feature_guide/structured_output.md:21
|
||||
msgid "Online Inference"
|
||||
msgstr "在线推理"
|
||||
|
||||
#: ../../user_guide/feature_guide/structured_output.md:23
|
||||
msgid ""
|
||||
"You can also generate structured outputs using the OpenAI's Completions and "
|
||||
"Chat API. The following parameters are supported, which must be added as "
|
||||
"extra parameters:"
|
||||
msgstr "你也可以使用 OpenAI 的 Completions 和 Chat API 生成结构化输出。支持以下参数,这些参数必须作为额外参数添加:"
|
||||
|
||||
#: ../../user_guide/feature_guide/structured_output.md:25
|
||||
msgid "`guided_choice`: the output will be exactly one of the choices."
|
||||
msgstr "`guided_choice`:输出将会是其中一个选项。"
|
||||
|
||||
#: ../../user_guide/feature_guide/structured_output.md:26
|
||||
msgid "`guided_regex`: the output will follow the regex pattern."
|
||||
msgstr "`guided_regex`:输出将遵循正则表达式模式。"
|
||||
|
||||
#: ../../user_guide/feature_guide/structured_output.md:27
|
||||
msgid "`guided_json`: the output will follow the JSON schema."
|
||||
msgstr "`guided_json`:输出将遵循 JSON 架构。"
|
||||
|
||||
#: ../../user_guide/feature_guide/structured_output.md:28
|
||||
msgid "`guided_grammar`: the output will follow the context free grammar."
|
||||
msgstr "`guided_grammar`:输出将遵循上下文无关文法。"
|
||||
|
||||
#: ../../user_guide/feature_guide/structured_output.md:30
|
||||
msgid ""
|
||||
"Structured outputs are supported by default in the OpenAI-Compatible Server."
|
||||
" You can choose to specify the backend to use by setting the `--guided-"
|
||||
"decoding-backend` flag to vllm serve. The default backend is `auto`, which "
|
||||
"will try to choose an appropriate backend based on the details of the "
|
||||
"request. You may also choose a specific backend, along with some options."
|
||||
msgstr ""
|
||||
"OpenAI 兼容服务器默认支持结构化输出。你可以通过设置 `--guided-decoding-backend` 标志为 vllm serve "
|
||||
"来指定要使用的后端。默认后端为 `auto`,它会根据请求的详细信息尝试选择合适的后端。你也可以选择特定的后端,并设置一些选项。"
|
||||
|
||||
#: ../../user_guide/feature_guide/structured_output.md:32
|
||||
msgid ""
|
||||
"Now let´s see an example for each of the cases, starting with the "
|
||||
"guided_choice, as it´s the easiest one:"
|
||||
msgstr "现在让我们来看每种情况的示例,首先是 guided_choice,因为它是最简单的:"
|
||||
|
||||
#: ../../user_guide/feature_guide/structured_output.md:51
|
||||
msgid ""
|
||||
"The next example shows how to use the guided_regex. The idea is to generate "
|
||||
"an email address, given a simple regex template:"
|
||||
msgstr "下一个例子展示了如何使用 guided_regex。其思路是基于一个简单的正则表达式模板生成一个电子邮件地址:"
|
||||
|
||||
#: ../../user_guide/feature_guide/structured_output.md:67
|
||||
msgid ""
|
||||
"One of the most relevant features in structured text generation is the "
|
||||
"option to generate a valid JSON with pre-defined fields and formats. For "
|
||||
"this we can use the guided_json parameter in two different ways:"
|
||||
msgstr ""
|
||||
"在结构化文本生成中,最相关的特性之一是能够生成具有预定义字段和格式的有效 JSON。为此,我们可以通过两种不同的方式使用 guided_json 参数:"
|
||||
|
||||
#: ../../user_guide/feature_guide/structured_output.md:69
|
||||
msgid "Using a JSON Schema."
|
||||
msgstr "使用 JSON 架构。"
|
||||
|
||||
#: ../../user_guide/feature_guide/structured_output.md:70
|
||||
msgid "Defining a Pydantic model and then extracting the JSON Schema from it."
|
||||
msgstr "定义一个 Pydantic 模型,然后从中提取 JSON Schema。"
|
||||
|
||||
#: ../../user_guide/feature_guide/structured_output.md:72
|
||||
msgid ""
|
||||
"The next example shows how to use the guided_json parameter with a Pydantic "
|
||||
"model:"
|
||||
msgstr "下一个示例展示了如何将 guided_json 参数与 Pydantic 模型一起使用:"
|
||||
|
||||
#: ../../user_guide/feature_guide/structured_output.md:104
|
||||
msgid ""
|
||||
"Finally we have the guided_grammar option, which is probably the most "
|
||||
"difficult to use, but it´s really powerful. It allows us to define complete "
|
||||
"languages like SQL queries. It works by using a context free EBNF grammar. "
|
||||
"As an example, we can use to define a specific format of simplified SQL "
|
||||
"queries:"
|
||||
msgstr ""
|
||||
"最后,我们有 guided_grammar 选项,这可能是最难使用的,但它非常强大。它允许我们定义完整的语言,比如 SQL 查询。它通过使用上下文无关的"
|
||||
" EBNF 语法来实现。例如,我们可以用它来定义一种简化 SQL 查询的特定格式:"
|
||||
|
||||
#: ../../user_guide/feature_guide/structured_output.md:134
|
||||
msgid ""
|
||||
"Find more examples [here](https://github.com/vllm-"
|
||||
"project/vllm/blob/main/examples/offline_inference/structured_outputs.py)."
|
||||
msgstr ""
|
||||
"在[这里](https://github.com/vllm-"
|
||||
"project/vllm/blob/main/examples/offline_inference/structured_outputs.py)可以找到更多示例。"
|
||||
|
||||
#: ../../user_guide/feature_guide/structured_output.md:136
|
||||
msgid "Offline Inference"
|
||||
msgstr "离线推理"
|
||||
|
||||
#: ../../user_guide/feature_guide/structured_output.md:138
|
||||
msgid ""
|
||||
"To use Structured Output, we'll need to configure the guided decoding using "
|
||||
"the class `GuidedDecodingParams` inside `SamplingParams`. The main available"
|
||||
" options inside `GuidedDecodingParams` are:"
|
||||
msgstr ""
|
||||
"要使用结构化输出,我们需要在 `SamplingParams` 内通过 `GuidedDecodingParams` "
|
||||
"类配置引导解码。`GuidedDecodingParams` 中主要可用的选项有:"
|
||||
|
||||
#: ../../user_guide/feature_guide/structured_output.md:140
|
||||
msgid "json"
|
||||
msgstr "json"
|
||||
|
||||
#: ../../user_guide/feature_guide/structured_output.md:141
|
||||
msgid "regex"
|
||||
msgstr "正则表达式"
|
||||
|
||||
#: ../../user_guide/feature_guide/structured_output.md:142
|
||||
msgid "choice"
|
||||
msgstr "选择"
|
||||
|
||||
#: ../../user_guide/feature_guide/structured_output.md:143
|
||||
msgid "grammar"
|
||||
msgstr "语法"
|
||||
|
||||
#: ../../user_guide/feature_guide/structured_output.md:145
|
||||
msgid "One example for the usage of the choice parameter is shown below:"
|
||||
msgstr "choice 参数用法的一个示例如下:"
|
||||
|
||||
#: ../../user_guide/feature_guide/structured_output.md:163
|
||||
msgid ""
|
||||
"Find more examples of other usages [here](https://github.com/vllm-"
|
||||
"project/vllm/blob/main/examples/offline_inference/structured_outputs.py)."
|
||||
msgstr ""
|
||||
"查看更多其他用法的示例 [在这里](https://github.com/vllm-"
|
||||
"project/vllm/blob/main/examples/offline_inference/structured_outputs.py)。"
|
||||
1660
docs/source/locale/zh_CN/LC_MESSAGES/user_guide/release_notes.po
Normal file
1660
docs/source/locale/zh_CN/LC_MESSAGES/user_guide/release_notes.po
Normal file
File diff suppressed because it is too large
Load Diff
@@ -0,0 +1,30 @@
|
||||
# Translations template for PROJECT.
|
||||
# Copyright (C) 2025 ORGANIZATION
|
||||
# This file is distributed under the same license as the PROJECT project.
|
||||
# FIRST AUTHOR <EMAIL@ADDRESS>, 2025.
|
||||
#
|
||||
#, fuzzy
|
||||
msgid ""
|
||||
msgstr ""
|
||||
"Project-Id-Version: PROJECT VERSION\n"
|
||||
"Report-Msgid-Bugs-To: EMAIL@ADDRESS\n"
|
||||
"POT-Creation-Date: 2025-07-18 09:01+0800\n"
|
||||
"PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n"
|
||||
"Last-Translator: FULL NAME <EMAIL@ADDRESS>\n"
|
||||
"Language-Team: LANGUAGE <LL@li.org>\n"
|
||||
"MIME-Version: 1.0\n"
|
||||
"Content-Type: text/plain; charset=utf-8\n"
|
||||
"Content-Transfer-Encoding: 8bit\n"
|
||||
"Generated-By: Babel 2.17.0\n"
|
||||
|
||||
#: ../../user_guide/support_matrix/index.md:5
|
||||
msgid "Support Matrix"
|
||||
msgstr "支持矩阵"
|
||||
|
||||
#: ../../user_guide/support_matrix/index.md:1
|
||||
msgid "Features and models"
|
||||
msgstr "特性与模型"
|
||||
|
||||
#: ../../user_guide/support_matrix/index.md:3
|
||||
msgid "This section provides a detailed supported matrix by vLLM Ascend."
|
||||
msgstr "本节提供了 vLLM Ascend 的详细支持矩阵。"
|
||||
@@ -0,0 +1,264 @@
|
||||
# SOME DESCRIPTIVE TITLE.
|
||||
# Copyright (C) 2025, vllm-ascend team
|
||||
# This file is distributed under the same license as the vllm-ascend
|
||||
# package.
|
||||
# FIRST AUTHOR <EMAIL@ADDRESS>, 2025.
|
||||
#
|
||||
#, fuzzy
|
||||
msgid ""
|
||||
msgstr ""
|
||||
"Project-Id-Version: vllm-ascend\n"
|
||||
"Report-Msgid-Bugs-To: \n"
|
||||
"POT-Creation-Date: 2025-07-18 09:01+0800\n"
|
||||
"PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n"
|
||||
"Last-Translator: FULL NAME <EMAIL@ADDRESS>\n"
|
||||
"Language-Team: zh_CN <LL@li.org>\n"
|
||||
"Language: zh_CN\n"
|
||||
"MIME-Version: 1.0\n"
|
||||
"Content-Type: text/plain; charset=utf-8\n"
|
||||
"Content-Transfer-Encoding: 8bit\n"
|
||||
"Plural-Forms: nplurals=1; plural=0;\n"
|
||||
"Generated-By: Babel 2.17.0\n"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md:1
|
||||
msgid "Feature Support"
|
||||
msgstr "功能支持"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md:3
|
||||
msgid ""
|
||||
"The feature support principle of vLLM Ascend is: **aligned with the vLLM**. "
|
||||
"We are also actively collaborating with the community to accelerate support."
|
||||
msgstr "vLLM Ascend 的特性支持原则是:**与 vLLM 保持一致**。我们也在积极与社区合作,加快支持进度。"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md:5
|
||||
msgid ""
|
||||
"You can check the [support status of vLLM V1 Engine][v1_user_guide]. Below "
|
||||
"is the feature support status of vLLM Ascend:"
|
||||
msgstr "你可以查看 [vLLM V1 引擎的支持状态][v1_user_guide]。下面是 vLLM Ascend 的功能支持情况:"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md
|
||||
msgid "Feature"
|
||||
msgstr "特性"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md
|
||||
msgid "vLLM V0 Engine"
|
||||
msgstr "vLLM V0 引擎"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md
|
||||
msgid "vLLM V1 Engine"
|
||||
msgstr "vLLM V1 引擎"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md
|
||||
msgid "Next Step"
|
||||
msgstr "下一步"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md
|
||||
msgid "Chunked Prefill"
|
||||
msgstr "分块预填充"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md
|
||||
msgid "🟢 Functional"
|
||||
msgstr "🟢 功能性"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md
|
||||
msgid "Functional, see detail note: [Chunked Prefill][cp]"
|
||||
msgstr "功能性,详见说明:[分块预填充][cp]"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md
|
||||
msgid "Automatic Prefix Caching"
|
||||
msgstr "自动前缀缓存"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md
|
||||
msgid "Functional, see detail note: [vllm-ascend#732][apc]"
|
||||
msgstr "可用,请参见详细说明:[vllm-ascend#732][apc]"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md
|
||||
msgid "LoRA"
|
||||
msgstr "LoRA"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md
|
||||
msgid "[vllm-ascend#396][multilora], [vllm-ascend#893][v1 multilora]"
|
||||
msgstr "[vllm-ascend#396][multilora],[vllm-ascend#893][v1 multilora]"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md
|
||||
msgid "Prompt adapter"
|
||||
msgstr "提示适配器"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md
|
||||
msgid "🔴 No plan"
|
||||
msgstr "🔴 无计划"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md
|
||||
msgid "This feature has been deprecated by vllm."
|
||||
msgstr "此功能已被 vllm 弃用。"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md
|
||||
msgid "Speculative decoding"
|
||||
msgstr "猜测式解码"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md
|
||||
msgid "Basic support"
|
||||
msgstr "基础支持"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md
|
||||
msgid "Pooling"
|
||||
msgstr "池化"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md
|
||||
msgid "🟡 Planned"
|
||||
msgstr "🟡 计划中"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md
|
||||
msgid "CI needed and adapting more models; V1 support rely on vLLM support."
|
||||
msgstr "需要持续集成(CI)并适配更多模型;V1 的支持依赖于 vLLM 的支持。"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md
|
||||
msgid "Enc-dec"
|
||||
msgstr "Enc-dec(编码-解码)"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md
|
||||
msgid "🔴 NO plan"
|
||||
msgstr "🔴 没有计划"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md
|
||||
msgid "Plan in 2025.06.30"
|
||||
msgstr "2025.06.30 的计划"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md
|
||||
msgid "Multi Modality"
|
||||
msgstr "多模态"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md
|
||||
msgid "[Tutorial][multimodal], optimizing and adapting more models"
|
||||
msgstr "[教程][multimodal],优化和适配更多模型"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md
|
||||
msgid "LogProbs"
|
||||
msgstr "LogProbs"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md
|
||||
msgid "CI needed"
|
||||
msgstr "需要持续集成(CI)"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md
|
||||
msgid "Prompt logProbs"
|
||||
msgstr "提示 logProbs"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md
|
||||
msgid "Async output"
|
||||
msgstr "异步输出"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md
|
||||
msgid "Multi step scheduler"
|
||||
msgstr "多步调度器"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md
|
||||
msgid "🔴 Deprecated"
|
||||
msgstr "🔴 已弃用"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md
|
||||
msgid "[vllm#8779][v1_rfc], replaced by [vLLM V1 Scheduler][v1_scheduler]"
|
||||
msgstr "[vllm#8779][v1_rfc],已被 [vLLM V1 调度器][v1_scheduler] 替代"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md
|
||||
msgid "Best of"
|
||||
msgstr "精选"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md
|
||||
msgid "[vllm#13361][best_of], CI needed"
|
||||
msgstr "[vllm#13361][best_of],需要持续集成(CI)"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md
|
||||
msgid "Beam search"
|
||||
msgstr "束搜索"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md
|
||||
msgid "Guided Decoding"
|
||||
msgstr "引导解码"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md
|
||||
msgid "[vllm-ascend#177][guided_decoding]"
|
||||
msgstr "[vllm-ascend#177][guided_decoding]"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md
|
||||
msgid "Tensor Parallel"
|
||||
msgstr "张量并行"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md
|
||||
msgid "Pipeline Parallel"
|
||||
msgstr "流水线并行"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md
|
||||
msgid "Expert Parallel"
|
||||
msgstr "专家并行"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md
|
||||
msgid "CI needed; No plan on V0 support"
|
||||
msgstr "需要持续集成;没有支持V0的计划"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md
|
||||
msgid "Data Parallel"
|
||||
msgstr "数据并行"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md
|
||||
msgid "CI needed; No plan on V0 support"
|
||||
msgstr "需要 CI;暂无 V0 支持计划"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md
|
||||
msgid "Prefill Decode Disaggregation"
|
||||
msgstr "预填充 解码 拆分"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md
|
||||
msgid "1P1D available, working on xPyD and V1 support."
|
||||
msgstr "1P1D 已可用,正在开发 xPyD 和 V1 支持。"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md
|
||||
msgid "Quantization"
|
||||
msgstr "量化"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md
|
||||
msgid "W8A8 available, CI needed; working on more quantization method support"
|
||||
msgstr "W8A8 已可用,需要持续集成(CI);正在开发对更多量化方法的支持。"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md
|
||||
msgid "Graph Mode"
|
||||
msgstr "图模式"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md
|
||||
msgid "🔵 Experimental"
|
||||
msgstr "🔵 实验性"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md
|
||||
msgid "Experimental, see detail note: [vllm-ascend#767][graph_mode]"
|
||||
msgstr "实验性功能,详见说明:[vllm-ascend#767][graph_mode]"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md
|
||||
msgid "Sleep Mode"
|
||||
msgstr "睡眠模式"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md
|
||||
msgid "level=1 available, CI needed, working on V1 support"
|
||||
msgstr "level=1 可用,需要CI,正在开发 V1 支持"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md:33
|
||||
msgid "🟢 Functional: Fully operational, with ongoing optimizations."
|
||||
msgstr "🟢 功能性:完全可用,正在持续优化中。"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md:34
|
||||
msgid ""
|
||||
"🔵 Experimental: Experimental support, interfaces and functions may change."
|
||||
msgstr "🔵 实验性:实验性支持,接口和功能可能会发生变化。"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md:35
|
||||
msgid "🚧 WIP: Under active development, will be supported soon."
|
||||
msgstr "🚧 WIP:正在积极开发中,很快将会支持。"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md:36
|
||||
msgid ""
|
||||
"🟡 Planned: Scheduled for future implementation (some may have open "
|
||||
"PRs/RFCs)."
|
||||
msgstr "🟡 计划中:已安排将来实现(其中一些可能已有开放的PR/RFC)。"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_features.md:37
|
||||
msgid "🔴 NO plan / Deprecated: No plan for V0 or deprecated by vLLM v1."
|
||||
msgstr "🔴 没有计划 / 已弃用:V0 没有计划或已被 vLLM v1 弃用。"
|
||||
@@ -0,0 +1,214 @@
|
||||
# SOME DESCRIPTIVE TITLE.
|
||||
# Copyright (C) 2025, vllm-ascend team
|
||||
# This file is distributed under the same license as the vllm-ascend
|
||||
# package.
|
||||
# FIRST AUTHOR <EMAIL@ADDRESS>, 2025.
|
||||
#
|
||||
#, fuzzy
|
||||
msgid ""
|
||||
msgstr ""
|
||||
"Project-Id-Version: vllm-ascend\n"
|
||||
"Report-Msgid-Bugs-To: \n"
|
||||
"POT-Creation-Date: 2025-07-18 09:01+0800\n"
|
||||
"PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n"
|
||||
"Last-Translator: FULL NAME <EMAIL@ADDRESS>\n"
|
||||
"Language-Team: zh_CN <LL@li.org>\n"
|
||||
"Language: zh_CN\n"
|
||||
"MIME-Version: 1.0\n"
|
||||
"Content-Type: text/plain; charset=utf-8\n"
|
||||
"Content-Transfer-Encoding: 8bit\n"
|
||||
"Plural-Forms: nplurals=1; plural=0;\n"
|
||||
"Generated-By: Babel 2.17.0\n"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_models.md:1
|
||||
msgid "Model Support"
|
||||
msgstr "模型支持"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_models.md:3
|
||||
msgid "Text-only Language Models"
|
||||
msgstr "纯文本语言模型"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_models.md:5
|
||||
#: ../../user_guide/support_matrix/supported_models.md:38
|
||||
msgid "Generative Models"
|
||||
msgstr "生成模型"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_models.md
|
||||
msgid "Model"
|
||||
msgstr "模型"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_models.md
|
||||
msgid "Supported"
|
||||
msgstr "支持"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_models.md
|
||||
msgid "Note"
|
||||
msgstr "注释"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_models.md
|
||||
msgid "DeepSeek v3"
|
||||
msgstr "DeepSeek v3"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_models.md
|
||||
msgid "✅"
|
||||
msgstr "✅"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_models.md
|
||||
msgid "DeepSeek R1"
|
||||
msgstr "DeepSeek R1"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_models.md
|
||||
msgid "DeepSeek Distill (Qwen/LLama)"
|
||||
msgstr "DeepSeek 精炼(Qwen/LLama)"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_models.md
|
||||
msgid "Qwen3"
|
||||
msgstr "Qwen3"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_models.md
|
||||
msgid "Qwen3-Moe"
|
||||
msgstr "Qwen3-Moe"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_models.md
|
||||
msgid "Qwen2.5"
|
||||
msgstr "Qwen2.5"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_models.md
|
||||
msgid "QwQ-32B"
|
||||
msgstr "QwQ-32B"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_models.md
|
||||
msgid "LLama3.1/3.2"
|
||||
msgstr "LLama3.1/3.2"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_models.md
|
||||
msgid "Internlm"
|
||||
msgstr "Internlm"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_models.md
|
||||
msgid "Baichuan"
|
||||
msgstr "百川"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_models.md
|
||||
msgid "Phi-4-mini"
|
||||
msgstr "Phi-4-mini"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_models.md
|
||||
msgid "MiniCPM"
|
||||
msgstr "MiniCPM"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_models.md
|
||||
msgid "MiniCPM3"
|
||||
msgstr "MiniCPM3"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_models.md
|
||||
msgid "LLama4"
|
||||
msgstr "LLama4"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_models.md
|
||||
msgid "Mistral"
|
||||
msgstr "Mistral"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_models.md
|
||||
msgid "Need test"
|
||||
msgstr "需要测试"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_models.md
|
||||
msgid "DeepSeek v2.5"
|
||||
msgstr "DeepSeek v2.5"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_models.md
|
||||
msgid "Gemma-2"
|
||||
msgstr "Gemma-2"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_models.md
|
||||
msgid "Mllama"
|
||||
msgstr "Mllama"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_models.md
|
||||
msgid "Gemma-3"
|
||||
msgstr "Gemma-3"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_models.md
|
||||
msgid "❌"
|
||||
msgstr "❌"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_models.md
|
||||
msgid "[#496](https://github.com/vllm-project/vllm-ascend/issues/496)"
|
||||
msgstr "[#496](https://github.com/vllm-project/vllm-ascend/issues/496)"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_models.md
|
||||
msgid "ChatGLM"
|
||||
msgstr "ChatGLM"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_models.md
|
||||
msgid "[#554](https://github.com/vllm-project/vllm-ascend/issues/554)"
|
||||
msgstr "[#554](https://github.com/vllm-project/vllm-ascend/issues/554)"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_models.md:29
|
||||
msgid "Pooling Models"
|
||||
msgstr "池化模型"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_models.md
|
||||
msgid "XLM-RoBERTa-based"
|
||||
msgstr "基于XLM-RoBERTa"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_models.md
|
||||
msgid "Molmo"
|
||||
msgstr "Molmo"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_models.md:36
|
||||
msgid "Multimodal Language Models"
|
||||
msgstr "多模态语言模型"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_models.md
|
||||
msgid "Qwen2-VL"
|
||||
msgstr "Qwen2-VL"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_models.md
|
||||
msgid "Qwen2.5-VL"
|
||||
msgstr "Qwen2.5-VL"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_models.md
|
||||
msgid "LLaVA 1.5"
|
||||
msgstr "LLaVA 1.5"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_models.md
|
||||
msgid "LLaVA 1.6"
|
||||
msgstr "LLaVA 1.6"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_models.md
|
||||
msgid "[#553](https://github.com/vllm-project/vllm-ascend/issues/553)"
|
||||
msgstr "[#553](https://github.com/vllm-project/vllm-ascend/issues/553)"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_models.md
|
||||
msgid "InternVL2"
|
||||
msgstr "InternVL2"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_models.md
|
||||
msgid "InternVL2.5"
|
||||
msgstr "InternVL2.5"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_models.md
|
||||
msgid "Qwen2-Audio"
|
||||
msgstr "Qwen2-Audio"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_models.md
|
||||
msgid "LLaVA-Next"
|
||||
msgstr "LLaVA-Next"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_models.md
|
||||
msgid "LLaVA-Next-Video"
|
||||
msgstr "LLaVA-Next-Video"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_models.md
|
||||
msgid "Phi-3-Vison/Phi-3.5-Vison"
|
||||
msgstr "Phi-3-Vison/Phi-3.5-Vison"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_models.md
|
||||
msgid "GLM-4v"
|
||||
msgstr "GLM-4v"
|
||||
|
||||
#: ../../user_guide/support_matrix/supported_models.md
|
||||
msgid "Ultravox"
|
||||
msgstr "Ultravox"
|
||||
Reference in New Issue
Block a user