vllm-ascend vnpu v1
This commit is contained in:
22
.github.backup/ISSUE_TEMPLATE/100-documentation.yml
Normal file
22
.github.backup/ISSUE_TEMPLATE/100-documentation.yml
Normal file
@@ -0,0 +1,22 @@
|
||||
name: 📚 Documentation
|
||||
description: Report an issue related to https://vllm-ascend.readthedocs.org
|
||||
title: "[Doc]: "
|
||||
labels: ["documentation"]
|
||||
|
||||
body:
|
||||
- type: textarea
|
||||
attributes:
|
||||
label: 📚 The doc issue
|
||||
description: >
|
||||
A clear and concise description of what content in https://vllm-ascend.readthedocs.org/ is an issue.
|
||||
validations:
|
||||
required: true
|
||||
- type: textarea
|
||||
attributes:
|
||||
label: Suggest a potential alternative/fix
|
||||
description: >
|
||||
Tell us how we could improve the documentation in this regard.
|
||||
- type: markdown
|
||||
attributes:
|
||||
value: >
|
||||
Thanks for contributing 🎉!
|
||||
37
.github.backup/ISSUE_TEMPLATE/110-user-story.yml
Normal file
37
.github.backup/ISSUE_TEMPLATE/110-user-story.yml
Normal file
@@ -0,0 +1,37 @@
|
||||
name: 📚 User Story
|
||||
description: Apply for an user story to be displayed on https://vllm-ascend.readthedocs.io/en/latest/community/user_stories/index.html
|
||||
title: "[User Story]: "
|
||||
labels: ["user-story"]
|
||||
|
||||
body:
|
||||
- type: textarea
|
||||
attributes:
|
||||
label: 📚 Title
|
||||
description: >
|
||||
A clear title about what your user story is about.
|
||||
validations:
|
||||
required: true
|
||||
- type: textarea
|
||||
attributes:
|
||||
label: About / Introduction
|
||||
description: >
|
||||
A brief introduction about the background of your use case, like your scenario, hardware size etc.
|
||||
- type: textarea
|
||||
attributes:
|
||||
label: Bussiness Challenges
|
||||
description: >
|
||||
Tell us how what kind of challenge you faced in this user story.
|
||||
- type: textarea
|
||||
attributes:
|
||||
label: Solving challenges with vLLM Ascend and benefits
|
||||
description: >
|
||||
Tell us how vLLM Ascend helped you overcome the challenges, including details like how you use it, what version you used, hardware info, etc. And what kind of benefit do you get from using vLLM Ascend
|
||||
- type: textarea
|
||||
attributes:
|
||||
label: Extra Info
|
||||
description: >
|
||||
Any extra infomation you want to include in this story
|
||||
- type: markdown
|
||||
attributes:
|
||||
value: >
|
||||
Thanks for contributing 🎉!
|
||||
42
.github.backup/ISSUE_TEMPLATE/200-installation.yml
Normal file
42
.github.backup/ISSUE_TEMPLATE/200-installation.yml
Normal file
@@ -0,0 +1,42 @@
|
||||
name: 🛠️ Installation
|
||||
description: Report an issue here when you hit errors during installation.
|
||||
title: "[Installation]: "
|
||||
labels: ["installation"]
|
||||
|
||||
body:
|
||||
- type: markdown
|
||||
attributes:
|
||||
value: >
|
||||
#### Before submitting an issue, please make sure the issue hasn't been already addressed by searching through [the existing and past issues](https://github.com/vllm-project/vllm-ascend/issues?q=is%3Aissue+sort%3Acreated-desc+).
|
||||
- type: textarea
|
||||
attributes:
|
||||
label: Your current environment
|
||||
description: |
|
||||
Please run the following and paste the output below.
|
||||
```sh
|
||||
npu-smi info
|
||||
cat /usr/local/Ascend/ascend-toolkit/latest/"$(uname -i)"-linux/ascend_toolkit_install.info
|
||||
wget https://raw.githubusercontent.com/vllm-project/vllm/main/collect_env.py
|
||||
# For security purposes, please feel free to check the contents of collect_env.py before running it.
|
||||
python collect_env.py
|
||||
```
|
||||
It is suggested to download and execute the latest script, as vllm might frequently update the diagnosis information needed for accurately and quickly responding to issues.
|
||||
value: |
|
||||
```text
|
||||
The output of `python collect_env.py`
|
||||
```
|
||||
validations:
|
||||
required: true
|
||||
- type: textarea
|
||||
attributes:
|
||||
label: How you are installing vllm and vllm-ascend
|
||||
description: |
|
||||
Paste the full command you are trying to execute.
|
||||
value: |
|
||||
```sh
|
||||
pip install -vvv vllm vllm-ascend
|
||||
```
|
||||
- type: markdown
|
||||
attributes:
|
||||
value: >
|
||||
Thanks for contributing 🎉!
|
||||
40
.github.backup/ISSUE_TEMPLATE/300-usage.yml
Normal file
40
.github.backup/ISSUE_TEMPLATE/300-usage.yml
Normal file
@@ -0,0 +1,40 @@
|
||||
name: 💻 Usage
|
||||
description: Raise an issue here if you don't know how to use vllm on Ascend.
|
||||
title: "[Usage]: "
|
||||
labels: ["usage"]
|
||||
|
||||
body:
|
||||
- type: markdown
|
||||
attributes:
|
||||
value: >
|
||||
#### Before submitting an issue, please make sure the issue hasn't been already addressed by searching through [the existing and past issues](https://github.com/vllm-project/vllm-ascend/issues?q=is%3Aissue+sort%3Acreated-desc+).
|
||||
- type: textarea
|
||||
attributes:
|
||||
label: Your current environment
|
||||
description: |
|
||||
Please run the following and paste the output below.
|
||||
```sh
|
||||
npu-smi info
|
||||
cat /usr/local/Ascend/ascend-toolkit/latest/"$(uname -i)"-linux/ascend_toolkit_install.info
|
||||
wget https://raw.githubusercontent.com/vllm-project/vllm/main/collect_env.py
|
||||
# For security purposes, please feel free to check the contents of collect_env.py before running it.
|
||||
python collect_env.py
|
||||
```
|
||||
It is suggested to download and execute the latest script, as vllm might frequently update the diagnosis information needed for accurately and quickly responding to issues.
|
||||
value: |
|
||||
```text
|
||||
The output of above commands
|
||||
```
|
||||
validations:
|
||||
required: true
|
||||
- type: textarea
|
||||
attributes:
|
||||
label: How would you like to use vllm on ascend
|
||||
description: |
|
||||
A detailed description of how you want to use vllm on ascend.
|
||||
value: |
|
||||
I want to run inference of a [specific model](put link here). I don't know how to integrate it with vllm.
|
||||
- type: markdown
|
||||
attributes:
|
||||
value: >
|
||||
Thanks for contributing 🎉!
|
||||
81
.github.backup/ISSUE_TEMPLATE/400-bug-report.yml
Normal file
81
.github.backup/ISSUE_TEMPLATE/400-bug-report.yml
Normal file
@@ -0,0 +1,81 @@
|
||||
name: 🐛 Bug report
|
||||
description: Raise an issue here if you find a bug.
|
||||
title: "[Bug]: "
|
||||
labels: ["bug"]
|
||||
|
||||
body:
|
||||
- type: markdown
|
||||
attributes:
|
||||
value: >
|
||||
#### Before submitting an issue, please make sure the issue hasn't been already addressed by searching through [the existing and past issues](https://github.com/vllm-project/vllm-ascend/issues?q=is%3Aissue+sort%3Acreated-desc+).
|
||||
- type: textarea
|
||||
attributes:
|
||||
label: Your current environment
|
||||
description: |
|
||||
Please run the following and paste the output below.
|
||||
```sh
|
||||
wget https://raw.githubusercontent.com/vllm-project/vllm-ascend/main/collect_env.py
|
||||
# For security purposes, please feel free to check the contents of collect_env.py before running it.
|
||||
python collect_env.py
|
||||
```
|
||||
It is suggested to download and execute the latest script, as vllm might frequently update the diagnosis information needed for accurately and quickly responding to issues.
|
||||
value: |
|
||||
<details>
|
||||
<summary>The output of `python collect_env.py`</summary>
|
||||
|
||||
```text
|
||||
Your output of above commands here
|
||||
```
|
||||
|
||||
</details>
|
||||
validations:
|
||||
required: true
|
||||
- type: textarea
|
||||
attributes:
|
||||
label: 🐛 Describe the bug
|
||||
description: |
|
||||
Please provide a clear and concise description of what the bug is.
|
||||
|
||||
If relevant, add a minimal example so that we can reproduce the error by running the code. It is very important for the snippet to be as succinct (minimal) as possible, so please take time to trim down any irrelevant code to help us debug efficiently. We are going to copy-paste your code and we expect to get the same result as you did: avoid any external data, and include the relevant imports, etc. For example:
|
||||
|
||||
```python
|
||||
from vllm import LLM, SamplingParams
|
||||
|
||||
prompts = [
|
||||
"Hello, my name is",
|
||||
"The president of the United States is",
|
||||
"The capital of France is",
|
||||
"The future of AI is",
|
||||
]
|
||||
sampling_params = SamplingParams(temperature=0.8, top_p=0.95)
|
||||
|
||||
llm = LLM(model="facebook/opt-125m")
|
||||
|
||||
outputs = llm.generate(prompts, sampling_params)
|
||||
|
||||
# Print the outputs.
|
||||
for output in outputs:
|
||||
prompt = output.prompt
|
||||
generated_text = output.outputs[0].text
|
||||
print(f"Prompt: {prompt!r}, Generated text: {generated_text!r}")
|
||||
```
|
||||
|
||||
If the code is too long (hopefully, it isn't), feel free to put it in a public gist and link it in the issue: https://gist.github.com.
|
||||
|
||||
Please also paste or describe the results you observe instead of the expected results. If you observe an error, please paste the error message including the **full** traceback of the exception. It may be relevant to wrap error messages in ```` ```triple quotes blocks``` ````.
|
||||
|
||||
Please set the environment variable `export VLLM_LOGGING_LEVEL=DEBUG` to turn on more logging to help debugging potential issues.
|
||||
|
||||
If you experienced crashes or hangs, it would be helpful to run vllm with `export VLLM_TRACE_FUNCTION=1` . All the function calls in vllm will be recorded. Inspect these log files, and tell which function crashes or hangs.
|
||||
placeholder: |
|
||||
A clear and concise description of what the bug is.
|
||||
|
||||
```python
|
||||
# Sample code to reproduce the problem
|
||||
```
|
||||
|
||||
```
|
||||
The error message you got, with the full traceback.
|
||||
```
|
||||
validations:
|
||||
required: true
|
||||
31
.github.backup/ISSUE_TEMPLATE/500-feature-request.yml
Normal file
31
.github.backup/ISSUE_TEMPLATE/500-feature-request.yml
Normal file
@@ -0,0 +1,31 @@
|
||||
name: 🚀 Feature request
|
||||
description: Submit a proposal/request for a new vllm-ascend feature
|
||||
title: "[Feature]: "
|
||||
labels: ["feature request"]
|
||||
|
||||
body:
|
||||
- type: markdown
|
||||
attributes:
|
||||
value: >
|
||||
#### Before submitting an issue, please make sure the issue hasn't been already addressed by searching through [the existing and past issues](https://github.com/vllm-project/vllm-ascend/issues?q=is%3Aissue+sort%3Acreated-desc+).
|
||||
- type: textarea
|
||||
attributes:
|
||||
label: 🚀 The feature, motivation and pitch
|
||||
description: >
|
||||
A clear and concise description of the feature proposal. Please outline the motivation for the proposal. Is your feature request related to a specific problem? e.g., *"I'm working on X and would like Y to be possible"*. If this is related to another GitHub issue, please link here too.
|
||||
validations:
|
||||
required: true
|
||||
- type: textarea
|
||||
attributes:
|
||||
label: Alternatives
|
||||
description: >
|
||||
A description of any alternative solutions or features you've considered, if any.
|
||||
- type: textarea
|
||||
attributes:
|
||||
label: Additional context
|
||||
description: >
|
||||
Add any other context or screenshots about the feature request.
|
||||
- type: markdown
|
||||
attributes:
|
||||
value: >
|
||||
Thanks for contributing 🎉!
|
||||
33
.github.backup/ISSUE_TEMPLATE/600-new-model.yml
Normal file
33
.github.backup/ISSUE_TEMPLATE/600-new-model.yml
Normal file
@@ -0,0 +1,33 @@
|
||||
name: 🤗 Support request for new model supported from huggingface/modelscope/modelers on Ascend
|
||||
description: Submit a proposal/request for a new model from huggingface/modelscope/modelers on Ascend
|
||||
title: "[New Model]: "
|
||||
labels: ["new model"]
|
||||
|
||||
body:
|
||||
- type: markdown
|
||||
attributes:
|
||||
value: >
|
||||
#### Before submitting an issue, please make sure the issue hasn't been already addressed by searching through [the existing and past issues](https://github.com/vllm-project/vllm-ascend/issues?q=is%3Aissue+sort%3Acreated-desc+).
|
||||
|
||||
#### We also highly recommend you read https://vllm-ascend.readthedocs.io/en/latest/user_guide/supported_models.html first to know which model already supported.
|
||||
- type: textarea
|
||||
attributes:
|
||||
label: The model to consider.
|
||||
description: >
|
||||
A huggingface/modelscope/modelers url, pointing to the model, e.g. https://huggingface.co/openai-community/gpt2 .
|
||||
validations:
|
||||
required: true
|
||||
- type: textarea
|
||||
attributes:
|
||||
label: The closest model vllm already supports.
|
||||
description: >
|
||||
Here is the list of models already supported by vllm: https://vllm-ascend.readthedocs.io/en/latest/user_guide/supported_models.html . Which model is the most similar to the model you want to add support for?
|
||||
- type: textarea
|
||||
attributes:
|
||||
label: What's your difficulty of supporting the model you want?
|
||||
description: >
|
||||
For example, any new operators or new architecture?
|
||||
- type: markdown
|
||||
attributes:
|
||||
value: >
|
||||
Thanks for contributing 🎉!
|
||||
54
.github.backup/ISSUE_TEMPLATE/700-performance-discussion.yml
Normal file
54
.github.backup/ISSUE_TEMPLATE/700-performance-discussion.yml
Normal file
@@ -0,0 +1,54 @@
|
||||
name: ⚡ Discussion on the performance of vllm-ascend
|
||||
description: Submit a proposal/discussion about the performance of vllm-ascend
|
||||
title: "[Performance]: "
|
||||
labels: ["performance"]
|
||||
|
||||
body:
|
||||
- type: markdown
|
||||
attributes:
|
||||
value: >
|
||||
#### Before submitting an issue, please make sure the issue hasn't been already addressed by searching through [the existing and past issues](https://github.com/vllm-project/vllm-ascend/issues?q=is%3Aissue+sort%3Acreated-desc+).
|
||||
- type: textarea
|
||||
attributes:
|
||||
label: Proposal to improve performance
|
||||
description: >
|
||||
How do you plan to improve vllm-ascend's performance?
|
||||
validations:
|
||||
required: false
|
||||
- type: textarea
|
||||
attributes:
|
||||
label: Report of performance regression
|
||||
description: >
|
||||
Please provide detailed description of performance comparison to confirm the regression. You may want to run the benchmark script at https://github.com/vllm-project/vllm/tree/main/benchmarks .
|
||||
validations:
|
||||
required: false
|
||||
- type: textarea
|
||||
attributes:
|
||||
label: Misc discussion on performance
|
||||
description: >
|
||||
Anything about the performance.
|
||||
validations:
|
||||
required: false
|
||||
- type: textarea
|
||||
attributes:
|
||||
label: Your current environment (if you think it is necessary)
|
||||
description: |
|
||||
Please run the following and paste the output below.
|
||||
```sh
|
||||
npu-smi info
|
||||
cat /usr/local/Ascend/ascend-toolkit/latest/"$(uname -i)"-linux/ascend_toolkit_install.info
|
||||
wget https://raw.githubusercontent.com/vllm-project/vllm/main/collect_env.py
|
||||
# For security purposes, please feel free to check the contents of collect_env.py before running it.
|
||||
python collect_env.py
|
||||
```
|
||||
It is suggested to download and execute the latest script, as vllm might frequently update the diagnosis information needed for accurately and quickly responding to issues.
|
||||
value: |
|
||||
```text
|
||||
The output of `python collect_env.py`
|
||||
```
|
||||
validations:
|
||||
required: false
|
||||
- type: markdown
|
||||
attributes:
|
||||
value: >
|
||||
Thanks for contributing 🎉!
|
||||
49
.github.backup/ISSUE_TEMPLATE/750-RFC.yml
Normal file
49
.github.backup/ISSUE_TEMPLATE/750-RFC.yml
Normal file
@@ -0,0 +1,49 @@
|
||||
name: 💬 Request for comments (RFC).
|
||||
description: Ask for feedback on major architectural changes or design choices.
|
||||
title: "[RFC]: "
|
||||
labels: ["RFC"]
|
||||
|
||||
body:
|
||||
- type: markdown
|
||||
attributes:
|
||||
value: >
|
||||
#### Please take a look at previous [RFCs](https://github.com/vllm-project/vllm-ascend/issues?q=label%3ARFC+sort%3Aupdated-desc) for reference.
|
||||
- type: textarea
|
||||
attributes:
|
||||
label: Motivation.
|
||||
description: >
|
||||
The motivation of the RFC.
|
||||
validations:
|
||||
required: true
|
||||
- type: textarea
|
||||
attributes:
|
||||
label: Proposed Change.
|
||||
description: >
|
||||
The proposed change of the RFC.
|
||||
validations:
|
||||
required: true
|
||||
- type: textarea
|
||||
attributes:
|
||||
label: Feedback Period.
|
||||
description: >
|
||||
The feedback period of the RFC. Usually at least one week.
|
||||
validations:
|
||||
required: false
|
||||
- type: textarea
|
||||
attributes:
|
||||
label: CC List.
|
||||
description: >
|
||||
The list of people you want to CC.
|
||||
validations:
|
||||
required: false
|
||||
- type: textarea
|
||||
attributes:
|
||||
label: Any Other Things.
|
||||
description: >
|
||||
Any other things you would like to mention, such as feature branch request.
|
||||
validations:
|
||||
required: false
|
||||
- type: markdown
|
||||
attributes:
|
||||
value: >
|
||||
Thanks for contributing 🎉!
|
||||
21
.github.backup/ISSUE_TEMPLATE/800-others.yml
Normal file
21
.github.backup/ISSUE_TEMPLATE/800-others.yml
Normal file
@@ -0,0 +1,21 @@
|
||||
name: 🎲 Others
|
||||
description: Submit a discussion as you like. Note that developers are heavily overloaded and we mainly rely on community users to answer these issues.
|
||||
title: "[Misc]: "
|
||||
labels: ["misc"]
|
||||
|
||||
body:
|
||||
- type: markdown
|
||||
attributes:
|
||||
value: >
|
||||
#### Before submitting an issue, please make sure the issue hasn't been already addressed by searching through [the existing and past issues](https://github.com/vllm-project/vllm-ascend/issues?q=is%3Aissue+sort%3Acreated-desc+).
|
||||
- type: textarea
|
||||
attributes:
|
||||
label: Anything you want to discuss about vllm on ascend.
|
||||
description: >
|
||||
Anything you want to discuss about vllm on ascend.
|
||||
validations:
|
||||
required: true
|
||||
- type: markdown
|
||||
attributes:
|
||||
value: >
|
||||
Thanks for contributing 🎉!
|
||||
104
.github.backup/ISSUE_TEMPLATE/900-release-checklist.yml
Normal file
104
.github.backup/ISSUE_TEMPLATE/900-release-checklist.yml
Normal file
@@ -0,0 +1,104 @@
|
||||
name: Release Checklist
|
||||
description: Generate a release checklist issue when prepare a new release.(Used for release team)
|
||||
title: "[Release]: Release checklist for v"
|
||||
|
||||
body:
|
||||
- type: textarea
|
||||
attributes:
|
||||
description: >
|
||||
Brief info for the new release.
|
||||
label: Release Checklist
|
||||
value: >
|
||||
**Release Version**:
|
||||
|
||||
**Release Branch**:
|
||||
|
||||
**Release Date**:
|
||||
|
||||
**Release Manager**:
|
||||
- type: textarea
|
||||
attributes:
|
||||
description: >
|
||||
Release notes.
|
||||
label: Prepare Release Note
|
||||
value: >
|
||||
- [ ] Create a new issue for release feedback
|
||||
|
||||
- [ ] Upgrade vllm version to the new version for CI and Dockerfile
|
||||
|
||||
- [ ] Write the release note PR.
|
||||
|
||||
- [ ] Update the feedback issue link in docs/source/faqs.md
|
||||
|
||||
- [ ] Add release note to docs/source/user_guide/release_notes.md
|
||||
|
||||
- [ ] Update release version in README.md and README.zh.md
|
||||
|
||||
- [ ] Update version info in docs/source/community/versioning_policy.md
|
||||
|
||||
- [ ] Update contributor info in docs/source/community/contributors.md
|
||||
|
||||
- [ ] Update package version in docs/conf.py
|
||||
- type: textarea
|
||||
attributes:
|
||||
description: >
|
||||
Make sure the code is merged.
|
||||
label: PR need Merge
|
||||
value: >
|
||||
- [ ] PR link1
|
||||
|
||||
- [ ] PR link2
|
||||
|
||||
- [ ] ...
|
||||
- type: textarea
|
||||
attributes:
|
||||
description: >
|
||||
Make sure the new Feature/Function is tested
|
||||
label: Functional Test
|
||||
value: >
|
||||
- [ ] Feature1
|
||||
|
||||
- [ ] Bug1
|
||||
|
||||
- [ ] ...
|
||||
- type: textarea
|
||||
attributes:
|
||||
description: >
|
||||
Make sure the doc is updated.
|
||||
label: Doc Test
|
||||
value: >
|
||||
- [ ] Tutorial is updated.
|
||||
|
||||
- [ ] User Guide is updated.
|
||||
|
||||
- [ ] Developer Guide is updated.
|
||||
- type: textarea
|
||||
attributes:
|
||||
description: >
|
||||
Make sure the artifacts is ready
|
||||
label: Prepare Artifacts
|
||||
value: >
|
||||
- [ ] Docker image is ready.
|
||||
|
||||
- [ ] Wheel package is ready.
|
||||
- type: textarea
|
||||
attributes:
|
||||
description: >
|
||||
Start to release.
|
||||
label: Release Step
|
||||
value: >
|
||||
- [ ] Release note PR is merged.
|
||||
|
||||
- [ ] Post the release on GitHub release page.
|
||||
|
||||
- [ ] Generate official doc page on https://app.readthedocs.org/dashboard/
|
||||
|
||||
- [ ] Wait for the wheel package to be available on https://pypi.org/project/vllm-ascend
|
||||
|
||||
- [ ] Wait for the docker image to be available on https://quay.io/ascend/vllm-ascend
|
||||
|
||||
- [ ] Upload 310p wheel to Github release page
|
||||
|
||||
- [ ] Broadcast the release news (By message, blog , etc)
|
||||
|
||||
- [ ] Close this issue
|
||||
1
.github.backup/ISSUE_TEMPLATE/config.yml
Normal file
1
.github.backup/ISSUE_TEMPLATE/config.yml
Normal file
@@ -0,0 +1 @@
|
||||
blank_issues_enabled: false
|
||||
Reference in New Issue
Block a user