### What this PR does / why we need it?
**Scope of Changes**:
| File Path |
| :--- |
|` vllm_ascend/quantization/compressed_tensors/compressed_tensors.py`|
|` vllm_ascend/quantization/quant_config.py`|
|` vllm_ascend/quantization/utils.py`|
|` vllm_ascend/quantization/w4a16.py`|
|` vllm_ascend/quantization/w4a4_flatquant_dynamic.py`|
|` vllm_ascend/quantization/w4a8_dynamic.py`|
|` vllm_ascend/quantization/w8a16.py`|
|` vllm_ascend/quantization/w8a8.py`|
|` vllm_ascend/quantization/w8a8_dynamic.py`|
|` vllm_ascend/quantization/w8a8_pdmix.py`|
|` vllm_ascend/quantization/w8a8mxfp8.py`|
|` vllm_ascend/sample/rejection_sampler.py`|
|` vllm_ascend/sample/sampler.py`|
|` vllm_ascend/worker/block_table.py`|
### Does this PR introduce _any_ user-facing change?
### How was this patch tested?
- vLLM version: v0.13.0
- vLLM main:
2c24bc6996
Signed-off-by: MrZ20 <2609716663@qq.com>
This commit is contained in:
@@ -51,10 +51,7 @@ line-length = 120
|
||||
# Folder to be modified
|
||||
exclude = [
|
||||
"tests/**",
|
||||
# (7)
|
||||
"vllm_ascend/quantization/**",
|
||||
"vllm_ascend/sample/*.py",
|
||||
"vllm_ascend/worker/block_table.py",
|
||||
|
||||
# (8)
|
||||
"vllm_ascend/ops/__init__.py",
|
||||
"vllm_ascend/ops/activation.py",
|
||||
@@ -66,6 +63,7 @@ exclude = [
|
||||
"vllm_ascend/ops/vocab_parallel_embedding.py",
|
||||
"vllm_ascend/ops/weight_prefetch.py",
|
||||
"vllm_ascend/spec_decode/**",
|
||||
|
||||
# (10)
|
||||
"vllm_ascend/ops/*linear*.py",
|
||||
"vllm_ascend/worker/worker.py",
|
||||
@@ -76,6 +74,7 @@ exclude = [
|
||||
"vllm_ascend/worker/v2/**",
|
||||
"vllm_ascend/worker/npu_input_batch.py",
|
||||
"vllm_ascend/ops/rotary_embedding.py",
|
||||
|
||||
# (11)
|
||||
"vllm_ascend/ops/fused_moe/**",
|
||||
]
|
||||
|
||||
Reference in New Issue
Block a user