Files
xc-llm-ascend/pytest.ini
Mengqing Cao df16c4f2bc [CI/UT] Ignore vllm/tests/test_vllm_port.py (#887)
Ignore `vllm/tests/test_vllm_port.py` in ut as no related to
vllm-ascend, and it is breaking CI

Signed-off-by: MengqingCao <cmq0113@163.com>
2025-05-16 18:52:59 +08:00

69 lines
2.7 KiB
INI

[pytest]
minversion = 6.0
norecursedirs =
vllm-empty/tests/prefix_caching
vllm-empty/tests/weight_loading
vllm-empty/tests/samplers
vllm-empty/tests/kernels
vllm-empty/tests/quantization
vllm-empty/tests/tool_use
vllm-empty/tests/runai_model_streamer_test
vllm-empty/tests/kv_transfer
vllm-empty/tests/plugins
vllm-empty/tests/plugins_tests
vllm-empty/tests/prompt_adapter
vllm-empty/tests/compile
vllm-empty/tests/lora
vllm-empty/tests/models
vllm-empty/tests/mistral_tool_use
vllm-empty/tests/standalone_tests
vllm-empty/tests/async_engine
vllm-empty/tests/mq_llm_engine
vllm-empty/tests/tokenization
vllm-empty/tests/core
vllm-empty/tests/tracing
vllm-empty/tests/engine
vllm-empty/tests/tensorizer_loader
vllm-empty/tests/entrypoints
vllm-empty/tests/model_executor
vllm-empty/tests/encoder_decoder
vllm-empty/tests/v1
vllm-empty/tests/spec_decode
vllm-empty/tests/multi_step
vllm-empty/tests/vllm_test_utils
vllm-empty/tests/tpu
vllm-empty/tests/distributed
vllm-empty/tests/basic_correctness
vllm-empty/tests/worker
vllm-empty/tests/metrics
vllm-empty/tests/neuron
; fastsafetensors not support npu now
vllm-empty/tests/fastsafetensors_loader
addopts = --ignore=vllm-empty/tests/test_utils.py
--ignore=vllm-empty/tests/test_config.py
--ignore=vllm-empty/tests/test_scalartype.py
--ignore=vllm-empty/tests/test_version.py
--ignore=vllm-empty/tests/test_embedded_commit.py
--ignore=vllm-empty/tests/test_inputs.py
--ignore=vllm-empty/tests/test_sharded_state_loader.py
--ignore=vllm-empty/tests/test_logger.py
--ignore=vllm-empty/tests/test_logits_processor.py
--ignore=vllm-empty/tests/test_regression.py
--ignore=vllm-empty/tests/models/decoder_only/language/test_aqlm.py
--ignore=vllm-empty/tests/models/decoder_only/language/test_gptq_marlin.py
--ignore=vllm-empty/tests/models/decoder_only/language/test_gptq_marlin_24.py
--ignore=vllm-empty/tests/multimodal/test_processing.py
--ignore=vllm-empty/tests/multimodal/test_processor_kwargs.py
--ignore=vllm-empty/tests/multimodal/test_utils.py
; Both V1 and V0 engine will be run in detokenizer/test_stop_reason
; VLLM_USE_V1=1 is not supported with device type=npu.
--ignore=vllm-empty/tests/detokenizer/test_stop_reason.py
; oom on llama-2-7b-hf
--ignore=vllm-empty/tests/detokenizer/test_stop_strings.py
; no need to run on vllm-ascend
--ignore=vllm-empty/tests/test_vllm_port.py
testpaths =
vllm-empty/tests