From e5c46bf169aa01e5f6a7c17c018f40470939ee48 Mon Sep 17 00:00:00 2001 From: wangxiyuan Date: Wed, 14 Jan 2026 11:23:38 +0800 Subject: [PATCH] [CI] Fix lint CI (#5880) Quick fix for lint CI - vLLM version: v0.13.0 - vLLM main: https://github.com/vllm-project/vllm/commit/bde38c11df0ea066a740efe9b77fff5418be45df Signed-off-by: wangxiyuan --- .../quantization/llm-compressor/w8a8_int8_dynamic_moe.py | 7 ++----- 1 file changed, 2 insertions(+), 5 deletions(-) diff --git a/examples/quantization/llm-compressor/w8a8_int8_dynamic_moe.py b/examples/quantization/llm-compressor/w8a8_int8_dynamic_moe.py index d5f5d887..14ebd6e1 100644 --- a/examples/quantization/llm-compressor/w8a8_int8_dynamic_moe.py +++ b/examples/quantization/llm-compressor/w8a8_int8_dynamic_moe.py @@ -1,14 +1,11 @@ import torch -from transformers import AutoModelForCausalLM, AutoTokenizer - from llmcompressor import oneshot from llmcompressor.modifiers.quantization import QuantizationModifier +from transformers import AutoModelForCausalLM, AutoTokenizer MODEL_ID = "Qwen/Qwen3-30B-A3B-Instruct-2507" -model = AutoModelForCausalLM.from_pretrained( - MODEL_ID, dtype=torch.bfloat16, trust_remote_code=True -) +model = AutoModelForCausalLM.from_pretrained(MODEL_ID, dtype=torch.bfloat16, trust_remote_code=True) tokenizer = AutoTokenizer.from_pretrained(MODEL_ID) recipe = QuantizationModifier(