diff --git a/examples/offline_inference_npu_v1.py b/examples/offline_inference_npu_v1.py index 72486f0..9f01c3a 100644 --- a/examples/offline_inference_npu_v1.py +++ b/examples/offline_inference_npu_v1.py @@ -17,6 +17,7 @@ # limitations under the License. # +# isort: skip_file import os os.environ["VLLM_USE_MODELSCOPE"] = "True" @@ -24,7 +25,6 @@ os.environ["VLLM_WORKER_MULTIPROC_METHOD"] = "spawn" from vllm import LLM, SamplingParams - if __name__ == "__main__": prompts = [ "Hello, my name is",