Improve process creation (#1534)
This commit is contained in:
@@ -4,7 +4,7 @@ import subprocess
|
||||
import unittest
|
||||
from unittest import mock
|
||||
|
||||
from sglang.srt.utils import prepare_model, prepare_tokenizer
|
||||
from sglang.srt.utils import prepare_model_and_tokenizer
|
||||
|
||||
|
||||
class TestDownloadFromModelScope(unittest.TestCase):
|
||||
@@ -21,25 +21,17 @@ class TestDownloadFromModelScope(unittest.TestCase):
|
||||
def tearDownClass(cls):
|
||||
pass
|
||||
|
||||
def test_prepare_model(self):
|
||||
def test_prepare_model_and_tokenizer(self):
|
||||
from modelscope.utils.file_utils import get_model_cache_root
|
||||
|
||||
model_cache_root = get_model_cache_root()
|
||||
if os.path.exists(model_cache_root):
|
||||
shutil.rmtree(model_cache_root)
|
||||
with mock.patch.dict(os.environ, self.with_modelscope_environ, clear=True):
|
||||
model_path = prepare_model(self.model)
|
||||
model_path, tokenizer_path = prepare_model_and_tokenizer(
|
||||
self.model, self.model
|
||||
)
|
||||
assert os.path.exists(os.path.join(model_path, "pytorch_model.bin"))
|
||||
|
||||
def test_prepare_tokenizer(self):
|
||||
from modelscope.utils.file_utils import get_model_cache_root
|
||||
|
||||
model_cache_root = get_model_cache_root()
|
||||
if os.path.exists(model_cache_root):
|
||||
shutil.rmtree(model_cache_root)
|
||||
with mock.patch.dict(os.environ, self.with_modelscope_environ, clear=True):
|
||||
tokenizer_path = prepare_tokenizer(self.model)
|
||||
assert not os.path.exists(os.path.join(tokenizer_path, "pytorch_model.bin"))
|
||||
assert os.path.exists(os.path.join(tokenizer_path, "config.json"))
|
||||
|
||||
|
||||
|
||||
@@ -26,7 +26,7 @@ class TestTritonAttnBackend(unittest.TestCase):
|
||||
)
|
||||
|
||||
if is_in_ci():
|
||||
assert output_throughput > 155, f"{output_throughput=}"
|
||||
assert output_throughput > 154, f"{output_throughput=}"
|
||||
|
||||
def test_mmlu(self):
|
||||
model = DEFAULT_MODEL_NAME_FOR_TEST
|
||||
|
||||
Reference in New Issue
Block a user