chore: add hf_xet dep (#6243)
This commit is contained in:
@@ -142,7 +142,7 @@ if __name__ == "__main__":
|
||||
parser.add_argument(
|
||||
"--model",
|
||||
default="meta-llama/Llama-4-Maverick-17B-128E-Instruct",
|
||||
help="Model name or ID",
|
||||
help="Model name or ID, only used for model name",
|
||||
)
|
||||
parser.add_argument(
|
||||
"--max-concurrency", type=int, default=144, help="Maximum concurrent requests"
|
||||
|
||||
@@ -31,3 +31,6 @@ pip install -e lmms-eval/
|
||||
|
||||
# Install FlashMLA for attention backend tests
|
||||
pip install git+https://github.com/deepseek-ai/FlashMLA.git
|
||||
|
||||
# Install hf_xet
|
||||
pip install huggingface_hub[hf_xet]
|
||||
|
||||
@@ -123,3 +123,6 @@ gdrcopy_copybw
|
||||
echo "=== Verify NVSHMEM ==="
|
||||
nvshmem-info -a
|
||||
# /opt/nvshmem/bin/perftest/device/pt-to-pt/shmem_put_bw
|
||||
|
||||
# Install hf_xet
|
||||
pip install huggingface_hub[hf_xet]
|
||||
|
||||
Reference in New Issue
Block a user