2025-04-19 17:38:18 +08:00
|
|
|
export HCCL_IF_IP=${local_ip}
|
|
|
|
|
export GLOO_SOCKET_IFNAME=${ifname}
|
|
|
|
|
export TP_SOCKET_IFNAME=${ifname}
|
|
|
|
|
export HCCL_SOCKET_IFNAME=${ifname}
|
|
|
|
|
|
|
|
|
|
# dp_size = node_size * dp_per_node
|
|
|
|
|
node_size=1
|
|
|
|
|
node_rank=0
|
2025-04-23 16:23:25 +08:00
|
|
|
dp_per_node=4
|
2025-04-19 17:38:18 +08:00
|
|
|
master_addr=127.0.0.1
|
|
|
|
|
master_port=12345
|
|
|
|
|
|
|
|
|
|
rm -rf ./.torchair_cache/
|
|
|
|
|
rm -rf ./dynamo_*
|
|
|
|
|
rm -rf /root/ascend/log/debug/plog/*
|
|
|
|
|
|
|
|
|
|
torchrun --nproc_per_node ${dp_per_node} --nnodes ${node_size} \
|
|
|
|
|
--node_rank ${node_rank} --master_addr ${master_addr} --master_port ${master_port} \
|
|
|
|
|
data_parallel.py
|