[ModelRunner] Fix cuda hard code in model runner (#155)
### What this PR does / why we need it? 1. Fix cuda hard code in model runner. 2. Fix tutorials doc rendering error. ### Does this PR introduce _any_ user-facing change? no. ### How was this patch tested? no. Signed-off-by: Shanshan Shen <467638484@qq.com>
This commit is contained in:
@@ -237,7 +237,10 @@ docker run \
|
|||||||
```
|
```
|
||||||
|
|
||||||
Choose one machine as head node, the other are worker nodes, then start ray on each machine:
|
Choose one machine as head node, the other are worker nodes, then start ray on each machine:
|
||||||
:::{note} Check out your `nic_name` by command `ip addr` :::
|
|
||||||
|
:::{note}
|
||||||
|
Check out your `nic_name` by command `ip addr`.
|
||||||
|
:::
|
||||||
|
|
||||||
```shell
|
```shell
|
||||||
# Head node
|
# Head node
|
||||||
|
|||||||
@@ -1113,8 +1113,8 @@ class NPUModelRunner(NPUModelRunnerBase[ModelInputForNPUWithSamplingMetadata]):
|
|||||||
|
|
||||||
if (self.observability_config is not None
|
if (self.observability_config is not None
|
||||||
and self.observability_config.collect_model_forward_time):
|
and self.observability_config.collect_model_forward_time):
|
||||||
model_forward_start = torch.cuda.Event(enable_timing=True)
|
model_forward_start = torch_npu.npu.Event(enable_timing=True)
|
||||||
model_forward_end = torch.cuda.Event(enable_timing=True)
|
model_forward_end = torch_npu.npu.Event(enable_timing=True)
|
||||||
model_forward_start.record()
|
model_forward_start.record()
|
||||||
|
|
||||||
if not bypass_model_exec:
|
if not bypass_model_exec:
|
||||||
|
|||||||
Reference in New Issue
Block a user