### What this PR does / why we need it?
In graph + RL scenario, we only capture the graph once, and the weight
address is expected to be the same across iterations. However, when
calling .contiguous() on weight tensors, a new memory address may be
allocated, causing the graph to capture incorrect weight addresses.
This PR modifies the weight update logic in AscendMLAImpl and
AscendSFAImpl to use copy_() instead of reassignment, ensuring the
weight addresses remain consistent across iterations.
detailed in #7473
### Does this PR introduce _any_ user-facing change?
### How was this patch tested?
Signed-off-by: Debonex <719893090@qq.com>