[Doc] Sensitive word modification (#8303)
<!-- Thanks for sending a pull request! BEFORE SUBMITTING, PLEASE READ https://docs.vllm.ai/en/latest/contributing/overview.html --> ### What this PR does / why we need it? This PR updates the documentation to replace specific hardware terms (e.g., HBM, 910B, 310P) with more generic or branded terms (e.g., on-chip memory, Atlas inference products) to comply with sensitive word requirements. ### Does this PR introduce _any_ user-facing change? no ### How was this patch tested? --------- Signed-off-by: herizhen <1270637059@qq.com> Signed-off-by: herizhen <59841270+herizhen@users.noreply.github.com>
This commit is contained in:
@@ -283,8 +283,8 @@ Send Dataset A to Instance 1 on Node 1 and record the Time to First Token
|
||||
### Preparation for Step 2
|
||||
|
||||
Before Step 2, send a fully random Dataset B to Instance 1. Due to the
|
||||
unified HBM/DRAM KV Cache with LRU (Least Recently Used) eviction policy,
|
||||
Dataset B's cache evicts Dataset A's cache from HBM, leaving Dataset A's
|
||||
unified on-chip memory/DRAM KV Cache with LRU (Least Recently Used) eviction policy,
|
||||
Dataset B's cache evicts Dataset A's cache from on-chip memory, leaving Dataset A's
|
||||
cache only in Node 1's DRAM.
|
||||
|
||||
### Step 2: Local DRAM Hit
|
||||
|
||||
Reference in New Issue
Block a user