[Misc] Fix metrics, weight update lock, request logging (#2543)

This commit is contained in:
Lianmin Zheng
2024-12-22 06:25:57 -08:00
parent 7d672d277b
commit 8496701934
11 changed files with 411 additions and 314 deletions

View File

@@ -184,26 +184,35 @@ class MHATokenToKVPool(BaseTokenToKVPool):
device: str,
):
super().__init__(size, dtype, device)
self.head_num = head_num
self.head_dim = head_dim
self.layer_num = layer_num
self._create_buffers()
def _create_buffers(self):
# [size, head_num, head_dim] for each layer
# The padded slot 0 is used for writing dummy outputs from padded tokens.
self.k_buffer = [
torch.empty(
(size + 1, head_num, head_dim),
(self.size + 1, self.head_num, self.head_dim),
dtype=self.store_dtype,
device=device,
device=self.device,
)
for _ in range(layer_num)
for _ in range(self.layer_num)
]
self.v_buffer = [
torch.empty(
(size + 1, head_num, head_dim),
(self.size + 1, self.head_num, self.head_dim),
dtype=self.store_dtype,
device=device,
device=self.device,
)
for _ in range(layer_num)
for _ in range(self.layer_num)
]
def _clear_buffers(self):
del self.k_buffer
del self.v_buffer
def get_key_buffer(self, layer_id: int):
if self.store_dtype != self.dtype:
return self.k_buffer[layer_id].view(self.dtype)
@@ -245,7 +254,6 @@ def copy_two_array(loc, dst_1, src_1, dst_2, src_2, dtype, store_dtype):
class MLATokenToKVPool(BaseTokenToKVPool):
def __init__(
self,
size: int,
@@ -298,7 +306,6 @@ class MLATokenToKVPool(BaseTokenToKVPool):
class DoubleSparseTokenToKVPool(BaseTokenToKVPool):
def __init__(
self,
size: int,