[1/N] Remove CacheConfig import in all model files (#1658)

This commit is contained in:
Byron Hsu
2024-10-14 09:06:34 -07:00
committed by GitHub
parent 02bc95796d
commit 56503d9bc9
30 changed files with 64 additions and 91 deletions

View File

@@ -18,7 +18,6 @@ from typing import Iterable, Optional, Tuple
import torch
from torch import nn
from transformers import LlamaConfig
from vllm.config import CacheConfig
from vllm.model_executor.model_loader.weight_utils import default_weight_loader
from sglang.srt.layers.logits_processor import LogitsProcessorOutput
@@ -32,7 +31,7 @@ class LlamaForClassification(nn.Module):
self,
config: LlamaConfig,
quant_config: Optional[QuantizationConfig] = None,
cache_config: Optional[CacheConfig] = None,
cache_config=None,
) -> None:
super().__init__()
self.config = config