From ef68eb28d8d45be6e0defe82245e16be9362e375 Mon Sep 17 00:00:00 2001 From: Cyrus Leung Date: Wed, 8 Jan 2025 13:40:09 +0800 Subject: [PATCH] [Bug] Fix pickling of `ModelConfig` when RunAI Model Streamer is used (#11825) Signed-off-by: DarkLight1337 --- vllm/config.py | 12 ++++++------ 1 file changed, 6 insertions(+), 6 deletions(-) diff --git a/vllm/config.py b/vllm/config.py index a9b6d6b19127f..44426489f686a 100644 --- a/vllm/config.py +++ b/vllm/config.py @@ -381,16 +381,16 @@ def maybe_pull_model_tokenizer_for_s3(self, model: str, """ if is_s3(model) or is_s3(tokenizer): if is_s3(model): - self.s3_model = S3Model() - self.s3_model.pull_files(model, allow_pattern=["*config.json"]) + s3_model = S3Model() + s3_model.pull_files(model, allow_pattern=["*config.json"]) self.model_weights = self.model - self.model = self.s3_model.dir + self.model = s3_model.dir if is_s3(tokenizer): - self.s3_tokenizer = S3Model() - self.s3_tokenizer.pull_files( + s3_tokenizer = S3Model() + s3_tokenizer.pull_files( model, ignore_pattern=["*.pt", "*.safetensors", "*.bin"]) - self.tokenizer = self.s3_tokenizer.dir + self.tokenizer = s3_tokenizer.dir def _init_multimodal_config( self, limit_mm_per_prompt: Optional[Mapping[str, int]]