|
38 | 38 | get_sentence_transformer_tokenizer_config, is_encoder_decoder,
|
39 | 39 | try_get_generation_config, uses_mrope)
|
40 | 40 | from vllm.transformers_utils.s3_utils import S3Model
|
41 |
| -from vllm.transformers_utils.utils import is_s3 |
| 41 | +from vllm.transformers_utils.utils import is_s3, maybe_model_redirect |
42 | 42 | from vllm.utils import (GiB_bytes, LayerBlockType, cuda_device_count_stateless,
|
43 | 43 | get_cpu_memory, random_uuid, resolve_obj_by_qualname)
|
44 | 44 |
|
@@ -266,9 +266,13 @@ def __init__(
|
266 | 266 | override_generation_config: Optional[dict[str, Any]] = None,
|
267 | 267 | model_impl: Union[str, ModelImpl] = ModelImpl.AUTO,
|
268 | 268 | ) -> None:
|
269 |
| - self.model = model |
| 269 | + self.model = maybe_model_redirect(model) |
| 270 | + self.tokenizer = maybe_model_redirect(tokenizer) |
| 271 | + |
270 | 272 | self.hf_config_path = hf_config_path
|
271 |
| - self.tokenizer = tokenizer |
| 273 | + if isinstance(hf_config_path, str): |
| 274 | + self.hf_config_path = maybe_model_redirect(hf_config_path) |
| 275 | + |
272 | 276 | self.tokenizer_mode = tokenizer_mode
|
273 | 277 | self.trust_remote_code = trust_remote_code
|
274 | 278 | self.allowed_local_media_path = allowed_local_media_path
|
|
0 commit comments