File tree Expand file tree Collapse file tree 1 file changed +0
-5
lines changed
python/ray/llm/_internal/serve/deployments/llm/vllm Expand file tree Collapse file tree 1 file changed +0
-5
lines changed Original file line number Diff line number Diff line change 4
4
5
5
from starlette .datastructures import State
6
6
from starlette .requests import Request
7
- from transformers .dynamic_module_utils import init_hf_modules
8
7
from vllm .engine .arg_utils import AsyncEngineArgs
9
8
from vllm .entrypoints .openai .cli_args import FrontendArgs
10
9
from vllm .entrypoints .openai .protocol import ErrorResponse as VLLMErrorResponse
@@ -110,10 +109,6 @@ def __init__(
110
109
"""
111
110
super ().__init__ (llm_config )
112
111
113
- # Ensure transformers_modules is initialized early in worker processes.
114
- # This is critical for models with trust_remote_code=True to avoid pickle errors.
115
- init_hf_modules ()
116
-
117
112
self .llm_config = llm_config
118
113
119
114
if vllm is None :
You can’t perform that action at this time.
0 commit comments