Skip to content

Commit 4fd5813

Browse files
authored
[TRTLLM-8933][chore] remove unused update_executor_config function (#8678)
Signed-off-by: junq <[email protected]>
1 parent c9b0879 commit 4fd5813

File tree

3 files changed

+19
-93
lines changed

3 files changed

+19
-93
lines changed

tensorrt_llm/_torch/pyexecutor/config.py

Lines changed: 0 additions & 54 deletions
Original file line numberDiff line numberDiff line change
@@ -3,11 +3,8 @@
33

44
from tensorrt_llm._torch.models.checkpoints.base_checkpoint_loader import \
55
BaseCheckpointLoader
6-
from tensorrt_llm.bindings.executor import ExecutorConfig
76

87
from ...llmapi.llm_args import LoadFormat, SamplerType
9-
from ...logger import logger
10-
from ...mapping import Mapping
118
from ..model_config import MoeLoadBalancerConfig
129
from .resource_manager import BaseResourceManager
1310

@@ -117,57 +114,6 @@ class PyTorchConfig:
117114
_limit_torch_cuda_mem_fraction: bool = True
118115

119116

120-
EXETENDED_EXECUTOR_CONFIG_FIELDS = [
121-
'backend',
122-
'pytorch_backend_config',
123-
'max_seq_len',
124-
'mapping',
125-
'hf_model_dir',
126-
'mm_encoder_only',
127-
]
128-
129-
130-
def update_executor_config(
131-
executor_config: ExecutorConfig,
132-
backend: Optional[str] = None,
133-
pytorch_backend_config: Optional[PyTorchConfig] = None,
134-
mapping: Optional[Mapping] = None,
135-
speculative_config: Optional["DecodingBaseConfig"] = None,
136-
hf_model_dir: Optional[str] = None,
137-
max_input_len: Optional[int] = None,
138-
max_seq_len: Optional[int] = None,
139-
checkpoint_format: Optional[str] = None,
140-
checkpoint_loader: Optional[BaseCheckpointLoader] = None,
141-
mm_encoder_only: bool = False):
142-
if backend is None:
143-
return
144-
145-
for field_name in EXETENDED_EXECUTOR_CONFIG_FIELDS:
146-
if hasattr(executor_config, field_name):
147-
raise AttributeError(
148-
f"{field_name} should be dynamically assigned.")
149-
setattr(executor_config, field_name, None)
150-
151-
executor_config.backend = backend
152-
executor_config.pytorch_backend_config = pytorch_backend_config
153-
executor_config.mapping = mapping
154-
executor_config.speculative_config = speculative_config
155-
executor_config.mm_encoder_only = mm_encoder_only
156-
157-
logger.info(f"{executor_config.pytorch_backend_config}")
158-
159-
executor_config.hf_model_dir = hf_model_dir
160-
161-
if max_input_len is not None:
162-
executor_config.max_input_len = max_input_len
163-
164-
if max_seq_len is not None:
165-
executor_config.max_seq_len = max_seq_len
166-
167-
executor_config.checkpoint_loader = _construct_checkpoint_loader(
168-
backend, checkpoint_loader, checkpoint_format)
169-
170-
171117
def _construct_checkpoint_loader(
172118
backend: str, checkpoint_loader: Optional[BaseCheckpointLoader],
173119
checkpoint_format: Optional[str]) -> Optional[BaseCheckpointLoader]:

tests/unittest/executor/test_base_worker.py

Lines changed: 6 additions & 32 deletions
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,6 @@
66
import torch
77

88
from tensorrt_llm._utils import mpi_comm, mpi_rank, mpi_world_size
9-
from tensorrt_llm.bindings import executor as tllm
109
from tensorrt_llm.llmapi.mpi_session import MpiPoolSession
1110

1211
# isort: off
@@ -15,7 +14,6 @@
1514
from utils.util import skip_single_gpu
1615
# isort: on
1716

18-
from tensorrt_llm._torch.pyexecutor.config import update_executor_config
1917
from tensorrt_llm.executor.base_worker import BaseWorker
2018
from tensorrt_llm.executor.request import GenerationRequest
2119
from tensorrt_llm.llmapi.llm_args import TorchLlmArgs
@@ -28,12 +26,16 @@
2826
class FakeWorker(BaseWorker):
2927

3028
def __init__(self, engine: str, tp_size: int = 1):
31-
llm_args, executor_config = create_fake_executor_config(engine, tp_size)
29+
llm_args = TorchLlmArgs(
30+
model=model_path,
31+
tensor_parallel_size=tp_size,
32+
backend='pytorch',
33+
enable_iter_perf_stats=True,
34+
)
3235
super().__init__(
3336
engine=engine,
3437
llm_args=llm_args,
3538
hf_model_dir=engine,
36-
executor_config=executor_config,
3739
)
3840
# Note: BaseWorker doesn't call setup_engine() automatically,
3941
# unlike GenerationExecutorWorker, so we need to call it manually
@@ -116,34 +118,6 @@ def test_fetch_responses_timeout(self, timeout: float):
116118
assert timeout / 2 <= elapsed <= timeout * 2, f"Latency out of expected range: {elapsed}"
117119

118120

119-
def create_fake_executor_config(model_path, tp_size=1):
120-
# Use TorchLlmArgs for PyTorch backend tests
121-
llm_args = TorchLlmArgs(
122-
model=model_path,
123-
tensor_parallel_size=tp_size,
124-
backend='pytorch',
125-
enable_iter_perf_stats=True,
126-
)
127-
128-
executor_config = tllm.ExecutorConfig(1)
129-
executor_config.max_batch_size = 1
130-
executor_config.model_world_size = tp_size
131-
132-
update_executor_config(
133-
executor_config,
134-
pytorch_backend_config=llm_args.get_pytorch_backend_config(),
135-
mapping=llm_args.parallel_config.to_mapping(),
136-
speculative_config=llm_args.speculative_config,
137-
hf_model_dir=model_path,
138-
max_input_len=20,
139-
max_seq_len=40,
140-
checkpoint_format=llm_args.checkpoint_format,
141-
checkpoint_loader=llm_args.checkpoint_loader,
142-
)
143-
144-
return llm_args, executor_config
145-
146-
147121
class TestRpcWorkerBaseTP2:
148122

149123
def setup_method(self):

tests/unittest/executor/test_rpc_worker.py

Lines changed: 13 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -6,12 +6,12 @@
66
from concurrent.futures import ProcessPoolExecutor
77

88
import pytest
9-
from test_base_worker import create_fake_executor_config
109

1110
from tensorrt_llm.executor.request import GenerationRequest
1211
from tensorrt_llm.executor.rpc import RPCClient
1312
from tensorrt_llm.executor.rpc.rpc_common import get_unique_ipc_addr
1413
from tensorrt_llm.executor.rpc_worker import RpcWorker
14+
from tensorrt_llm.llmapi.llm_args import TorchLlmArgs
1515
from tensorrt_llm.llmapi.mpi_session import MpiPoolSession
1616
from tensorrt_llm.sampling_params import SamplingParams
1717

@@ -28,8 +28,12 @@
2828
class TestRpcWorkerTP1:
2929

3030
def setup_method(self):
31-
self.llm_args, self.executor_config = create_fake_executor_config(
32-
model_path)
31+
self.llm_args = TorchLlmArgs(
32+
model=model_path,
33+
tensor_parallel_size=1,
34+
backend='pytorch',
35+
enable_iter_perf_stats=True,
36+
)
3337
self.pool, self.addr = self.create_worker_pool()
3438
self.client = self.create_rpc_client(self.addr)
3539
self.client.setup_engine().remote()
@@ -50,7 +54,6 @@ def create_worker_pool(self):
5054
RpcWorker.main_task,
5155
engine=model_path,
5256
rpc_addr=addr,
53-
executor_config=self.executor_config,
5457
llm_args=self.llm_args,
5558
hf_model_dir=model_path,
5659
)
@@ -200,8 +203,12 @@ async def consume_stats():
200203
class TestRpcWorkerTP2:
201204

202205
def setup_method(self):
203-
self.llm_args, self.executor_config = create_fake_executor_config(
204-
model_path, tp_size=2)
206+
self.llm_args = TorchLlmArgs(
207+
model=model_path,
208+
tensor_parallel_size=2,
209+
backend='pytorch',
210+
enable_iter_perf_stats=True,
211+
)
205212
self.session, self.addr, self.futures = self.create_worker_session()
206213
self.client = self.create_rpc_client(self.addr)
207214
self.client.setup_engine().remote()
@@ -218,7 +225,6 @@ def create_worker_session(self):
218225
futures = session.submit(RpcWorker.main_task,
219226
engine=model_path,
220227
rpc_addr=addr,
221-
executor_config=self.executor_config,
222228
llm_args=self.llm_args,
223229
hf_model_dir=model_path,
224230
model_world_size=2)

0 commit comments

Comments
 (0)