|
67 | 67 | from hyperparam_utils import get_recommended_lr, get_lr_with_warmup |
68 | 68 | from logger import StructuredLogger |
69 | 69 | from checkpoint_manager import CheckpointManager, find_latest_run |
| 70 | + from inspect_eval import run_inspect_evaluation, INSPECT_AVAILABLE |
70 | 71 | except ImportError: |
71 | 72 | TrainingConfig = None |
72 | 73 | DataLoader = None |
|
76 | 77 | StructuredLogger = None |
77 | 78 | CheckpointManager = None |
78 | 79 | find_latest_run = None |
| 80 | + run_inspect_evaluation = None |
| 81 | + INSPECT_AVAILABLE = False |
79 | 82 |
|
80 | 83 |
|
81 | 84 | def prepare_training_data( |
@@ -177,6 +180,7 @@ async def run_evaluations( |
177 | 180 | tasks: list, |
178 | 181 | renderer_name: str, |
179 | 182 | threshold: float, |
| 183 | + service_client: Optional[Any] = None, |
180 | 184 | training_client: Optional[Any] = None, |
181 | 185 | evalops_client: Optional[Any] = None, |
182 | 186 | test_suite_id: Optional[str] = None, |
@@ -399,6 +403,7 @@ async def async_main(config_path: str, resume: bool = False) -> None: |
399 | 403 | tasks=tasks, |
400 | 404 | renderer_name=renderer_name, |
401 | 405 | threshold=eval_threshold, |
| 406 | + service_client=service_client, |
402 | 407 | training_client=training_client, |
403 | 408 | evalops_client=evalops_client, |
404 | 409 | test_suite_id=test_suite_id, |
|
0 commit comments