Skip to content

Commit 9ef355b

Browse files
authored
fix test compatibility as AdamW became subclass of Adam (#20574)
Co-authored-by: Haifeng Jin <[email protected]>
1 parent 7bd9279 commit 9ef355b

File tree

1 file changed

+2
-2
lines changed

1 file changed

+2
-2
lines changed

tests/tests_pytorch/callbacks/test_lr_monitor.py

+2-2
Original file line numberDiff line numberDiff line change
@@ -548,10 +548,10 @@ def finetune_function(self, pl_module, epoch: int, optimizer):
548548
"""Called when the epoch begins."""
549549
if epoch == 1 and isinstance(optimizer, torch.optim.SGD):
550550
self.unfreeze_and_add_param_group(pl_module.backbone[0], optimizer, lr=0.1)
551-
if epoch == 2 and isinstance(optimizer, torch.optim.Adam):
551+
if epoch == 2 and type(optimizer) is torch.optim.Adam:
552552
self.unfreeze_and_add_param_group(pl_module.layer, optimizer, lr=0.1)
553553

554-
if epoch == 3 and isinstance(optimizer, torch.optim.Adam):
554+
if epoch == 3 and type(optimizer) is torch.optim.Adam:
555555
assert len(optimizer.param_groups) == 2
556556
self.unfreeze_and_add_param_group(pl_module.backbone[1], optimizer, lr=0.1)
557557
assert len(optimizer.param_groups) == 3

0 commit comments

Comments
 (0)