Skip to content
This repository was archived by the owner on Jun 3, 2025. It is now read-only.

Commit fc208cb

Browse files
authored
[Cherry Pick][Transformers] Initialize with grad_sampler when no scaler (#630) (#631)
1 parent 7fe4554 commit fc208cb

File tree

1 file changed

+2
-1
lines changed
  • src/sparseml/transformers/sparsification

1 file changed

+2
-1
lines changed

src/sparseml/transformers/sparsification/trainer.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -231,13 +231,14 @@ def create_optimizer(self):
231231
self.manager,
232232
steps_per_epoch=self.manager_steps_per_epoch,
233233
loggers=self.manager_loggers,
234-
grad_sampler=self.grad_sampler,
234+
initialize_kwargs={"grad_sampler": self.grad_sampler},
235235
)
236236
if not self.manager.initialized:
237237
self.manager.initialize(
238238
self.model,
239239
loggers=self.manager_loggers,
240240
distillation_teacher=self.teacher,
241+
grad_sampler=self.grad_sampler,
241242
)
242243
self.manager_initialized = True
243244
_LOGGER.info(

0 commit comments

Comments
 (0)