diff --git a/environments/community/pytorch_optimizer_coding/FOB/pytorch_fob/engine/run.py b/environments/community/pytorch_optimizer_coding/FOB/pytorch_fob/engine/run.py index af2915a9..7535e314 100644 --- a/environments/community/pytorch_optimizer_coding/FOB/pytorch_fob/engine/run.py +++ b/environments/community/pytorch_optimizer_coding/FOB/pytorch_fob/engine/run.py @@ -220,7 +220,7 @@ class Run: gradient_clip_algorithm=self.engine.gradient_clip_alg, precision=precision_with_fallback(self.engine.precision), # type: ignore accelerator=self.engine.accelerator, - log_every_n_steps=self.engine.logging_inteval, + log_every_n_steps=self.engine.logging_interval, ) def get_tester(self) -> Trainer: @@ -333,7 +333,7 @@ class Run: ) if self.engine.log_extra: self._callbacks["extra"] = LogTrainingStats( - log_every_n_steps=self.engine.logging_inteval, + log_every_n_steps=self.engine.logging_interval, **( self.engine.log_extra if isinstance(self.engine.log_extra, dict)