diff --git a/zeus/optimizer/__init__.py b/zeus/optimizer/__init__.py index 6516c098..15136074 100644 --- a/zeus/optimizer/__init__.py +++ b/zeus/optimizer/__init__.py @@ -14,5 +14,4 @@ """A collection of optimizers for various knobs.""" -from zeus.optimizer.power_limit import GlobalPowerLimitOptimizer -from zeus.optimizer.power_limit import HFGlobalPowerLimitOptimizer +from zeus.optimizer.power_limit import GlobalPowerLimitOptimizer, HFGlobalPowerLimitOptimizer diff --git a/zeus/optimizer/power_limit.py b/zeus/optimizer/power_limit.py index 76e5d0cd..3ce8d604 100644 --- a/zeus/optimizer/power_limit.py +++ b/zeus/optimizer/power_limit.py @@ -511,7 +511,7 @@ def __init__( pl_step: int = 25, profile_path: str | Path | None = None, ) -> None: - r"""[Wrapped for Hugging Face Trainer Callback] Initialize the optimizer. + r"""Initialize the optimizer. GPU indices to profile and optimize for are taken from `monitor.gpu_indices`. @@ -551,7 +551,7 @@ def on_epoch_end( model: PreTrainedModel, **kwargs, ) -> None: - """[Wrapped for Hugging Face Trainer Callback] Mark the end of a training epoch.""" + """Mark the end of a training epoch.""" self.optimizer.on_epoch_end() def on_step_begin( @@ -562,5 +562,5 @@ def on_step_begin( model: PreTrainedModel, **kwargs, ) -> None: - """[Wrapped for Hugging Face Trainer Callback] Mark the beginning of a training step.""" + """Mark the beginning of a training step.""" self.optimizer.on_step_begin()