You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: pytorch_accelerated/trainer.py
+5-6Lines changed: 5 additions & 6 deletions
Original file line number
Diff line number
Diff line change
@@ -428,7 +428,7 @@ def train(
428
428
:param num_epochs: the number of epochs to train for
429
429
:param eval_dataset: the dataset to use during evaluation epochs, if this is not provided, evaluation is skipped.
430
430
:param per_device_batch_size: the batch size to use per device
431
-
:param max_num_train_steps: the maximum number of steps across all processes to train for. If provided, this will override num_epochs
431
+
:param max_num_train_steps: the maximum number of steps across all processes to train for. If both max_num_train_steps and num_epochs are provided, the smaller of the two limits is used.
432
432
:param gradient_accumulation_steps: accumulate gradients to the specified number of steps to simulate a bigger batch size. By default, this is set to ``1``
433
433
:param gradient_clip_value: if specified, the gradients of the model's parameters will be clipped to the range ``[-gradient_clip_value, gradient_clip_value]``
434
434
:param create_scheduler_fn: a function which accepts an optimizer as an argument and returns a learning rate scheduler
0 commit comments