diff --git a/docs/source/en/tutorials/basic_training.md b/docs/source/en/tutorials/basic_training.md index 4f7c29cfe513..62b94de3a3fa 100644 --- a/docs/source/en/tutorials/basic_training.md +++ b/docs/source/en/tutorials/basic_training.md @@ -340,6 +340,7 @@ Now you can wrap all these components together in a training loop with 🤗 Acce ... loss = F.mse_loss(noise_pred, noise) ... accelerator.backward(loss) +... if (step + 1) % config.gradient_accumulation_steps == 0: ... accelerator.clip_grad_norm_(model.parameters(), 1.0) ... optimizer.step() ... lr_scheduler.step()