From 35d9ca43d45c60ad8b59f0dccab5c4e12baeb0e3 Mon Sep 17 00:00:00 2001 From: RandomGamingDev Date: Wed, 24 Jul 2024 22:11:53 -0400 Subject: [PATCH] added line allowing gradient accumulation to work for basic_training example --- docs/source/en/tutorials/basic_training.md | 1 + 1 file changed, 1 insertion(+) diff --git a/docs/source/en/tutorials/basic_training.md b/docs/source/en/tutorials/basic_training.md index 4f7c29cfe513..62b94de3a3fa 100644 --- a/docs/source/en/tutorials/basic_training.md +++ b/docs/source/en/tutorials/basic_training.md @@ -340,6 +340,7 @@ Now you can wrap all these components together in a training loop with 🤗 Acce ... loss = F.mse_loss(noise_pred, noise) ... accelerator.backward(loss) +... if (step + 1) % config.gradient_accumulation_steps == 0: ... accelerator.clip_grad_norm_(model.parameters(), 1.0) ... optimizer.step() ... lr_scheduler.step()