We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 52338f0 commit 8a6589cCopy full SHA for 8a6589c
paddlenlp/trainer/trainer.py
@@ -973,6 +973,7 @@ def train(
973
self.timers and self.timers("optimizer-step").start()
974
975
if self.args.gradient_accumulation_steps > 1 and self._enable_delay_scale_loss():
976
+ paddle.device.synchronize()
977
for p in model._layers.parameters():
978
with paddle.no_grad():
979
if hasattr(p, "main_grad") and p.main_grad is not None:
0 commit comments