We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent ea2926c commit 7c3ab53Copy full SHA for 7c3ab53
paddlenlp/trainer/training_args.py
@@ -1026,6 +1026,11 @@ def __post_init__(self):
1026
enable_dp_comm_overlap and enable_sharding_comm_overlap
1027
), "dp_comm_overlap and sharding_comm_overlap cannot be enabled at the same time"
1028
1029
+ if enable_sharding_comm_overlap and not self.amp_master_grad:
1030
+ raise ValueError(
1031
+ "If `enable_sharding_comm_overlap` in pipeline_parallel_configs, `amp_master_grad` must be True."
1032
+ )
1033
+
1034
dygraph_pp_configs = {
1035
"delay_scale_loss": True if "enable_delay_scale_loss" in pipeline_parallel_config else False,
1036
"dp_comm_overlap": enable_dp_comm_overlap,
0 commit comments