File tree Expand file tree Collapse file tree 2 files changed +12
-1
lines changed Expand file tree Collapse file tree 2 files changed +12
-1
lines changed Original file line number Diff line number Diff line change 3535 scale_qk_by_layer_num : True
3636 fused_softmax_with_triangular : True
3737 use_flash_attn : False
38-
38+ sequence_parallel : False # TODO make sequence_parallel as an independent parallel and be set id Distributed
3939
4040Data :
4141 Train :
Original file line number Diff line number Diff line change 1616import os
1717import sys
1818
19+ from sympy import sequence
20+
1921import paddle
2022import paddle .distributed as dist
2123import paddle .distributed .auto_parallel as auto
@@ -42,6 +44,15 @@ def process_dist_configs(config):
4244 mp_degree = configs .setdefault ("mp_degree" , 1 )
4345 pp_degree = configs .setdefault ("pp_degree" , 1 )
4446
47+ # disenable sequence parallel is mp_degree < 2.
48+ sequence_parallel = config ["Model" ]["sequence_parallel" ]
49+ if mp_degree < 2 and sequence_parallel :
50+ config ["Model" ]["sequence_parallel" ] = False
51+ logger .warning (
52+ "sequence_parallel is turn off since mp_degree < 2."
53+ )
54+
55+
4556 # sharding default
4657 sharding_config = configs ["sharding" ]
4758 sharding_degree = sharding_config .setdefault ("sharding_degree" , 1 )
You can’t perform that action at this time.
0 commit comments