You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: llm/auto_parallel/llama/run_pretrain_auto.py
+2-10Lines changed: 2 additions & 10 deletions
Original file line number
Diff line number
Diff line change
@@ -255,14 +255,6 @@ class ModelArguments:
255
255
"help": "Pre-training from existing paddlenlp model weights. Default False and model will train from scratch. If set True, the model_name_or_path argument must exist in the paddlenlp models."
256
256
},
257
257
)
258
-
sequence_parallel: bool=field(
259
-
default=False,
260
-
metadata={"help": "whether to use sequence parallel"},
261
-
)
262
-
fuse_sequence_parallel_allreduce: bool=field(
263
-
default=False,
264
-
metadata={"help": "whether to use fuse sequence parallel allreduce"},
Copy file name to clipboardExpand all lines: llm/auto_parallel/qwen/run_pretrain_3D_auto.py
+2-10Lines changed: 2 additions & 10 deletions
Original file line number
Diff line number
Diff line change
@@ -239,14 +239,6 @@ class ModelArguments:
239
239
"help": "Pre-training from existing paddlenlp model weights. Default False and model will train from scratch. If set True, the model_name_or_path argument must exist in the paddlenlp models."
240
240
},
241
241
)
242
-
sequence_parallel: bool=field(
243
-
default=False,
244
-
metadata={"help": "whether to use sequence parallel"},
245
-
)
246
-
fuse_sequence_parallel_allreduce: bool=field(
247
-
default=False,
248
-
metadata={"help": "whether to use fuse sequence parallel allreduce"},
0 commit comments