Skip to content

Commit 85ba573

Browse files
[AutoParallel] Support qwen for auto_parallel (#8312)
* run 4096 * support 3d semi auto run of qwen model * [AutoParallel] Add qwen auto parallel. * Delete useless files. * Fix code style problem. * Fix problems. * Polish code. * Add CI-cases. * Add bf16 ci testcase. * Fix some problems. --------- Co-authored-by: pangengzheng <[email protected]> Co-authored-by: pangengzheng <pangengzheng.baidu.com>
1 parent d4edd19 commit 85ba573

File tree

6 files changed

+2151
-0
lines changed

6 files changed

+2151
-0
lines changed
Lines changed: 45 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,45 @@
1+
{
2+
"model_name_or_path": "qwen/qwen-14b",
3+
"tokenizer_name_or_path": "qwen/qwen-14b",
4+
"input_dir": "./data",
5+
"output_dir": "./checkpoints/qwen_pretrain_ckpts",
6+
"per_device_train_batch_size": 1,
7+
"gradient_accumulation_steps": 2,
8+
"per_device_eval_batch_size": 16,
9+
"data_parallel_degree": 2,
10+
"tensor_parallel_degree": 2,
11+
"pipeline_parallel_degree": 2,
12+
"virtual_pp_degree": 1,
13+
"sequence_parallel": 0,
14+
"use_flash_attention": false,
15+
"use_fused_rms_norm": false,
16+
"use_fused_rope": false,
17+
"max_seq_length": 4096,
18+
"learning_rate": 3e-05,
19+
"min_learning_rate": 3e-06,
20+
"scale_loss": 1024,
21+
"warmup_steps": 30,
22+
"logging_steps": 1,
23+
"max_steps": 10000,
24+
"save_steps": 1000,
25+
"eval_steps": 10000,
26+
"weight_decay": 0.01,
27+
"bf16": true,
28+
"fp16_opt_level": "O2",
29+
"warmup_ratio": 0.01,
30+
"max_grad_norm": 1.0,
31+
"dataloader_num_workers": 1,
32+
"continue_training": 0,
33+
"do_train": true,
34+
"do_eval": true,
35+
"do_predict": true,
36+
"disable_tqdm": true,
37+
"recompute": true,
38+
"recompute_granularity": "core_attn",
39+
"recompute_use_reentrant": true,
40+
"distributed_dataloader": 0,
41+
"save_total_limit": 2,
42+
"parallel_mode": "auto",
43+
"enable_auto_parallel": 1,
44+
"to_static": 0
45+
}

0 commit comments

Comments
 (0)