Skip to content

Commit 41d51a3

Browse files
author
Lin Wang
committed
lin's update
1 parent d327b74 commit 41d51a3

16 files changed

+301389
-0
lines changed

LIN/dpo_negative_output/dpo_log_history.json

Lines changed: 761 additions & 0 deletions
Large diffs are not rendered by default.
Lines changed: 38 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,38 @@
1+
{
2+
"activation_function": "gelu_new",
3+
"architectures": [
4+
"GPT2LMHeadModel"
5+
],
6+
"attn_pdrop": 0.1,
7+
"bos_token_id": 50256,
8+
"embd_pdrop": 0.1,
9+
"eos_token_id": 50256,
10+
"initializer_range": 0.02,
11+
"layer_norm_epsilon": 1e-05,
12+
"model_type": "gpt2",
13+
"n_ctx": 1024,
14+
"n_embd": 768,
15+
"n_head": 12,
16+
"n_inner": null,
17+
"n_layer": 12,
18+
"n_positions": 1024,
19+
"reorder_and_upcast_attn": false,
20+
"resid_pdrop": 0.1,
21+
"scale_attn_by_inverse_layer_idx": false,
22+
"scale_attn_weights": true,
23+
"summary_activation": null,
24+
"summary_first_dropout": 0.1,
25+
"summary_proj_to_labels": true,
26+
"summary_type": "cls_index",
27+
"summary_use_proj": true,
28+
"task_specific_params": {
29+
"text-generation": {
30+
"do_sample": true,
31+
"max_length": 50
32+
}
33+
},
34+
"torch_dtype": "float32",
35+
"transformers_version": "4.51.3",
36+
"use_cache": true,
37+
"vocab_size": 50257
38+
}
Lines changed: 6 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,6 @@
1+
{
2+
"_from_model_config": true,
3+
"bos_token_id": 50256,
4+
"eos_token_id": 50256,
5+
"transformers_version": "4.51.3"
6+
}

0 commit comments

Comments
 (0)