File size: 3,341 Bytes
d0258f9
 
 
 
174b104
 
d0258f9
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
174b104
d0258f9
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
174b104
d0258f9
174b104
d0258f9
174b104
d0258f9
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
{
    "dataset_name": "nbtpj/summ_ds_train",
    "dataset_config_name": null,
    "train_split_name": "merge36_cnndmsamsumxsum",
    "text_template": "{text}\nTL;DR: {summary}",
    "label_col": "summary",
    "freeze_role2": false,
    "only_train_role1": false,
    "model_name_or_path": "gpt2",
    "ref_role1_name_or_path": "gpt2",
    "ref_role2_name_or_path": "gpt2",
    "pretrained_role2_name_or_path": "none",
    "config_name": null,
    "vectorizer_path": "/common/home/users/m/mq.nguyen.2023/testcode/SAC_LM/module9_clmv2/vectorizer/wikitext103_tfidf_full.joblib",
    "tokenizer_name": null,
    "use_slow_tokenizer": false,
    "per_device_train_batch_size": 4,
    "per_device_query_rollout_batch_size": 32,
    "per_device_eval_batch_size": 4,
    "vllm_vram_ratio": 0.3,
    "learning_rate": 3e-07,
    "grad_norm": 0.5,
    "weight_decay": 1e-05,
    "max_train_steps": 40000,
    "max_train_rollouts": 100000,
    "gradient_accumulation_steps": 1,
    "lr_scheduler_type": "constant",
    "num_warmup_steps": 200,
    "seed": 0,
    "model_type": null,
    "block_size": 1024,
    "mini_epoch": 1,
    "rollout_game": "baseline3v2",
    "rl_algo": "on_policy",
    "constraint_type": "kl",
    "clamp_update": false,
    "rl_w": 1.0,
    "lm_w": 1.0,
    "n_generate": 4,
    "n_augment": 0,
    "gradient_checkpoint": false,
    "group_relative_norm": true,
    "sample_config": {
        "do_sample": true,
        "min_new_tokens": 20,
        "temperature": 1.0
    },
    "inference_config": {
        "do_sample": true,
        "temperature": 0.0,
        "min_new_tokens": 5,
        "max_new_tokens": 250
    },
    "rollout_config": {
        "accuracy_w": 12.960656597279739,
        "len_pen": 1.0,
        "accuracy_w2": 30.20289640158668,
        "len_pen2": 1.0,
        "threshold": 0.09062263471261954,
        "similarity_fn": "rouge"
    },
    "ent_coef": 0.0001,
    "beta_coef": "0.0",
    "prompt_0": "{text}",
    "prompt_1": "{text}\nTL;DR: ",
    "prompt_2": "Given the text: {role1_output}\nReconstruct the summarized text to the detailed:",
    "prompt_eval": "{text}\nTL;DR:",
    "epsilon": 0.2,
    "a2g_norm": true,
    "vllm_sleep": true,
    "lora": false,
    "need_attn_mask": true,
    "gamma": 0.95,
    "trust_remote_code": true,
    "test_glue": false,
    "test_clm": false,
    "causal_model": true,
    "test_gen": true,
    "log_rollout_txt": true,
    "trunc_eval": 256,
    "trunc_evals": [
        "cnndm___12",
        "samsum___12",
        "xsum___12",
        "gigaword___200",
        "duc___50"
    ],
    "use_deepspeed": false,
    "zero_config": 2,
    "log_interval": "5m",
    "eval_interval": "100",
    "checkpoint_interval": "100",
    "lm_fraction": -1.0,
    "push_to_hub": null,
    "keep_eval_size": false,
    "mixed_precision": "bf16",
    "tune_metrics": [
        "cnndm/rouge1___1.0",
        "cnndm/bertscore_f1___0.25",
        "samsum/rouge1___1.0",
        "samsum/bertscore_f1___0.25",
        "xsum/rouge1___1.0",
        "xsum/bertscore_f1___0.25"
    ],
    "base_path": "/common/home/users/m/mq.nguyen.2023/testcode/SAC_LM/module9_clmv2",
    "script": "/common/home/users/m/mq.nguyen.2023/testcode/SAC_LM/module9_clmv2/execute/utune/debug.py",
    "train_from_raw": true,
    "world_size": 1,
    "cpu_per_worker": 7,
    "gpu_per_worker": 1
}