LemTenku commited on
Commit
b20ccdc
·
verified ·
1 Parent(s): 3e1501c

Upload fft-layer-config (2) (1) (1).yml

Browse files
Files changed (1) hide show
  1. fft-layer-config (2) (1) (1).yml +103 -0
fft-layer-config (2) (1) (1).yml ADDED
@@ -0,0 +1,103 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ base_model: unsloth/Meta-Llama-3.1-8B-Instruct
2
+ load_in_8bit: false
3
+ load_in_4bit: false
4
+ strict: false
5
+
6
+ # Layer unfreezing configuration
7
+ unfreezing:
8
+ enabled: true
9
+ mode: "selective"
10
+ layer_selection:
11
+ patterns:
12
+ - "model.layers.1"
13
+ - "model.layers.2"
14
+ - "model.layers.3"
15
+ - "model.layers.4"
16
+ - "model.layers.5"
17
+ - "model.layers.6"
18
+ - "model.layers.7"
19
+ - "model.layers.8"
20
+ - "model.layers.9"
21
+ - "model.layers.10"
22
+ - "model.layers.11"
23
+ component_types:
24
+ - "input_layernorm"
25
+ - "self_attn.q_proj"
26
+ - "self_attn.k_proj"
27
+ - "self_attn.v_proj"
28
+ - "self_attn.o_proj"
29
+ - "post_attention_layernorm"
30
+ - "mlp.gate_proj"
31
+ - "mlp.up_proj"
32
+ - "mlp.down_proj"
33
+
34
+ # Pretraining dataset configuration
35
+ pretraining_dataset:
36
+ - path: LemTenku/adasdxacac
37
+ text_column: text
38
+ type: pretrain
39
+ trust_remote_code: true
40
+
41
+ # Required for streaming datasets
42
+ max_steps: 14932
43
+ batching_strategy: stream
44
+
45
+ datasets:
46
+ - path: LemTenku/adasdxacac
47
+ type: pretrain
48
+ text_column: text
49
+
50
+ dataset_prepared_path: last_run_prepared
51
+ val_set_size: 0.05
52
+ output_dir: ./outputs/out
53
+
54
+ sequence_len: 8192
55
+ sample_packing: true
56
+ pad_to_sequence_len: true
57
+
58
+ wandb_project:
59
+ wandb_entity:
60
+ wandb_watch:
61
+ wandb_name:
62
+ wandb_log_model:
63
+
64
+ gradient_accumulation_steps: 8
65
+ micro_batch_size: 1
66
+ num_epochs: 3
67
+ optimizer: paged_adamw_8bit
68
+ lr_scheduler: constant
69
+ learning_rate: 4e-5
70
+
71
+ train_on_inputs: false
72
+ group_by_length: false
73
+ bf16: auto
74
+ fp16:
75
+ tf32: false
76
+
77
+ gradient_checkpointing: true
78
+ gradient_checkpointing_kwargs:
79
+ use_reentrant: false
80
+ early_stopping_patience:
81
+ resume_from_checkpoint:
82
+ logging_steps: 1
83
+ xformers_attention:
84
+ flash_attention: true
85
+
86
+ warmup_steps: 0
87
+ evaluation_strategy: "no"
88
+ saves_per_epoch: 1
89
+ debug:
90
+ deepspeed:
91
+ weight_decay: 0.0
92
+ fsdp:
93
+ fsdp_config:
94
+
95
+ special_tokens:
96
+ bos_token: "<|begin_of_text|>"
97
+ eos_token: "<|eot_id|>"
98
+ pad_token: "<|end_of_text|>"
99
+
100
+ tokens:
101
+ - "<|begin_of_text|>"
102
+ - "<|eot_id|>"
103
+ - "<|end_of_text|>"