Pretam commited on
Commit
48a2caf
·
verified ·
1 Parent(s): 42288e3

Upload folder using huggingface_hub

Browse files
Files changed (3) hide show
  1. best.pt +3 -0
  2. config.yaml +123 -0
  3. last.pt +3 -0
best.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e39e97b6c4104b76524104a1e387f1dae669df253b05ed426142165886dbdb19
3
+ size 9102066094
config.yaml ADDED
@@ -0,0 +1,123 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ backbone: FacebookAI/xlm-roberta-large
2
+ base: hierarchical_necti
3
+ data:
4
+ data_path: /home/pretam-pg/DepNeCTI/DepNeCTI-XLMR/Trankit_Data
5
+ granularity: Finegrain
6
+ max_length: 128
7
+ num_workers: 4
8
+ use_context: true
9
+ dataset: necti_finegrain
10
+ dim_model: 1024
11
+ freeze_bert: false
12
+ gpu:
13
+ gpus: 1
14
+ precision: 16
15
+ use_gpu: true
16
+ label_hierarchy:
17
+ avyayibhava_subtypes:
18
+ - 45
19
+ - 46
20
+ - 47
21
+ - 48
22
+ - 49
23
+ - 50
24
+ - 51
25
+ - 52
26
+ - 53
27
+ bahuvrihi_subtypes:
28
+ - 20
29
+ - 21
30
+ - 22
31
+ - 23
32
+ - 24
33
+ - 25
34
+ - 26
35
+ - 27
36
+ - 28
37
+ - 29
38
+ - 30
39
+ - 31
40
+ - 32
41
+ - 33
42
+ - 34
43
+ dvandva_subtypes:
44
+ - 35
45
+ - 36
46
+ - 37
47
+ - 38
48
+ - 39
49
+ - 40
50
+ - 41
51
+ - 42
52
+ - 43
53
+ - 44
54
+ tatpurusa_subtypes:
55
+ - 0
56
+ - 1
57
+ - 2
58
+ - 3
59
+ - 4
60
+ - 5
61
+ - 6
62
+ - 7
63
+ - 8
64
+ - 9
65
+ - 10
66
+ - 11
67
+ - 12
68
+ - 13
69
+ - 14
70
+ - 15
71
+ - 16
72
+ - 17
73
+ - 18
74
+ - 19
75
+ logging:
76
+ experiment_name: local-refinement-v1
77
+ logger: wandb
78
+ project_name: hierarchical-diffusion-necti
79
+ loss_type: l2
80
+ noise_schedule: cosine
81
+ objective: pred_x0
82
+ output:
83
+ eval_every_n_epochs: 1
84
+ log_every_n_steps: 50
85
+ output_dir: saved_models/hierarchical_necti
86
+ save_best: true
87
+ save_last: true
88
+ sampling_steps: 100
89
+ snr_scale: 1.0
90
+ stage1:
91
+ depth: 6
92
+ mlp_ratio: 4.0
93
+ num_classes: 6
94
+ num_heads: 16
95
+ stage2:
96
+ depth: 4
97
+ mlp_ratio: 4.0
98
+ num_classes: 56
99
+ num_heads: 16
100
+ window_size: 7
101
+ time_steps: 1000
102
+ training:
103
+ batch_size: 16
104
+ freeze_stage1_for_stage2: true
105
+ lr_bert: 2e-5
106
+ lr_scheduler_type: cosine
107
+ lr_stage1: 1e-4
108
+ lr_stage2: 1e-4
109
+ max_epochs: 50
110
+ max_grad_norm: 1.0
111
+ max_steps: 100000
112
+ min_delta: 0.0001
113
+ monitor: val_USS
114
+ optimizer_type: AdamW
115
+ patience: 10
116
+ stage1_epochs: 30
117
+ stage1_weight: 1.0
118
+ stage2_epochs: 20
119
+ stage2_weight: 1.0
120
+ strategy: joint
121
+ warmup_ratio: 0.05
122
+ warmup_steps: 500
123
+ weight_decay: 0.01
last.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7ac1e1c8880b1c933c042c2f34a5fb5602cdf7069aef40c9f9468a5d8632d701
3
+ size 9102066094