sezginer commited on
Commit
3f5b716
·
verified ·
1 Parent(s): 590faf3

Upload folder using huggingface_hub

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. encoder-decoder/16_16_8/1st_stage.ckpt +3 -0
  2. encoder-decoder/16_16_8/2nd_stage.ckpt +3 -0
  3. encoder-decoder/16_16_8/3rd_stage.ckpt +3 -0
  4. encoder-decoder/8_8_8/1st_stage.ckpt +3 -0
  5. encoder-decoder/8_8_8/2nd_stage.ckpt +3 -0
  6. encoder-decoder/8_8_8/3rd_stage.ckpt +3 -0
  7. report_generation/16_16_8/checkpoint-38000/README.md +202 -0
  8. report_generation/16_16_8/checkpoint-38000/adapter_config.json +37 -0
  9. report_generation/16_16_8/checkpoint-38000/adapter_model.safetensors +3 -0
  10. report_generation/16_16_8/checkpoint-38000/config.json +56 -0
  11. report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt +3 -0
  12. report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_10_mp_rank_00_optim_states.pt +3 -0
  13. report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_11_mp_rank_00_optim_states.pt +3 -0
  14. report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_12_mp_rank_00_optim_states.pt +3 -0
  15. report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_13_mp_rank_00_optim_states.pt +3 -0
  16. report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_14_mp_rank_00_optim_states.pt +3 -0
  17. report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_15_mp_rank_00_optim_states.pt +3 -0
  18. report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_16_mp_rank_00_optim_states.pt +3 -0
  19. report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_17_mp_rank_00_optim_states.pt +3 -0
  20. report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_18_mp_rank_00_optim_states.pt +3 -0
  21. report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_19_mp_rank_00_optim_states.pt +3 -0
  22. report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt +3 -0
  23. report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_20_mp_rank_00_optim_states.pt +3 -0
  24. report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_21_mp_rank_00_optim_states.pt +3 -0
  25. report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_22_mp_rank_00_optim_states.pt +3 -0
  26. report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_23_mp_rank_00_optim_states.pt +3 -0
  27. report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_24_mp_rank_00_optim_states.pt +3 -0
  28. report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_25_mp_rank_00_optim_states.pt +3 -0
  29. report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_26_mp_rank_00_optim_states.pt +3 -0
  30. report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_27_mp_rank_00_optim_states.pt +3 -0
  31. report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_28_mp_rank_00_optim_states.pt +3 -0
  32. report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_29_mp_rank_00_optim_states.pt +3 -0
  33. report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt +3 -0
  34. report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_30_mp_rank_00_optim_states.pt +3 -0
  35. report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_31_mp_rank_00_optim_states.pt +3 -0
  36. report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_32_mp_rank_00_optim_states.pt +3 -0
  37. report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_33_mp_rank_00_optim_states.pt +3 -0
  38. report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_34_mp_rank_00_optim_states.pt +3 -0
  39. report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_35_mp_rank_00_optim_states.pt +3 -0
  40. report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_36_mp_rank_00_optim_states.pt +3 -0
  41. report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_37_mp_rank_00_optim_states.pt +3 -0
  42. report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_38_mp_rank_00_optim_states.pt +3 -0
  43. report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_39_mp_rank_00_optim_states.pt +3 -0
  44. report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_3_mp_rank_00_optim_states.pt +3 -0
  45. report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_4_mp_rank_00_optim_states.pt +3 -0
  46. report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_5_mp_rank_00_optim_states.pt +3 -0
  47. report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_6_mp_rank_00_optim_states.pt +3 -0
  48. report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_7_mp_rank_00_optim_states.pt +3 -0
  49. report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_8_mp_rank_00_optim_states.pt +3 -0
  50. report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_9_mp_rank_00_optim_states.pt +3 -0
encoder-decoder/16_16_8/1st_stage.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d24fcc33ad8e39fd5d8c4485ee57f208f5ebe1074645d6a4d9b89c2b43253415
3
+ size 417168514
encoder-decoder/16_16_8/2nd_stage.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7e25b31a98a0f6c5b2e5859152424651d92b978a88274af6df3a4dd08a5ce937
3
+ size 417168225
encoder-decoder/16_16_8/3rd_stage.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a3d28cf1c8a17aa30df2c6d6819540744afa397e1b6f22f0a3247320699fd9a5
3
+ size 417168225
encoder-decoder/8_8_8/1st_stage.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d6aabfb3c675045db59f552d7d4fdc0e7ecfd0ccab047299fe62fcd5a564d161
3
+ size 375316226
encoder-decoder/8_8_8/2nd_stage.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bfcc0a0cdf0f1bfaa9b01acfea1f6b020926120de1623eddcfd23e2c03fd6cb7
3
+ size 375315993
encoder-decoder/8_8_8/3rd_stage.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6978a2fcc30ebc4b1313329821ab5288b60409b29684d8adb0e070f9a9fbe46c
3
+ size 375315993
report_generation/16_16_8/checkpoint-38000/README.md ADDED
@@ -0,0 +1,202 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: meta-llama/Meta-Llama-3.1-8B-Instruct
3
+ library_name: peft
4
+ ---
5
+
6
+ # Model Card for Model ID
7
+
8
+ <!-- Provide a quick summary of what the model is/does. -->
9
+
10
+
11
+
12
+ ## Model Details
13
+
14
+ ### Model Description
15
+
16
+ <!-- Provide a longer summary of what this model is. -->
17
+
18
+
19
+
20
+ - **Developed by:** [More Information Needed]
21
+ - **Funded by [optional]:** [More Information Needed]
22
+ - **Shared by [optional]:** [More Information Needed]
23
+ - **Model type:** [More Information Needed]
24
+ - **Language(s) (NLP):** [More Information Needed]
25
+ - **License:** [More Information Needed]
26
+ - **Finetuned from model [optional]:** [More Information Needed]
27
+
28
+ ### Model Sources [optional]
29
+
30
+ <!-- Provide the basic links for the model. -->
31
+
32
+ - **Repository:** [More Information Needed]
33
+ - **Paper [optional]:** [More Information Needed]
34
+ - **Demo [optional]:** [More Information Needed]
35
+
36
+ ## Uses
37
+
38
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
39
+
40
+ ### Direct Use
41
+
42
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
43
+
44
+ [More Information Needed]
45
+
46
+ ### Downstream Use [optional]
47
+
48
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
49
+
50
+ [More Information Needed]
51
+
52
+ ### Out-of-Scope Use
53
+
54
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
55
+
56
+ [More Information Needed]
57
+
58
+ ## Bias, Risks, and Limitations
59
+
60
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
61
+
62
+ [More Information Needed]
63
+
64
+ ### Recommendations
65
+
66
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
67
+
68
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
69
+
70
+ ## How to Get Started with the Model
71
+
72
+ Use the code below to get started with the model.
73
+
74
+ [More Information Needed]
75
+
76
+ ## Training Details
77
+
78
+ ### Training Data
79
+
80
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
81
+
82
+ [More Information Needed]
83
+
84
+ ### Training Procedure
85
+
86
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
87
+
88
+ #### Preprocessing [optional]
89
+
90
+ [More Information Needed]
91
+
92
+
93
+ #### Training Hyperparameters
94
+
95
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
96
+
97
+ #### Speeds, Sizes, Times [optional]
98
+
99
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
100
+
101
+ [More Information Needed]
102
+
103
+ ## Evaluation
104
+
105
+ <!-- This section describes the evaluation protocols and provides the results. -->
106
+
107
+ ### Testing Data, Factors & Metrics
108
+
109
+ #### Testing Data
110
+
111
+ <!-- This should link to a Dataset Card if possible. -->
112
+
113
+ [More Information Needed]
114
+
115
+ #### Factors
116
+
117
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
118
+
119
+ [More Information Needed]
120
+
121
+ #### Metrics
122
+
123
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
124
+
125
+ [More Information Needed]
126
+
127
+ ### Results
128
+
129
+ [More Information Needed]
130
+
131
+ #### Summary
132
+
133
+
134
+
135
+ ## Model Examination [optional]
136
+
137
+ <!-- Relevant interpretability work for the model goes here -->
138
+
139
+ [More Information Needed]
140
+
141
+ ## Environmental Impact
142
+
143
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
144
+
145
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
146
+
147
+ - **Hardware Type:** [More Information Needed]
148
+ - **Hours used:** [More Information Needed]
149
+ - **Cloud Provider:** [More Information Needed]
150
+ - **Compute Region:** [More Information Needed]
151
+ - **Carbon Emitted:** [More Information Needed]
152
+
153
+ ## Technical Specifications [optional]
154
+
155
+ ### Model Architecture and Objective
156
+
157
+ [More Information Needed]
158
+
159
+ ### Compute Infrastructure
160
+
161
+ [More Information Needed]
162
+
163
+ #### Hardware
164
+
165
+ [More Information Needed]
166
+
167
+ #### Software
168
+
169
+ [More Information Needed]
170
+
171
+ ## Citation [optional]
172
+
173
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
174
+
175
+ **BibTeX:**
176
+
177
+ [More Information Needed]
178
+
179
+ **APA:**
180
+
181
+ [More Information Needed]
182
+
183
+ ## Glossary [optional]
184
+
185
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
186
+
187
+ [More Information Needed]
188
+
189
+ ## More Information [optional]
190
+
191
+ [More Information Needed]
192
+
193
+ ## Model Card Authors [optional]
194
+
195
+ [More Information Needed]
196
+
197
+ ## Model Card Contact
198
+
199
+ [More Information Needed]
200
+ ### Framework versions
201
+
202
+ - PEFT 0.14.0
report_generation/16_16_8/checkpoint-38000/adapter_config.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "meta-llama/Meta-Llama-3.1-8B-Instruct",
5
+ "bias": "none",
6
+ "eva_config": null,
7
+ "exclude_modules": null,
8
+ "fan_in_fan_out": false,
9
+ "inference_mode": true,
10
+ "init_lora_weights": true,
11
+ "layer_replication": null,
12
+ "layers_pattern": null,
13
+ "layers_to_transform": null,
14
+ "loftq_config": {},
15
+ "lora_alpha": 256,
16
+ "lora_bias": false,
17
+ "lora_dropout": 0.05,
18
+ "megatron_config": null,
19
+ "megatron_core": "megatron.core",
20
+ "modules_to_save": null,
21
+ "peft_type": "LORA",
22
+ "r": 128,
23
+ "rank_pattern": {},
24
+ "revision": null,
25
+ "target_modules": [
26
+ "o_proj",
27
+ "v_proj",
28
+ "down_proj",
29
+ "q_proj",
30
+ "gate_proj",
31
+ "k_proj",
32
+ "up_proj"
33
+ ],
34
+ "task_type": "CAUSAL_LM",
35
+ "use_dora": false,
36
+ "use_rslora": false
37
+ }
report_generation/16_16_8/checkpoint-38000/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4971aab21859c6f6939ff8dc7616101944031729ed001d0569ceaf10403915e7
3
+ size 671150064
report_generation/16_16_8/checkpoint-38000/config.json ADDED
@@ -0,0 +1,56 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "meta-llama/Meta-Llama-3.1-8B-Instruct",
3
+ "architectures": [
4
+ "LlamaForCausalLM"
5
+ ],
6
+ "attention_bias": false,
7
+ "attention_dropout": 0.0,
8
+ "bos_token_id": 128000,
9
+ "eos_token_id": [
10
+ 128001,
11
+ 128008,
12
+ 128009
13
+ ],
14
+ "freeze_mm_mlp_adapter": false,
15
+ "hidden_act": "silu",
16
+ "hidden_size": 4096,
17
+ "image_aspect_ratio": "pad",
18
+ "initializer_range": 0.02,
19
+ "intermediate_size": 14336,
20
+ "max_position_embeddings": 131072,
21
+ "mlp_bias": false,
22
+ "mm_context_size": 18,
23
+ "mm_hidden_size": 18,
24
+ "mm_patch_merge_type": "flat",
25
+ "mm_projector_lr": 2e-05,
26
+ "mm_projector_type": "attn_pool+mlp2x_gelu",
27
+ "mm_use_im_patch_token": false,
28
+ "mm_use_im_start_end": false,
29
+ "mm_vision_select_feature": "patch",
30
+ "mm_vision_select_layer": -2,
31
+ "mm_vision_tower": "openai/clip-vit-large-patch14-336",
32
+ "model_type": "llava_llama",
33
+ "num_attention_heads": 32,
34
+ "num_hidden_layers": 32,
35
+ "num_key_value_heads": 8,
36
+ "pad_token_id": 128256,
37
+ "pretraining_tp": 1,
38
+ "rms_norm_eps": 1e-05,
39
+ "rope_scaling": {
40
+ "factor": 8.0,
41
+ "high_freq_factor": 4.0,
42
+ "low_freq_factor": 1.0,
43
+ "original_max_position_embeddings": 8192,
44
+ "rope_type": "llama3"
45
+ },
46
+ "rope_theta": 500000.0,
47
+ "tie_word_embeddings": false,
48
+ "tokenizer_model_max_length": 128000,
49
+ "tokenizer_padding_side": "right",
50
+ "torch_dtype": "bfloat16",
51
+ "transformers_version": "4.44.0",
52
+ "tune_mm_mlp_adapter": false,
53
+ "use_cache": false,
54
+ "use_mm_proj": true,
55
+ "vocab_size": 128261
56
+ }
report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b9bd7d7c6ea5939377008b543778e5ad76bde1f9ee5a2044377b2425cbf561e4
3
+ size 263338482
report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_10_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6dc170a1e68c8b681a98fbb4bd2f53b3a3ad9836ecf7eb392dd3bcd207affe6c
3
+ size 263338562
report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_11_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:33d52f110e057bd6b5380581d01852242b683020abeb4ff9cfea31d7afa8cbcf
3
+ size 263338562
report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_12_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:089e976e5d19310250a22fc15b15c1eb35b1df7130544cffdc983ed7b76b8566
3
+ size 263338562
report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_13_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a69ab33ac46038c9aa83e3b4a2ff79fb87d0bc142158e1bc06e8fc864bb32418
3
+ size 263338562
report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_14_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bbab51f7135e51044e9dde5a4ab297c2d425a2cebef52b112048a23750e9337c
3
+ size 263338562
report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_15_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eb01064e0df46743ac406cd44a88303d5323ecd00904baf2a1f10567ddb7149c
3
+ size 263338562
report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_16_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:81141a68b25a5ea814c0d6e7b63688cecfcfe3fce5f8606eead85376a37ac3f9
3
+ size 263338562
report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_17_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fe05a6509d1b6659de693933e20a3e4d523af8048c8ab96802c5519b58ee382f
3
+ size 263338562
report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_18_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4f160ed2731475e45b91bf0122b9f6c3150d0c63ea08f426b1b08fd9336f9c1a
3
+ size 263338562
report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_19_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:502b1896fb83180f0231d3248694f520aaa0cff927ef6743d3c4adcc87dadafd
3
+ size 263338562
report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:629e892e1d63353322019d3a5e9a5523f14d9ba959815fd22ac9c34fb8d02e8c
3
+ size 263338482
report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_20_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0e500fb964304917b11176f99727bbb6b0422efb270a01c9b0afe7c80a72bbe8
3
+ size 263338562
report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_21_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e44b525d7484c2b0ebfdaacb0d935b1e9adfab8cdd2d7c5dd17e7cda0ecf2884
3
+ size 263338562
report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_22_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:633d4916220abb89711bcede2bf2b746cf49ea233901671ea6acfcee7b852a3d
3
+ size 263338562
report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_23_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:93d05dc1b28b1c209e1ca2c5133217d757464f48cf8a204cd88bee63588de7b8
3
+ size 263338562
report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_24_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:50c398f34656c2d3873738f8f3aab9ae5c337dc2ad96c00cab4f2f68382024b8
3
+ size 263338562
report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_25_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6d9f050ca8adcff35ef557e175f65222d9fc70ea7bc335453dc8c9fb6636fe6b
3
+ size 263338562
report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_26_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:98f146bebcdd17f49f2c0963800debd5344955e2d58e661fdc0ac95727e18760
3
+ size 263338562
report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_27_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:500602c3cdabe49825c612dafd50a4bb1c32f461b88f899b5ff8baac72654462
3
+ size 263338562
report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_28_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:07403d8fa847ca06dc5d7777d6645f08e7c64c0659ba7baa627a9d1148a7988f
3
+ size 263338562
report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_29_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:888b09039e41ea3ed8c7010bbd69b7e1b16a15df53d73091ca7c70c069dd0d21
3
+ size 263338562
report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0212eaf0e2e597f8c812fee4a09e7aacda727964f960ecffa21fef8eb73f33c2
3
+ size 263338482
report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_30_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3355d35ad33360c45f60ea12bda00fa1632639517e3a266723dd5af92acd7906
3
+ size 263338562
report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_31_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a35c96ec5260c6de131bb9482cf807461231fbd924760854bc60cd72b1397c11
3
+ size 263338562
report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_32_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c7d3185d670e81d8d983e99ee343d9652aef043fdb020a39450ea9f252318194
3
+ size 263338562
report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_33_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:00cab5986377357d72d552215251100325c70084cd821a55f64161bb348a4807
3
+ size 263338562
report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_34_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5fef6932f1e2216bcd8b8294e9eeeee06db04006af5c69a9ffc8ee3c721a72d2
3
+ size 263338562
report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_35_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e2de5bfa16170bbf66ad8cb89964da9c40c12b13eb4e1898500f638cde4e95d3
3
+ size 263338562
report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_36_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3fe3d988068afb77a8bd804a4a57c1032c48994549952954b7dbf40deac99430
3
+ size 263338562
report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_37_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:57aa6014a419a2adad640e9e0bba88ba5597cc3922fb367daa1fdcc4a1cd9930
3
+ size 263338562
report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_38_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3266602c318db45aee52a5712a1d2069f449d5a79eae2fb9cf323bc86cfc50a5
3
+ size 263338562
report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_39_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:209872c12b98bfd8ddca026f085299b7a022f53869b82223122af5ea119eba9a
3
+ size 263338562
report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_3_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9980952a3be56f2cb5ef1dc941cb1ad47bc015ae60fd062b4c56375a63445243
3
+ size 263338482
report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_4_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:86ea03f758040a94595e89fddc3536750b66c2d7383ca5e5ed3d692b152c9843
3
+ size 263338482
report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_5_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dc96cf92ff2cf10a75ccb32c5732cc2e6d043eb44f4cc08597080a1da79f2bc3
3
+ size 263338482
report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_6_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:439161bfdf5220318f2197fa7891888f97e48f6414c4dd4697062f1f9885c99a
3
+ size 263338482
report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_7_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e77d5b661ce48eae80938448dd8ae1a250165871b3c37e20a360ab7b04c44f8
3
+ size 263338482
report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_8_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9d8043aa268b73910d56fd011fcd0e53a0e38941f9ae9c0e86a680e58772d883
3
+ size 263338482
report_generation/16_16_8/checkpoint-38000/global_step38000/bf16_zero_pp_rank_9_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:be22751c986057a4e04a82d87385e2af6ffb97b17f1903e7fd69a838efb68f61
3
+ size 263338482