TimeRobber commited on
Commit
340102f
·
1 Parent(s): d24e4aa

Upload part 2

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +49 -0
  2. checkpoint_1007000/state.param_states.decoder.layers_17.pre_mlp_layer_norm.scale.v/.zarray +1 -0
  3. checkpoint_1007000/state.param_states.decoder.layers_17.pre_mlp_layer_norm.scale.v/0 +0 -0
  4. checkpoint_1007000/target.decoder.layers_12.pre_cross_attention_layer_norm.scale/.zarray +1 -0
  5. checkpoint_1007000/target.decoder.layers_12.pre_cross_attention_layer_norm.scale/0 +0 -0
  6. checkpoint_1007000/target.decoder.layers_15.mlp.wi_0.kernel/.zarray +1 -0
  7. checkpoint_1007000/target.decoder.layers_15.mlp.wi_0.kernel/.zarray.__tmp1010806816001982373.~1663960300885807~ +1 -0
  8. checkpoint_1007000/target.decoder.layers_15.mlp.wi_0.kernel/.zarray.__tmp10571021281251091362.~1663960300940635~ +1 -0
  9. checkpoint_1007000/target.decoder.layers_15.mlp.wi_0.kernel/.zarray.__tmp10679081918811478228.~1663960301110307~ +1 -0
  10. checkpoint_1007000/target.decoder.layers_15.mlp.wi_0.kernel/.zarray.__tmp12954582408526729884.~1663960300383989~ +1 -0
  11. checkpoint_1007000/target.decoder.layers_15.mlp.wi_0.kernel/.zarray.__tmp13941922830588999623.~1663960300685772~ +1 -0
  12. checkpoint_1007000/target.decoder.layers_15.mlp.wi_0.kernel/.zarray.__tmp17009882858474760064.~1663960300502914~ +1 -0
  13. checkpoint_1007000/target.decoder.layers_15.mlp.wi_0.kernel/.zarray.__tmp5599219680940660583.~1663960300501108~ +1 -0
  14. checkpoint_1007000/target.decoder.layers_15.mlp.wi_0.kernel/0.0 +3 -0
  15. checkpoint_1007000/target.decoder.layers_15.mlp.wi_0.kernel/0.1 +3 -0
  16. checkpoint_1007000/target.decoder.layers_15.mlp.wi_0.kernel/0.2 +3 -0
  17. checkpoint_1007000/target.decoder.layers_15.mlp.wi_0.kernel/0.3 +3 -0
  18. checkpoint_1007000/target.decoder.layers_15.mlp.wi_0.kernel/0.4 +3 -0
  19. checkpoint_1007000/target.decoder.layers_15.mlp.wi_0.kernel/0.5 +3 -0
  20. checkpoint_1007000/target.decoder.layers_15.mlp.wi_0.kernel/0.6 +3 -0
  21. checkpoint_1007000/target.decoder.layers_15.mlp.wi_0.kernel/0.7 +3 -0
  22. checkpoint_1007000/target.decoder.layers_16.encoder_decoder_attention.key.kernel/.zarray +1 -0
  23. checkpoint_1007000/target.decoder.layers_16.encoder_decoder_attention.key.kernel/.zarray.__tmp1277066131995913345.~1663960300272122~ +1 -0
  24. checkpoint_1007000/target.decoder.layers_16.encoder_decoder_attention.key.kernel/0.0 +3 -0
  25. checkpoint_1007000/target.decoder.layers_16.encoder_decoder_attention.key.kernel/0.1 +3 -0
  26. checkpoint_1007000/target.decoder.layers_16.encoder_decoder_attention.key.kernel/0.2 +3 -0
  27. checkpoint_1007000/target.decoder.layers_16.encoder_decoder_attention.key.kernel/0.3 +3 -0
  28. checkpoint_1007000/target.decoder.layers_16.encoder_decoder_attention.key.kernel/0.4 +3 -0
  29. checkpoint_1007000/target.decoder.layers_16.encoder_decoder_attention.key.kernel/0.5 +3 -0
  30. checkpoint_1007000/target.decoder.layers_16.encoder_decoder_attention.key.kernel/0.6 +3 -0
  31. checkpoint_1007000/target.decoder.layers_16.encoder_decoder_attention.key.kernel/0.7 +3 -0
  32. checkpoint_1007000/target.decoder.layers_18.self_attention.key.kernel/.zarray +1 -0
  33. checkpoint_1007000/target.decoder.layers_18.self_attention.key.kernel/.zarray.__tmp12475133037183008814.~1663960300420670~ +1 -0
  34. checkpoint_1007000/target.decoder.layers_18.self_attention.key.kernel/.zarray.__tmp17331114784864401939.~1663960300465552~ +1 -0
  35. checkpoint_1007000/target.decoder.layers_18.self_attention.key.kernel/.zarray.__tmp4384522759500365203.~1663960300688851~ +1 -0
  36. checkpoint_1007000/target.decoder.layers_18.self_attention.key.kernel/.zarray.__tmp4785258170125338790.~1663960300499566~ +1 -0
  37. checkpoint_1007000/target.decoder.layers_18.self_attention.key.kernel/.zarray.__tmp4930767467751954311.~1663960300872912~ +1 -0
  38. checkpoint_1007000/target.decoder.layers_18.self_attention.key.kernel/.zarray.__tmp532225422100881862.~1663960300495791~ +1 -0
  39. checkpoint_1007000/target.decoder.layers_18.self_attention.key.kernel/.zarray.__tmp6571638690794785116.~1663960300628674~ +1 -0
  40. checkpoint_1007000/target.decoder.layers_18.self_attention.key.kernel/0.0 +3 -0
  41. checkpoint_1007000/target.decoder.layers_18.self_attention.key.kernel/0.1 +3 -0
  42. checkpoint_1007000/target.decoder.layers_18.self_attention.key.kernel/0.2 +3 -0
  43. checkpoint_1007000/target.decoder.layers_18.self_attention.key.kernel/0.3 +3 -0
  44. checkpoint_1007000/target.decoder.layers_18.self_attention.key.kernel/0.4 +3 -0
  45. checkpoint_1007000/target.decoder.layers_18.self_attention.key.kernel/0.5 +3 -0
  46. checkpoint_1007000/target.decoder.layers_18.self_attention.key.kernel/0.6 +3 -0
  47. checkpoint_1007000/target.decoder.layers_18.self_attention.key.kernel/0.7 +3 -0
  48. checkpoint_1007000/target.decoder.layers_6.pre_cross_attention_layer_norm.scale/.zarray +1 -0
  49. checkpoint_1007000/target.decoder.layers_6.pre_cross_attention_layer_norm.scale/0 +0 -0
  50. checkpoint_1007000/target.encoder.layers_1.attention.out.kernel/.zarray +1 -0
.gitattributes CHANGED
@@ -122,3 +122,52 @@ checkpoint_1007000/target.encoder.layers_4.attention.key.kernel/0.6 filter=lfs d
122
  checkpoint_1007000/target.encoder.layers_4.attention.key.kernel/0.5 filter=lfs diff=lfs merge=lfs -text
123
  checkpoint_1007000/target.encoder.layers_4.attention.key.kernel/0.3 filter=lfs diff=lfs merge=lfs -text
124
  checkpoint_1007000/target.encoder.layers_4.attention.key.kernel/0.2 filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
122
  checkpoint_1007000/target.encoder.layers_4.attention.key.kernel/0.5 filter=lfs diff=lfs merge=lfs -text
123
  checkpoint_1007000/target.encoder.layers_4.attention.key.kernel/0.3 filter=lfs diff=lfs merge=lfs -text
124
  checkpoint_1007000/target.encoder.layers_4.attention.key.kernel/0.2 filter=lfs diff=lfs merge=lfs -text
125
+ checkpoint_1007000/target.encoder.layers_4.attention.key.kernel/0.4 filter=lfs diff=lfs merge=lfs -text
126
+ checkpoint_1007000/target.encoder.layers_12.mlp.wo.kernel/6.0 filter=lfs diff=lfs merge=lfs -text
127
+ checkpoint_1007000/target.encoder.layers_12.mlp.wo.kernel/1.0 filter=lfs diff=lfs merge=lfs -text
128
+ checkpoint_1007000/target.encoder.layers_12.mlp.wo.kernel/3.0 filter=lfs diff=lfs merge=lfs -text
129
+ checkpoint_1007000/target.encoder.layers_12.mlp.wo.kernel/0.0 filter=lfs diff=lfs merge=lfs -text
130
+ checkpoint_1007000/target.encoder.layers_12.mlp.wo.kernel/7.0 filter=lfs diff=lfs merge=lfs -text
131
+ checkpoint_1007000/target.encoder.layers_12.mlp.wo.kernel/5.0 filter=lfs diff=lfs merge=lfs -text
132
+ checkpoint_1007000/target.encoder.layers_12.mlp.wo.kernel/2.0 filter=lfs diff=lfs merge=lfs -text
133
+ checkpoint_1007000/target.encoder.layers_12.mlp.wo.kernel/4.0 filter=lfs diff=lfs merge=lfs -text
134
+ checkpoint_1007000/target.encoder.layers_1.attention.out.kernel/6.0 filter=lfs diff=lfs merge=lfs -text
135
+ checkpoint_1007000/target.encoder.layers_1.attention.out.kernel/1.0 filter=lfs diff=lfs merge=lfs -text
136
+ checkpoint_1007000/target.encoder.layers_1.attention.out.kernel/3.0 filter=lfs diff=lfs merge=lfs -text
137
+ checkpoint_1007000/target.encoder.layers_1.attention.out.kernel/0.0 filter=lfs diff=lfs merge=lfs -text
138
+ checkpoint_1007000/target.encoder.layers_1.attention.out.kernel/7.0 filter=lfs diff=lfs merge=lfs -text
139
+ checkpoint_1007000/target.encoder.layers_1.attention.out.kernel/5.0 filter=lfs diff=lfs merge=lfs -text
140
+ checkpoint_1007000/target.encoder.layers_1.attention.out.kernel/2.0 filter=lfs diff=lfs merge=lfs -text
141
+ checkpoint_1007000/target.encoder.layers_1.attention.out.kernel/4.0 filter=lfs diff=lfs merge=lfs -text
142
+ checkpoint_1007000/target.decoder.layers_18.self_attention.key.kernel/0.1 filter=lfs diff=lfs merge=lfs -text
143
+ checkpoint_1007000/target.decoder.layers_18.self_attention.key.kernel/0.7 filter=lfs diff=lfs merge=lfs -text
144
+ checkpoint_1007000/target.decoder.layers_18.self_attention.key.kernel/0.0 filter=lfs diff=lfs merge=lfs -text
145
+ checkpoint_1007000/target.decoder.layers_18.self_attention.key.kernel/0.6 filter=lfs diff=lfs merge=lfs -text
146
+ checkpoint_1007000/target.decoder.layers_18.self_attention.key.kernel/0.5 filter=lfs diff=lfs merge=lfs -text
147
+ checkpoint_1007000/target.decoder.layers_18.self_attention.key.kernel/0.3 filter=lfs diff=lfs merge=lfs -text
148
+ checkpoint_1007000/target.decoder.layers_18.self_attention.key.kernel/0.2 filter=lfs diff=lfs merge=lfs -text
149
+ checkpoint_1007000/target.decoder.layers_18.self_attention.key.kernel/0.4 filter=lfs diff=lfs merge=lfs -text
150
+ checkpoint_1007000/target.decoder.layers_16.encoder_decoder_attention.key.kernel/0.1 filter=lfs diff=lfs merge=lfs -text
151
+ checkpoint_1007000/target.decoder.layers_16.encoder_decoder_attention.key.kernel/0.7 filter=lfs diff=lfs merge=lfs -text
152
+ checkpoint_1007000/target.decoder.layers_16.encoder_decoder_attention.key.kernel/0.0 filter=lfs diff=lfs merge=lfs -text
153
+ checkpoint_1007000/target.decoder.layers_16.encoder_decoder_attention.key.kernel/0.6 filter=lfs diff=lfs merge=lfs -text
154
+ checkpoint_1007000/target.decoder.layers_16.encoder_decoder_attention.key.kernel/0.5 filter=lfs diff=lfs merge=lfs -text
155
+ checkpoint_1007000/target.decoder.layers_16.encoder_decoder_attention.key.kernel/0.3 filter=lfs diff=lfs merge=lfs -text
156
+ checkpoint_1007000/target.decoder.layers_16.encoder_decoder_attention.key.kernel/0.2 filter=lfs diff=lfs merge=lfs -text
157
+ checkpoint_1007000/target.decoder.layers_16.encoder_decoder_attention.key.kernel/0.4 filter=lfs diff=lfs merge=lfs -text
158
+ checkpoint_1007000/target.encoder.layers_12.mlp.wi_1.kernel/0.1 filter=lfs diff=lfs merge=lfs -text
159
+ checkpoint_1007000/target.encoder.layers_12.mlp.wi_1.kernel/0.7 filter=lfs diff=lfs merge=lfs -text
160
+ checkpoint_1007000/target.encoder.layers_12.mlp.wi_1.kernel/0.0 filter=lfs diff=lfs merge=lfs -text
161
+ checkpoint_1007000/target.encoder.layers_12.mlp.wi_1.kernel/0.6 filter=lfs diff=lfs merge=lfs -text
162
+ checkpoint_1007000/target.encoder.layers_12.mlp.wi_1.kernel/0.5 filter=lfs diff=lfs merge=lfs -text
163
+ checkpoint_1007000/target.encoder.layers_12.mlp.wi_1.kernel/0.3 filter=lfs diff=lfs merge=lfs -text
164
+ checkpoint_1007000/target.encoder.layers_12.mlp.wi_1.kernel/0.2 filter=lfs diff=lfs merge=lfs -text
165
+ checkpoint_1007000/target.encoder.layers_12.mlp.wi_1.kernel/0.4 filter=lfs diff=lfs merge=lfs -text
166
+ checkpoint_1007000/target.decoder.layers_15.mlp.wi_0.kernel/0.1 filter=lfs diff=lfs merge=lfs -text
167
+ checkpoint_1007000/target.decoder.layers_15.mlp.wi_0.kernel/0.7 filter=lfs diff=lfs merge=lfs -text
168
+ checkpoint_1007000/target.decoder.layers_15.mlp.wi_0.kernel/0.0 filter=lfs diff=lfs merge=lfs -text
169
+ checkpoint_1007000/target.decoder.layers_15.mlp.wi_0.kernel/0.6 filter=lfs diff=lfs merge=lfs -text
170
+ checkpoint_1007000/target.decoder.layers_15.mlp.wi_0.kernel/0.5 filter=lfs diff=lfs merge=lfs -text
171
+ checkpoint_1007000/target.decoder.layers_15.mlp.wi_0.kernel/0.3 filter=lfs diff=lfs merge=lfs -text
172
+ checkpoint_1007000/target.decoder.layers_15.mlp.wi_0.kernel/0.2 filter=lfs diff=lfs merge=lfs -text
173
+ checkpoint_1007000/target.decoder.layers_15.mlp.wi_0.kernel/0.4 filter=lfs diff=lfs merge=lfs -text
checkpoint_1007000/state.param_states.decoder.layers_17.pre_mlp_layer_norm.scale.v/.zarray ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096],"zarr_format":2}
checkpoint_1007000/state.param_states.decoder.layers_17.pre_mlp_layer_norm.scale.v/0 ADDED
Binary file (14.7 kB). View file
 
checkpoint_1007000/target.decoder.layers_12.pre_cross_attention_layer_norm.scale/.zarray ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096],"zarr_format":2}
checkpoint_1007000/target.decoder.layers_12.pre_cross_attention_layer_norm.scale/0 ADDED
Binary file (14.8 kB). View file
 
checkpoint_1007000/target.decoder.layers_15.mlp.wi_0.kernel/.zarray ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096,1280],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,10240],"zarr_format":2}
checkpoint_1007000/target.decoder.layers_15.mlp.wi_0.kernel/.zarray.__tmp1010806816001982373.~1663960300885807~ ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096,1280],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,10240],"zarr_format":2}
checkpoint_1007000/target.decoder.layers_15.mlp.wi_0.kernel/.zarray.__tmp10571021281251091362.~1663960300940635~ ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096,1280],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,10240],"zarr_format":2}
checkpoint_1007000/target.decoder.layers_15.mlp.wi_0.kernel/.zarray.__tmp10679081918811478228.~1663960301110307~ ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096,1280],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,10240],"zarr_format":2}
checkpoint_1007000/target.decoder.layers_15.mlp.wi_0.kernel/.zarray.__tmp12954582408526729884.~1663960300383989~ ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096,1280],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,10240],"zarr_format":2}
checkpoint_1007000/target.decoder.layers_15.mlp.wi_0.kernel/.zarray.__tmp13941922830588999623.~1663960300685772~ ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096,1280],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,10240],"zarr_format":2}
checkpoint_1007000/target.decoder.layers_15.mlp.wi_0.kernel/.zarray.__tmp17009882858474760064.~1663960300502914~ ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096,1280],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,10240],"zarr_format":2}
checkpoint_1007000/target.decoder.layers_15.mlp.wi_0.kernel/.zarray.__tmp5599219680940660583.~1663960300501108~ ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096,1280],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,10240],"zarr_format":2}
checkpoint_1007000/target.decoder.layers_15.mlp.wi_0.kernel/0.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:990b236e489cc6d911faa0f90bc752396c2cae9748057dc31dac9126cb563e8e
3
+ size 19534603
checkpoint_1007000/target.decoder.layers_15.mlp.wi_0.kernel/0.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:05cb06f49efbde11a75eba505cc6a7c84a72a845bbab3d57e9cb0e97a6428e63
3
+ size 19534140
checkpoint_1007000/target.decoder.layers_15.mlp.wi_0.kernel/0.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d48ba3568b6c8ff8a442c2fda27667707f3d263b4c0ca82d20f5a27b6e06b238
3
+ size 19535270
checkpoint_1007000/target.decoder.layers_15.mlp.wi_0.kernel/0.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d95d8059c942282cdc96730062a5ade9c67eab1bb01e7f99a2939aee3a7432d3
3
+ size 19535020
checkpoint_1007000/target.decoder.layers_15.mlp.wi_0.kernel/0.4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ffbd98d9d1f21c1d4e343fef6a4ad15e1d54b8328de6a5fff6222675a05b537c
3
+ size 19533987
checkpoint_1007000/target.decoder.layers_15.mlp.wi_0.kernel/0.5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8a4a77ac02c0ba94b63af2f1d4009a245a39f263d03957e47f04f9ab91bd67f5
3
+ size 19533724
checkpoint_1007000/target.decoder.layers_15.mlp.wi_0.kernel/0.6 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7a26ba1818290a41872b08befd5580915b586838930ac5180b1f6cb1add95786
3
+ size 19533830
checkpoint_1007000/target.decoder.layers_15.mlp.wi_0.kernel/0.7 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a133321b47b93f5515d92d341dacaf6ae681c6ee60969ccf19acbce09babcaa9
3
+ size 19534548
checkpoint_1007000/target.decoder.layers_16.encoder_decoder_attention.key.kernel/.zarray ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096,512],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,4096],"zarr_format":2}
checkpoint_1007000/target.decoder.layers_16.encoder_decoder_attention.key.kernel/.zarray.__tmp1277066131995913345.~1663960300272122~ ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096,512],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,4096],"zarr_format":2}
checkpoint_1007000/target.decoder.layers_16.encoder_decoder_attention.key.kernel/0.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:577949af17df08e0fd40da573f2c547271de23625c1ed2500730c713201754d0
3
+ size 7819044
checkpoint_1007000/target.decoder.layers_16.encoder_decoder_attention.key.kernel/0.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b3ad6bfb4318c47b026ed5fd4c3d12f7546564906353790135f1845bdcb06645
3
+ size 7816282
checkpoint_1007000/target.decoder.layers_16.encoder_decoder_attention.key.kernel/0.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2f7a959ff6e24fe3d945f26074cdc959dfca502a4db5fd40914f62c6f2d3cdd6
3
+ size 7817590
checkpoint_1007000/target.decoder.layers_16.encoder_decoder_attention.key.kernel/0.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6124fa7fa6600f78cecb985219317d46b77931384eaba69d44254877ee3e3abf
3
+ size 7817469
checkpoint_1007000/target.decoder.layers_16.encoder_decoder_attention.key.kernel/0.4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2493e048cdf234a2d2a01f4061e9afa21e3a747b697d8c0716ebe894b08b7e52
3
+ size 7817969
checkpoint_1007000/target.decoder.layers_16.encoder_decoder_attention.key.kernel/0.5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:452c1f1edc6076a05d8d97cdc74b3902e2921ade4a3be6336d273b8bb1b0c12d
3
+ size 7818125
checkpoint_1007000/target.decoder.layers_16.encoder_decoder_attention.key.kernel/0.6 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d62975dd5d5b44aed04c7569c5bb2babce70f3b0d1af4fbd8da39c4ec404afd2
3
+ size 7817556
checkpoint_1007000/target.decoder.layers_16.encoder_decoder_attention.key.kernel/0.7 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1acad59365b73e991d77ecb94a3e37d2d46f73c0b620a29221921d166ca70a90
3
+ size 7817542
checkpoint_1007000/target.decoder.layers_18.self_attention.key.kernel/.zarray ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096,512],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,4096],"zarr_format":2}
checkpoint_1007000/target.decoder.layers_18.self_attention.key.kernel/.zarray.__tmp12475133037183008814.~1663960300420670~ ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096,512],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,4096],"zarr_format":2}
checkpoint_1007000/target.decoder.layers_18.self_attention.key.kernel/.zarray.__tmp17331114784864401939.~1663960300465552~ ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096,512],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,4096],"zarr_format":2}
checkpoint_1007000/target.decoder.layers_18.self_attention.key.kernel/.zarray.__tmp4384522759500365203.~1663960300688851~ ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096,512],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,4096],"zarr_format":2}
checkpoint_1007000/target.decoder.layers_18.self_attention.key.kernel/.zarray.__tmp4785258170125338790.~1663960300499566~ ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096,512],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,4096],"zarr_format":2}
checkpoint_1007000/target.decoder.layers_18.self_attention.key.kernel/.zarray.__tmp4930767467751954311.~1663960300872912~ ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096,512],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,4096],"zarr_format":2}
checkpoint_1007000/target.decoder.layers_18.self_attention.key.kernel/.zarray.__tmp532225422100881862.~1663960300495791~ ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096,512],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,4096],"zarr_format":2}
checkpoint_1007000/target.decoder.layers_18.self_attention.key.kernel/.zarray.__tmp6571638690794785116.~1663960300628674~ ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096,512],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,4096],"zarr_format":2}
checkpoint_1007000/target.decoder.layers_18.self_attention.key.kernel/0.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b4663c7fcdf4632ef2e5960c53ab11772edf115e1ba221c4fdf8fa158cc93eb
3
+ size 7813531
checkpoint_1007000/target.decoder.layers_18.self_attention.key.kernel/0.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c745f8461ebffc69a912d0ccc79abd7666e4a9de2e910851fda28584897de0e1
3
+ size 7812098
checkpoint_1007000/target.decoder.layers_18.self_attention.key.kernel/0.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6246abe4c51f7d6f4afc1fd79c9c269acb6b3abed9abbe2bfceb4a404bfad2c2
3
+ size 7811109
checkpoint_1007000/target.decoder.layers_18.self_attention.key.kernel/0.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:32cb6ff98744097a524183c63906023ec99872b7fdfed51ffd1f3419e8d6501e
3
+ size 7812397
checkpoint_1007000/target.decoder.layers_18.self_attention.key.kernel/0.4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:50438f17281d7e534f3ff9e8526d2587240299cde3c149255117c50566c55e9e
3
+ size 7830058
checkpoint_1007000/target.decoder.layers_18.self_attention.key.kernel/0.5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d13a6f75680d0aa35162d8ea28d34a67f7d26305388fe4f7cdc7a28d38349856
3
+ size 7814582
checkpoint_1007000/target.decoder.layers_18.self_attention.key.kernel/0.6 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9cbcdaf64414fe7d8122031c7b66ae80e3b9a9c242d8cf2646053717cd447488
3
+ size 7811384
checkpoint_1007000/target.decoder.layers_18.self_attention.key.kernel/0.7 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fb382cdb78f0f70dcd1ba332788a73eab52caebddf292bab652fe3189baebef8
3
+ size 7817572
checkpoint_1007000/target.decoder.layers_6.pre_cross_attention_layer_norm.scale/.zarray ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096],"zarr_format":2}
checkpoint_1007000/target.decoder.layers_6.pre_cross_attention_layer_norm.scale/0 ADDED
Binary file (14.8 kB). View file
 
checkpoint_1007000/target.encoder.layers_1.attention.out.kernel/.zarray ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[512,4096],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,4096],"zarr_format":2}