TimeRobber commited on
Commit
2db1d34
·
1 Parent(s): 6d3453f

Upload part 15

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +47 -0
  2. checkpoint_1007000/state.param_states.decoder.layers_0.pre_mlp_layer_norm.scale.v/.zarray +1 -0
  3. checkpoint_1007000/state.param_states.decoder.layers_0.pre_mlp_layer_norm.scale.v/0 +0 -0
  4. checkpoint_1007000/state.param_states.decoder.layers_3.pre_mlp_layer_norm.scale.v/.zarray +1 -0
  5. checkpoint_1007000/state.param_states.decoder.layers_3.pre_mlp_layer_norm.scale.v/0 +0 -0
  6. checkpoint_1007000/state.param_states.encoder.layers_15.pre_mlp_layer_norm.scale.v/.zarray +1 -0
  7. checkpoint_1007000/state.param_states.encoder.layers_15.pre_mlp_layer_norm.scale.v/0 +0 -0
  8. checkpoint_1007000/state.param_states.encoder.layers_9.pre_attention_layer_norm.scale.v/.zarray +1 -0
  9. checkpoint_1007000/state.param_states.encoder.layers_9.pre_attention_layer_norm.scale.v/0 +0 -0
  10. checkpoint_1007000/target.decoder.layers_1.encoder_decoder_attention.out.kernel/.zarray +1 -0
  11. checkpoint_1007000/target.decoder.layers_1.encoder_decoder_attention.out.kernel/.zarray.__tmp10452781550683565952.~1663960299959818~ +1 -0
  12. checkpoint_1007000/target.decoder.layers_1.encoder_decoder_attention.out.kernel/.zarray.__tmp6795471601700893938.~1663960300550924~ +1 -0
  13. checkpoint_1007000/target.decoder.layers_1.encoder_decoder_attention.out.kernel/0.0 +3 -0
  14. checkpoint_1007000/target.decoder.layers_1.encoder_decoder_attention.out.kernel/1.0 +3 -0
  15. checkpoint_1007000/target.decoder.layers_1.encoder_decoder_attention.out.kernel/2.0 +3 -0
  16. checkpoint_1007000/target.decoder.layers_1.encoder_decoder_attention.out.kernel/3.0 +3 -0
  17. checkpoint_1007000/target.decoder.layers_1.encoder_decoder_attention.out.kernel/4.0 +3 -0
  18. checkpoint_1007000/target.decoder.layers_1.encoder_decoder_attention.out.kernel/5.0 +3 -0
  19. checkpoint_1007000/target.decoder.layers_1.encoder_decoder_attention.out.kernel/6.0 +3 -0
  20. checkpoint_1007000/target.decoder.layers_1.encoder_decoder_attention.out.kernel/7.0 +3 -0
  21. checkpoint_1007000/target.decoder.layers_4.self_attention.value.kernel/.zarray +1 -0
  22. checkpoint_1007000/target.decoder.layers_4.self_attention.value.kernel/.zarray.__tmp12400406934700200612.~1663960300730888~ +1 -0
  23. checkpoint_1007000/target.decoder.layers_4.self_attention.value.kernel/.zarray.__tmp13585271832861140024.~1663960300775802~ +1 -0
  24. checkpoint_1007000/target.decoder.layers_4.self_attention.value.kernel/.zarray.__tmp14092873880176463808.~1663960300548948~ +1 -0
  25. checkpoint_1007000/target.decoder.layers_4.self_attention.value.kernel/.zarray.__tmp16326833665509948487.~1663960300628583~ +1 -0
  26. checkpoint_1007000/target.decoder.layers_4.self_attention.value.kernel/.zarray.__tmp1645611234738431017.~1663960300724381~ +1 -0
  27. checkpoint_1007000/target.decoder.layers_4.self_attention.value.kernel/.zarray.__tmp17896257401658942111.~1663960300471318~ +1 -0
  28. checkpoint_1007000/target.decoder.layers_4.self_attention.value.kernel/.zarray.__tmp8260727994482058377.~1663960300554681~ +1 -0
  29. checkpoint_1007000/target.decoder.layers_4.self_attention.value.kernel/0.0 +3 -0
  30. checkpoint_1007000/target.decoder.layers_4.self_attention.value.kernel/0.1 +3 -0
  31. checkpoint_1007000/target.decoder.layers_4.self_attention.value.kernel/0.2 +3 -0
  32. checkpoint_1007000/target.decoder.layers_4.self_attention.value.kernel/0.3 +3 -0
  33. checkpoint_1007000/target.decoder.layers_4.self_attention.value.kernel/0.4 +3 -0
  34. checkpoint_1007000/target.decoder.layers_4.self_attention.value.kernel/0.5 +3 -0
  35. checkpoint_1007000/target.decoder.layers_4.self_attention.value.kernel/0.6 +3 -0
  36. checkpoint_1007000/target.decoder.layers_4.self_attention.value.kernel/0.7 +3 -0
  37. checkpoint_1007000/target.encoder.layers_10.attention.key.kernel/.zarray +1 -0
  38. checkpoint_1007000/target.encoder.layers_10.attention.key.kernel/.zarray.__tmp14925359356772131659.~1663960300724602~ +1 -0
  39. checkpoint_1007000/target.encoder.layers_10.attention.key.kernel/.zarray.__tmp2542588794593353836.~1663960300625998~ +1 -0
  40. checkpoint_1007000/target.encoder.layers_10.attention.key.kernel/.zarray.__tmp3850820632510219419.~1663960301069089~ +1 -0
  41. checkpoint_1007000/target.encoder.layers_10.attention.key.kernel/.zarray.__tmp4598496975983557628.~1663960300885331~ +1 -0
  42. checkpoint_1007000/target.encoder.layers_10.attention.key.kernel/.zarray.__tmp706916633014621655.~1663960300888363~ +1 -0
  43. checkpoint_1007000/target.encoder.layers_10.attention.key.kernel/.zarray.__tmp730787871119168432.~1663960300549618~ +1 -0
  44. checkpoint_1007000/target.encoder.layers_10.attention.key.kernel/.zarray.__tmp7721123098801844376.~1663960300775518~ +1 -0
  45. checkpoint_1007000/target.encoder.layers_10.attention.key.kernel/0.0 +3 -0
  46. checkpoint_1007000/target.encoder.layers_10.attention.key.kernel/0.1 +3 -0
  47. checkpoint_1007000/target.encoder.layers_10.attention.key.kernel/0.2 +3 -0
  48. checkpoint_1007000/target.encoder.layers_10.attention.key.kernel/0.3 +3 -0
  49. checkpoint_1007000/target.encoder.layers_10.attention.key.kernel/0.4 +3 -0
  50. checkpoint_1007000/target.encoder.layers_10.attention.key.kernel/0.5 +3 -0
.gitattributes CHANGED
@@ -756,3 +756,50 @@ checkpoint_1007000/target.decoder.layers_6.self_attention.out.kernel/5.0 filter=
756
  checkpoint_1007000/target.decoder.layers_6.self_attention.out.kernel/2.0 filter=lfs diff=lfs merge=lfs -text
757
  checkpoint_1007000/target.decoder.layers_6.self_attention.out.kernel/4.0 filter=lfs diff=lfs merge=lfs -text
758
  checkpoint_1007000/target.encoder.layers_10.mlp.wi_1.kernel/0.1 filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
756
  checkpoint_1007000/target.decoder.layers_6.self_attention.out.kernel/2.0 filter=lfs diff=lfs merge=lfs -text
757
  checkpoint_1007000/target.decoder.layers_6.self_attention.out.kernel/4.0 filter=lfs diff=lfs merge=lfs -text
758
  checkpoint_1007000/target.encoder.layers_10.mlp.wi_1.kernel/0.1 filter=lfs diff=lfs merge=lfs -text
759
+ checkpoint_1007000/target.encoder.layers_10.mlp.wi_1.kernel/0.7 filter=lfs diff=lfs merge=lfs -text
760
+ checkpoint_1007000/target.encoder.layers_10.mlp.wi_1.kernel/0.0 filter=lfs diff=lfs merge=lfs -text
761
+ checkpoint_1007000/target.encoder.layers_10.mlp.wi_1.kernel/0.6 filter=lfs diff=lfs merge=lfs -text
762
+ checkpoint_1007000/target.encoder.layers_10.mlp.wi_1.kernel/0.5 filter=lfs diff=lfs merge=lfs -text
763
+ checkpoint_1007000/target.encoder.layers_10.mlp.wi_1.kernel/0.3 filter=lfs diff=lfs merge=lfs -text
764
+ checkpoint_1007000/target.encoder.layers_10.mlp.wi_1.kernel/0.2 filter=lfs diff=lfs merge=lfs -text
765
+ checkpoint_1007000/target.encoder.layers_10.mlp.wi_1.kernel/0.4 filter=lfs diff=lfs merge=lfs -text
766
+ checkpoint_1007000/target.encoder.layers_3.mlp.wo.kernel/6.0 filter=lfs diff=lfs merge=lfs -text
767
+ checkpoint_1007000/target.encoder.layers_3.mlp.wo.kernel/1.0 filter=lfs diff=lfs merge=lfs -text
768
+ checkpoint_1007000/target.encoder.layers_3.mlp.wo.kernel/3.0 filter=lfs diff=lfs merge=lfs -text
769
+ checkpoint_1007000/target.encoder.layers_3.mlp.wo.kernel/0.0 filter=lfs diff=lfs merge=lfs -text
770
+ checkpoint_1007000/target.encoder.layers_3.mlp.wo.kernel/7.0 filter=lfs diff=lfs merge=lfs -text
771
+ checkpoint_1007000/target.encoder.layers_3.mlp.wo.kernel/5.0 filter=lfs diff=lfs merge=lfs -text
772
+ checkpoint_1007000/target.encoder.layers_3.mlp.wo.kernel/2.0 filter=lfs diff=lfs merge=lfs -text
773
+ checkpoint_1007000/target.encoder.layers_3.mlp.wo.kernel/4.0 filter=lfs diff=lfs merge=lfs -text
774
+ checkpoint_1007000/target.decoder.layers_4.self_attention.value.kernel/0.1 filter=lfs diff=lfs merge=lfs -text
775
+ checkpoint_1007000/target.decoder.layers_4.self_attention.value.kernel/0.7 filter=lfs diff=lfs merge=lfs -text
776
+ checkpoint_1007000/target.decoder.layers_4.self_attention.value.kernel/0.0 filter=lfs diff=lfs merge=lfs -text
777
+ checkpoint_1007000/target.decoder.layers_4.self_attention.value.kernel/0.6 filter=lfs diff=lfs merge=lfs -text
778
+ checkpoint_1007000/target.decoder.layers_4.self_attention.value.kernel/0.5 filter=lfs diff=lfs merge=lfs -text
779
+ checkpoint_1007000/target.decoder.layers_4.self_attention.value.kernel/0.3 filter=lfs diff=lfs merge=lfs -text
780
+ checkpoint_1007000/target.decoder.layers_4.self_attention.value.kernel/0.2 filter=lfs diff=lfs merge=lfs -text
781
+ checkpoint_1007000/target.decoder.layers_4.self_attention.value.kernel/0.4 filter=lfs diff=lfs merge=lfs -text
782
+ checkpoint_1007000/target.encoder.layers_18.attention.out.kernel/6.0 filter=lfs diff=lfs merge=lfs -text
783
+ checkpoint_1007000/target.encoder.layers_18.attention.out.kernel/1.0 filter=lfs diff=lfs merge=lfs -text
784
+ checkpoint_1007000/target.encoder.layers_18.attention.out.kernel/3.0 filter=lfs diff=lfs merge=lfs -text
785
+ checkpoint_1007000/target.encoder.layers_18.attention.out.kernel/0.0 filter=lfs diff=lfs merge=lfs -text
786
+ checkpoint_1007000/target.encoder.layers_18.attention.out.kernel/7.0 filter=lfs diff=lfs merge=lfs -text
787
+ checkpoint_1007000/target.encoder.layers_18.attention.out.kernel/5.0 filter=lfs diff=lfs merge=lfs -text
788
+ checkpoint_1007000/target.encoder.layers_18.attention.out.kernel/2.0 filter=lfs diff=lfs merge=lfs -text
789
+ checkpoint_1007000/target.encoder.layers_18.attention.out.kernel/4.0 filter=lfs diff=lfs merge=lfs -text
790
+ checkpoint_1007000/target.encoder.layers_10.attention.key.kernel/0.1 filter=lfs diff=lfs merge=lfs -text
791
+ checkpoint_1007000/target.encoder.layers_10.attention.key.kernel/0.7 filter=lfs diff=lfs merge=lfs -text
792
+ checkpoint_1007000/target.encoder.layers_10.attention.key.kernel/0.0 filter=lfs diff=lfs merge=lfs -text
793
+ checkpoint_1007000/target.encoder.layers_10.attention.key.kernel/0.6 filter=lfs diff=lfs merge=lfs -text
794
+ checkpoint_1007000/target.encoder.layers_10.attention.key.kernel/0.5 filter=lfs diff=lfs merge=lfs -text
795
+ checkpoint_1007000/target.encoder.layers_10.attention.key.kernel/0.3 filter=lfs diff=lfs merge=lfs -text
796
+ checkpoint_1007000/target.encoder.layers_10.attention.key.kernel/0.2 filter=lfs diff=lfs merge=lfs -text
797
+ checkpoint_1007000/target.encoder.layers_10.attention.key.kernel/0.4 filter=lfs diff=lfs merge=lfs -text
798
+ checkpoint_1007000/target.decoder.layers_1.encoder_decoder_attention.out.kernel/6.0 filter=lfs diff=lfs merge=lfs -text
799
+ checkpoint_1007000/target.decoder.layers_1.encoder_decoder_attention.out.kernel/1.0 filter=lfs diff=lfs merge=lfs -text
800
+ checkpoint_1007000/target.decoder.layers_1.encoder_decoder_attention.out.kernel/3.0 filter=lfs diff=lfs merge=lfs -text
801
+ checkpoint_1007000/target.decoder.layers_1.encoder_decoder_attention.out.kernel/0.0 filter=lfs diff=lfs merge=lfs -text
802
+ checkpoint_1007000/target.decoder.layers_1.encoder_decoder_attention.out.kernel/7.0 filter=lfs diff=lfs merge=lfs -text
803
+ checkpoint_1007000/target.decoder.layers_1.encoder_decoder_attention.out.kernel/5.0 filter=lfs diff=lfs merge=lfs -text
804
+ checkpoint_1007000/target.decoder.layers_1.encoder_decoder_attention.out.kernel/2.0 filter=lfs diff=lfs merge=lfs -text
805
+ checkpoint_1007000/target.decoder.layers_1.encoder_decoder_attention.out.kernel/4.0 filter=lfs diff=lfs merge=lfs -text
checkpoint_1007000/state.param_states.decoder.layers_0.pre_mlp_layer_norm.scale.v/.zarray ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096],"zarr_format":2}
checkpoint_1007000/state.param_states.decoder.layers_0.pre_mlp_layer_norm.scale.v/0 ADDED
Binary file (14.8 kB). View file
 
checkpoint_1007000/state.param_states.decoder.layers_3.pre_mlp_layer_norm.scale.v/.zarray ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096],"zarr_format":2}
checkpoint_1007000/state.param_states.decoder.layers_3.pre_mlp_layer_norm.scale.v/0 ADDED
Binary file (14.9 kB). View file
 
checkpoint_1007000/state.param_states.encoder.layers_15.pre_mlp_layer_norm.scale.v/.zarray ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096],"zarr_format":2}
checkpoint_1007000/state.param_states.encoder.layers_15.pre_mlp_layer_norm.scale.v/0 ADDED
Binary file (15.1 kB). View file
 
checkpoint_1007000/state.param_states.encoder.layers_9.pre_attention_layer_norm.scale.v/.zarray ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096],"zarr_format":2}
checkpoint_1007000/state.param_states.encoder.layers_9.pre_attention_layer_norm.scale.v/0 ADDED
Binary file (15.3 kB). View file
 
checkpoint_1007000/target.decoder.layers_1.encoder_decoder_attention.out.kernel/.zarray ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[512,4096],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,4096],"zarr_format":2}
checkpoint_1007000/target.decoder.layers_1.encoder_decoder_attention.out.kernel/.zarray.__tmp10452781550683565952.~1663960299959818~ ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[512,4096],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,4096],"zarr_format":2}
checkpoint_1007000/target.decoder.layers_1.encoder_decoder_attention.out.kernel/.zarray.__tmp6795471601700893938.~1663960300550924~ ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[512,4096],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,4096],"zarr_format":2}
checkpoint_1007000/target.decoder.layers_1.encoder_decoder_attention.out.kernel/0.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a7729f3e197ab9a0676935d7ad3395cd2b8985f19705b74ee1cded8f2123a0a1
3
+ size 7818512
checkpoint_1007000/target.decoder.layers_1.encoder_decoder_attention.out.kernel/1.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a3807185840914e8a06f0d2b5346b30d9cec6fcbe705a790326a9b9a0eee7355
3
+ size 7816960
checkpoint_1007000/target.decoder.layers_1.encoder_decoder_attention.out.kernel/2.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5ac640421454bddc0aa6e81189ff274d8763a3dfa1399fd4e044a8275fc6850c
3
+ size 7815017
checkpoint_1007000/target.decoder.layers_1.encoder_decoder_attention.out.kernel/3.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7cf927729dc1c55b247746fbe155bc5d54ad0bbbb12cad99d5ada04945f149a6
3
+ size 7818290
checkpoint_1007000/target.decoder.layers_1.encoder_decoder_attention.out.kernel/4.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:58c8cf3f1a83c773ea050d42fb6a72368da9c2ffe1f0c8e2b4b951e6914841b9
3
+ size 7816718
checkpoint_1007000/target.decoder.layers_1.encoder_decoder_attention.out.kernel/5.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f3ca425d8de5dcfe927a0c0855c608964462092aa376eadae0294466c62a157d
3
+ size 7819024
checkpoint_1007000/target.decoder.layers_1.encoder_decoder_attention.out.kernel/6.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cea01f95ab69a5c632899b3f264630a44f093b2cb151dd2dd5a35d7ba8ec99d6
3
+ size 7818568
checkpoint_1007000/target.decoder.layers_1.encoder_decoder_attention.out.kernel/7.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fdbae2e923416cb3f1cc977031bcb93f91f5059485260d3084081eae719df64a
3
+ size 7819994
checkpoint_1007000/target.decoder.layers_4.self_attention.value.kernel/.zarray ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096,512],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,4096],"zarr_format":2}
checkpoint_1007000/target.decoder.layers_4.self_attention.value.kernel/.zarray.__tmp12400406934700200612.~1663960300730888~ ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096,512],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,4096],"zarr_format":2}
checkpoint_1007000/target.decoder.layers_4.self_attention.value.kernel/.zarray.__tmp13585271832861140024.~1663960300775802~ ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096,512],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,4096],"zarr_format":2}
checkpoint_1007000/target.decoder.layers_4.self_attention.value.kernel/.zarray.__tmp14092873880176463808.~1663960300548948~ ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096,512],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,4096],"zarr_format":2}
checkpoint_1007000/target.decoder.layers_4.self_attention.value.kernel/.zarray.__tmp16326833665509948487.~1663960300628583~ ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096,512],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,4096],"zarr_format":2}
checkpoint_1007000/target.decoder.layers_4.self_attention.value.kernel/.zarray.__tmp1645611234738431017.~1663960300724381~ ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096,512],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,4096],"zarr_format":2}
checkpoint_1007000/target.decoder.layers_4.self_attention.value.kernel/.zarray.__tmp17896257401658942111.~1663960300471318~ ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096,512],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,4096],"zarr_format":2}
checkpoint_1007000/target.decoder.layers_4.self_attention.value.kernel/.zarray.__tmp8260727994482058377.~1663960300554681~ ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096,512],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,4096],"zarr_format":2}
checkpoint_1007000/target.decoder.layers_4.self_attention.value.kernel/0.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:77748ef6dc0f7d65ea20d04e82103ae7587df94cd285e10e5aab7ab8e5f5d14a
3
+ size 7830170
checkpoint_1007000/target.decoder.layers_4.self_attention.value.kernel/0.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:38b5fb3c57410d351306c47ab7440bc4604e376143de8e19cdd0245991703db8
3
+ size 7823920
checkpoint_1007000/target.decoder.layers_4.self_attention.value.kernel/0.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:355eec214a7d7b10547613e0fe583e489285b8ea1f9cf2b0e35373f5084f5c6a
3
+ size 7818936
checkpoint_1007000/target.decoder.layers_4.self_attention.value.kernel/0.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4e674fb4aea32e6323e3b5c31ffcaad2395f2a58da63b2046ec8b19a9fd0b220
3
+ size 7824757
checkpoint_1007000/target.decoder.layers_4.self_attention.value.kernel/0.4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0b6575d14ffadfa5cbbd82084d6ba9aa43156b749c4d156622dad65fe5557051
3
+ size 7827117
checkpoint_1007000/target.decoder.layers_4.self_attention.value.kernel/0.5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e56a96cf7e98c431be7dbacee4986ca71ed0b082c4c9d7fdfc10e364d69ecff5
3
+ size 7825619
checkpoint_1007000/target.decoder.layers_4.self_attention.value.kernel/0.6 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:133878bd936f06776015059e7fa1546fdfc12f18015dc7edab3bc6ddf00c02fd
3
+ size 7826228
checkpoint_1007000/target.decoder.layers_4.self_attention.value.kernel/0.7 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:02e59470fd96a7eb02f895c28f0cc1a8ab4ac390dd987dee5d08532e0e680c8c
3
+ size 7829635
checkpoint_1007000/target.encoder.layers_10.attention.key.kernel/.zarray ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096,512],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,4096],"zarr_format":2}
checkpoint_1007000/target.encoder.layers_10.attention.key.kernel/.zarray.__tmp14925359356772131659.~1663960300724602~ ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096,512],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,4096],"zarr_format":2}
checkpoint_1007000/target.encoder.layers_10.attention.key.kernel/.zarray.__tmp2542588794593353836.~1663960300625998~ ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096,512],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,4096],"zarr_format":2}
checkpoint_1007000/target.encoder.layers_10.attention.key.kernel/.zarray.__tmp3850820632510219419.~1663960301069089~ ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096,512],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,4096],"zarr_format":2}
checkpoint_1007000/target.encoder.layers_10.attention.key.kernel/.zarray.__tmp4598496975983557628.~1663960300885331~ ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096,512],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,4096],"zarr_format":2}
checkpoint_1007000/target.encoder.layers_10.attention.key.kernel/.zarray.__tmp706916633014621655.~1663960300888363~ ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096,512],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,4096],"zarr_format":2}
checkpoint_1007000/target.encoder.layers_10.attention.key.kernel/.zarray.__tmp730787871119168432.~1663960300549618~ ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096,512],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,4096],"zarr_format":2}
checkpoint_1007000/target.encoder.layers_10.attention.key.kernel/.zarray.__tmp7721123098801844376.~1663960300775518~ ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096,512],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,4096],"zarr_format":2}
checkpoint_1007000/target.encoder.layers_10.attention.key.kernel/0.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:075d6ded78fbd6347d61b525ee034cbc705c7a028240d2154fe889998404bdcc
3
+ size 7804423
checkpoint_1007000/target.encoder.layers_10.attention.key.kernel/0.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f55a2ca7fd4d8203113a63e105cef25f2ad4fb5f5c52b237734cf413b71585e6
3
+ size 7802069
checkpoint_1007000/target.encoder.layers_10.attention.key.kernel/0.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b80861db6e3f638bced104eb05fd49d6674249d9ae22ae9b042bafdd14ce0d49
3
+ size 7802968
checkpoint_1007000/target.encoder.layers_10.attention.key.kernel/0.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:66c85b4a3fd1c5d3ab065d882f96a1833f01cfc752c4bcb321a429974699a87f
3
+ size 7800188
checkpoint_1007000/target.encoder.layers_10.attention.key.kernel/0.4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:72badae23b381c9b925191b27d2971c7e3bcdce6739270ae754722971b56d0b1
3
+ size 7802995
checkpoint_1007000/target.encoder.layers_10.attention.key.kernel/0.5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3186a9f6765e579aeb5b580ed6bf92ff6920d223aa966a36865ffa3ff2977c66
3
+ size 7802939