stas commited on
Commit
5a89037
·
1 Parent(s): da8ef78

small model

Browse files
Files changed (28) hide show
  1. 2l-2gpus/000-module.3.input_layernorm/events.out.tfevents.1637214166.hope.523883.4 +3 -0
  2. 2l-2gpus/000-module.3.mlp.dense_4h_to_h/events.out.tfevents.1637214166.hope.523883.12 +3 -0
  3. 2l-2gpus/000-module.3.mlp.dense_h_to_4h/events.out.tfevents.1637214166.hope.523883.11 +3 -0
  4. 2l-2gpus/000-module.3.mlp/events.out.tfevents.1637214166.hope.523883.13 +3 -0
  5. 2l-2gpus/000-module.3.post_attention_layernorm/events.out.tfevents.1637214166.hope.523883.10 +3 -0
  6. 2l-2gpus/000-module.3.self_attention.attention_dropout/events.out.tfevents.1637214166.hope.523883.7 +3 -0
  7. 2l-2gpus/000-module.3.self_attention.dense/events.out.tfevents.1637214166.hope.523883.8 +3 -0
  8. 2l-2gpus/000-module.3.self_attention.query_key_value/events.out.tfevents.1637214166.hope.523883.5 +3 -0
  9. 2l-2gpus/000-module.3.self_attention.scale_mask_softmax/events.out.tfevents.1637214166.hope.523883.6 +3 -0
  10. 2l-2gpus/000-module.3.self_attention/events.out.tfevents.1637214166.hope.523883.9 +3 -0
  11. 2l-2gpus/000-module.3/events.out.tfevents.1637214166.hope.523883.14 +3 -0
  12. 2l-2gpus/000-module.4.input_layernorm/events.out.tfevents.1637214166.hope.523883.15 +3 -0
  13. 2l-2gpus/000-module.4.mlp.dense_4h_to_h/events.out.tfevents.1637214166.hope.523883.23 +3 -0
  14. 2l-2gpus/000-module.4.mlp.dense_h_to_4h/events.out.tfevents.1637214166.hope.523883.22 +3 -0
  15. 2l-2gpus/000-module.4.mlp/events.out.tfevents.1637214166.hope.523883.24 +3 -0
  16. 2l-2gpus/000-module.4.post_attention_layernorm/events.out.tfevents.1637214166.hope.523883.21 +3 -0
  17. 2l-2gpus/000-module.4.self_attention.attention_dropout/events.out.tfevents.1637214166.hope.523883.18 +3 -0
  18. 2l-2gpus/000-module.4.self_attention.dense/events.out.tfevents.1637214166.hope.523883.19 +3 -0
  19. 2l-2gpus/000-module.4.self_attention.query_key_value/events.out.tfevents.1637214166.hope.523883.16 +3 -0
  20. 2l-2gpus/000-module.4.self_attention.scale_mask_softmax/events.out.tfevents.1637214166.hope.523883.17 +3 -0
  21. 2l-2gpus/000-module.4.self_attention/events.out.tfevents.1637214166.hope.523883.20 +3 -0
  22. 2l-2gpus/000-module.4/events.out.tfevents.1637214166.hope.523883.25 +3 -0
  23. 2l-2gpus/000-module.6/events.out.tfevents.1637214166.hope.523883.26 +3 -0
  24. 2l-2gpus/000-module.tied_modules.embed.embedding_dropout/events.out.tfevents.1637214166.hope.523883.2 +3 -0
  25. 2l-2gpus/000-module.tied_modules.embed.position_embeddings/events.out.tfevents.1637214166.hope.523883.1 +3 -0
  26. 2l-2gpus/000-module.tied_modules.embed.word_embeddings/events.out.tfevents.1637214165.hope.523883.0 +3 -0
  27. 2l-2gpus/000-module.tied_modules.embed/events.out.tfevents.1637214166.hope.523883.3 +3 -0
  28. 2l-2gpus/000-module/events.out.tfevents.1637214166.hope.523883.27 +3 -0
2l-2gpus/000-module.3.input_layernorm/events.out.tfevents.1637214166.hope.523883.4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:88d3ff5eaf998ed951c1560975666f4f4014ba140f363812759f0c85b8872c11
3
+ size 667936
2l-2gpus/000-module.3.mlp.dense_4h_to_h/events.out.tfevents.1637214166.hope.523883.12 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:263a551f00eeb1e634ce8058f6d81e16fd89b433292c14911baf6ede43f156e1
3
+ size 681880
2l-2gpus/000-module.3.mlp.dense_h_to_4h/events.out.tfevents.1637214166.hope.523883.11 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d5bfe704983fa96f818deb5349ca4f9c416071d9b7003b13e0c6dda75f02485e
3
+ size 681880
2l-2gpus/000-module.3.mlp/events.out.tfevents.1637214166.hope.523883.13 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9ee85de4317718c096d4ed48c77570c1f84502793448ba1be501e7c8d20d5663
3
+ size 400768
2l-2gpus/000-module.3.post_attention_layernorm/events.out.tfevents.1637214166.hope.523883.10 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fb1fa0059e4452b4965afca592479f03a1205099476ae4532b38d0fce8f6b529
3
+ size 667936
2l-2gpus/000-module.3.self_attention.attention_dropout/events.out.tfevents.1637214166.hope.523883.7 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c7c86b1c52755755278620de1747efbefe108decca05cbbb1fc4713171e4bb31
3
+ size 293152
2l-2gpus/000-module.3.self_attention.dense/events.out.tfevents.1637214166.hope.523883.8 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:77d6fbd47d7cb462461c2fcbd203bd4c087e37328d9832ad2df045ab563a4800
3
+ size 681880
2l-2gpus/000-module.3.self_attention.query_key_value/events.out.tfevents.1637214166.hope.523883.5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bda28db630d314b9a8fa481d83fba98d82b5730b1f8ec9ee8bbe4222f610c7d2
3
+ size 626068
2l-2gpus/000-module.3.self_attention.scale_mask_softmax/events.out.tfevents.1637214166.hope.523883.6 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e7aa57772d48d7775a7423469cb9a615e6e8bb6c44434dd30776275399c2c8d8
3
+ size 293152
2l-2gpus/000-module.3.self_attention/events.out.tfevents.1637214166.hope.523883.9 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ac0d3604320d191fdaeb79a003213a4bbb7979fbd2e66ff720e5a608104619f5
3
+ size 400768
2l-2gpus/000-module.3/events.out.tfevents.1637214166.hope.523883.14 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:04d6abefb735e388856c3ce066b868e54386688212ea3bc24ceec27194295515
3
+ size 386824
2l-2gpus/000-module.4.input_layernorm/events.out.tfevents.1637214166.hope.523883.15 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9426f9acd88b3aab117d5928c33ea59245f9eaf0869a4c8d07cf99afac3d7b71
3
+ size 667936
2l-2gpus/000-module.4.mlp.dense_4h_to_h/events.out.tfevents.1637214166.hope.523883.23 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a622f475834a5ebf815bed5f7fc5a15f75b0d138cd4003a2773859d56a1843c1
3
+ size 681880
2l-2gpus/000-module.4.mlp.dense_h_to_4h/events.out.tfevents.1637214166.hope.523883.22 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:30a03578103bcfb1cadfc9fa0c2a972952e3c9eb2a83930d655f748af4761dfc
3
+ size 681880
2l-2gpus/000-module.4.mlp/events.out.tfevents.1637214166.hope.523883.24 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:884c06810b7b4f79e6efa4f5e97880979f37ecf96b8ec53c9c40a11dfd982ceb
3
+ size 400768
2l-2gpus/000-module.4.post_attention_layernorm/events.out.tfevents.1637214166.hope.523883.21 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0adeee80200d9ff25dc905f00f4a4f0a77c3e571a144f7f6786e5966a78a015a
3
+ size 667936
2l-2gpus/000-module.4.self_attention.attention_dropout/events.out.tfevents.1637214166.hope.523883.18 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:081c066cc83c3d03ab5e5ea9cfa0b0739595f5db13ba53cebdbad319f26fdc6f
3
+ size 293152
2l-2gpus/000-module.4.self_attention.dense/events.out.tfevents.1637214166.hope.523883.19 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4f6eaf1f9ab76bc27d07c14fb8b72179bb062502b1ba6cc01f1ccab4bd4f1513
3
+ size 681880
2l-2gpus/000-module.4.self_attention.query_key_value/events.out.tfevents.1637214166.hope.523883.16 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7352001fd7672a8a80fb2958a453f109f9a560c4e8751ac621ffb57a6ffd3563
3
+ size 626068
2l-2gpus/000-module.4.self_attention.scale_mask_softmax/events.out.tfevents.1637214166.hope.523883.17 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8f87028df24556698dc4e275552af628b65b0fb9f604ad61ed5d1c872cd53a55
3
+ size 293152
2l-2gpus/000-module.4.self_attention/events.out.tfevents.1637214166.hope.523883.20 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:731d8a9e3d4d5ae4b7df12de4586e6edd5459af4b0cf7eaee137f1262a363707
3
+ size 400768
2l-2gpus/000-module.4/events.out.tfevents.1637214166.hope.523883.25 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:621cd765ad7f6aaee95cb03c87ca7836585890ad0a8047864ead9c27c5d59746
3
+ size 386824
2l-2gpus/000-module.6/events.out.tfevents.1637214166.hope.523883.26 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:07573617c57894ab7ffba27fcc1249622f9985d8594630129196035d95f77cc0
3
+ size 486736
2l-2gpus/000-module.tied_modules.embed.embedding_dropout/events.out.tfevents.1637214166.hope.523883.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:873414c3f398c15f022b627028623dae509a85686dd1c3229831e4e3f7355d5f
3
+ size 201016
2l-2gpus/000-module.tied_modules.embed.position_embeddings/events.out.tfevents.1637214166.hope.523883.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4a35e9486ff4ae69f01ff1410fba9017045e035465e11c1bf33be241bfd71889
3
+ size 244540
2l-2gpus/000-module.tied_modules.embed.word_embeddings/events.out.tfevents.1637214165.hope.523883.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6032e735e28a9680cb42800604eee4f9e99da966bd3d783df505e06e9120c815
3
+ size 244540
2l-2gpus/000-module.tied_modules.embed/events.out.tfevents.1637214166.hope.523883.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d8a49ab5719db0eb9afe38bbde475848e21c13b6abcda713abc877c03aeff9e7
3
+ size 146860
2l-2gpus/000-module/events.out.tfevents.1637214166.hope.523883.27 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e9abe11ccc53257222f7ea74fe483fb202bc0ecadce0fb75096a40f818dd9536
3
+ size 146860