marco-molinari commited on
Commit
f8a14f6
·
verified ·
1 Parent(s): 228d75f

Upload folder using huggingface_hub

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. 1/.gpt_neox.layers.0/0_12287.safetensors +3 -0
  2. 1/.gpt_neox.layers.0/config.json +1 -0
  3. 1/.gpt_neox.layers.1/0_12287.safetensors +3 -0
  4. 1/.gpt_neox.layers.1/config.json +1 -0
  5. 1/.gpt_neox.layers.10/0_12287.safetensors +3 -0
  6. 1/.gpt_neox.layers.10/config.json +1 -0
  7. 1/.gpt_neox.layers.2/0_12287.safetensors +3 -0
  8. 1/.gpt_neox.layers.2/config.json +1 -0
  9. 1/.gpt_neox.layers.3/0_12287.safetensors +3 -0
  10. 1/.gpt_neox.layers.3/config.json +1 -0
  11. 1/.gpt_neox.layers.4/0_12287.safetensors +3 -0
  12. 1/.gpt_neox.layers.4/config.json +1 -0
  13. 1/.gpt_neox.layers.5/0_12287.safetensors +3 -0
  14. 1/.gpt_neox.layers.5/config.json +1 -0
  15. 1/.gpt_neox.layers.6/0_12287.safetensors +3 -0
  16. 1/.gpt_neox.layers.6/config.json +1 -0
  17. 1/.gpt_neox.layers.7/0_12287.safetensors +3 -0
  18. 1/.gpt_neox.layers.7/config.json +1 -0
  19. 1/.gpt_neox.layers.8/0_12287.safetensors +3 -0
  20. 1/.gpt_neox.layers.8/config.json +1 -0
  21. 1/.gpt_neox.layers.9/0_12287.safetensors +3 -0
  22. 1/.gpt_neox.layers.9/config.json +1 -0
  23. 2/.gpt_neox.layers.0/0_12287.safetensors +3 -0
  24. 2/.gpt_neox.layers.0/config.json +1 -0
  25. 2/.gpt_neox.layers.1/0_12287.safetensors +3 -0
  26. 2/.gpt_neox.layers.1/config.json +1 -0
  27. 2/.gpt_neox.layers.10/0_12287.safetensors +3 -0
  28. 2/.gpt_neox.layers.10/config.json +1 -0
  29. 2/.gpt_neox.layers.2/0_12287.safetensors +3 -0
  30. 2/.gpt_neox.layers.2/config.json +1 -0
  31. 2/.gpt_neox.layers.3/0_12287.safetensors +3 -0
  32. 2/.gpt_neox.layers.3/config.json +1 -0
  33. 2/.gpt_neox.layers.4/0_12287.safetensors +3 -0
  34. 2/.gpt_neox.layers.4/config.json +1 -0
  35. 2/.gpt_neox.layers.5/0_12287.safetensors +3 -0
  36. 2/.gpt_neox.layers.5/config.json +1 -0
  37. 2/.gpt_neox.layers.6/0_12287.safetensors +3 -0
  38. 2/.gpt_neox.layers.6/config.json +1 -0
  39. 2/.gpt_neox.layers.7/0_12287.safetensors +3 -0
  40. 2/.gpt_neox.layers.7/config.json +1 -0
  41. 2/.gpt_neox.layers.8/0_12287.safetensors +3 -0
  42. 2/.gpt_neox.layers.8/config.json +1 -0
  43. 2/.gpt_neox.layers.9/0_12287.safetensors +3 -0
  44. 2/.gpt_neox.layers.9/config.json +1 -0
  45. 3/.gpt_neox.layers.0/0_12287.safetensors +3 -0
  46. 3/.gpt_neox.layers.0/config.json +1 -0
  47. 3/.gpt_neox.layers.1/0_12287.safetensors +3 -0
  48. 3/.gpt_neox.layers.1/config.json +1 -0
  49. 3/.gpt_neox.layers.10/0_12287.safetensors +3 -0
  50. 3/.gpt_neox.layers.10/config.json +1 -0
1/.gpt_neox.layers.0/0_12287.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fc0a06b7f2a9d81782caf0b2f40f3b0532913d06bd5bb3dce3a3666fe8376aca
3
+ size 367821056
1/.gpt_neox.layers.0/config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"dataset_repo": "gngdb/subset_the_pile_deduplicated", "dataset_split": "train[:1%]", "dataset_name": "", "dataset_row": "text", "batch_size": 8, "ctx_len": 512, "n_tokens": 100000, "n_splits": 1, "model_name": "EleutherAI/pythia-160m"}
1/.gpt_neox.layers.1/0_12287.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e230f789eaff15c8e60b33745530c0e1f352130ceb13b50ed49683b5403ebfcb
3
+ size 367821056
1/.gpt_neox.layers.1/config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"dataset_repo": "gngdb/subset_the_pile_deduplicated", "dataset_split": "train[:1%]", "dataset_name": "", "dataset_row": "text", "batch_size": 8, "ctx_len": 512, "n_tokens": 100000, "n_splits": 1, "model_name": "EleutherAI/pythia-160m"}
1/.gpt_neox.layers.10/0_12287.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9164803a20200179501ef1f1edb1a43da86a79f369ce7c34b5a258f1b0e36560
3
+ size 367821056
1/.gpt_neox.layers.10/config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"dataset_repo": "gngdb/subset_the_pile_deduplicated", "dataset_split": "train[:1%]", "dataset_name": "", "dataset_row": "text", "batch_size": 8, "ctx_len": 512, "n_tokens": 100000, "n_splits": 1, "model_name": "EleutherAI/pythia-160m"}
1/.gpt_neox.layers.2/0_12287.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:43ff5bcb14be88fc2bf7da197af3d0a171f6e243d36decd6edb097183c29afb6
3
+ size 367821056
1/.gpt_neox.layers.2/config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"dataset_repo": "gngdb/subset_the_pile_deduplicated", "dataset_split": "train[:1%]", "dataset_name": "", "dataset_row": "text", "batch_size": 8, "ctx_len": 512, "n_tokens": 100000, "n_splits": 1, "model_name": "EleutherAI/pythia-160m"}
1/.gpt_neox.layers.3/0_12287.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:31524d314a2007c65378e681348e365cf6c4b890f203627528af22d21e24db88
3
+ size 367821056
1/.gpt_neox.layers.3/config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"dataset_repo": "gngdb/subset_the_pile_deduplicated", "dataset_split": "train[:1%]", "dataset_name": "", "dataset_row": "text", "batch_size": 8, "ctx_len": 512, "n_tokens": 100000, "n_splits": 1, "model_name": "EleutherAI/pythia-160m"}
1/.gpt_neox.layers.4/0_12287.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6678fc4583529131bd0640ece248c2f624338044aa27dca51958167b093c1606
3
+ size 367821056
1/.gpt_neox.layers.4/config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"dataset_repo": "gngdb/subset_the_pile_deduplicated", "dataset_split": "train[:1%]", "dataset_name": "", "dataset_row": "text", "batch_size": 8, "ctx_len": 512, "n_tokens": 100000, "n_splits": 1, "model_name": "EleutherAI/pythia-160m"}
1/.gpt_neox.layers.5/0_12287.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:898666d4a88ed05e7314b79ba28ba5ba8f207818619b31337c7838d9f065af67
3
+ size 367821056
1/.gpt_neox.layers.5/config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"dataset_repo": "gngdb/subset_the_pile_deduplicated", "dataset_split": "train[:1%]", "dataset_name": "", "dataset_row": "text", "batch_size": 8, "ctx_len": 512, "n_tokens": 100000, "n_splits": 1, "model_name": "EleutherAI/pythia-160m"}
1/.gpt_neox.layers.6/0_12287.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b32e7c6431282d5e57363ac11cc6dad0a98ab2d4d4a5a82ece71aab7f9a3f7c6
3
+ size 367821056
1/.gpt_neox.layers.6/config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"dataset_repo": "gngdb/subset_the_pile_deduplicated", "dataset_split": "train[:1%]", "dataset_name": "", "dataset_row": "text", "batch_size": 8, "ctx_len": 512, "n_tokens": 100000, "n_splits": 1, "model_name": "EleutherAI/pythia-160m"}
1/.gpt_neox.layers.7/0_12287.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0f870cbc19c3da8f973b6bfc71048a4e2a708dac51aea92307237d0b40ef67b2
3
+ size 367821056
1/.gpt_neox.layers.7/config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"dataset_repo": "gngdb/subset_the_pile_deduplicated", "dataset_split": "train[:1%]", "dataset_name": "", "dataset_row": "text", "batch_size": 8, "ctx_len": 512, "n_tokens": 100000, "n_splits": 1, "model_name": "EleutherAI/pythia-160m"}
1/.gpt_neox.layers.8/0_12287.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c12a4bf52c0f068d39c444d9327fba16b8186ef6fe80eee8ad79ae3c7d67e58d
3
+ size 367821056
1/.gpt_neox.layers.8/config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"dataset_repo": "gngdb/subset_the_pile_deduplicated", "dataset_split": "train[:1%]", "dataset_name": "", "dataset_row": "text", "batch_size": 8, "ctx_len": 512, "n_tokens": 100000, "n_splits": 1, "model_name": "EleutherAI/pythia-160m"}
1/.gpt_neox.layers.9/0_12287.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:43db3b66a4570ca1d1e257a4206ebc9c0c88c4e5bbc6d6e666da4fbee8fdc553
3
+ size 367821056
1/.gpt_neox.layers.9/config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"dataset_repo": "gngdb/subset_the_pile_deduplicated", "dataset_split": "train[:1%]", "dataset_name": "", "dataset_row": "text", "batch_size": 8, "ctx_len": 512, "n_tokens": 100000, "n_splits": 1, "model_name": "EleutherAI/pythia-160m"}
2/.gpt_neox.layers.0/0_12287.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c073fc592bec4bb0d570f48fab988345855669e67cafd5061e9f04b699d91155
3
+ size 367821056
2/.gpt_neox.layers.0/config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"dataset_repo": "gngdb/subset_the_pile_deduplicated", "dataset_split": "train[:1%]", "dataset_name": "", "dataset_row": "text", "batch_size": 8, "ctx_len": 512, "n_tokens": 100000, "n_splits": 1, "model_name": "EleutherAI/pythia-160m"}
2/.gpt_neox.layers.1/0_12287.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4db5626fb4cffe873f4cf34ba6ce2e11d148265b830414e636891ea813e85027
3
+ size 367821056
2/.gpt_neox.layers.1/config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"dataset_repo": "gngdb/subset_the_pile_deduplicated", "dataset_split": "train[:1%]", "dataset_name": "", "dataset_row": "text", "batch_size": 8, "ctx_len": 512, "n_tokens": 100000, "n_splits": 1, "model_name": "EleutherAI/pythia-160m"}
2/.gpt_neox.layers.10/0_12287.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:72195ae6ffe5750a2d257ced74a53b002ae7bba22c193d442d5ef93f5bb95ad8
3
+ size 367821056
2/.gpt_neox.layers.10/config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"dataset_repo": "gngdb/subset_the_pile_deduplicated", "dataset_split": "train[:1%]", "dataset_name": "", "dataset_row": "text", "batch_size": 8, "ctx_len": 512, "n_tokens": 100000, "n_splits": 1, "model_name": "EleutherAI/pythia-160m"}
2/.gpt_neox.layers.2/0_12287.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6788ded064c8cb6d47fab1fb7aeefdb5fb9e8eb6b77179b2e667bdaf3bab0079
3
+ size 367821056
2/.gpt_neox.layers.2/config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"dataset_repo": "gngdb/subset_the_pile_deduplicated", "dataset_split": "train[:1%]", "dataset_name": "", "dataset_row": "text", "batch_size": 8, "ctx_len": 512, "n_tokens": 100000, "n_splits": 1, "model_name": "EleutherAI/pythia-160m"}
2/.gpt_neox.layers.3/0_12287.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ce70e75e30c9d533148005833046a65c8954e42079557b6c761019bec02d9603
3
+ size 367821056
2/.gpt_neox.layers.3/config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"dataset_repo": "gngdb/subset_the_pile_deduplicated", "dataset_split": "train[:1%]", "dataset_name": "", "dataset_row": "text", "batch_size": 8, "ctx_len": 512, "n_tokens": 100000, "n_splits": 1, "model_name": "EleutherAI/pythia-160m"}
2/.gpt_neox.layers.4/0_12287.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:47244f69121fe3e8bc701b948a33f1d98ec5d8bbfe677851c1713892e989ab3b
3
+ size 367821056
2/.gpt_neox.layers.4/config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"dataset_repo": "gngdb/subset_the_pile_deduplicated", "dataset_split": "train[:1%]", "dataset_name": "", "dataset_row": "text", "batch_size": 8, "ctx_len": 512, "n_tokens": 100000, "n_splits": 1, "model_name": "EleutherAI/pythia-160m"}
2/.gpt_neox.layers.5/0_12287.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:603d8c3c4faa8b4074295b6d6028ecbbc87102e983d347a069f8d9a4ef4fa1a1
3
+ size 367821056
2/.gpt_neox.layers.5/config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"dataset_repo": "gngdb/subset_the_pile_deduplicated", "dataset_split": "train[:1%]", "dataset_name": "", "dataset_row": "text", "batch_size": 8, "ctx_len": 512, "n_tokens": 100000, "n_splits": 1, "model_name": "EleutherAI/pythia-160m"}
2/.gpt_neox.layers.6/0_12287.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7a1c82bed6d08ff38d9c727f4c8b64ac4d3908faf324f1e6926d4449a03af0e4
3
+ size 367821056
2/.gpt_neox.layers.6/config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"dataset_repo": "gngdb/subset_the_pile_deduplicated", "dataset_split": "train[:1%]", "dataset_name": "", "dataset_row": "text", "batch_size": 8, "ctx_len": 512, "n_tokens": 100000, "n_splits": 1, "model_name": "EleutherAI/pythia-160m"}
2/.gpt_neox.layers.7/0_12287.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ffc1b3add314886e737630b25a28bd95ca225587327532987fe92b06d5e5baee
3
+ size 367821056
2/.gpt_neox.layers.7/config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"dataset_repo": "gngdb/subset_the_pile_deduplicated", "dataset_split": "train[:1%]", "dataset_name": "", "dataset_row": "text", "batch_size": 8, "ctx_len": 512, "n_tokens": 100000, "n_splits": 1, "model_name": "EleutherAI/pythia-160m"}
2/.gpt_neox.layers.8/0_12287.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3d3e52057129311b899169f8c32c4079571784ac94c1d2375212079cafde8c56
3
+ size 367821056
2/.gpt_neox.layers.8/config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"dataset_repo": "gngdb/subset_the_pile_deduplicated", "dataset_split": "train[:1%]", "dataset_name": "", "dataset_row": "text", "batch_size": 8, "ctx_len": 512, "n_tokens": 100000, "n_splits": 1, "model_name": "EleutherAI/pythia-160m"}
2/.gpt_neox.layers.9/0_12287.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cb729ad816fedc53923cd106a8fd837c213f8a9a33b6603250b90e19e2903c2f
3
+ size 367821056
2/.gpt_neox.layers.9/config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"dataset_repo": "gngdb/subset_the_pile_deduplicated", "dataset_split": "train[:1%]", "dataset_name": "", "dataset_row": "text", "batch_size": 8, "ctx_len": 512, "n_tokens": 100000, "n_splits": 1, "model_name": "EleutherAI/pythia-160m"}
3/.gpt_neox.layers.0/0_12287.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eb0f9232fe5df7fa5946bc85ea834d8221a6686898d29cdfe87f91835619f825
3
+ size 367821056
3/.gpt_neox.layers.0/config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"dataset_repo": "gngdb/subset_the_pile_deduplicated", "dataset_split": "train[:1%]", "dataset_name": "", "dataset_row": "text", "batch_size": 8, "ctx_len": 512, "n_tokens": 100000, "n_splits": 1, "model_name": "EleutherAI/pythia-160m"}
3/.gpt_neox.layers.1/0_12287.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:469af5397dfe73c0b506e3071d0a378824eb3cd0180372806ee70b2809cd9ccf
3
+ size 367821056
3/.gpt_neox.layers.1/config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"dataset_repo": "gngdb/subset_the_pile_deduplicated", "dataset_split": "train[:1%]", "dataset_name": "", "dataset_row": "text", "batch_size": 8, "ctx_len": 512, "n_tokens": 100000, "n_splits": 1, "model_name": "EleutherAI/pythia-160m"}
3/.gpt_neox.layers.10/0_12287.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:72195ae6ffe5750a2d257ced74a53b002ae7bba22c193d442d5ef93f5bb95ad8
3
+ size 367821056
3/.gpt_neox.layers.10/config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"dataset_repo": "gngdb/subset_the_pile_deduplicated", "dataset_split": "train[:1%]", "dataset_name": "", "dataset_row": "text", "batch_size": 8, "ctx_len": 512, "n_tokens": 100000, "n_splits": 1, "model_name": "EleutherAI/pythia-160m"}