Theia-4869 commited on
Commit
98d6198
·
1 Parent(s): 9e94652

add alpaca_llamaPeft_normBias_QF_512_70B

Browse files
Files changed (22) hide show
  1. finetune/sg/alpaca_llamaPeft_normBias_QF_512_70B/epoch0/consolidated.00-of-01.model.pth +3 -0
  2. finetune/sg/alpaca_llamaPeft_normBias_QF_512_70B/epoch0/consolidated.00-of-01.optimizer.pth +3 -0
  3. finetune/sg/alpaca_llamaPeft_normBias_QF_512_70B/epoch0/consolidated.00-of-01.other.pth +3 -0
  4. finetune/sg/alpaca_llamaPeft_normBias_QF_512_70B/epoch0/rank-specific-00000-of-00002.pth +3 -0
  5. finetune/sg/alpaca_llamaPeft_normBias_QF_512_70B/epoch0/rank-specific-00001-of-00002.pth +3 -0
  6. finetune/sg/alpaca_llamaPeft_normBias_QF_512_70B/epoch1/consolidated.00-of-01.model.pth +3 -0
  7. finetune/sg/alpaca_llamaPeft_normBias_QF_512_70B/epoch1/consolidated.00-of-01.optimizer.pth +3 -0
  8. finetune/sg/alpaca_llamaPeft_normBias_QF_512_70B/epoch1/consolidated.00-of-01.other.pth +3 -0
  9. finetune/sg/alpaca_llamaPeft_normBias_QF_512_70B/epoch1/rank-specific-00000-of-00002.pth +3 -0
  10. finetune/sg/alpaca_llamaPeft_normBias_QF_512_70B/epoch1/rank-specific-00001-of-00002.pth +3 -0
  11. finetune/sg/alpaca_llamaPeft_normBias_QF_512_70B/epoch2/consolidated.00-of-01.model.pth +3 -0
  12. finetune/sg/alpaca_llamaPeft_normBias_QF_512_70B/epoch2/consolidated.00-of-01.optimizer.pth +3 -0
  13. finetune/sg/alpaca_llamaPeft_normBias_QF_512_70B/epoch2/consolidated.00-of-01.other.pth +3 -0
  14. finetune/sg/alpaca_llamaPeft_normBias_QF_512_70B/epoch2/rank-specific-00000-of-00002.pth +3 -0
  15. finetune/sg/alpaca_llamaPeft_normBias_QF_512_70B/epoch2/rank-specific-00001-of-00002.pth +3 -0
  16. finetune/sg/alpaca_llamaPeft_normBias_QF_512_70B/epoch3/consolidated.00-of-01.model.pth +3 -0
  17. finetune/sg/alpaca_llamaPeft_normBias_QF_512_70B/epoch3/consolidated.00-of-01.optimizer.pth +3 -0
  18. finetune/sg/alpaca_llamaPeft_normBias_QF_512_70B/epoch3/consolidated.00-of-01.other.pth +3 -0
  19. finetune/sg/alpaca_llamaPeft_normBias_QF_512_70B/epoch3/rank-specific-00000-of-00002.pth +3 -0
  20. finetune/sg/alpaca_llamaPeft_normBias_QF_512_70B/epoch3/rank-specific-00001-of-00002.pth +3 -0
  21. finetune/sg/alpaca_llamaPeft_normBias_QF_512_70B/log.txt +4 -0
  22. finetune/sg/alpaca_llamaPeft_normBias_QF_512_70B/output.log +0 -0
finetune/sg/alpaca_llamaPeft_normBias_QF_512_70B/epoch0/consolidated.00-of-01.model.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:57840f6a94b317e5216383955e671d5c485aa459e9a77c46399004b14aab3449
3
+ size 16308187
finetune/sg/alpaca_llamaPeft_normBias_QF_512_70B/epoch0/consolidated.00-of-01.optimizer.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:98e48515afe7b737d914e6a945fddc635b1f3bd53d716da73568279b19d4e9fc
3
+ size 64801559
finetune/sg/alpaca_llamaPeft_normBias_QF_512_70B/epoch0/consolidated.00-of-01.other.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:15ad86707b40e58d9f1e5b304e7f138c74be797e45cdcd6c9c3e67d1ddea2a8b
3
+ size 1687
finetune/sg/alpaca_llamaPeft_normBias_QF_512_70B/epoch0/rank-specific-00000-of-00002.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:773a3b4cf6877fcfb087e3efb3b631fc40b16a6fdef5b9afb3cd6854bc59509f
3
+ size 537
finetune/sg/alpaca_llamaPeft_normBias_QF_512_70B/epoch0/rank-specific-00001-of-00002.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9baaa2fa1f453e7261aad9ed636db8f4395edcee0ada8daad1f8078b40d5f61c
3
+ size 537
finetune/sg/alpaca_llamaPeft_normBias_QF_512_70B/epoch1/consolidated.00-of-01.model.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:95c28c92604c98733f7d11de93aabb7b91bf51cf6d5d1b4a7648f88735df9be8
3
+ size 16308187
finetune/sg/alpaca_llamaPeft_normBias_QF_512_70B/epoch1/consolidated.00-of-01.optimizer.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:726e9afa5e22acbf9e238aaa616537ff8bc766c7a407d50f49fd184a85596d3f
3
+ size 64801559
finetune/sg/alpaca_llamaPeft_normBias_QF_512_70B/epoch1/consolidated.00-of-01.other.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e3ffbc6452c328bba47a43a12dbd4bc293c231de8f57a6bd819aa611ed703d60
3
+ size 1687
finetune/sg/alpaca_llamaPeft_normBias_QF_512_70B/epoch1/rank-specific-00000-of-00002.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:773a3b4cf6877fcfb087e3efb3b631fc40b16a6fdef5b9afb3cd6854bc59509f
3
+ size 537
finetune/sg/alpaca_llamaPeft_normBias_QF_512_70B/epoch1/rank-specific-00001-of-00002.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9baaa2fa1f453e7261aad9ed636db8f4395edcee0ada8daad1f8078b40d5f61c
3
+ size 537
finetune/sg/alpaca_llamaPeft_normBias_QF_512_70B/epoch2/consolidated.00-of-01.model.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:71c909bd4009747dcecd359334b72083ce7d70ba611d0835b3f9e805633df345
3
+ size 16308187
finetune/sg/alpaca_llamaPeft_normBias_QF_512_70B/epoch2/consolidated.00-of-01.optimizer.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8af183f8bb45d2e6aecf19b37dd23b380900961ab1df49a422c47a88086dd99b
3
+ size 64801559
finetune/sg/alpaca_llamaPeft_normBias_QF_512_70B/epoch2/consolidated.00-of-01.other.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4e1407e17b55be720204de47db0e9d89c18253e4bd99ce9beecf96812ad9220b
3
+ size 1687
finetune/sg/alpaca_llamaPeft_normBias_QF_512_70B/epoch2/rank-specific-00000-of-00002.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:773a3b4cf6877fcfb087e3efb3b631fc40b16a6fdef5b9afb3cd6854bc59509f
3
+ size 537
finetune/sg/alpaca_llamaPeft_normBias_QF_512_70B/epoch2/rank-specific-00001-of-00002.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9baaa2fa1f453e7261aad9ed636db8f4395edcee0ada8daad1f8078b40d5f61c
3
+ size 537
finetune/sg/alpaca_llamaPeft_normBias_QF_512_70B/epoch3/consolidated.00-of-01.model.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1ca963a2250f78056df0749f6d91daf572f7e87a00398a2ea04fb8e0d4fb2981
3
+ size 16308187
finetune/sg/alpaca_llamaPeft_normBias_QF_512_70B/epoch3/consolidated.00-of-01.optimizer.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d8f6388325fba6658d8f5f383cb6affc164930411672f0944d7f249e10a03b78
3
+ size 64801559
finetune/sg/alpaca_llamaPeft_normBias_QF_512_70B/epoch3/consolidated.00-of-01.other.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:99c45bf04236026e6ce60807e96c85d649fb32f7065883be1052037d816478dc
3
+ size 1687
finetune/sg/alpaca_llamaPeft_normBias_QF_512_70B/epoch3/rank-specific-00000-of-00002.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:773a3b4cf6877fcfb087e3efb3b631fc40b16a6fdef5b9afb3cd6854bc59509f
3
+ size 537
finetune/sg/alpaca_llamaPeft_normBias_QF_512_70B/epoch3/rank-specific-00001-of-00002.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9baaa2fa1f453e7261aad9ed636db8f4395edcee0ada8daad1f8078b40d5f61c
3
+ size 537
finetune/sg/alpaca_llamaPeft_normBias_QF_512_70B/log.txt ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {"train_lr": 2.49923076923077e-05, "train_closs": 0.7980487976441016, "train_grad_norm": 0.5980251384056532, "epoch": 0, "val_lr": 2.49923076923077e-05, "val_closs": 0.7980487976441016, "val_grad_norm": 0.5980251384056532}
2
+ {"train_lr": 4.6109080828728024e-05, "train_closs": 0.7623572307492678, "train_grad_norm": 0.45453824085914174, "epoch": 1, "val_lr": 4.6109080828728024e-05, "val_closs": 0.7623572307492678, "val_grad_norm": 0.45453824085914174}
3
+ {"train_lr": 2.750346153846151e-05, "train_closs": 0.750338752788993, "train_grad_norm": 0.46191218195511746, "epoch": 2, "val_lr": 2.750346153846151e-05, "val_closs": 0.750338752788993, "val_grad_norm": 0.46191218195511746}
4
+ {"train_lr": 8.894380709733404e-06, "train_closs": 0.742047518081963, "train_grad_norm": 0.47685301401064945, "epoch": 3, "val_lr": 8.894380709733404e-06, "val_closs": 0.742047518081963, "val_grad_norm": 0.47685301401064945}
finetune/sg/alpaca_llamaPeft_normBias_QF_512_70B/output.log ADDED
The diff for this file is too large to render. See raw diff