Theia-4869 commited on
Commit
bfabc95
·
1 Parent(s): f0741c0

add alpaca_llamaPeft_normBiasLora_QF_512_70B

Browse files
Files changed (47) hide show
  1. finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch0/consolidated.00-of-01.model.pth +3 -0
  2. finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch0/consolidated.00-of-01.optimizer.pth +3 -0
  3. finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch0/consolidated.00-of-01.other.pth +3 -0
  4. finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch0/rank-specific-00000-of-00008.pth +3 -0
  5. finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch0/rank-specific-00001-of-00008.pth +3 -0
  6. finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch0/rank-specific-00002-of-00008.pth +3 -0
  7. finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch0/rank-specific-00003-of-00008.pth +3 -0
  8. finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch0/rank-specific-00004-of-00008.pth +3 -0
  9. finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch0/rank-specific-00005-of-00008.pth +3 -0
  10. finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch0/rank-specific-00006-of-00008.pth +3 -0
  11. finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch0/rank-specific-00007-of-00008.pth +3 -0
  12. finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch1/consolidated.00-of-01.model.pth +3 -0
  13. finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch1/consolidated.00-of-01.optimizer.pth +3 -0
  14. finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch1/consolidated.00-of-01.other.pth +3 -0
  15. finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch1/rank-specific-00000-of-00008.pth +3 -0
  16. finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch1/rank-specific-00001-of-00008.pth +3 -0
  17. finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch1/rank-specific-00002-of-00008.pth +3 -0
  18. finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch1/rank-specific-00003-of-00008.pth +3 -0
  19. finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch1/rank-specific-00004-of-00008.pth +3 -0
  20. finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch1/rank-specific-00005-of-00008.pth +3 -0
  21. finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch1/rank-specific-00006-of-00008.pth +3 -0
  22. finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch1/rank-specific-00007-of-00008.pth +3 -0
  23. finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch2/consolidated.00-of-01.model.pth +3 -0
  24. finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch2/consolidated.00-of-01.optimizer.pth +3 -0
  25. finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch2/consolidated.00-of-01.other.pth +3 -0
  26. finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch2/rank-specific-00000-of-00008.pth +3 -0
  27. finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch2/rank-specific-00001-of-00008.pth +3 -0
  28. finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch2/rank-specific-00002-of-00008.pth +3 -0
  29. finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch2/rank-specific-00003-of-00008.pth +3 -0
  30. finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch2/rank-specific-00004-of-00008.pth +3 -0
  31. finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch2/rank-specific-00005-of-00008.pth +3 -0
  32. finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch2/rank-specific-00006-of-00008.pth +3 -0
  33. finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch2/rank-specific-00007-of-00008.pth +3 -0
  34. finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch3.zip +3 -0
  35. finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch3/consolidated.00-of-01.model.pth +3 -0
  36. finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch3/consolidated.00-of-01.optimizer.pth +3 -0
  37. finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch3/consolidated.00-of-01.other.pth +3 -0
  38. finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch3/rank-specific-00000-of-00008.pth +3 -0
  39. finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch3/rank-specific-00001-of-00008.pth +3 -0
  40. finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch3/rank-specific-00002-of-00008.pth +3 -0
  41. finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch3/rank-specific-00003-of-00008.pth +3 -0
  42. finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch3/rank-specific-00004-of-00008.pth +3 -0
  43. finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch3/rank-specific-00005-of-00008.pth +3 -0
  44. finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch3/rank-specific-00006-of-00008.pth +3 -0
  45. finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch3/rank-specific-00007-of-00008.pth +3 -0
  46. finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/log.txt +4 -0
  47. finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/output.log +0 -0
finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch0/consolidated.00-of-01.model.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:845d95e7a891af4aa67d67010d0345372ba47972219b6584a84276a2d24f7b4c
3
+ size 297203748
finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch0/consolidated.00-of-01.optimizer.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bdfa2eb02a223643fd49768731adaf8b51dc7ae98c0a461826366103f80c8bd5
3
+ size 652420087
finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch0/consolidated.00-of-01.other.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c5889268866aaf511f8591d23816d7faa96bed89b07daff596fe31445d6abc6
3
+ size 1751
finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch0/rank-specific-00000-of-00008.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ec2932635da1a4de71c34aa8fcbcba91dfb0ac1ddc7859f8f87280546b7e786a
3
+ size 537
finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch0/rank-specific-00001-of-00008.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:88973b3c418b507bcde1467ec3902218b83d95fe4e022aca11b09c3f86cde7ac
3
+ size 537
finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch0/rank-specific-00002-of-00008.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eee15a274ea5f27c0360c85bd878d6e0f2072076cae26311c52798f7d836643a
3
+ size 537
finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch0/rank-specific-00003-of-00008.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61651d612914693bf494e5609388a6f9239090c45b3abcc9c4fa5c7a814c7a7e
3
+ size 537
finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch0/rank-specific-00004-of-00008.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8cd6ad8f3d2bcfa25c957717227143e64751970f9b367b28b205a5084a8f476a
3
+ size 537
finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch0/rank-specific-00005-of-00008.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bf049e1944a87da00e6860d1884d0eb312dc5a389a832a4e76a582493ec26972
3
+ size 537
finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch0/rank-specific-00006-of-00008.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8174e84cf8a0553f73baf42bd13d65974b85944a834fa7f75433c0be044e2f04
3
+ size 537
finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch0/rank-specific-00007-of-00008.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fb6f9198ace60febfc0ad5d85588a3d4021799762f521c1a6b87adc99c8889ce
3
+ size 537
finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch1/consolidated.00-of-01.model.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b4dde1230986c4e5b21b516ceed0f0cf400e13c18d318e74a21cd0e791b0de0
3
+ size 297203748
finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch1/consolidated.00-of-01.optimizer.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61f0a81d25eacf47d9d027a432e926afe87659daf833caaf6d7942497739b40e
3
+ size 652420087
finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch1/consolidated.00-of-01.other.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:96368f4fb4ce9d163fe1194a6040a4cdcf0e6cddb3d702548f866177a4ce8d80
3
+ size 1751
finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch1/rank-specific-00000-of-00008.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ec2932635da1a4de71c34aa8fcbcba91dfb0ac1ddc7859f8f87280546b7e786a
3
+ size 537
finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch1/rank-specific-00001-of-00008.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:88973b3c418b507bcde1467ec3902218b83d95fe4e022aca11b09c3f86cde7ac
3
+ size 537
finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch1/rank-specific-00002-of-00008.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eee15a274ea5f27c0360c85bd878d6e0f2072076cae26311c52798f7d836643a
3
+ size 537
finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch1/rank-specific-00003-of-00008.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61651d612914693bf494e5609388a6f9239090c45b3abcc9c4fa5c7a814c7a7e
3
+ size 537
finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch1/rank-specific-00004-of-00008.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8cd6ad8f3d2bcfa25c957717227143e64751970f9b367b28b205a5084a8f476a
3
+ size 537
finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch1/rank-specific-00005-of-00008.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bf049e1944a87da00e6860d1884d0eb312dc5a389a832a4e76a582493ec26972
3
+ size 537
finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch1/rank-specific-00006-of-00008.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8174e84cf8a0553f73baf42bd13d65974b85944a834fa7f75433c0be044e2f04
3
+ size 537
finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch1/rank-specific-00007-of-00008.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fb6f9198ace60febfc0ad5d85588a3d4021799762f521c1a6b87adc99c8889ce
3
+ size 537
finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch2/consolidated.00-of-01.model.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3ff2e36a1e758f232f7ad1986a7524026242d1aab6a62cea8587e04b70019032
3
+ size 297203748
finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch2/consolidated.00-of-01.optimizer.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5b46b67179745b5e1b07f2b8b7601c259152124a2ea43ce07085fd617858c399
3
+ size 652420087
finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch2/consolidated.00-of-01.other.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:307dbfce529fa5fb62614ed0ef32f8a5808cf46d71b5204dd27bb6d5866f4280
3
+ size 1751
finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch2/rank-specific-00000-of-00008.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ec2932635da1a4de71c34aa8fcbcba91dfb0ac1ddc7859f8f87280546b7e786a
3
+ size 537
finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch2/rank-specific-00001-of-00008.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:88973b3c418b507bcde1467ec3902218b83d95fe4e022aca11b09c3f86cde7ac
3
+ size 537
finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch2/rank-specific-00002-of-00008.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eee15a274ea5f27c0360c85bd878d6e0f2072076cae26311c52798f7d836643a
3
+ size 537
finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch2/rank-specific-00003-of-00008.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61651d612914693bf494e5609388a6f9239090c45b3abcc9c4fa5c7a814c7a7e
3
+ size 537
finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch2/rank-specific-00004-of-00008.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8cd6ad8f3d2bcfa25c957717227143e64751970f9b367b28b205a5084a8f476a
3
+ size 537
finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch2/rank-specific-00005-of-00008.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bf049e1944a87da00e6860d1884d0eb312dc5a389a832a4e76a582493ec26972
3
+ size 537
finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch2/rank-specific-00006-of-00008.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8174e84cf8a0553f73baf42bd13d65974b85944a834fa7f75433c0be044e2f04
3
+ size 537
finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch2/rank-specific-00007-of-00008.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fb6f9198ace60febfc0ad5d85588a3d4021799762f521c1a6b87adc99c8889ce
3
+ size 537
finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch3.zip ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:44e41d4fb0f7a593a9f8aea8b04e6df9babf3299c4fd36d98dd8d0b45f243ec2
3
+ size 353563693
finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch3/consolidated.00-of-01.model.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d35ec2de84035df505ee9afdfda40ee31647126e02c67504a29af31edb2f3f1
3
+ size 297203748
finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch3/consolidated.00-of-01.optimizer.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b63a86201d584128bf1d9add0a1c18fb2b8470ec81839c424b98d7009b70a948
3
+ size 652420087
finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch3/consolidated.00-of-01.other.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e7c4e36a1edeada21386a34ca7b8f14e4c60081320427fd35a68427cd71307f1
3
+ size 1751
finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch3/rank-specific-00000-of-00008.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ec2932635da1a4de71c34aa8fcbcba91dfb0ac1ddc7859f8f87280546b7e786a
3
+ size 537
finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch3/rank-specific-00001-of-00008.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:88973b3c418b507bcde1467ec3902218b83d95fe4e022aca11b09c3f86cde7ac
3
+ size 537
finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch3/rank-specific-00002-of-00008.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eee15a274ea5f27c0360c85bd878d6e0f2072076cae26311c52798f7d836643a
3
+ size 537
finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch3/rank-specific-00003-of-00008.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61651d612914693bf494e5609388a6f9239090c45b3abcc9c4fa5c7a814c7a7e
3
+ size 537
finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch3/rank-specific-00004-of-00008.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8cd6ad8f3d2bcfa25c957717227143e64751970f9b367b28b205a5084a8f476a
3
+ size 537
finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch3/rank-specific-00005-of-00008.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bf049e1944a87da00e6860d1884d0eb312dc5a389a832a4e76a582493ec26972
3
+ size 537
finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch3/rank-specific-00006-of-00008.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8174e84cf8a0553f73baf42bd13d65974b85944a834fa7f75433c0be044e2f04
3
+ size 537
finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/epoch3/rank-specific-00007-of-00008.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fb6f9198ace60febfc0ad5d85588a3d4021799762f521c1a6b87adc99c8889ce
3
+ size 537
finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/log.txt ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {"train_lr": 2.49692118226601e-05, "train_closs": 0.8056433233203082, "train_grad_norm": 0.5908145041668357, "epoch": 0, "val_lr": 2.49692118226601e-05, "val_closs": 0.8056433233203082, "val_grad_norm": 0.5908145041668357}
2
+ {"train_lr": 4.611427498140404e-05, "train_closs": 0.7642887281919164, "train_grad_norm": 0.44820075342661053, "epoch": 1, "val_lr": 4.611427498140404e-05, "val_closs": 0.7642887281919164, "val_grad_norm": 0.44820075342661053}
3
+ {"train_lr": 2.7513854679802933e-05, "train_closs": 0.750088836140403, "train_grad_norm": 0.4348251323027564, "epoch": 2, "val_lr": 2.7513854679802933e-05, "val_closs": 0.750088836140403, "val_grad_norm": 0.4348251323027564}
4
+ {"train_lr": 8.899579698398978e-06, "train_closs": 0.7437510005037272, "train_grad_norm": 0.4512289605410815, "epoch": 3, "val_lr": 8.899579698398978e-06, "val_closs": 0.7437510005037272, "val_grad_norm": 0.4512289605410815}
finetune/sg/alpaca_llamaPeft_normBiasLora_QF_512_70B/output.log ADDED
The diff for this file is too large to render. See raw diff