Upload folder using huggingface_hub
Browse files- checkpoint-124.pth +3 -0
- events.out.tfevents.1765750679.phe108c-yuewang-02.1031892.0 +3 -0
- events.out.tfevents.1765778016.phe108c-yuewang-02.1063886.0 +3 -0
- events.out.tfevents.1765784424.phe108c-yuewang-02.1092414.0 +3 -0
- events.out.tfevents.1765787935.phe108c-yuewang-02.1097973.0 +3 -0
- events.out.tfevents.1765805852.phe108c-yuewang-02.1108454.0 +3 -0
- events.out.tfevents.1765819581.phe108c-yuewang-02.1114159.0 +3 -0
- events.out.tfevents.1765916962.phe108c-yuewang-02.2912013.0 +3 -0
- events.out.tfevents.1765962508.phe108c-yuewang-02.250959.0 +3 -0
- events.out.tfevents.1765964368.phe108c-yuewang-02.597867.0 +3 -0
- events.out.tfevents.1765967041.phe108c-yuewang-02.523103.0 +3 -0
- events.out.tfevents.1765972816.phe108c-yuewang-02.4092656.0 +3 -0
- events.out.tfevents.1765974399.phe108c-yuewang-02.787186.0 +3 -0
- events.out.tfevents.1766040655.phe108c-yuewang-02.3974176.0 +3 -0
- events.out.tfevents.1766086752.phe108c-yuewang-02.2102202.0 +3 -0
- events.out.tfevents.1766101210.phe108c-yuewang-02.2357724.0 +3 -0
- events.out.tfevents.1766127019.phe108c-yuewang-02.4152936.0 +3 -0
- events.out.tfevents.1766134288.phe108c-yuewang-02.391862.0 +3 -0
- log.txt +125 -0
- run.yaml +88 -0
- train_split.json +0 -0
- val_split.json +1 -0
checkpoint-124.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:adf3b84b5e5d380e413917b9aa8f15750c08e2e985f6da073ce9bfbf2fa9b7b1
|
| 3 |
+
size 1099294898
|
events.out.tfevents.1765750679.phe108c-yuewang-02.1031892.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:84beb71e8bef80980bc0c1e8c400f39caa0184e97fe2aca72dbeae2ce5ff7e8b
|
| 3 |
+
size 148332
|
events.out.tfevents.1765778016.phe108c-yuewang-02.1063886.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b2572c827b9935a7fd7e96d9334b7eed483e7a4bc879929eaeef8d39f5abe0ae
|
| 3 |
+
size 13678
|
events.out.tfevents.1765784424.phe108c-yuewang-02.1092414.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3462774d2f6f4ce716483f7f9737e287c7d4baa8adcb5c9e30371d301a84f8bf
|
| 3 |
+
size 3410
|
events.out.tfevents.1765787935.phe108c-yuewang-02.1097973.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f74228e4fb9f7747d7d9f84f0b1b55f1f36b66f79da0b45e2f028e4c0da93cef
|
| 3 |
+
size 69293
|
events.out.tfevents.1765805852.phe108c-yuewang-02.1108454.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d9df9576db70092d2042445c73bc9490fc39b6ca3840f95ccdc6f94b01da55b1
|
| 3 |
+
size 70011
|
events.out.tfevents.1765819581.phe108c-yuewang-02.1114159.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9ced62c2fb56d4df0839f7c1de6872eae206319dd15bcbd958655a7ebe821e4a
|
| 3 |
+
size 544808
|
events.out.tfevents.1765916962.phe108c-yuewang-02.2912013.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ef1e3470d64a0af0d0df6199d4f9cbe27f120ad3a1766995cb793ac9e0d65c5b
|
| 3 |
+
size 171276
|
events.out.tfevents.1765962508.phe108c-yuewang-02.250959.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:108c1e787121fd397844e0f7d2aac334e77ae4fb9ea2451c5862b62bec942e72
|
| 3 |
+
size 396
|
events.out.tfevents.1765964368.phe108c-yuewang-02.597867.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:273d0fe64002681657667384ed12c0a7bb4f5f345a8ff2a6c13593c3a887896d
|
| 3 |
+
size 88
|
events.out.tfevents.1765967041.phe108c-yuewang-02.523103.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c12f64a74bfed59302246650aefe114922bbd95a9b76608ce8d746ecdf0e1b62
|
| 3 |
+
size 10868
|
events.out.tfevents.1765972816.phe108c-yuewang-02.4092656.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6857e8f94514d4d2c180e7c60735095ae886af52ceac2c0d8e715f7b69384621
|
| 3 |
+
size 2860
|
events.out.tfevents.1765974399.phe108c-yuewang-02.787186.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ec1d422f1cef5b6391508a4cf2a88a1ef6a92b8119af2b2c4422a19315393728
|
| 3 |
+
size 259512
|
events.out.tfevents.1766040655.phe108c-yuewang-02.3974176.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5ce8b8396714ed229f88bec2db758c5a13ab3f4478387fd85ac0a9bb438e516d
|
| 3 |
+
size 160188
|
events.out.tfevents.1766086752.phe108c-yuewang-02.2102202.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0e936a9ae34794a5cffd44e3a8a44aa9cdd8b3b093f1c95beebad07b6fad8265
|
| 3 |
+
size 63276
|
events.out.tfevents.1766101210.phe108c-yuewang-02.2357724.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4d283e2df4bf584b7429d2758bb3c809aed27ea7135de4a7ce46203bcc562029
|
| 3 |
+
size 127132
|
events.out.tfevents.1766127019.phe108c-yuewang-02.4152936.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d99d3a853c1078d31e3bd90a94b716024cfaad037db27bad4c154fa296d35150
|
| 3 |
+
size 13948
|
events.out.tfevents.1766134288.phe108c-yuewang-02.391862.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:001473cda22d08b783f9e73571a0daee007310e573793f94dc7711c99434cc0f
|
| 3 |
+
size 226456
|
log.txt
ADDED
|
@@ -0,0 +1,125 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{"train_lr": 0.00019564004345400005, "train_loss": 0.3029090781689445, "epoch": 0, "val_lr": 0.00010000000000000006, "val_loss": 0.20378573006018996}
|
| 2 |
+
{"train_lr": 0.00048637343608965073, "train_loss": 0.1901836911292664, "epoch": 1, "val_lr": 0.0004487973013898667, "val_loss": 0.18067710906267165}
|
| 3 |
+
{"train_lr": 0.0004998696082222175, "train_loss": 0.17816189841009103, "epoch": 2, "val_lr": 0.0004999143342294313, "val_loss": 0.17080191090175018}
|
| 4 |
+
{"train_lr": 0.0004995895072293546, "train_loss": 0.15774387331829595, "epoch": 3, "val_lr": 0.0004996726615235633, "val_loss": 0.12231620099572908}
|
| 5 |
+
{"train_lr": 0.0004991485594922281, "train_loss": 0.11175334399146974, "epoch": 4, "val_lr": 0.0004992700923113186, "val_loss": 0.10168942175204715}
|
| 6 |
+
{"train_lr": 0.000498547062354982, "train_loss": 0.0965861452922021, "epoch": 5, "val_lr": 0.0004987070201088922, "val_loss": 0.08977454705607324}
|
| 7 |
+
{"train_lr": 0.0004977853648358597, "train_loss": 0.08411245128968473, "epoch": 6, "val_lr": 0.0004979836008699763, "val_loss": 0.07585888654584923}
|
| 8 |
+
{"train_lr": 0.0004968640124725848, "train_loss": 0.07101845985501036, "epoch": 7, "val_lr": 0.0004971005611665038, "val_loss": 0.06519926275264044}
|
| 9 |
+
{"train_lr": 0.0004957835503316204, "train_loss": 0.064371204305105, "epoch": 8, "val_lr": 0.0004960579065175926, "val_loss": 0.061161860093474386}
|
| 10 |
+
{"train_lr": 0.00049454470905488, "train_loss": 0.060200666000887956, "epoch": 9, "val_lr": 0.0004948468761150538, "val_loss": 0.05733869946919955}
|
| 11 |
+
{"train_lr": 0.0004931482322768172, "train_loss": 0.05695078827936596, "epoch": 10, "val_lr": 0.0004934981331595625, "val_loss": 0.05281244964158583}
|
| 12 |
+
{"train_lr": 0.0004915950874001617, "train_loss": 0.05440806119826721, "epoch": 11, "val_lr": 0.0004919824024869898, "val_loss": 0.05236734223792389}
|
| 13 |
+
{"train_lr": 0.0004898863189075323, "train_loss": 0.05319481884278281, "epoch": 12, "val_lr": 0.0004903105349971374, "val_loss": 0.05266624510904232}
|
| 14 |
+
{"train_lr": 0.0004880228557725985, "train_loss": 0.05180120371913614, "epoch": 13, "val_lr": 0.0004884843962028363, "val_loss": 0.05038926356691655}
|
| 15 |
+
{"train_lr": 0.00048600605160856875, "train_loss": 0.050377890708376415, "epoch": 14, "val_lr": 0.000486503724771813, "val_loss": 0.04809159743308555}
|
| 16 |
+
{"train_lr": 0.00048383709037651707, "train_loss": 0.049341674697004324, "epoch": 15, "val_lr": 0.0004843728643456271, "val_loss": 0.04898137157627692}
|
| 17 |
+
{"train_lr": 0.00048151750704065376, "train_loss": 0.04828970470076605, "epoch": 16, "val_lr": 0.0004820872719566832, "val_loss": 0.04582030848988632}
|
| 18 |
+
{"train_lr": 0.00047904868262758124, "train_loss": 0.04746105371095358, "epoch": 17, "val_lr": 0.0004796561867071299, "val_loss": 0.046291777344110115}
|
| 19 |
+
{"train_lr": 0.0004764322268321257, "train_loss": 0.046354372404379604, "epoch": 18, "val_lr": 0.00047707271926729345, "val_loss": 0.04550358132837573}
|
| 20 |
+
{"train_lr": 0.00047366992437445206, "train_loss": 0.04631630348176203, "epoch": 19, "val_lr": 0.00047434512192096024, "val_loss": 0.044303551486264615}
|
| 21 |
+
{"train_lr": 0.00047076356999288126, "train_loss": 0.044630916993501385, "epoch": 20, "val_lr": 0.0004714737347557191, "val_loss": 0.04328328773106744}
|
| 22 |
+
{"train_lr": 0.000467714621679718, "train_loss": 0.043510656242827815, "epoch": 21, "val_lr": 0.00046844592492022695, "val_loss": 0.043618555585023505}
|
| 23 |
+
{"train_lr": 0.0004645256362715261, "train_loss": 0.04274718968581208, "epoch": 22, "val_lr": 0.00046530222453661365, "val_loss": 0.04072705302678514}
|
| 24 |
+
{"train_lr": 0.00046119826652148675, "train_loss": 0.04158026777635636, "epoch": 23, "val_lr": 0.0004620080338788992, "val_loss": 0.039345405019819736}
|
| 25 |
+
{"train_lr": 0.00045773486228217617, "train_loss": 0.04112308661733077, "epoch": 24, "val_lr": 0.0004585791000421254, "val_loss": 0.04036848139294908}
|
| 26 |
+
{"train_lr": 0.00045413766264156073, "train_loss": 0.04041471443990466, "epoch": 25, "val_lr": 0.00045501146756788053, "val_loss": 0.03908891554176808}
|
| 27 |
+
{"train_lr": 0.00045040871313989535, "train_loss": 0.040147826829208776, "epoch": 26, "val_lr": 0.0004513138575405235, "val_loss": 0.03813537592440844}
|
| 28 |
+
{"train_lr": 0.00044655071230848936, "train_loss": 0.039322634453001694, "epoch": 27, "val_lr": 0.0004474868983232679, "val_loss": 0.0382390329313855}
|
| 29 |
+
{"train_lr": 0.0004425660737458544, "train_loss": 0.03864359832081936, "epoch": 28, "val_lr": 0.00044353150487087085, "val_loss": 0.03713711851014523}
|
| 30 |
+
{"train_lr": 0.000438457395267602, "train_loss": 0.03818666758431093, "epoch": 29, "val_lr": 0.0004394525610181687, "val_loss": 0.03699907187372446}
|
| 31 |
+
{"train_lr": 0.0004342270013093638, "train_loss": 0.03759662008126303, "epoch": 30, "val_lr": 0.00043525215678922604, "val_loss": 0.03622726344208165}
|
| 32 |
+
{"train_lr": 0.0004298782177121366, "train_loss": 0.03709144374584913, "epoch": 31, "val_lr": 0.0004309313797143709, "val_loss": 0.0362949859260059}
|
| 33 |
+
{"train_lr": 0.00042541330648248807, "train_loss": 0.03704276450347565, "epoch": 32, "val_lr": 0.0004264929103607731, "val_loss": 0.03630306718498468}
|
| 34 |
+
{"train_lr": 0.00042083532955537885, "train_loss": 0.03612849534275468, "epoch": 33, "val_lr": 0.00042194127825292196, "val_loss": 0.03573022988866868}
|
| 35 |
+
{"train_lr": 0.0004161473908507008, "train_loss": 0.03549579941874152, "epoch": 34, "val_lr": 0.0004172802929047031, "val_loss": 0.0337221956165928}
|
| 36 |
+
{"train_lr": 0.0004113523070013329, "train_loss": 0.035088212288574554, "epoch": 35, "val_lr": 0.00041249087910818116, "val_loss": 0.03479692658279524}
|
| 37 |
+
{"train_lr": 0.0004064533191198741, "train_loss": 0.034415727810464376, "epoch": 36, "val_lr": 0.0004076367699445188, "val_loss": 0.03286332018461412}
|
| 38 |
+
{"train_lr": 0.00040145348832920267, "train_loss": 0.03416300144759015, "epoch": 37, "val_lr": 0.00040266163356827945, "val_loss": 0.03268003067551333}
|
| 39 |
+
{"train_lr": 0.0003963559138950938, "train_loss": 0.03316745480847089, "epoch": 38, "val_lr": 0.0003975862941054119, "val_loss": 0.03240224858367346}
|
| 40 |
+
{"train_lr": 0.00039116424326958324, "train_loss": 0.032830451658156075, "epoch": 39, "val_lr": 0.00039241506916727956, "val_loss": 0.03241762261660326}
|
| 41 |
+
{"train_lr": 0.0003858812365114264, "train_loss": 0.03221809167793154, "epoch": 40, "val_lr": 0.0003871535046846055, "val_loss": 0.030558354837277275}
|
| 42 |
+
{"train_lr": 0.0003805109682349588, "train_loss": 0.0316272711980638, "epoch": 41, "val_lr": 0.0003818037542649857, "val_loss": 0.0298328703108031}
|
| 43 |
+
{"train_lr": 0.00037505651830543516, "train_loss": 0.030875583239354694, "epoch": 42, "val_lr": 0.00037636906354619396, "val_loss": 0.02985021250146463}
|
| 44 |
+
{"train_lr": 0.0003695215196280058, "train_loss": 0.030450353346133528, "epoch": 43, "val_lr": 0.0003708529348886704, "val_loss": 0.02946670093815627}
|
| 45 |
+
{"train_lr": 0.00036390953986632644, "train_loss": 0.029792958058218758, "epoch": 44, "val_lr": 0.00036526101124799925, "val_loss": 0.028769544291696172}
|
| 46 |
+
{"train_lr": 0.00035822386312613303, "train_loss": 0.029045628254025534, "epoch": 45, "val_lr": 0.0003595919218602794, "val_loss": 0.028720535781054247}
|
| 47 |
+
{"train_lr": 0.00035246864244515197, "train_loss": 0.02842943575366767, "epoch": 46, "val_lr": 0.00035385171834479594, "val_loss": 0.02763190899249606}
|
| 48 |
+
{"train_lr": 0.00034664744493014834, "train_loss": 0.027837987167432028, "epoch": 47, "val_lr": 0.00034804587662652675, "val_loss": 0.026703940858463133}
|
| 49 |
+
{"train_lr": 0.0003407639194330738, "train_loss": 0.02754394838467286, "epoch": 48, "val_lr": 0.00034218017350477737, "val_loss": 0.0257092342276859}
|
| 50 |
+
{"train_lr": 0.00033482187370514893, "train_loss": 0.02691245303462513, "epoch": 49, "val_lr": 0.00033624979313522395, "val_loss": 0.025880782399326563}
|
| 51 |
+
{"train_lr": 0.0003288254193979696, "train_loss": 0.026448398734446408, "epoch": 50, "val_lr": 0.00033026512825549525, "val_loss": 0.025931137435138224}
|
| 52 |
+
{"train_lr": 0.00032277780083401965, "train_loss": 0.025860479277777966, "epoch": 51, "val_lr": 0.0003242294003332557, "val_loss": 0.024530567467212678}
|
| 53 |
+
{"train_lr": 0.0003166832772149368, "train_loss": 0.025295751232208095, "epoch": 52, "val_lr": 0.0003181465224205295, "val_loss": 0.02443710269947206}
|
| 54 |
+
{"train_lr": 0.00031054590880512834, "train_loss": 0.024831028110069087, "epoch": 53, "val_lr": 0.0003120199681636746, "val_loss": 0.024005142320704654}
|
| 55 |
+
{"train_lr": 0.00030436953575710196, "train_loss": 0.024337971479333836, "epoch": 54, "val_lr": 0.0003058537002949272, "val_loss": 0.023419170529719015}
|
| 56 |
+
{"train_lr": 0.00029815805927156566, "train_loss": 0.02390463634179727, "epoch": 55, "val_lr": 0.00029959978892800203, "val_loss": 0.02288346139427561}
|
| 57 |
+
{"train_lr": 0.0002919157893170792, "train_loss": 0.023330022025673264, "epoch": 56, "val_lr": 0.0002934112533794024, "val_loss": 0.0228662821684238}
|
| 58 |
+
{"train_lr": 0.000285646125863976, "train_loss": 0.02302058585830379, "epoch": 57, "val_lr": 0.00028714861572367255, "val_loss": 0.021559586439281703}
|
| 59 |
+
{"train_lr": 0.00027935389395657595, "train_loss": 0.02267140933978619, "epoch": 58, "val_lr": 0.0002808629371486902, "val_loss": 0.021895042590312358}
|
| 60 |
+
{"train_lr": 0.00027304247236984726, "train_loss": 0.022264836005202026, "epoch": 59, "val_lr": 0.0002745547464698992, "val_loss": 0.021240073793207206}
|
| 61 |
+
{"train_lr": 0.0002667164136507622, "train_loss": 0.02188992994808349, "epoch": 60, "val_lr": 0.00026823044849693084, "val_loss": 0.020826252639323237}
|
| 62 |
+
{"train_lr": 0.0002603792240263071, "train_loss": 0.021416045910589848, "epoch": 61, "val_lr": 0.0002618978397730599, "val_loss": 0.020783679086593836}
|
| 63 |
+
{"train_lr": 0.0002540355193130702, "train_loss": 0.02100502312555527, "epoch": 62, "val_lr": 0.00025555314829048964, "val_loss": 0.020066257754695557}
|
| 64 |
+
{"train_lr": 0.00024768923113325603, "train_loss": 0.020786816299805052, "epoch": 63, "val_lr": 0.00024920760104002606, "val_loss": 0.01962972852215171}
|
| 65 |
+
{"train_lr": 0.00024134440660235424, "train_loss": 0.020371674011462393, "epoch": 64, "val_lr": 0.00024286279144732108, "val_loss": 0.019635748700238764}
|
| 66 |
+
{"train_lr": 0.0002350054246168891, "train_loss": 0.01984649458765861, "epoch": 65, "val_lr": 0.00023652184380847327, "val_loss": 0.018980687991115113}
|
| 67 |
+
{"train_lr": 0.00022867565801970824, "train_loss": 0.019436031269265304, "epoch": 66, "val_lr": 0.00023018832132361074, "val_loss": 0.01885488977847946}
|
| 68 |
+
{"train_lr": 0.00022236003826934538, "train_loss": 0.019223793374951954, "epoch": 67, "val_lr": 0.00022384425604489309, "val_loss": 0.018089972862381804}
|
| 69 |
+
{"train_lr": 0.00021606188272490744, "train_loss": 0.01876881279790378, "epoch": 68, "val_lr": 0.00021756575153400524, "val_loss": 0.018229862333100755}
|
| 70 |
+
{"train_lr": 0.0002097857536270519, "train_loss": 0.01860100810686817, "epoch": 69, "val_lr": 0.00021128399468234633, "val_loss": 0.01770009936808492}
|
| 71 |
+
{"train_lr": 0.00020353563942233506, "train_loss": 0.01837872951752221, "epoch": 70, "val_lr": 0.00020502761050491806, "val_loss": 0.01706261415818026}
|
| 72 |
+
{"train_lr": 0.0001973154945066465, "train_loss": 0.017986078369934398, "epoch": 71, "val_lr": 0.00019880030871354622, "val_loss": 0.017128126863390207}
|
| 73 |
+
{"train_lr": 0.0001911290314158904, "train_loss": 0.017612919353283737, "epoch": 72, "val_lr": 0.0001926054816676948, "val_loss": 0.01704678402841091}
|
| 74 |
+
{"train_lr": 0.00018498069568769108, "train_loss": 0.017327215939184584, "epoch": 73, "val_lr": 0.00018644681189779387, "val_loss": 0.01635143433616856}
|
| 75 |
+
{"train_lr": 0.00017887421629145811, "train_loss": 0.017028148669741214, "epoch": 74, "val_lr": 0.00018033149006968616, "val_loss": 0.016331549726366516}
|
| 76 |
+
{"train_lr": 0.00017281357449716822, "train_loss": 0.016653849102066836, "epoch": 75, "val_lr": 0.00017425877343425064, "val_loss": 0.015566111665219068}
|
| 77 |
+
{"train_lr": 0.0001668027037119927, "train_loss": 0.01647036786889259, "epoch": 76, "val_lr": 0.00016818973307559858, "val_loss": 0.014950288312796217}
|
| 78 |
+
{"train_lr": 0.00016084546491516838, "train_loss": 0.01609618966705212, "epoch": 77, "val_lr": 0.0001622641764016449, "val_loss": 0.015607096098392503}
|
| 79 |
+
{"train_lr": 0.00015494559384151709, "train_loss": 0.015878241306616584, "epoch": 78, "val_lr": 0.00015635066862102265, "val_loss": 0.01478586053579218}
|
| 80 |
+
{"train_lr": 0.0001491071436286576, "train_loss": 0.015681709726296386, "epoch": 79, "val_lr": 0.0001504992200667879, "val_loss": 0.015376914625729972}
|
| 81 |
+
{"train_lr": 0.00014333368103339436, "train_loss": 0.01536582656565532, "epoch": 80, "val_lr": 0.00014470812040117537, "val_loss": 0.014833954703062773}
|
| 82 |
+
{"train_lr": 0.00013762877671113355, "train_loss": 0.015042718994185609, "epoch": 81, "val_lr": 0.00013898601217372442, "val_loss": 0.014250776116987543}
|
| 83 |
+
{"train_lr": 0.0001319963676977656, "train_loss": 0.014874645587631222, "epoch": 82, "val_lr": 0.00013333594729889647, "val_loss": 0.014122673421211186}
|
| 84 |
+
{"train_lr": 0.00012644000894844172, "train_loss": 0.01462754887003495, "epoch": 83, "val_lr": 0.0001277610660149649, "val_loss": 0.014263364846729451}
|
| 85 |
+
{"train_lr": 0.00012096340874629102, "train_loss": 0.014381824359157271, "epoch": 84, "val_lr": 0.00012226808773231843, "val_loss": 0.013550933956199128}
|
| 86 |
+
{"train_lr": 0.0001155698280035565, "train_loss": 0.014335416555139847, "epoch": 85, "val_lr": 0.00011685318803230731, "val_loss": 0.013834573171381665}
|
| 87 |
+
{"train_lr": 0.00011026296158263907, "train_loss": 0.014103251056881707, "epoch": 86, "val_lr": 0.00011152624273383104, "val_loss": 0.013314086787372094}
|
| 88 |
+
{"train_lr": 0.00010504600853888364, "train_loss": 0.013863970035279755, "epoch": 87, "val_lr": 0.00010628492357509662, "val_loss": 0.013290417570620775}
|
| 89 |
+
{"train_lr": 9.992253733922532e-05, "train_loss": 0.013815163830919484, "epoch": 88, "val_lr": 0.00010113833361545515, "val_loss": 0.012893795895035424}
|
| 90 |
+
{"train_lr": 9.489576647573349e-05, "train_loss": 0.013514626766281467, "epoch": 89, "val_lr": 9.608814468650322e-05, "val_loss": 0.013136028532042272}
|
| 91 |
+
{"train_lr": 8.996904642878931e-05, "train_loss": 0.0134097098326199, "epoch": 90, "val_lr": 9.111810643178455e-05, "val_loss": 0.01242462544189405}
|
| 92 |
+
{"train_lr": 8.514539285366458e-05, "train_loss": 0.013135068401647795, "epoch": 91, "val_lr": 8.628824012057681e-05, "val_loss": 0.012393929177939187}
|
| 93 |
+
{"train_lr": 8.04279432727628e-05, "train_loss": 0.013013286120520594, "epoch": 92, "val_lr": 8.154721108420788e-05, "val_loss": 0.012394825573934893}
|
| 94 |
+
{"train_lr": 7.581986828398433e-05, "train_loss": 0.012972520542583937, "epoch": 93, "val_lr": 7.69105678142417e-05, "val_loss": 0.012417257270284407}
|
| 95 |
+
{"train_lr": 7.132404085957154e-05, "train_loss": 0.01284948122561654, "epoch": 94, "val_lr": 7.238882556203876e-05, "val_loss": 0.012092059523227715}
|
| 96 |
+
{"train_lr": 6.694325050301429e-05, "train_loss": 0.01272580515895986, "epoch": 95, "val_lr": 6.798187606991025e-05, "val_loss": 0.012048265095189458}
|
| 97 |
+
{"train_lr": 6.26805137222791e-05, "train_loss": 0.01264511175966782, "epoch": 96, "val_lr": 6.367159987954019e-05, "val_loss": 0.011939053134920522}
|
| 98 |
+
{"train_lr": 5.853855307613423e-05, "train_loss": 0.012489690558386555, "epoch": 97, "val_lr": 5.9517780084338156e-05, "val_loss": 0.011728202640604708}
|
| 99 |
+
{"train_lr": 5.452006353841531e-05, "train_loss": 0.012315334338378815, "epoch": 98, "val_lr": 5.546804239881225e-05, "val_loss": 0.011587674211114063}
|
| 100 |
+
{"train_lr": 5.0627196885295775e-05, "train_loss": 0.012222715360107703, "epoch": 99, "val_lr": 5.1545734750216244e-05, "val_loss": 0.011885711507871747}
|
| 101 |
+
{"train_lr": 4.6862939276060094e-05, "train_loss": 0.0120737815875928, "epoch": 100, "val_lr": 4.775105820496612e-05, "val_loss": 0.01174918009576598}
|
| 102 |
+
{"train_lr": 4.3229659440581875e-05, "train_loss": 0.012053680395158752, "epoch": 101, "val_lr": 4.408571933548691e-05, "val_loss": 0.011422525994479656}
|
| 103 |
+
{"train_lr": 3.9729663788073124e-05, "train_loss": 0.011984574670019234, "epoch": 102, "val_lr": 4.05546541795973e-05, "val_loss": 0.011494848364206955}
|
| 104 |
+
{"train_lr": 3.636514259023374e-05, "train_loss": 0.011839345053557375, "epoch": 103, "val_lr": 3.715585419200682e-05, "val_loss": 0.011368920762590535}
|
| 105 |
+
{"train_lr": 3.3138255278447185e-05, "train_loss": 0.011952498970095888, "epoch": 104, "val_lr": 3.389811365138092e-05, "val_loss": 0.011258657481597706}
|
| 106 |
+
{"train_lr": 3.005119280906492e-05, "train_loss": 0.011793950572450647, "epoch": 105, "val_lr": 3.077516253091211e-05, "val_loss": 0.011036837981339163}
|
| 107 |
+
{"train_lr": 2.7105964216265776e-05, "train_loss": 0.01165330332612068, "epoch": 106, "val_lr": 2.7796658418673295e-05, "val_loss": 0.011326960379993844}
|
| 108 |
+
{"train_lr": 2.4304168241662768e-05, "train_loss": 0.011600216138117635, "epoch": 107, "val_lr": 2.496039061000059e-05, "val_loss": 0.011227262035507949}
|
| 109 |
+
{"train_lr": 2.1647883702788352e-05, "train_loss": 0.011528690113152568, "epoch": 108, "val_lr": 2.2268171500970164e-05, "val_loss": 0.011150405061925994}
|
| 110 |
+
{"train_lr": 1.913884253928374e-05, "train_loss": 0.011567669638481713, "epoch": 109, "val_lr": 1.9723835954023598e-05, "val_loss": 0.010844607260004859}
|
| 111 |
+
{"train_lr": 1.67786382543702e-05, "train_loss": 0.01138995437179139, "epoch": 110, "val_lr": 1.7328584238632968e-05, "val_loss": 0.010685282611032767}
|
| 112 |
+
{"train_lr": 1.456843174835642e-05, "train_loss": 0.011354313507038319, "epoch": 111, "val_lr": 1.5073475525714953e-05, "val_loss": 0.011157728767730007}
|
| 113 |
+
{"train_lr": 1.251018386556775e-05, "train_loss": 0.011371533288274711, "epoch": 112, "val_lr": 1.2987145024074648e-05, "val_loss": 0.010850955787602635}
|
| 114 |
+
{"train_lr": 1.0604939279716262e-05, "train_loss": 0.011391660562431898, "epoch": 113, "val_lr": 1.1044898371185302e-05, "val_loss": 0.010716146725826547}
|
| 115 |
+
{"train_lr": 8.853953268125544e-06, "train_loss": 0.01128660032884765, "epoch": 114, "val_lr": 9.257403022359512e-06, "val_loss": 0.01041673325933516}
|
| 116 |
+
{"train_lr": 7.25839576126639e-06, "train_loss": 0.01125125261874781, "epoch": 115, "val_lr": 7.62434821019335e-06, "val_loss": 0.0105993557566156}
|
| 117 |
+
{"train_lr": 5.819260897595943e-06, "train_loss": 0.011226332090388182, "epoch": 116, "val_lr": 6.147901616568134e-06, "val_loss": 0.010801995811983944}
|
| 118 |
+
{"train_lr": 4.537500139906067e-06, "train_loss": 0.011306854250873922, "epoch": 117, "val_lr": 4.82821837662118e-06, "val_loss": 0.010578870405590841}
|
| 119 |
+
{"train_lr": 3.413925883331067e-06, "train_loss": 0.011253086827925814, "epoch": 118, "val_lr": 3.667352208154465e-06, "val_loss": 0.010513919106555907}
|
| 120 |
+
{"train_lr": 2.449249230041731e-06, "train_loss": 0.011142525816522475, "epoch": 119, "val_lr": 2.6651165730630095e-06, "val_loss": 0.010430092779026666}
|
| 121 |
+
{"train_lr": 1.6441586857713777e-06, "train_loss": 0.011191111990599493, "epoch": 120, "val_lr": 1.820522041858372e-06, "val_loss": 0.01031723869304488}
|
| 122 |
+
{"train_lr": 9.990316580090408e-07, "train_loss": 0.01121511724456812, "epoch": 121, "val_lr": 1.1373589087718278e-06, "val_loss": 0.010718939803540706}
|
| 123 |
+
{"train_lr": 5.144216248127181e-07, "train_loss": 0.011190643205175104, "epoch": 122, "val_lr": 6.143148911205126e-07, "val_loss": 0.010613977236673237}
|
| 124 |
+
{"train_lr": 1.9059472917665377e-07, "train_loss": 0.01114721703897224, "epoch": 123, "val_lr": 2.5178795675100504e-07, "val_loss": 0.010474107173220145}
|
| 125 |
+
{"train_lr": 2.773946210712383e-08, "train_loss": 0.01110242478758527, "epoch": 124, "val_lr": 5.051507677919919e-08, "val_loss": 0.01055344306708624}
|
run.yaml
ADDED
|
@@ -0,0 +1,88 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
!!python/object:icrt.util.args.ExperimentConfig
|
| 2 |
+
dataset_cfg: !!python/object:icrt.util.args.DatasetConfig
|
| 3 |
+
action_noise: 0.0
|
| 4 |
+
dataset_fraction: 1.0
|
| 5 |
+
dataset_json: ./config/dataset_config_libero_90_ver3.json
|
| 6 |
+
goal_conditioned: false
|
| 7 |
+
non_overlapping: 64
|
| 8 |
+
num_repeat_traj: 1
|
| 9 |
+
num_weighted_steps: 30
|
| 10 |
+
proprio_noise: 0.005
|
| 11 |
+
rebalance_tasks: true
|
| 12 |
+
shuffle_repeat_traj: true
|
| 13 |
+
skip_step: false
|
| 14 |
+
sort_by_lang: true
|
| 15 |
+
task_barrier: true
|
| 16 |
+
task_names: null
|
| 17 |
+
vision_aug: true
|
| 18 |
+
visual_trace_noise: 0.005
|
| 19 |
+
device: cuda
|
| 20 |
+
dist_on_itp: false
|
| 21 |
+
dist_url: env://
|
| 22 |
+
load_config: null
|
| 23 |
+
local_rank: -1
|
| 24 |
+
logging_cfg: !!python/object:icrt.util.args.LoggingConfig
|
| 25 |
+
log_dir: /data/tientoan/icrt_output/output_libero_90/no_prompt_loss_0_ver3
|
| 26 |
+
log_name: no_prompt_loss_0_ver3
|
| 27 |
+
output_dir: /data/tientoan/icrt_output/output_libero_90/no_prompt_loss_0_ver3
|
| 28 |
+
model_cfg: !!python/object:icrt.util.args.ModelConfig
|
| 29 |
+
policy_cfg: !!python/object:icrt.util.args.PolicyConfig
|
| 30 |
+
adapter_mlp_ratio: 4.0
|
| 31 |
+
adapter_num_heads: 8
|
| 32 |
+
camera_pos_emb: false
|
| 33 |
+
decoder_pred_head: mlp
|
| 34 |
+
kl_div_loss: false
|
| 35 |
+
llama_ckpt_dir: /home/mfu/checkpoints/llama-2/llama-2-7b
|
| 36 |
+
load_llama: true
|
| 37 |
+
lora_layer_idxs: null
|
| 38 |
+
lora_rank: 4
|
| 39 |
+
loss_w_action: 1.0
|
| 40 |
+
modality_pos_emb: false
|
| 41 |
+
multikv_attn_pool: false
|
| 42 |
+
no_prompt_loss: true
|
| 43 |
+
num_inference_diffusion_steps: null
|
| 44 |
+
num_train_diffusion_steps: null
|
| 45 |
+
phase: pretrain
|
| 46 |
+
pred_action_only: true
|
| 47 |
+
pretrained_path: null
|
| 48 |
+
remove_proprio: false
|
| 49 |
+
scale_loss: 1.0
|
| 50 |
+
scratch_llama_config: config/model_config/custom_transformer.json
|
| 51 |
+
separate_camera_adapter: true
|
| 52 |
+
step_weight: 1.0
|
| 53 |
+
vision_encoder_cfg: !!python/object:icrt.util.args.VisionEncoderConfig
|
| 54 |
+
vision_encoder: ./checkpoints/crossmae_rtx/cross-mae-rtx-vitb.pth
|
| 55 |
+
vision_lora: false
|
| 56 |
+
vision_lora_rank: 8
|
| 57 |
+
vision_nonpretrained: false
|
| 58 |
+
vision_unfreeze_all: false
|
| 59 |
+
vision_unfreeze_last_n: 0
|
| 60 |
+
optimizer_cfg: !!python/object:icrt.util.args.OptimizerConfig
|
| 61 |
+
blr: 0.001
|
| 62 |
+
lr: 0.0005
|
| 63 |
+
min_lr: 0.0
|
| 64 |
+
warmup_epochs: 1.25
|
| 65 |
+
weight_decay: 0.01
|
| 66 |
+
shared_cfg: !!python/object:icrt.util.args.SharedConfig
|
| 67 |
+
batch_size: 1
|
| 68 |
+
num_cameras: 2
|
| 69 |
+
num_pred_steps: 16
|
| 70 |
+
num_stages: 1
|
| 71 |
+
num_visual_trace_points: 5
|
| 72 |
+
random_mask_visual_trace: true
|
| 73 |
+
resume: /data/tientoan/icrt_output/output_libero_90/no_prompt_loss_0_ver3/checkpoint-124.pth
|
| 74 |
+
rot_6d: true
|
| 75 |
+
save_every: 1
|
| 76 |
+
scale_action: null
|
| 77 |
+
seed: 0
|
| 78 |
+
seq_length: 512
|
| 79 |
+
split_epoch: 1
|
| 80 |
+
start_epoch: 0
|
| 81 |
+
use_delta_action: true
|
| 82 |
+
train: true
|
| 83 |
+
trainer_cfg: !!python/object:icrt.util.args.TrainerConfig
|
| 84 |
+
accum_iter: 64
|
| 85 |
+
epochs: 125
|
| 86 |
+
num_workers: 20
|
| 87 |
+
pin_memory: true
|
| 88 |
+
world_size: 1
|
train_split.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
val_split.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
["STUDY_SCENE1_pick_up_the_book_and_place_it_in_the_left_compartment_of_the_caddy_21", "KITCHEN_SCENE10_put_the_black_bowl_in_the_top_drawer_of_the_cabinet_10", "STUDY_SCENE3_pick_up_the_book_and_place_it_in_the_front_compartment_of_the_caddy_6", "KITCHEN_SCENE2_stack_the_black_bowl_at_the_front_on_the_black_bowl_in_the_middle_34", "KITCHEN_SCENE2_put_the_black_bowl_at_the_front_on_the_plate_44", "KITCHEN_SCENE5_close_the_top_drawer_of_the_cabinet_6", "LIVING_ROOM_SCENE2_pick_up_the_milk_and_put_it_in_the_basket_6", "STUDY_SCENE3_pick_up_the_book_and_place_it_in_the_right_compartment_of_the_caddy_49", "LIVING_ROOM_SCENE3_pick_up_the_cream_cheese_and_put_it_in_the_tray_2", "LIVING_ROOM_SCENE3_pick_up_the_butter_and_put_it_in_the_tray_25", "KITCHEN_SCENE4_put_the_wine_bottle_in_the_bottom_drawer_of_the_cabinet_45", "KITCHEN_SCENE5_put_the_black_bowl_on_the_plate_46", "KITCHEN_SCENE6_put_the_yellow_and_white_mug_to_the_front_of_the_white_mug_43", "KITCHEN_SCENE4_put_the_wine_bottle_in_the_bottom_drawer_of_the_cabinet_35", "KITCHEN_SCENE10_put_the_chocolate_pudding_in_the_top_drawer_of_the_cabinet_and_close_it_45", "LIVING_ROOM_SCENE4_stack_the_left_bowl_on_the_right_bowl_and_place_them_in_the_tray_49", "KITCHEN_SCENE4_put_the_black_bowl_on_top_of_the_cabinet_31", "KITCHEN_SCENE4_put_the_wine_bottle_in_the_bottom_drawer_of_the_cabinet_42", "LIVING_ROOM_SCENE1_pick_up_the_alphabet_soup_and_put_it_in_the_basket_3", "KITCHEN_SCENE4_put_the_black_bowl_in_the_bottom_drawer_of_the_cabinet_31", "KITCHEN_SCENE1_open_the_bottom_drawer_of_the_cabinet_3", "LIVING_ROOM_SCENE4_stack_the_left_bowl_on_the_right_bowl_and_place_them_in_the_tray_15", "LIVING_ROOM_SCENE1_pick_up_the_alphabet_soup_and_put_it_in_the_basket_10", "KITCHEN_SCENE3_put_the_frying_pan_on_the_stove_48", "LIVING_ROOM_SCENE6_put_the_chocolate_pudding_to_the_right_of_the_plate_10", "KITCHEN_SCENE10_close_the_top_drawer_of_the_cabinet_and_put_the_black_bowl_on_top_of_it_48", "LIVING_ROOM_SCENE4_stack_the_left_bowl_on_the_right_bowl_and_place_them_in_the_tray_23", "LIVING_ROOM_SCENE6_put_the_chocolate_pudding_to_the_left_of_the_plate_49", "LIVING_ROOM_SCENE4_pick_up_the_chocolate_pudding_and_put_it_in_the_tray_43", "LIVING_ROOM_SCENE6_put_the_chocolate_pudding_to_the_right_of_the_plate_24", "LIVING_ROOM_SCENE6_put_the_chocolate_pudding_to_the_right_of_the_plate_0", "LIVING_ROOM_SCENE4_stack_the_left_bowl_on_the_right_bowl_and_place_them_in_the_tray_27", "KITCHEN_SCENE5_put_the_black_bowl_on_top_of_the_cabinet_38", "LIVING_ROOM_SCENE3_pick_up_the_ketchup_and_put_it_in_the_tray_45", "KITCHEN_SCENE10_put_the_butter_at_the_front_in_the_top_drawer_of_the_cabinet_and_close_it_26", "LIVING_ROOM_SCENE6_put_the_white_mug_on_the_plate_20", "LIVING_ROOM_SCENE4_pick_up_the_salad_dressing_and_put_it_in_the_tray_3", "KITCHEN_SCENE9_put_the_white_bowl_on_top_of_the_cabinet_9", "LIVING_ROOM_SCENE3_pick_up_the_tomato_sauce_and_put_it_in_the_tray_30", "KITCHEN_SCENE9_turn_on_the_stove_27", "LIVING_ROOM_SCENE1_pick_up_the_ketchup_and_put_it_in_the_basket_24", "LIVING_ROOM_SCENE4_pick_up_the_salad_dressing_and_put_it_in_the_tray_24", "KITCHEN_SCENE3_turn_on_the_stove_13", "STUDY_SCENE3_pick_up_the_book_and_place_it_in_the_right_compartment_of_the_caddy_3", "LIVING_ROOM_SCENE1_pick_up_the_cream_cheese_box_and_put_it_in_the_basket_47", "LIVING_ROOM_SCENE2_pick_up_the_tomato_sauce_and_put_it_in_the_basket_8", "KITCHEN_SCENE10_put_the_black_bowl_in_the_top_drawer_of_the_cabinet_43", "KITCHEN_SCENE8_put_the_right_moka_pot_on_the_stove_15", "LIVING_ROOM_SCENE1_pick_up_the_alphabet_soup_and_put_it_in_the_basket_21", "STUDY_SCENE2_pick_up_the_book_and_place_it_in_the_front_compartment_of_the_caddy_14", "KITCHEN_SCENE3_turn_on_the_stove_34", "KITCHEN_SCENE2_open_the_top_drawer_of_the_cabinet_8", "LIVING_ROOM_SCENE2_pick_up_the_tomato_sauce_and_put_it_in_the_basket_46", "KITCHEN_SCENE9_put_the_white_bowl_on_top_of_the_cabinet_17", "KITCHEN_SCENE6_close_the_microwave_47", "LIVING_ROOM_SCENE3_pick_up_the_alphabet_soup_and_put_it_in_the_tray_12", "KITCHEN_SCENE10_put_the_butter_at_the_front_in_the_top_drawer_of_the_cabinet_and_close_it_19", "LIVING_ROOM_SCENE5_put_the_red_mug_on_the_right_plate_44", "KITCHEN_SCENE9_turn_on_the_stove_and_put_the_frying_pan_on_it_16", "KITCHEN_SCENE1_open_the_top_drawer_of_the_cabinet_47", "KITCHEN_SCENE3_turn_on_the_stove_7", "STUDY_SCENE3_pick_up_the_book_and_place_it_in_the_left_compartment_of_the_caddy_45", "KITCHEN_SCENE7_put_the_white_bowl_on_the_plate_1", "LIVING_ROOM_SCENE3_pick_up_the_alphabet_soup_and_put_it_in_the_tray_16", "KITCHEN_SCENE2_stack_the_middle_black_bowl_on_the_back_black_bowl_33", "KITCHEN_SCENE9_put_the_white_bowl_on_top_of_the_cabinet_41", "KITCHEN_SCENE5_put_the_black_bowl_in_the_top_drawer_of_the_cabinet_16", "STUDY_SCENE4_pick_up_the_book_in_the_middle_and_place_it_on_the_cabinet_shelf_11", "KITCHEN_SCENE9_put_the_frying_pan_under_the_cabinet_shelf_36", "KITCHEN_SCENE10_put_the_chocolate_pudding_in_the_top_drawer_of_the_cabinet_and_close_it_27", "STUDY_SCENE3_pick_up_the_red_mug_and_place_it_to_the_right_of_the_caddy_19", "STUDY_SCENE1_pick_up_the_book_and_place_it_in_the_front_compartment_of_the_caddy_25", "LIVING_ROOM_SCENE6_put_the_red_mug_on_the_plate_12", "KITCHEN_SCENE5_put_the_ketchup_in_the_top_drawer_of_the_cabinet_29", "KITCHEN_SCENE7_put_the_white_bowl_to_the_right_of_the_plate_33", "LIVING_ROOM_SCENE1_pick_up_the_tomato_sauce_and_put_it_in_the_basket_25", "KITCHEN_SCENE7_open_the_microwave_27", "KITCHEN_SCENE5_put_the_black_bowl_in_the_top_drawer_of_the_cabinet_28", "KITCHEN_SCENE2_put_the_black_bowl_at_the_back_on_the_plate_11", "KITCHEN_SCENE9_put_the_frying_pan_under_the_cabinet_shelf_0", "KITCHEN_SCENE1_open_the_top_drawer_of_the_cabinet_9", "LIVING_ROOM_SCENE5_put_the_red_mug_on_the_right_plate_37", "LIVING_ROOM_SCENE4_pick_up_the_salad_dressing_and_put_it_in_the_tray_10", "LIVING_ROOM_SCENE4_stack_the_left_bowl_on_the_right_bowl_and_place_them_in_the_tray_35", "LIVING_ROOM_SCENE5_put_the_white_mug_on_the_left_plate_37", "STUDY_SCENE1_pick_up_the_book_and_place_it_in_the_left_compartment_of_the_caddy_22", "KITCHEN_SCENE7_put_the_white_bowl_to_the_right_of_the_plate_18", "KITCHEN_SCENE4_put_the_black_bowl_on_top_of_the_cabinet_47", "KITCHEN_SCENE9_turn_on_the_stove_and_put_the_frying_pan_on_it_17", "KITCHEN_SCENE4_put_the_wine_bottle_in_the_bottom_drawer_of_the_cabinet_40", "LIVING_ROOM_SCENE5_put_the_red_mug_on_the_left_plate_33", "KITCHEN_SCENE3_turn_on_the_stove_6", "KITCHEN_SCENE4_put_the_black_bowl_on_top_of_the_cabinet_34", "KITCHEN_SCENE9_turn_on_the_stove_and_put_the_frying_pan_on_it_29", "STUDY_SCENE4_pick_up_the_book_on_the_right_and_place_it_on_the_cabinet_shelf_47", "KITCHEN_SCENE4_close_the_bottom_drawer_of_the_cabinet_and_open_the_top_drawer_4", "LIVING_ROOM_SCENE5_put_the_white_mug_on_the_left_plate_48", "KITCHEN_SCENE2_stack_the_black_bowl_at_the_front_on_the_black_bowl_in_the_middle_38", "KITCHEN_SCENE1_open_the_bottom_drawer_of_the_cabinet_46", "LIVING_ROOM_SCENE2_pick_up_the_alphabet_soup_and_put_it_in_the_basket_12", "KITCHEN_SCENE10_close_the_top_drawer_of_the_cabinet_and_put_the_black_bowl_on_top_of_it_46", "LIVING_ROOM_SCENE4_pick_up_the_black_bowl_on_the_left_and_put_it_in_the_tray_21", "LIVING_ROOM_SCENE5_put_the_red_mug_on_the_right_plate_36", "KITCHEN_SCENE7_open_the_microwave_1", "LIVING_ROOM_SCENE3_pick_up_the_tomato_sauce_and_put_it_in_the_tray_46", "LIVING_ROOM_SCENE3_pick_up_the_alphabet_soup_and_put_it_in_the_tray_26", "KITCHEN_SCENE9_put_the_frying_pan_under_the_cabinet_shelf_8", "KITCHEN_SCENE10_close_the_top_drawer_of_the_cabinet_and_put_the_black_bowl_on_top_of_it_9", "LIVING_ROOM_SCENE6_put_the_chocolate_pudding_to_the_right_of_the_plate_46", "KITCHEN_SCENE1_put_the_black_bowl_on_the_plate_26", "LIVING_ROOM_SCENE4_stack_the_left_bowl_on_the_right_bowl_and_place_them_in_the_tray_24", "LIVING_ROOM_SCENE6_put_the_red_mug_on_the_plate_36", "LIVING_ROOM_SCENE2_pick_up_the_orange_juice_and_put_it_in_the_basket_1", "LIVING_ROOM_SCENE1_pick_up_the_alphabet_soup_and_put_it_in_the_basket_8", "KITCHEN_SCENE5_put_the_black_bowl_in_the_top_drawer_of_the_cabinet_42", "KITCHEN_SCENE10_put_the_chocolate_pudding_in_the_top_drawer_of_the_cabinet_and_close_it_28", "KITCHEN_SCENE3_put_the_moka_pot_on_the_stove_17", "STUDY_SCENE4_pick_up_the_book_on_the_right_and_place_it_on_the_cabinet_shelf_45", "KITCHEN_SCENE5_put_the_ketchup_in_the_top_drawer_of_the_cabinet_22", "LIVING_ROOM_SCENE6_put_the_chocolate_pudding_to_the_right_of_the_plate_1", "KITCHEN_SCENE7_put_the_white_bowl_on_the_plate_17", "KITCHEN_SCENE1_open_the_top_drawer_of_the_cabinet_and_put_the_bowl_in_it_25", "LIVING_ROOM_SCENE2_pick_up_the_tomato_sauce_and_put_it_in_the_basket_13", "KITCHEN_SCENE10_put_the_chocolate_pudding_in_the_top_drawer_of_the_cabinet_and_close_it_43", "KITCHEN_SCENE2_stack_the_black_bowl_at_the_front_on_the_black_bowl_in_the_middle_20", "STUDY_SCENE2_pick_up_the_book_and_place_it_in_the_right_compartment_of_the_caddy_5", "STUDY_SCENE3_pick_up_the_book_and_place_it_in_the_left_compartment_of_the_caddy_41", "KITCHEN_SCENE9_put_the_white_bowl_on_top_of_the_cabinet_11", "LIVING_ROOM_SCENE6_put_the_chocolate_pudding_to_the_right_of_the_plate_2", "KITCHEN_SCENE9_put_the_white_bowl_on_top_of_the_cabinet_43", "STUDY_SCENE4_pick_up_the_book_on_the_right_and_place_it_on_the_cabinet_shelf_34", "LIVING_ROOM_SCENE1_pick_up_the_ketchup_and_put_it_in_the_basket_29", "KITCHEN_SCENE2_stack_the_middle_black_bowl_on_the_back_black_bowl_19", "LIVING_ROOM_SCENE1_pick_up_the_alphabet_soup_and_put_it_in_the_basket_26", "KITCHEN_SCENE4_put_the_wine_bottle_in_the_bottom_drawer_of_the_cabinet_43", "LIVING_ROOM_SCENE5_put_the_red_mug_on_the_right_plate_0", "STUDY_SCENE2_pick_up_the_book_and_place_it_in_the_right_compartment_of_the_caddy_12", "LIVING_ROOM_SCENE3_pick_up_the_cream_cheese_and_put_it_in_the_tray_40", "LIVING_ROOM_SCENE4_pick_up_the_chocolate_pudding_and_put_it_in_the_tray_15", "KITCHEN_SCENE10_put_the_butter_at_the_back_in_the_top_drawer_of_the_cabinet_and_close_it_18", "LIVING_ROOM_SCENE2_pick_up_the_alphabet_soup_and_put_it_in_the_basket_6", "KITCHEN_SCENE2_put_the_middle_black_bowl_on_top_of_the_cabinet_26", "LIVING_ROOM_SCENE3_pick_up_the_ketchup_and_put_it_in_the_tray_27", "KITCHEN_SCENE2_open_the_top_drawer_of_the_cabinet_41", "KITCHEN_SCENE2_put_the_black_bowl_at_the_front_on_the_plate_46", "LIVING_ROOM_SCENE4_pick_up_the_chocolate_pudding_and_put_it_in_the_tray_36", "STUDY_SCENE1_pick_up_the_book_and_place_it_in_the_right_compartment_of_the_caddy_18", "STUDY_SCENE1_pick_up_the_book_and_place_it_in_the_right_compartment_of_the_caddy_7", "KITCHEN_SCENE1_put_the_black_bowl_on_the_plate_2", "LIVING_ROOM_SCENE5_put_the_red_mug_on_the_right_plate_38", "LIVING_ROOM_SCENE6_put_the_white_mug_on_the_plate_2", "KITCHEN_SCENE2_put_the_middle_black_bowl_on_the_plate_2", "KITCHEN_SCENE2_stack_the_middle_black_bowl_on_the_back_black_bowl_13", "STUDY_SCENE4_pick_up_the_book_on_the_right_and_place_it_under_the_cabinet_shelf_20", "LIVING_ROOM_SCENE6_put_the_white_mug_on_the_plate_5", "KITCHEN_SCENE9_turn_on_the_stove_46", "KITCHEN_SCENE10_put_the_black_bowl_in_the_top_drawer_of_the_cabinet_15", "LIVING_ROOM_SCENE5_put_the_red_mug_on_the_left_plate_14", "LIVING_ROOM_SCENE3_pick_up_the_butter_and_put_it_in_the_tray_29", "STUDY_SCENE2_pick_up_the_book_and_place_it_in_the_left_compartment_of_the_caddy_30", "KITCHEN_SCENE9_put_the_frying_pan_on_top_of_the_cabinet_49", "KITCHEN_SCENE9_turn_on_the_stove_and_put_the_frying_pan_on_it_41", "STUDY_SCENE4_pick_up_the_book_in_the_middle_and_place_it_on_the_cabinet_shelf_3", "KITCHEN_SCENE2_stack_the_black_bowl_at_the_front_on_the_black_bowl_in_the_middle_2", "KITCHEN_SCENE3_turn_on_the_stove_and_put_the_frying_pan_on_it_10", "LIVING_ROOM_SCENE1_pick_up_the_ketchup_and_put_it_in_the_basket_17", "KITCHEN_SCENE9_put_the_white_bowl_on_top_of_the_cabinet_27", "LIVING_ROOM_SCENE4_stack_the_left_bowl_on_the_right_bowl_and_place_them_in_the_tray_22", "LIVING_ROOM_SCENE1_pick_up_the_cream_cheese_box_and_put_it_in_the_basket_22", "LIVING_ROOM_SCENE5_put_the_red_mug_on_the_left_plate_1", "KITCHEN_SCENE2_open_the_top_drawer_of_the_cabinet_35", "KITCHEN_SCENE7_put_the_white_bowl_to_the_right_of_the_plate_36", "LIVING_ROOM_SCENE5_put_the_red_mug_on_the_left_plate_13", "STUDY_SCENE3_pick_up_the_red_mug_and_place_it_to_the_right_of_the_caddy_13", "LIVING_ROOM_SCENE1_pick_up_the_tomato_sauce_and_put_it_in_the_basket_3", "LIVING_ROOM_SCENE1_pick_up_the_cream_cheese_box_and_put_it_in_the_basket_36", "LIVING_ROOM_SCENE3_pick_up_the_ketchup_and_put_it_in_the_tray_35", "STUDY_SCENE2_pick_up_the_book_and_place_it_in_the_right_compartment_of_the_caddy_46", "KITCHEN_SCENE2_put_the_middle_black_bowl_on_top_of_the_cabinet_18", "LIVING_ROOM_SCENE2_pick_up_the_orange_juice_and_put_it_in_the_basket_37", "STUDY_SCENE2_pick_up_the_book_and_place_it_in_the_front_compartment_of_the_caddy_35", "LIVING_ROOM_SCENE3_pick_up_the_tomato_sauce_and_put_it_in_the_tray_36", "STUDY_SCENE2_pick_up_the_book_and_place_it_in_the_right_compartment_of_the_caddy_26", "LIVING_ROOM_SCENE6_put_the_chocolate_pudding_to_the_right_of_the_plate_35", "LIVING_ROOM_SCENE3_pick_up_the_tomato_sauce_and_put_it_in_the_tray_35", "KITCHEN_SCENE9_put_the_frying_pan_on_the_cabinet_shelf_40", "KITCHEN_SCENE8_turn_off_the_stove_27", "KITCHEN_SCENE10_put_the_chocolate_pudding_in_the_top_drawer_of_the_cabinet_and_close_it_47", "STUDY_SCENE4_pick_up_the_book_in_the_middle_and_place_it_on_the_cabinet_shelf_42", "KITCHEN_SCENE4_close_the_bottom_drawer_of_the_cabinet_4", "STUDY_SCENE2_pick_up_the_book_and_place_it_in_the_back_compartment_of_the_caddy_32", "KITCHEN_SCENE8_put_the_right_moka_pot_on_the_stove_23", "KITCHEN_SCENE2_stack_the_black_bowl_at_the_front_on_the_black_bowl_in_the_middle_28", "KITCHEN_SCENE3_put_the_frying_pan_on_the_stove_10", "STUDY_SCENE1_pick_up_the_book_and_place_it_in_the_front_compartment_of_the_caddy_20", "KITCHEN_SCENE7_put_the_white_bowl_on_the_plate_29", "LIVING_ROOM_SCENE3_pick_up_the_ketchup_and_put_it_in_the_tray_42", "LIVING_ROOM_SCENE4_pick_up_the_chocolate_pudding_and_put_it_in_the_tray_23"]
|