nanidafvck commited on
Commit
e53787a
·
verified ·
1 Parent(s): 0224148

Upload folder using huggingface_hub

Browse files
checkpoint-124.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:adf3b84b5e5d380e413917b9aa8f15750c08e2e985f6da073ce9bfbf2fa9b7b1
3
+ size 1099294898
events.out.tfevents.1765750679.phe108c-yuewang-02.1031892.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:84beb71e8bef80980bc0c1e8c400f39caa0184e97fe2aca72dbeae2ce5ff7e8b
3
+ size 148332
events.out.tfevents.1765778016.phe108c-yuewang-02.1063886.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b2572c827b9935a7fd7e96d9334b7eed483e7a4bc879929eaeef8d39f5abe0ae
3
+ size 13678
events.out.tfevents.1765784424.phe108c-yuewang-02.1092414.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3462774d2f6f4ce716483f7f9737e287c7d4baa8adcb5c9e30371d301a84f8bf
3
+ size 3410
events.out.tfevents.1765787935.phe108c-yuewang-02.1097973.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f74228e4fb9f7747d7d9f84f0b1b55f1f36b66f79da0b45e2f028e4c0da93cef
3
+ size 69293
events.out.tfevents.1765805852.phe108c-yuewang-02.1108454.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d9df9576db70092d2042445c73bc9490fc39b6ca3840f95ccdc6f94b01da55b1
3
+ size 70011
events.out.tfevents.1765819581.phe108c-yuewang-02.1114159.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9ced62c2fb56d4df0839f7c1de6872eae206319dd15bcbd958655a7ebe821e4a
3
+ size 544808
events.out.tfevents.1765916962.phe108c-yuewang-02.2912013.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef1e3470d64a0af0d0df6199d4f9cbe27f120ad3a1766995cb793ac9e0d65c5b
3
+ size 171276
events.out.tfevents.1765962508.phe108c-yuewang-02.250959.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:108c1e787121fd397844e0f7d2aac334e77ae4fb9ea2451c5862b62bec942e72
3
+ size 396
events.out.tfevents.1765964368.phe108c-yuewang-02.597867.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:273d0fe64002681657667384ed12c0a7bb4f5f345a8ff2a6c13593c3a887896d
3
+ size 88
events.out.tfevents.1765967041.phe108c-yuewang-02.523103.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c12f64a74bfed59302246650aefe114922bbd95a9b76608ce8d746ecdf0e1b62
3
+ size 10868
events.out.tfevents.1765972816.phe108c-yuewang-02.4092656.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6857e8f94514d4d2c180e7c60735095ae886af52ceac2c0d8e715f7b69384621
3
+ size 2860
events.out.tfevents.1765974399.phe108c-yuewang-02.787186.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ec1d422f1cef5b6391508a4cf2a88a1ef6a92b8119af2b2c4422a19315393728
3
+ size 259512
events.out.tfevents.1766040655.phe108c-yuewang-02.3974176.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5ce8b8396714ed229f88bec2db758c5a13ab3f4478387fd85ac0a9bb438e516d
3
+ size 160188
events.out.tfevents.1766086752.phe108c-yuewang-02.2102202.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0e936a9ae34794a5cffd44e3a8a44aa9cdd8b3b093f1c95beebad07b6fad8265
3
+ size 63276
events.out.tfevents.1766101210.phe108c-yuewang-02.2357724.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d283e2df4bf584b7429d2758bb3c809aed27ea7135de4a7ce46203bcc562029
3
+ size 127132
events.out.tfevents.1766127019.phe108c-yuewang-02.4152936.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d99d3a853c1078d31e3bd90a94b716024cfaad037db27bad4c154fa296d35150
3
+ size 13948
events.out.tfevents.1766134288.phe108c-yuewang-02.391862.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:001473cda22d08b783f9e73571a0daee007310e573793f94dc7711c99434cc0f
3
+ size 226456
log.txt ADDED
@@ -0,0 +1,125 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"train_lr": 0.00019564004345400005, "train_loss": 0.3029090781689445, "epoch": 0, "val_lr": 0.00010000000000000006, "val_loss": 0.20378573006018996}
2
+ {"train_lr": 0.00048637343608965073, "train_loss": 0.1901836911292664, "epoch": 1, "val_lr": 0.0004487973013898667, "val_loss": 0.18067710906267165}
3
+ {"train_lr": 0.0004998696082222175, "train_loss": 0.17816189841009103, "epoch": 2, "val_lr": 0.0004999143342294313, "val_loss": 0.17080191090175018}
4
+ {"train_lr": 0.0004995895072293546, "train_loss": 0.15774387331829595, "epoch": 3, "val_lr": 0.0004996726615235633, "val_loss": 0.12231620099572908}
5
+ {"train_lr": 0.0004991485594922281, "train_loss": 0.11175334399146974, "epoch": 4, "val_lr": 0.0004992700923113186, "val_loss": 0.10168942175204715}
6
+ {"train_lr": 0.000498547062354982, "train_loss": 0.0965861452922021, "epoch": 5, "val_lr": 0.0004987070201088922, "val_loss": 0.08977454705607324}
7
+ {"train_lr": 0.0004977853648358597, "train_loss": 0.08411245128968473, "epoch": 6, "val_lr": 0.0004979836008699763, "val_loss": 0.07585888654584923}
8
+ {"train_lr": 0.0004968640124725848, "train_loss": 0.07101845985501036, "epoch": 7, "val_lr": 0.0004971005611665038, "val_loss": 0.06519926275264044}
9
+ {"train_lr": 0.0004957835503316204, "train_loss": 0.064371204305105, "epoch": 8, "val_lr": 0.0004960579065175926, "val_loss": 0.061161860093474386}
10
+ {"train_lr": 0.00049454470905488, "train_loss": 0.060200666000887956, "epoch": 9, "val_lr": 0.0004948468761150538, "val_loss": 0.05733869946919955}
11
+ {"train_lr": 0.0004931482322768172, "train_loss": 0.05695078827936596, "epoch": 10, "val_lr": 0.0004934981331595625, "val_loss": 0.05281244964158583}
12
+ {"train_lr": 0.0004915950874001617, "train_loss": 0.05440806119826721, "epoch": 11, "val_lr": 0.0004919824024869898, "val_loss": 0.05236734223792389}
13
+ {"train_lr": 0.0004898863189075323, "train_loss": 0.05319481884278281, "epoch": 12, "val_lr": 0.0004903105349971374, "val_loss": 0.05266624510904232}
14
+ {"train_lr": 0.0004880228557725985, "train_loss": 0.05180120371913614, "epoch": 13, "val_lr": 0.0004884843962028363, "val_loss": 0.05038926356691655}
15
+ {"train_lr": 0.00048600605160856875, "train_loss": 0.050377890708376415, "epoch": 14, "val_lr": 0.000486503724771813, "val_loss": 0.04809159743308555}
16
+ {"train_lr": 0.00048383709037651707, "train_loss": 0.049341674697004324, "epoch": 15, "val_lr": 0.0004843728643456271, "val_loss": 0.04898137157627692}
17
+ {"train_lr": 0.00048151750704065376, "train_loss": 0.04828970470076605, "epoch": 16, "val_lr": 0.0004820872719566832, "val_loss": 0.04582030848988632}
18
+ {"train_lr": 0.00047904868262758124, "train_loss": 0.04746105371095358, "epoch": 17, "val_lr": 0.0004796561867071299, "val_loss": 0.046291777344110115}
19
+ {"train_lr": 0.0004764322268321257, "train_loss": 0.046354372404379604, "epoch": 18, "val_lr": 0.00047707271926729345, "val_loss": 0.04550358132837573}
20
+ {"train_lr": 0.00047366992437445206, "train_loss": 0.04631630348176203, "epoch": 19, "val_lr": 0.00047434512192096024, "val_loss": 0.044303551486264615}
21
+ {"train_lr": 0.00047076356999288126, "train_loss": 0.044630916993501385, "epoch": 20, "val_lr": 0.0004714737347557191, "val_loss": 0.04328328773106744}
22
+ {"train_lr": 0.000467714621679718, "train_loss": 0.043510656242827815, "epoch": 21, "val_lr": 0.00046844592492022695, "val_loss": 0.043618555585023505}
23
+ {"train_lr": 0.0004645256362715261, "train_loss": 0.04274718968581208, "epoch": 22, "val_lr": 0.00046530222453661365, "val_loss": 0.04072705302678514}
24
+ {"train_lr": 0.00046119826652148675, "train_loss": 0.04158026777635636, "epoch": 23, "val_lr": 0.0004620080338788992, "val_loss": 0.039345405019819736}
25
+ {"train_lr": 0.00045773486228217617, "train_loss": 0.04112308661733077, "epoch": 24, "val_lr": 0.0004585791000421254, "val_loss": 0.04036848139294908}
26
+ {"train_lr": 0.00045413766264156073, "train_loss": 0.04041471443990466, "epoch": 25, "val_lr": 0.00045501146756788053, "val_loss": 0.03908891554176808}
27
+ {"train_lr": 0.00045040871313989535, "train_loss": 0.040147826829208776, "epoch": 26, "val_lr": 0.0004513138575405235, "val_loss": 0.03813537592440844}
28
+ {"train_lr": 0.00044655071230848936, "train_loss": 0.039322634453001694, "epoch": 27, "val_lr": 0.0004474868983232679, "val_loss": 0.0382390329313855}
29
+ {"train_lr": 0.0004425660737458544, "train_loss": 0.03864359832081936, "epoch": 28, "val_lr": 0.00044353150487087085, "val_loss": 0.03713711851014523}
30
+ {"train_lr": 0.000438457395267602, "train_loss": 0.03818666758431093, "epoch": 29, "val_lr": 0.0004394525610181687, "val_loss": 0.03699907187372446}
31
+ {"train_lr": 0.0004342270013093638, "train_loss": 0.03759662008126303, "epoch": 30, "val_lr": 0.00043525215678922604, "val_loss": 0.03622726344208165}
32
+ {"train_lr": 0.0004298782177121366, "train_loss": 0.03709144374584913, "epoch": 31, "val_lr": 0.0004309313797143709, "val_loss": 0.0362949859260059}
33
+ {"train_lr": 0.00042541330648248807, "train_loss": 0.03704276450347565, "epoch": 32, "val_lr": 0.0004264929103607731, "val_loss": 0.03630306718498468}
34
+ {"train_lr": 0.00042083532955537885, "train_loss": 0.03612849534275468, "epoch": 33, "val_lr": 0.00042194127825292196, "val_loss": 0.03573022988866868}
35
+ {"train_lr": 0.0004161473908507008, "train_loss": 0.03549579941874152, "epoch": 34, "val_lr": 0.0004172802929047031, "val_loss": 0.0337221956165928}
36
+ {"train_lr": 0.0004113523070013329, "train_loss": 0.035088212288574554, "epoch": 35, "val_lr": 0.00041249087910818116, "val_loss": 0.03479692658279524}
37
+ {"train_lr": 0.0004064533191198741, "train_loss": 0.034415727810464376, "epoch": 36, "val_lr": 0.0004076367699445188, "val_loss": 0.03286332018461412}
38
+ {"train_lr": 0.00040145348832920267, "train_loss": 0.03416300144759015, "epoch": 37, "val_lr": 0.00040266163356827945, "val_loss": 0.03268003067551333}
39
+ {"train_lr": 0.0003963559138950938, "train_loss": 0.03316745480847089, "epoch": 38, "val_lr": 0.0003975862941054119, "val_loss": 0.03240224858367346}
40
+ {"train_lr": 0.00039116424326958324, "train_loss": 0.032830451658156075, "epoch": 39, "val_lr": 0.00039241506916727956, "val_loss": 0.03241762261660326}
41
+ {"train_lr": 0.0003858812365114264, "train_loss": 0.03221809167793154, "epoch": 40, "val_lr": 0.0003871535046846055, "val_loss": 0.030558354837277275}
42
+ {"train_lr": 0.0003805109682349588, "train_loss": 0.0316272711980638, "epoch": 41, "val_lr": 0.0003818037542649857, "val_loss": 0.0298328703108031}
43
+ {"train_lr": 0.00037505651830543516, "train_loss": 0.030875583239354694, "epoch": 42, "val_lr": 0.00037636906354619396, "val_loss": 0.02985021250146463}
44
+ {"train_lr": 0.0003695215196280058, "train_loss": 0.030450353346133528, "epoch": 43, "val_lr": 0.0003708529348886704, "val_loss": 0.02946670093815627}
45
+ {"train_lr": 0.00036390953986632644, "train_loss": 0.029792958058218758, "epoch": 44, "val_lr": 0.00036526101124799925, "val_loss": 0.028769544291696172}
46
+ {"train_lr": 0.00035822386312613303, "train_loss": 0.029045628254025534, "epoch": 45, "val_lr": 0.0003595919218602794, "val_loss": 0.028720535781054247}
47
+ {"train_lr": 0.00035246864244515197, "train_loss": 0.02842943575366767, "epoch": 46, "val_lr": 0.00035385171834479594, "val_loss": 0.02763190899249606}
48
+ {"train_lr": 0.00034664744493014834, "train_loss": 0.027837987167432028, "epoch": 47, "val_lr": 0.00034804587662652675, "val_loss": 0.026703940858463133}
49
+ {"train_lr": 0.0003407639194330738, "train_loss": 0.02754394838467286, "epoch": 48, "val_lr": 0.00034218017350477737, "val_loss": 0.0257092342276859}
50
+ {"train_lr": 0.00033482187370514893, "train_loss": 0.02691245303462513, "epoch": 49, "val_lr": 0.00033624979313522395, "val_loss": 0.025880782399326563}
51
+ {"train_lr": 0.0003288254193979696, "train_loss": 0.026448398734446408, "epoch": 50, "val_lr": 0.00033026512825549525, "val_loss": 0.025931137435138224}
52
+ {"train_lr": 0.00032277780083401965, "train_loss": 0.025860479277777966, "epoch": 51, "val_lr": 0.0003242294003332557, "val_loss": 0.024530567467212678}
53
+ {"train_lr": 0.0003166832772149368, "train_loss": 0.025295751232208095, "epoch": 52, "val_lr": 0.0003181465224205295, "val_loss": 0.02443710269947206}
54
+ {"train_lr": 0.00031054590880512834, "train_loss": 0.024831028110069087, "epoch": 53, "val_lr": 0.0003120199681636746, "val_loss": 0.024005142320704654}
55
+ {"train_lr": 0.00030436953575710196, "train_loss": 0.024337971479333836, "epoch": 54, "val_lr": 0.0003058537002949272, "val_loss": 0.023419170529719015}
56
+ {"train_lr": 0.00029815805927156566, "train_loss": 0.02390463634179727, "epoch": 55, "val_lr": 0.00029959978892800203, "val_loss": 0.02288346139427561}
57
+ {"train_lr": 0.0002919157893170792, "train_loss": 0.023330022025673264, "epoch": 56, "val_lr": 0.0002934112533794024, "val_loss": 0.0228662821684238}
58
+ {"train_lr": 0.000285646125863976, "train_loss": 0.02302058585830379, "epoch": 57, "val_lr": 0.00028714861572367255, "val_loss": 0.021559586439281703}
59
+ {"train_lr": 0.00027935389395657595, "train_loss": 0.02267140933978619, "epoch": 58, "val_lr": 0.0002808629371486902, "val_loss": 0.021895042590312358}
60
+ {"train_lr": 0.00027304247236984726, "train_loss": 0.022264836005202026, "epoch": 59, "val_lr": 0.0002745547464698992, "val_loss": 0.021240073793207206}
61
+ {"train_lr": 0.0002667164136507622, "train_loss": 0.02188992994808349, "epoch": 60, "val_lr": 0.00026823044849693084, "val_loss": 0.020826252639323237}
62
+ {"train_lr": 0.0002603792240263071, "train_loss": 0.021416045910589848, "epoch": 61, "val_lr": 0.0002618978397730599, "val_loss": 0.020783679086593836}
63
+ {"train_lr": 0.0002540355193130702, "train_loss": 0.02100502312555527, "epoch": 62, "val_lr": 0.00025555314829048964, "val_loss": 0.020066257754695557}
64
+ {"train_lr": 0.00024768923113325603, "train_loss": 0.020786816299805052, "epoch": 63, "val_lr": 0.00024920760104002606, "val_loss": 0.01962972852215171}
65
+ {"train_lr": 0.00024134440660235424, "train_loss": 0.020371674011462393, "epoch": 64, "val_lr": 0.00024286279144732108, "val_loss": 0.019635748700238764}
66
+ {"train_lr": 0.0002350054246168891, "train_loss": 0.01984649458765861, "epoch": 65, "val_lr": 0.00023652184380847327, "val_loss": 0.018980687991115113}
67
+ {"train_lr": 0.00022867565801970824, "train_loss": 0.019436031269265304, "epoch": 66, "val_lr": 0.00023018832132361074, "val_loss": 0.01885488977847946}
68
+ {"train_lr": 0.00022236003826934538, "train_loss": 0.019223793374951954, "epoch": 67, "val_lr": 0.00022384425604489309, "val_loss": 0.018089972862381804}
69
+ {"train_lr": 0.00021606188272490744, "train_loss": 0.01876881279790378, "epoch": 68, "val_lr": 0.00021756575153400524, "val_loss": 0.018229862333100755}
70
+ {"train_lr": 0.0002097857536270519, "train_loss": 0.01860100810686817, "epoch": 69, "val_lr": 0.00021128399468234633, "val_loss": 0.01770009936808492}
71
+ {"train_lr": 0.00020353563942233506, "train_loss": 0.01837872951752221, "epoch": 70, "val_lr": 0.00020502761050491806, "val_loss": 0.01706261415818026}
72
+ {"train_lr": 0.0001973154945066465, "train_loss": 0.017986078369934398, "epoch": 71, "val_lr": 0.00019880030871354622, "val_loss": 0.017128126863390207}
73
+ {"train_lr": 0.0001911290314158904, "train_loss": 0.017612919353283737, "epoch": 72, "val_lr": 0.0001926054816676948, "val_loss": 0.01704678402841091}
74
+ {"train_lr": 0.00018498069568769108, "train_loss": 0.017327215939184584, "epoch": 73, "val_lr": 0.00018644681189779387, "val_loss": 0.01635143433616856}
75
+ {"train_lr": 0.00017887421629145811, "train_loss": 0.017028148669741214, "epoch": 74, "val_lr": 0.00018033149006968616, "val_loss": 0.016331549726366516}
76
+ {"train_lr": 0.00017281357449716822, "train_loss": 0.016653849102066836, "epoch": 75, "val_lr": 0.00017425877343425064, "val_loss": 0.015566111665219068}
77
+ {"train_lr": 0.0001668027037119927, "train_loss": 0.01647036786889259, "epoch": 76, "val_lr": 0.00016818973307559858, "val_loss": 0.014950288312796217}
78
+ {"train_lr": 0.00016084546491516838, "train_loss": 0.01609618966705212, "epoch": 77, "val_lr": 0.0001622641764016449, "val_loss": 0.015607096098392503}
79
+ {"train_lr": 0.00015494559384151709, "train_loss": 0.015878241306616584, "epoch": 78, "val_lr": 0.00015635066862102265, "val_loss": 0.01478586053579218}
80
+ {"train_lr": 0.0001491071436286576, "train_loss": 0.015681709726296386, "epoch": 79, "val_lr": 0.0001504992200667879, "val_loss": 0.015376914625729972}
81
+ {"train_lr": 0.00014333368103339436, "train_loss": 0.01536582656565532, "epoch": 80, "val_lr": 0.00014470812040117537, "val_loss": 0.014833954703062773}
82
+ {"train_lr": 0.00013762877671113355, "train_loss": 0.015042718994185609, "epoch": 81, "val_lr": 0.00013898601217372442, "val_loss": 0.014250776116987543}
83
+ {"train_lr": 0.0001319963676977656, "train_loss": 0.014874645587631222, "epoch": 82, "val_lr": 0.00013333594729889647, "val_loss": 0.014122673421211186}
84
+ {"train_lr": 0.00012644000894844172, "train_loss": 0.01462754887003495, "epoch": 83, "val_lr": 0.0001277610660149649, "val_loss": 0.014263364846729451}
85
+ {"train_lr": 0.00012096340874629102, "train_loss": 0.014381824359157271, "epoch": 84, "val_lr": 0.00012226808773231843, "val_loss": 0.013550933956199128}
86
+ {"train_lr": 0.0001155698280035565, "train_loss": 0.014335416555139847, "epoch": 85, "val_lr": 0.00011685318803230731, "val_loss": 0.013834573171381665}
87
+ {"train_lr": 0.00011026296158263907, "train_loss": 0.014103251056881707, "epoch": 86, "val_lr": 0.00011152624273383104, "val_loss": 0.013314086787372094}
88
+ {"train_lr": 0.00010504600853888364, "train_loss": 0.013863970035279755, "epoch": 87, "val_lr": 0.00010628492357509662, "val_loss": 0.013290417570620775}
89
+ {"train_lr": 9.992253733922532e-05, "train_loss": 0.013815163830919484, "epoch": 88, "val_lr": 0.00010113833361545515, "val_loss": 0.012893795895035424}
90
+ {"train_lr": 9.489576647573349e-05, "train_loss": 0.013514626766281467, "epoch": 89, "val_lr": 9.608814468650322e-05, "val_loss": 0.013136028532042272}
91
+ {"train_lr": 8.996904642878931e-05, "train_loss": 0.0134097098326199, "epoch": 90, "val_lr": 9.111810643178455e-05, "val_loss": 0.01242462544189405}
92
+ {"train_lr": 8.514539285366458e-05, "train_loss": 0.013135068401647795, "epoch": 91, "val_lr": 8.628824012057681e-05, "val_loss": 0.012393929177939187}
93
+ {"train_lr": 8.04279432727628e-05, "train_loss": 0.013013286120520594, "epoch": 92, "val_lr": 8.154721108420788e-05, "val_loss": 0.012394825573934893}
94
+ {"train_lr": 7.581986828398433e-05, "train_loss": 0.012972520542583937, "epoch": 93, "val_lr": 7.69105678142417e-05, "val_loss": 0.012417257270284407}
95
+ {"train_lr": 7.132404085957154e-05, "train_loss": 0.01284948122561654, "epoch": 94, "val_lr": 7.238882556203876e-05, "val_loss": 0.012092059523227715}
96
+ {"train_lr": 6.694325050301429e-05, "train_loss": 0.01272580515895986, "epoch": 95, "val_lr": 6.798187606991025e-05, "val_loss": 0.012048265095189458}
97
+ {"train_lr": 6.26805137222791e-05, "train_loss": 0.01264511175966782, "epoch": 96, "val_lr": 6.367159987954019e-05, "val_loss": 0.011939053134920522}
98
+ {"train_lr": 5.853855307613423e-05, "train_loss": 0.012489690558386555, "epoch": 97, "val_lr": 5.9517780084338156e-05, "val_loss": 0.011728202640604708}
99
+ {"train_lr": 5.452006353841531e-05, "train_loss": 0.012315334338378815, "epoch": 98, "val_lr": 5.546804239881225e-05, "val_loss": 0.011587674211114063}
100
+ {"train_lr": 5.0627196885295775e-05, "train_loss": 0.012222715360107703, "epoch": 99, "val_lr": 5.1545734750216244e-05, "val_loss": 0.011885711507871747}
101
+ {"train_lr": 4.6862939276060094e-05, "train_loss": 0.0120737815875928, "epoch": 100, "val_lr": 4.775105820496612e-05, "val_loss": 0.01174918009576598}
102
+ {"train_lr": 4.3229659440581875e-05, "train_loss": 0.012053680395158752, "epoch": 101, "val_lr": 4.408571933548691e-05, "val_loss": 0.011422525994479656}
103
+ {"train_lr": 3.9729663788073124e-05, "train_loss": 0.011984574670019234, "epoch": 102, "val_lr": 4.05546541795973e-05, "val_loss": 0.011494848364206955}
104
+ {"train_lr": 3.636514259023374e-05, "train_loss": 0.011839345053557375, "epoch": 103, "val_lr": 3.715585419200682e-05, "val_loss": 0.011368920762590535}
105
+ {"train_lr": 3.3138255278447185e-05, "train_loss": 0.011952498970095888, "epoch": 104, "val_lr": 3.389811365138092e-05, "val_loss": 0.011258657481597706}
106
+ {"train_lr": 3.005119280906492e-05, "train_loss": 0.011793950572450647, "epoch": 105, "val_lr": 3.077516253091211e-05, "val_loss": 0.011036837981339163}
107
+ {"train_lr": 2.7105964216265776e-05, "train_loss": 0.01165330332612068, "epoch": 106, "val_lr": 2.7796658418673295e-05, "val_loss": 0.011326960379993844}
108
+ {"train_lr": 2.4304168241662768e-05, "train_loss": 0.011600216138117635, "epoch": 107, "val_lr": 2.496039061000059e-05, "val_loss": 0.011227262035507949}
109
+ {"train_lr": 2.1647883702788352e-05, "train_loss": 0.011528690113152568, "epoch": 108, "val_lr": 2.2268171500970164e-05, "val_loss": 0.011150405061925994}
110
+ {"train_lr": 1.913884253928374e-05, "train_loss": 0.011567669638481713, "epoch": 109, "val_lr": 1.9723835954023598e-05, "val_loss": 0.010844607260004859}
111
+ {"train_lr": 1.67786382543702e-05, "train_loss": 0.01138995437179139, "epoch": 110, "val_lr": 1.7328584238632968e-05, "val_loss": 0.010685282611032767}
112
+ {"train_lr": 1.456843174835642e-05, "train_loss": 0.011354313507038319, "epoch": 111, "val_lr": 1.5073475525714953e-05, "val_loss": 0.011157728767730007}
113
+ {"train_lr": 1.251018386556775e-05, "train_loss": 0.011371533288274711, "epoch": 112, "val_lr": 1.2987145024074648e-05, "val_loss": 0.010850955787602635}
114
+ {"train_lr": 1.0604939279716262e-05, "train_loss": 0.011391660562431898, "epoch": 113, "val_lr": 1.1044898371185302e-05, "val_loss": 0.010716146725826547}
115
+ {"train_lr": 8.853953268125544e-06, "train_loss": 0.01128660032884765, "epoch": 114, "val_lr": 9.257403022359512e-06, "val_loss": 0.01041673325933516}
116
+ {"train_lr": 7.25839576126639e-06, "train_loss": 0.01125125261874781, "epoch": 115, "val_lr": 7.62434821019335e-06, "val_loss": 0.0105993557566156}
117
+ {"train_lr": 5.819260897595943e-06, "train_loss": 0.011226332090388182, "epoch": 116, "val_lr": 6.147901616568134e-06, "val_loss": 0.010801995811983944}
118
+ {"train_lr": 4.537500139906067e-06, "train_loss": 0.011306854250873922, "epoch": 117, "val_lr": 4.82821837662118e-06, "val_loss": 0.010578870405590841}
119
+ {"train_lr": 3.413925883331067e-06, "train_loss": 0.011253086827925814, "epoch": 118, "val_lr": 3.667352208154465e-06, "val_loss": 0.010513919106555907}
120
+ {"train_lr": 2.449249230041731e-06, "train_loss": 0.011142525816522475, "epoch": 119, "val_lr": 2.6651165730630095e-06, "val_loss": 0.010430092779026666}
121
+ {"train_lr": 1.6441586857713777e-06, "train_loss": 0.011191111990599493, "epoch": 120, "val_lr": 1.820522041858372e-06, "val_loss": 0.01031723869304488}
122
+ {"train_lr": 9.990316580090408e-07, "train_loss": 0.01121511724456812, "epoch": 121, "val_lr": 1.1373589087718278e-06, "val_loss": 0.010718939803540706}
123
+ {"train_lr": 5.144216248127181e-07, "train_loss": 0.011190643205175104, "epoch": 122, "val_lr": 6.143148911205126e-07, "val_loss": 0.010613977236673237}
124
+ {"train_lr": 1.9059472917665377e-07, "train_loss": 0.01114721703897224, "epoch": 123, "val_lr": 2.5178795675100504e-07, "val_loss": 0.010474107173220145}
125
+ {"train_lr": 2.773946210712383e-08, "train_loss": 0.01110242478758527, "epoch": 124, "val_lr": 5.051507677919919e-08, "val_loss": 0.01055344306708624}
run.yaml ADDED
@@ -0,0 +1,88 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ !!python/object:icrt.util.args.ExperimentConfig
2
+ dataset_cfg: !!python/object:icrt.util.args.DatasetConfig
3
+ action_noise: 0.0
4
+ dataset_fraction: 1.0
5
+ dataset_json: ./config/dataset_config_libero_90_ver3.json
6
+ goal_conditioned: false
7
+ non_overlapping: 64
8
+ num_repeat_traj: 1
9
+ num_weighted_steps: 30
10
+ proprio_noise: 0.005
11
+ rebalance_tasks: true
12
+ shuffle_repeat_traj: true
13
+ skip_step: false
14
+ sort_by_lang: true
15
+ task_barrier: true
16
+ task_names: null
17
+ vision_aug: true
18
+ visual_trace_noise: 0.005
19
+ device: cuda
20
+ dist_on_itp: false
21
+ dist_url: env://
22
+ load_config: null
23
+ local_rank: -1
24
+ logging_cfg: !!python/object:icrt.util.args.LoggingConfig
25
+ log_dir: /data/tientoan/icrt_output/output_libero_90/no_prompt_loss_0_ver3
26
+ log_name: no_prompt_loss_0_ver3
27
+ output_dir: /data/tientoan/icrt_output/output_libero_90/no_prompt_loss_0_ver3
28
+ model_cfg: !!python/object:icrt.util.args.ModelConfig
29
+ policy_cfg: !!python/object:icrt.util.args.PolicyConfig
30
+ adapter_mlp_ratio: 4.0
31
+ adapter_num_heads: 8
32
+ camera_pos_emb: false
33
+ decoder_pred_head: mlp
34
+ kl_div_loss: false
35
+ llama_ckpt_dir: /home/mfu/checkpoints/llama-2/llama-2-7b
36
+ load_llama: true
37
+ lora_layer_idxs: null
38
+ lora_rank: 4
39
+ loss_w_action: 1.0
40
+ modality_pos_emb: false
41
+ multikv_attn_pool: false
42
+ no_prompt_loss: true
43
+ num_inference_diffusion_steps: null
44
+ num_train_diffusion_steps: null
45
+ phase: pretrain
46
+ pred_action_only: true
47
+ pretrained_path: null
48
+ remove_proprio: false
49
+ scale_loss: 1.0
50
+ scratch_llama_config: config/model_config/custom_transformer.json
51
+ separate_camera_adapter: true
52
+ step_weight: 1.0
53
+ vision_encoder_cfg: !!python/object:icrt.util.args.VisionEncoderConfig
54
+ vision_encoder: ./checkpoints/crossmae_rtx/cross-mae-rtx-vitb.pth
55
+ vision_lora: false
56
+ vision_lora_rank: 8
57
+ vision_nonpretrained: false
58
+ vision_unfreeze_all: false
59
+ vision_unfreeze_last_n: 0
60
+ optimizer_cfg: !!python/object:icrt.util.args.OptimizerConfig
61
+ blr: 0.001
62
+ lr: 0.0005
63
+ min_lr: 0.0
64
+ warmup_epochs: 1.25
65
+ weight_decay: 0.01
66
+ shared_cfg: !!python/object:icrt.util.args.SharedConfig
67
+ batch_size: 1
68
+ num_cameras: 2
69
+ num_pred_steps: 16
70
+ num_stages: 1
71
+ num_visual_trace_points: 5
72
+ random_mask_visual_trace: true
73
+ resume: /data/tientoan/icrt_output/output_libero_90/no_prompt_loss_0_ver3/checkpoint-124.pth
74
+ rot_6d: true
75
+ save_every: 1
76
+ scale_action: null
77
+ seed: 0
78
+ seq_length: 512
79
+ split_epoch: 1
80
+ start_epoch: 0
81
+ use_delta_action: true
82
+ train: true
83
+ trainer_cfg: !!python/object:icrt.util.args.TrainerConfig
84
+ accum_iter: 64
85
+ epochs: 125
86
+ num_workers: 20
87
+ pin_memory: true
88
+ world_size: 1
train_split.json ADDED
The diff for this file is too large to render. See raw diff
 
val_split.json ADDED
@@ -0,0 +1 @@
 
 
1
+ ["STUDY_SCENE1_pick_up_the_book_and_place_it_in_the_left_compartment_of_the_caddy_21", "KITCHEN_SCENE10_put_the_black_bowl_in_the_top_drawer_of_the_cabinet_10", "STUDY_SCENE3_pick_up_the_book_and_place_it_in_the_front_compartment_of_the_caddy_6", "KITCHEN_SCENE2_stack_the_black_bowl_at_the_front_on_the_black_bowl_in_the_middle_34", "KITCHEN_SCENE2_put_the_black_bowl_at_the_front_on_the_plate_44", "KITCHEN_SCENE5_close_the_top_drawer_of_the_cabinet_6", "LIVING_ROOM_SCENE2_pick_up_the_milk_and_put_it_in_the_basket_6", "STUDY_SCENE3_pick_up_the_book_and_place_it_in_the_right_compartment_of_the_caddy_49", "LIVING_ROOM_SCENE3_pick_up_the_cream_cheese_and_put_it_in_the_tray_2", "LIVING_ROOM_SCENE3_pick_up_the_butter_and_put_it_in_the_tray_25", "KITCHEN_SCENE4_put_the_wine_bottle_in_the_bottom_drawer_of_the_cabinet_45", "KITCHEN_SCENE5_put_the_black_bowl_on_the_plate_46", "KITCHEN_SCENE6_put_the_yellow_and_white_mug_to_the_front_of_the_white_mug_43", "KITCHEN_SCENE4_put_the_wine_bottle_in_the_bottom_drawer_of_the_cabinet_35", "KITCHEN_SCENE10_put_the_chocolate_pudding_in_the_top_drawer_of_the_cabinet_and_close_it_45", "LIVING_ROOM_SCENE4_stack_the_left_bowl_on_the_right_bowl_and_place_them_in_the_tray_49", "KITCHEN_SCENE4_put_the_black_bowl_on_top_of_the_cabinet_31", "KITCHEN_SCENE4_put_the_wine_bottle_in_the_bottom_drawer_of_the_cabinet_42", "LIVING_ROOM_SCENE1_pick_up_the_alphabet_soup_and_put_it_in_the_basket_3", "KITCHEN_SCENE4_put_the_black_bowl_in_the_bottom_drawer_of_the_cabinet_31", "KITCHEN_SCENE1_open_the_bottom_drawer_of_the_cabinet_3", "LIVING_ROOM_SCENE4_stack_the_left_bowl_on_the_right_bowl_and_place_them_in_the_tray_15", "LIVING_ROOM_SCENE1_pick_up_the_alphabet_soup_and_put_it_in_the_basket_10", "KITCHEN_SCENE3_put_the_frying_pan_on_the_stove_48", "LIVING_ROOM_SCENE6_put_the_chocolate_pudding_to_the_right_of_the_plate_10", "KITCHEN_SCENE10_close_the_top_drawer_of_the_cabinet_and_put_the_black_bowl_on_top_of_it_48", "LIVING_ROOM_SCENE4_stack_the_left_bowl_on_the_right_bowl_and_place_them_in_the_tray_23", "LIVING_ROOM_SCENE6_put_the_chocolate_pudding_to_the_left_of_the_plate_49", "LIVING_ROOM_SCENE4_pick_up_the_chocolate_pudding_and_put_it_in_the_tray_43", "LIVING_ROOM_SCENE6_put_the_chocolate_pudding_to_the_right_of_the_plate_24", "LIVING_ROOM_SCENE6_put_the_chocolate_pudding_to_the_right_of_the_plate_0", "LIVING_ROOM_SCENE4_stack_the_left_bowl_on_the_right_bowl_and_place_them_in_the_tray_27", "KITCHEN_SCENE5_put_the_black_bowl_on_top_of_the_cabinet_38", "LIVING_ROOM_SCENE3_pick_up_the_ketchup_and_put_it_in_the_tray_45", "KITCHEN_SCENE10_put_the_butter_at_the_front_in_the_top_drawer_of_the_cabinet_and_close_it_26", "LIVING_ROOM_SCENE6_put_the_white_mug_on_the_plate_20", "LIVING_ROOM_SCENE4_pick_up_the_salad_dressing_and_put_it_in_the_tray_3", "KITCHEN_SCENE9_put_the_white_bowl_on_top_of_the_cabinet_9", "LIVING_ROOM_SCENE3_pick_up_the_tomato_sauce_and_put_it_in_the_tray_30", "KITCHEN_SCENE9_turn_on_the_stove_27", "LIVING_ROOM_SCENE1_pick_up_the_ketchup_and_put_it_in_the_basket_24", "LIVING_ROOM_SCENE4_pick_up_the_salad_dressing_and_put_it_in_the_tray_24", "KITCHEN_SCENE3_turn_on_the_stove_13", "STUDY_SCENE3_pick_up_the_book_and_place_it_in_the_right_compartment_of_the_caddy_3", "LIVING_ROOM_SCENE1_pick_up_the_cream_cheese_box_and_put_it_in_the_basket_47", "LIVING_ROOM_SCENE2_pick_up_the_tomato_sauce_and_put_it_in_the_basket_8", "KITCHEN_SCENE10_put_the_black_bowl_in_the_top_drawer_of_the_cabinet_43", "KITCHEN_SCENE8_put_the_right_moka_pot_on_the_stove_15", "LIVING_ROOM_SCENE1_pick_up_the_alphabet_soup_and_put_it_in_the_basket_21", "STUDY_SCENE2_pick_up_the_book_and_place_it_in_the_front_compartment_of_the_caddy_14", "KITCHEN_SCENE3_turn_on_the_stove_34", "KITCHEN_SCENE2_open_the_top_drawer_of_the_cabinet_8", "LIVING_ROOM_SCENE2_pick_up_the_tomato_sauce_and_put_it_in_the_basket_46", "KITCHEN_SCENE9_put_the_white_bowl_on_top_of_the_cabinet_17", "KITCHEN_SCENE6_close_the_microwave_47", "LIVING_ROOM_SCENE3_pick_up_the_alphabet_soup_and_put_it_in_the_tray_12", "KITCHEN_SCENE10_put_the_butter_at_the_front_in_the_top_drawer_of_the_cabinet_and_close_it_19", "LIVING_ROOM_SCENE5_put_the_red_mug_on_the_right_plate_44", "KITCHEN_SCENE9_turn_on_the_stove_and_put_the_frying_pan_on_it_16", "KITCHEN_SCENE1_open_the_top_drawer_of_the_cabinet_47", "KITCHEN_SCENE3_turn_on_the_stove_7", "STUDY_SCENE3_pick_up_the_book_and_place_it_in_the_left_compartment_of_the_caddy_45", "KITCHEN_SCENE7_put_the_white_bowl_on_the_plate_1", "LIVING_ROOM_SCENE3_pick_up_the_alphabet_soup_and_put_it_in_the_tray_16", "KITCHEN_SCENE2_stack_the_middle_black_bowl_on_the_back_black_bowl_33", "KITCHEN_SCENE9_put_the_white_bowl_on_top_of_the_cabinet_41", "KITCHEN_SCENE5_put_the_black_bowl_in_the_top_drawer_of_the_cabinet_16", "STUDY_SCENE4_pick_up_the_book_in_the_middle_and_place_it_on_the_cabinet_shelf_11", "KITCHEN_SCENE9_put_the_frying_pan_under_the_cabinet_shelf_36", "KITCHEN_SCENE10_put_the_chocolate_pudding_in_the_top_drawer_of_the_cabinet_and_close_it_27", "STUDY_SCENE3_pick_up_the_red_mug_and_place_it_to_the_right_of_the_caddy_19", "STUDY_SCENE1_pick_up_the_book_and_place_it_in_the_front_compartment_of_the_caddy_25", "LIVING_ROOM_SCENE6_put_the_red_mug_on_the_plate_12", "KITCHEN_SCENE5_put_the_ketchup_in_the_top_drawer_of_the_cabinet_29", "KITCHEN_SCENE7_put_the_white_bowl_to_the_right_of_the_plate_33", "LIVING_ROOM_SCENE1_pick_up_the_tomato_sauce_and_put_it_in_the_basket_25", "KITCHEN_SCENE7_open_the_microwave_27", "KITCHEN_SCENE5_put_the_black_bowl_in_the_top_drawer_of_the_cabinet_28", "KITCHEN_SCENE2_put_the_black_bowl_at_the_back_on_the_plate_11", "KITCHEN_SCENE9_put_the_frying_pan_under_the_cabinet_shelf_0", "KITCHEN_SCENE1_open_the_top_drawer_of_the_cabinet_9", "LIVING_ROOM_SCENE5_put_the_red_mug_on_the_right_plate_37", "LIVING_ROOM_SCENE4_pick_up_the_salad_dressing_and_put_it_in_the_tray_10", "LIVING_ROOM_SCENE4_stack_the_left_bowl_on_the_right_bowl_and_place_them_in_the_tray_35", "LIVING_ROOM_SCENE5_put_the_white_mug_on_the_left_plate_37", "STUDY_SCENE1_pick_up_the_book_and_place_it_in_the_left_compartment_of_the_caddy_22", "KITCHEN_SCENE7_put_the_white_bowl_to_the_right_of_the_plate_18", "KITCHEN_SCENE4_put_the_black_bowl_on_top_of_the_cabinet_47", "KITCHEN_SCENE9_turn_on_the_stove_and_put_the_frying_pan_on_it_17", "KITCHEN_SCENE4_put_the_wine_bottle_in_the_bottom_drawer_of_the_cabinet_40", "LIVING_ROOM_SCENE5_put_the_red_mug_on_the_left_plate_33", "KITCHEN_SCENE3_turn_on_the_stove_6", "KITCHEN_SCENE4_put_the_black_bowl_on_top_of_the_cabinet_34", "KITCHEN_SCENE9_turn_on_the_stove_and_put_the_frying_pan_on_it_29", "STUDY_SCENE4_pick_up_the_book_on_the_right_and_place_it_on_the_cabinet_shelf_47", "KITCHEN_SCENE4_close_the_bottom_drawer_of_the_cabinet_and_open_the_top_drawer_4", "LIVING_ROOM_SCENE5_put_the_white_mug_on_the_left_plate_48", "KITCHEN_SCENE2_stack_the_black_bowl_at_the_front_on_the_black_bowl_in_the_middle_38", "KITCHEN_SCENE1_open_the_bottom_drawer_of_the_cabinet_46", "LIVING_ROOM_SCENE2_pick_up_the_alphabet_soup_and_put_it_in_the_basket_12", "KITCHEN_SCENE10_close_the_top_drawer_of_the_cabinet_and_put_the_black_bowl_on_top_of_it_46", "LIVING_ROOM_SCENE4_pick_up_the_black_bowl_on_the_left_and_put_it_in_the_tray_21", "LIVING_ROOM_SCENE5_put_the_red_mug_on_the_right_plate_36", "KITCHEN_SCENE7_open_the_microwave_1", "LIVING_ROOM_SCENE3_pick_up_the_tomato_sauce_and_put_it_in_the_tray_46", "LIVING_ROOM_SCENE3_pick_up_the_alphabet_soup_and_put_it_in_the_tray_26", "KITCHEN_SCENE9_put_the_frying_pan_under_the_cabinet_shelf_8", "KITCHEN_SCENE10_close_the_top_drawer_of_the_cabinet_and_put_the_black_bowl_on_top_of_it_9", "LIVING_ROOM_SCENE6_put_the_chocolate_pudding_to_the_right_of_the_plate_46", "KITCHEN_SCENE1_put_the_black_bowl_on_the_plate_26", "LIVING_ROOM_SCENE4_stack_the_left_bowl_on_the_right_bowl_and_place_them_in_the_tray_24", "LIVING_ROOM_SCENE6_put_the_red_mug_on_the_plate_36", "LIVING_ROOM_SCENE2_pick_up_the_orange_juice_and_put_it_in_the_basket_1", "LIVING_ROOM_SCENE1_pick_up_the_alphabet_soup_and_put_it_in_the_basket_8", "KITCHEN_SCENE5_put_the_black_bowl_in_the_top_drawer_of_the_cabinet_42", "KITCHEN_SCENE10_put_the_chocolate_pudding_in_the_top_drawer_of_the_cabinet_and_close_it_28", "KITCHEN_SCENE3_put_the_moka_pot_on_the_stove_17", "STUDY_SCENE4_pick_up_the_book_on_the_right_and_place_it_on_the_cabinet_shelf_45", "KITCHEN_SCENE5_put_the_ketchup_in_the_top_drawer_of_the_cabinet_22", "LIVING_ROOM_SCENE6_put_the_chocolate_pudding_to_the_right_of_the_plate_1", "KITCHEN_SCENE7_put_the_white_bowl_on_the_plate_17", "KITCHEN_SCENE1_open_the_top_drawer_of_the_cabinet_and_put_the_bowl_in_it_25", "LIVING_ROOM_SCENE2_pick_up_the_tomato_sauce_and_put_it_in_the_basket_13", "KITCHEN_SCENE10_put_the_chocolate_pudding_in_the_top_drawer_of_the_cabinet_and_close_it_43", "KITCHEN_SCENE2_stack_the_black_bowl_at_the_front_on_the_black_bowl_in_the_middle_20", "STUDY_SCENE2_pick_up_the_book_and_place_it_in_the_right_compartment_of_the_caddy_5", "STUDY_SCENE3_pick_up_the_book_and_place_it_in_the_left_compartment_of_the_caddy_41", "KITCHEN_SCENE9_put_the_white_bowl_on_top_of_the_cabinet_11", "LIVING_ROOM_SCENE6_put_the_chocolate_pudding_to_the_right_of_the_plate_2", "KITCHEN_SCENE9_put_the_white_bowl_on_top_of_the_cabinet_43", "STUDY_SCENE4_pick_up_the_book_on_the_right_and_place_it_on_the_cabinet_shelf_34", "LIVING_ROOM_SCENE1_pick_up_the_ketchup_and_put_it_in_the_basket_29", "KITCHEN_SCENE2_stack_the_middle_black_bowl_on_the_back_black_bowl_19", "LIVING_ROOM_SCENE1_pick_up_the_alphabet_soup_and_put_it_in_the_basket_26", "KITCHEN_SCENE4_put_the_wine_bottle_in_the_bottom_drawer_of_the_cabinet_43", "LIVING_ROOM_SCENE5_put_the_red_mug_on_the_right_plate_0", "STUDY_SCENE2_pick_up_the_book_and_place_it_in_the_right_compartment_of_the_caddy_12", "LIVING_ROOM_SCENE3_pick_up_the_cream_cheese_and_put_it_in_the_tray_40", "LIVING_ROOM_SCENE4_pick_up_the_chocolate_pudding_and_put_it_in_the_tray_15", "KITCHEN_SCENE10_put_the_butter_at_the_back_in_the_top_drawer_of_the_cabinet_and_close_it_18", "LIVING_ROOM_SCENE2_pick_up_the_alphabet_soup_and_put_it_in_the_basket_6", "KITCHEN_SCENE2_put_the_middle_black_bowl_on_top_of_the_cabinet_26", "LIVING_ROOM_SCENE3_pick_up_the_ketchup_and_put_it_in_the_tray_27", "KITCHEN_SCENE2_open_the_top_drawer_of_the_cabinet_41", "KITCHEN_SCENE2_put_the_black_bowl_at_the_front_on_the_plate_46", "LIVING_ROOM_SCENE4_pick_up_the_chocolate_pudding_and_put_it_in_the_tray_36", "STUDY_SCENE1_pick_up_the_book_and_place_it_in_the_right_compartment_of_the_caddy_18", "STUDY_SCENE1_pick_up_the_book_and_place_it_in_the_right_compartment_of_the_caddy_7", "KITCHEN_SCENE1_put_the_black_bowl_on_the_plate_2", "LIVING_ROOM_SCENE5_put_the_red_mug_on_the_right_plate_38", "LIVING_ROOM_SCENE6_put_the_white_mug_on_the_plate_2", "KITCHEN_SCENE2_put_the_middle_black_bowl_on_the_plate_2", "KITCHEN_SCENE2_stack_the_middle_black_bowl_on_the_back_black_bowl_13", "STUDY_SCENE4_pick_up_the_book_on_the_right_and_place_it_under_the_cabinet_shelf_20", "LIVING_ROOM_SCENE6_put_the_white_mug_on_the_plate_5", "KITCHEN_SCENE9_turn_on_the_stove_46", "KITCHEN_SCENE10_put_the_black_bowl_in_the_top_drawer_of_the_cabinet_15", "LIVING_ROOM_SCENE5_put_the_red_mug_on_the_left_plate_14", "LIVING_ROOM_SCENE3_pick_up_the_butter_and_put_it_in_the_tray_29", "STUDY_SCENE2_pick_up_the_book_and_place_it_in_the_left_compartment_of_the_caddy_30", "KITCHEN_SCENE9_put_the_frying_pan_on_top_of_the_cabinet_49", "KITCHEN_SCENE9_turn_on_the_stove_and_put_the_frying_pan_on_it_41", "STUDY_SCENE4_pick_up_the_book_in_the_middle_and_place_it_on_the_cabinet_shelf_3", "KITCHEN_SCENE2_stack_the_black_bowl_at_the_front_on_the_black_bowl_in_the_middle_2", "KITCHEN_SCENE3_turn_on_the_stove_and_put_the_frying_pan_on_it_10", "LIVING_ROOM_SCENE1_pick_up_the_ketchup_and_put_it_in_the_basket_17", "KITCHEN_SCENE9_put_the_white_bowl_on_top_of_the_cabinet_27", "LIVING_ROOM_SCENE4_stack_the_left_bowl_on_the_right_bowl_and_place_them_in_the_tray_22", "LIVING_ROOM_SCENE1_pick_up_the_cream_cheese_box_and_put_it_in_the_basket_22", "LIVING_ROOM_SCENE5_put_the_red_mug_on_the_left_plate_1", "KITCHEN_SCENE2_open_the_top_drawer_of_the_cabinet_35", "KITCHEN_SCENE7_put_the_white_bowl_to_the_right_of_the_plate_36", "LIVING_ROOM_SCENE5_put_the_red_mug_on_the_left_plate_13", "STUDY_SCENE3_pick_up_the_red_mug_and_place_it_to_the_right_of_the_caddy_13", "LIVING_ROOM_SCENE1_pick_up_the_tomato_sauce_and_put_it_in_the_basket_3", "LIVING_ROOM_SCENE1_pick_up_the_cream_cheese_box_and_put_it_in_the_basket_36", "LIVING_ROOM_SCENE3_pick_up_the_ketchup_and_put_it_in_the_tray_35", "STUDY_SCENE2_pick_up_the_book_and_place_it_in_the_right_compartment_of_the_caddy_46", "KITCHEN_SCENE2_put_the_middle_black_bowl_on_top_of_the_cabinet_18", "LIVING_ROOM_SCENE2_pick_up_the_orange_juice_and_put_it_in_the_basket_37", "STUDY_SCENE2_pick_up_the_book_and_place_it_in_the_front_compartment_of_the_caddy_35", "LIVING_ROOM_SCENE3_pick_up_the_tomato_sauce_and_put_it_in_the_tray_36", "STUDY_SCENE2_pick_up_the_book_and_place_it_in_the_right_compartment_of_the_caddy_26", "LIVING_ROOM_SCENE6_put_the_chocolate_pudding_to_the_right_of_the_plate_35", "LIVING_ROOM_SCENE3_pick_up_the_tomato_sauce_and_put_it_in_the_tray_35", "KITCHEN_SCENE9_put_the_frying_pan_on_the_cabinet_shelf_40", "KITCHEN_SCENE8_turn_off_the_stove_27", "KITCHEN_SCENE10_put_the_chocolate_pudding_in_the_top_drawer_of_the_cabinet_and_close_it_47", "STUDY_SCENE4_pick_up_the_book_in_the_middle_and_place_it_on_the_cabinet_shelf_42", "KITCHEN_SCENE4_close_the_bottom_drawer_of_the_cabinet_4", "STUDY_SCENE2_pick_up_the_book_and_place_it_in_the_back_compartment_of_the_caddy_32", "KITCHEN_SCENE8_put_the_right_moka_pot_on_the_stove_23", "KITCHEN_SCENE2_stack_the_black_bowl_at_the_front_on_the_black_bowl_in_the_middle_28", "KITCHEN_SCENE3_put_the_frying_pan_on_the_stove_10", "STUDY_SCENE1_pick_up_the_book_and_place_it_in_the_front_compartment_of_the_caddy_20", "KITCHEN_SCENE7_put_the_white_bowl_on_the_plate_29", "LIVING_ROOM_SCENE3_pick_up_the_ketchup_and_put_it_in_the_tray_42", "LIVING_ROOM_SCENE4_pick_up_the_chocolate_pudding_and_put_it_in_the_tray_23"]