ChihHanShen commited on
Commit
a52f7a6
·
verified ·
1 Parent(s): 492e99f

Upload folder using huggingface_hub

Browse files
Files changed (20) hide show
  1. .gitattributes +1 -0
  2. checkpoints/finetune_task48_2000step/checkpoints/steps_1000_pytorch_model.pt +3 -0
  3. checkpoints/finetune_task48_2000step/checkpoints/steps_1500_pytorch_model.pt +3 -0
  4. checkpoints/finetune_task48_2000step/checkpoints/steps_2000_pytorch_model.pt +3 -0
  5. checkpoints/finetune_task48_2000step/checkpoints/steps_500_pytorch_model.pt +3 -0
  6. checkpoints/finetune_task48_2000step/config.yaml +48 -0
  7. checkpoints/finetune_task48_2000step/dataset_statistics.json +133 -0
  8. checkpoints/finetune_task48_2000step/final_model/pytorch_model.pt +3 -0
  9. checkpoints/finetune_task48_2000step/summary.jsonl +4 -0
  10. checkpoints/finetune_task48_2000step/wandb/wandb/debug-internal.log +12 -0
  11. checkpoints/finetune_task48_2000step/wandb/wandb/debug.log +0 -0
  12. checkpoints/finetune_task48_2000step/wandb/wandb/run-20260316_102706-w0ljmpp2/files/config.yaml +73 -0
  13. checkpoints/finetune_task48_2000step/wandb/wandb/run-20260316_102706-w0ljmpp2/files/output.log +234 -0
  14. checkpoints/finetune_task48_2000step/wandb/wandb/run-20260316_102706-w0ljmpp2/files/requirements.txt +190 -0
  15. checkpoints/finetune_task48_2000step/wandb/wandb/run-20260316_102706-w0ljmpp2/files/wandb-metadata.json +48 -0
  16. checkpoints/finetune_task48_2000step/wandb/wandb/run-20260316_102706-w0ljmpp2/files/wandb-summary.json +1 -0
  17. checkpoints/finetune_task48_2000step/wandb/wandb/run-20260316_102706-w0ljmpp2/logs/debug-core.log +19 -0
  18. checkpoints/finetune_task48_2000step/wandb/wandb/run-20260316_102706-w0ljmpp2/logs/debug-internal.log +12 -0
  19. checkpoints/finetune_task48_2000step/wandb/wandb/run-20260316_102706-w0ljmpp2/logs/debug.log +0 -0
  20. checkpoints/finetune_task48_2000step/wandb/wandb/run-20260316_102706-w0ljmpp2/run-w0ljmpp2.wandb +3 -0
.gitattributes CHANGED
@@ -38,3 +38,4 @@ checkpoints/wandb/wandb/run-20260316_085932-77uivys0/run-77uivys0.wandb filter=l
38
  checkpoints/pretrained_goal_2000step/wandb/wandb/run-20260316_073559-h1hybozy/run-h1hybozy.wandb filter=lfs diff=lfs merge=lfs -text
39
  checkpoints/finetune_task2_2000step/wandb/wandb/run-20260316_085932-77uivys0/run-77uivys0.wandb filter=lfs diff=lfs merge=lfs -text
40
  checkpoints/finetune_task33_2000step/wandb/wandb/run-20260316_092332-e1rd7b68/run-e1rd7b68.wandb filter=lfs diff=lfs merge=lfs -text
 
 
38
  checkpoints/pretrained_goal_2000step/wandb/wandb/run-20260316_073559-h1hybozy/run-h1hybozy.wandb filter=lfs diff=lfs merge=lfs -text
39
  checkpoints/finetune_task2_2000step/wandb/wandb/run-20260316_085932-77uivys0/run-77uivys0.wandb filter=lfs diff=lfs merge=lfs -text
40
  checkpoints/finetune_task33_2000step/wandb/wandb/run-20260316_092332-e1rd7b68/run-e1rd7b68.wandb filter=lfs diff=lfs merge=lfs -text
41
+ checkpoints/finetune_task48_2000step/wandb/wandb/run-20260316_102706-w0ljmpp2/run-w0ljmpp2.wandb filter=lfs diff=lfs merge=lfs -text
checkpoints/finetune_task48_2000step/checkpoints/steps_1000_pytorch_model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:19f6c920dede8cff547f3face56e4946aeb0e1e0f6a452f78091508095362565
3
+ size 8146438221
checkpoints/finetune_task48_2000step/checkpoints/steps_1500_pytorch_model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0a3274bfec6045a79745cf1f8797c75df703ae2416caec64b5ebfa4afe5db9be
3
+ size 8146438221
checkpoints/finetune_task48_2000step/checkpoints/steps_2000_pytorch_model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:81fb4977e54cbfba7e68e872b6b496d08e40a4fc433e5182fc3fadc8b5ef8eae
3
+ size 8146438221
checkpoints/finetune_task48_2000step/checkpoints/steps_500_pytorch_model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:879914e55b87d2586b1a0581363bfd875007ad40a30ed8a9a99fe29f4d9958a6
3
+ size 8146437392
checkpoints/finetune_task48_2000step/config.yaml ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ datasets:
2
+ vla_data:
3
+ CoT_prompt: Your task is {instruction}. To identify the key objects for your task.
4
+ Locate their bounding boxes in [x1,y1,x2,y2] format.
5
+ data_mix: libero_90_task_48
6
+ data_root_dir: playground/Datasets/LEROBOT_LIBERO_DATA
7
+ dataset_py: lerobot_datasets
8
+ per_device_batch_size: 1
9
+ sequential_step_sampling: false
10
+ video_backend: torchvision_av
11
+ framework:
12
+ action_model:
13
+ action_dim: 7
14
+ future_action_window_size: 7
15
+ past_action_window_size: 0
16
+ name: QwenFast
17
+ qwenvl:
18
+ base_vlm: playground/Pretrained_models/Qwen2.5-VL-3B-Instruct-Action
19
+ output_dir: ./results/Checkpoints/finetune_task48_2000step
20
+ run_id: finetune_task48_2000step
21
+ run_root_dir: ./results/Checkpoints
22
+ seed: 42
23
+ trainer:
24
+ eval_interval: 100
25
+ freeze_modules: qwen_vl_interface.model.model.visual,dino_encoder
26
+ gradient_accumulation_steps: 1
27
+ gradient_clipping: 1.0
28
+ is_resume: true
29
+ learning_rate:
30
+ action_model: 0.0001
31
+ base: 2.5e-05
32
+ qwen_vl_interface: 1.0e-05
33
+ logging_frequency: 100
34
+ lr_scheduler_type: cosine_with_min_lr
35
+ max_train_steps: 2000
36
+ num_warmup_steps: 5000
37
+ optimizer:
38
+ betas:
39
+ - 0.9
40
+ - 0.95
41
+ eps: 1.0e-08
42
+ weight_decay: 1.0e-08
43
+ pretrained_checkpoint: /content/starVLA_r/results/Checkpoints/Qwen2.5-VL-FAST-LIBERO-4in1/checkpoints/steps_30000_pytorch_model.pt
44
+ save_interval: 500
45
+ scheduler_specific_kwargs:
46
+ min_lr: 1.0e-06
47
+ wandb_entity: michellelin9102-usc
48
+ wandb_project: starVLA_Libero
checkpoints/finetune_task48_2000step/dataset_statistics.json ADDED
@@ -0,0 +1,133 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "franka": {
3
+ "action": {
4
+ "mean": [
5
+ 0.04281630367040634,
6
+ 0.16635118424892426,
7
+ -0.10021582990884781,
8
+ 0.0008796637994237244,
9
+ 0.009483124129474163,
10
+ -0.0018370545003563166,
11
+ 0.48476770520210266
12
+ ],
13
+ "std": [
14
+ 0.23814928531646729,
15
+ 0.48973721265792847,
16
+ 0.4945138692855835,
17
+ 0.03829231485724449,
18
+ 0.048698168247938156,
19
+ 0.031572677195072174,
20
+ 0.4997621178627014
21
+ ],
22
+ "max": [
23
+ 0.8142856955528259,
24
+ 0.9375,
25
+ 0.9375,
26
+ 0.1971428543329239,
27
+ 0.3385714292526245,
28
+ 0.1703571379184723,
29
+ 1.0
30
+ ],
31
+ "min": [
32
+ -0.7392857074737549,
33
+ -0.9375,
34
+ -0.9375,
35
+ -0.17571428418159485,
36
+ -0.21642857789993286,
37
+ -0.2442857176065445,
38
+ 0.0
39
+ ],
40
+ "q01": [
41
+ -0.5450624823570251,
42
+ -0.8571428656578064,
43
+ -0.9375,
44
+ -0.10607142746448517,
45
+ -0.11142857372760773,
46
+ -0.09857142716646194,
47
+ 0.0
48
+ ],
49
+ "q99": [
50
+ 0.6401785612106323,
51
+ 0.9375,
52
+ 0.9375,
53
+ 0.11249999701976776,
54
+ 0.1633823961019516,
55
+ 0.0798109620809555,
56
+ 1.0
57
+ ],
58
+ "mask": [
59
+ true,
60
+ true,
61
+ true,
62
+ true,
63
+ true,
64
+ true,
65
+ false
66
+ ]
67
+ },
68
+ "state": {
69
+ "mean": [
70
+ -0.0446692518889904,
71
+ -0.009891818277537823,
72
+ 0.5988359451293945,
73
+ 3.129826545715332,
74
+ -0.04531480744481087,
75
+ -0.08344470709562302,
76
+ 0.03376416116952896,
77
+ -0.03510740399360657
78
+ ],
79
+ "std": [
80
+ 0.03938385099172592,
81
+ 0.15407362580299377,
82
+ 0.07102718949317932,
83
+ 0.057501792907714844,
84
+ 0.13009506464004517,
85
+ 0.13546207547187805,
86
+ 0.005857877433300018,
87
+ 0.005015654023736715
88
+ ],
89
+ "max": [
90
+ 0.08624760061502457,
91
+ 0.32907697558403015,
92
+ 0.7157319784164429,
93
+ 3.3233284950256348,
94
+ 0.5133177638053894,
95
+ 0.3517029881477356,
96
+ 0.04034854471683502,
97
+ -0.017706822603940964
98
+ ],
99
+ "min": [
100
+ -0.13770002126693726,
101
+ -0.1906675398349762,
102
+ 0.47950780391693115,
103
+ 2.9164979457855225,
104
+ -0.6462156176567078,
105
+ -0.6787144541740417,
106
+ 0.013480174355208874,
107
+ -0.041075896471738815
108
+ ],
109
+ "q01": [
110
+ -0.12296133488416672,
111
+ -0.18305309116840363,
112
+ 0.48121508955955505,
113
+ 2.990795373916626,
114
+ -0.438663512468338,
115
+ -0.45049262046813965,
116
+ 0.017378225922584534,
117
+ -0.04001536965370178
118
+ ],
119
+ "q99": [
120
+ 0.0570642463862896,
121
+ 0.2863735556602478,
122
+ 0.6942169070243835,
123
+ 3.2584145069122314,
124
+ 0.28781232237815857,
125
+ 0.28255322575569153,
126
+ 0.039883945137262344,
127
+ -0.01976053975522518
128
+ ]
129
+ },
130
+ "num_transitions": 5252,
131
+ "num_trajectories": 44
132
+ }
133
+ }
checkpoints/finetune_task48_2000step/final_model/pytorch_model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4f765bd031428635d4adc7c885f0434fa5c853d7ec429e36e4e81039919d80dc
3
+ size 8146425390
checkpoints/finetune_task48_2000step/summary.jsonl ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {"steps": 500}
2
+ {"steps": 1000}
3
+ {"steps": 1500}
4
+ {"steps": 2000}
checkpoints/finetune_task48_2000step/wandb/wandb/debug-internal.log ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2026-03-16T10:27:07.464425119Z","level":"INFO","msg":"stream: starting","core version":"0.24.2"}
2
+ {"time":"2026-03-16T10:27:07.83079351Z","level":"INFO","msg":"stream: created new stream","id":"w0ljmpp2"}
3
+ {"time":"2026-03-16T10:27:07.833232151Z","level":"INFO","msg":"handler: started","stream_id":"w0ljmpp2"}
4
+ {"time":"2026-03-16T10:27:07.833403373Z","level":"INFO","msg":"stream: started","id":"w0ljmpp2"}
5
+ {"time":"2026-03-16T10:27:07.833477504Z","level":"INFO","msg":"writer: started","stream_id":"w0ljmpp2"}
6
+ {"time":"2026-03-16T10:27:07.833576847Z","level":"INFO","msg":"sender: started","stream_id":"w0ljmpp2"}
7
+ {"time":"2026-03-16T10:50:35.313300483Z","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
8
+ {"time":"2026-03-16T10:50:35.636464134Z","level":"INFO","msg":"handler: operation stats","stats":{}}
9
+ {"time":"2026-03-16T10:50:35.641152876Z","level":"INFO","msg":"stream: closing","id":"w0ljmpp2"}
10
+ {"time":"2026-03-16T10:50:35.641183005Z","level":"INFO","msg":"handler: closed","stream_id":"w0ljmpp2"}
11
+ {"time":"2026-03-16T10:50:35.641292294Z","level":"INFO","msg":"sender: closed","stream_id":"w0ljmpp2"}
12
+ {"time":"2026-03-16T10:50:35.641305178Z","level":"INFO","msg":"stream: closed","id":"w0ljmpp2"}
checkpoints/finetune_task48_2000step/wandb/wandb/debug.log ADDED
File without changes
checkpoints/finetune_task48_2000step/wandb/wandb/run-20260316_102706-w0ljmpp2/files/config.yaml ADDED
@@ -0,0 +1,73 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ _wandb:
2
+ value:
3
+ cli_version: 0.24.2
4
+ e:
5
+ xjkqn2u8dse3qnn37usztgd1uc0n91bo:
6
+ args:
7
+ - --config_yaml
8
+ - ./examples/LIBERO/train_files/my_libero_finetune.yaml
9
+ - --datasets.vla_data.data_mix
10
+ - libero_90_task_48
11
+ - --run_id
12
+ - finetune_task48_2000step
13
+ codePath: starVLA/training/train_starvla.py
14
+ codePathLocal: starVLA/training/train_starvla.py
15
+ cpu_count: 6
16
+ cpu_count_logical: 12
17
+ cudaVersion: "13.0"
18
+ disk:
19
+ /:
20
+ total: "253055008768"
21
+ used: "155059826688"
22
+ email: chihhans@usc.edu
23
+ executable: /usr/local/envs/starvla/bin/python3.10
24
+ git:
25
+ commit: e952c81219e9fac2c3183a27cd378e592c4c9ef0
26
+ remote: https://github.com/tliao730/starVLA_r
27
+ gpu: NVIDIA A100-SXM4-80GB
28
+ gpu_count: 1
29
+ gpu_nvidia:
30
+ - architecture: Ampere
31
+ cudaCores: 6912
32
+ memoryTotal: "85899345920"
33
+ name: NVIDIA A100-SXM4-80GB
34
+ uuid: GPU-1000e8c7-f9d7-74b0-8fdb-aad3f6d24e69
35
+ host: c89e62d63bf0
36
+ memory:
37
+ total: "179370471424"
38
+ os: Linux-6.6.113+-x86_64-with-glibc2.35
39
+ program: /content/starVLA_r/starVLA/training/train_starvla.py
40
+ python: CPython 3.10.20
41
+ root: ./results/Checkpoints/finetune_task48_2000step/wandb
42
+ startedAt: "2026-03-16T10:27:06.946692Z"
43
+ writerId: xjkqn2u8dse3qnn37usztgd1uc0n91bo
44
+ m: []
45
+ python_version: 3.10.20
46
+ t:
47
+ "1":
48
+ - 1
49
+ - 11
50
+ - 41
51
+ - 49
52
+ - 63
53
+ - 71
54
+ - 80
55
+ - 83
56
+ "2":
57
+ - 1
58
+ - 11
59
+ - 41
60
+ - 49
61
+ - 63
62
+ - 71
63
+ - 80
64
+ - 83
65
+ "3":
66
+ - 2
67
+ - 13
68
+ - 61
69
+ "4": 3.10.20
70
+ "5": 0.24.2
71
+ "6": 4.57.0
72
+ "12": 0.24.2
73
+ "13": linux-x86_64
checkpoints/finetune_task48_2000step/wandb/wandb/run-20260316_102706-w0ljmpp2/files/output.log ADDED
@@ -0,0 +1,234 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 03/16 [10:27:08] INFO | >> [RANK 0] ***** Training train_starvla.py:326
2
+ Configuration *****
3
+ INFO | >> [RANK 0] Total train_starvla.py:327
4
+ optimization steps = 2000
5
+ INFO | >> [RANK 0] Per device batch train_starvla.py:328
6
+ size = 1
7
+ INFO | >> [RANK 0] Gradient train_starvla.py:329
8
+ accumulation steps = 1
9
+ INFO | >> [RANK 0] Total batch size train_starvla.py:330
10
+ = 1
11
+ 20%|██ | 400/2000 [04:19<16:53, 1.58it/s, data_times=0.002, model_times=0.628]
12
+ 03/16 [10:28:15] INFO | >> [RANK 0] Step 100, Loss: train_starvla.py:242
13
+ {'action_dit_loss':
14
+ 9.314701080322266, 'mse_score':
15
+ np.float64(0.034079255802760465),
16
+ 'data_time':
17
+ 0.0013499019987648353,
18
+ 'model_time': 0.635298358,
19
+ 'learning_rate':
20
+ 2.0000000000000002e-07, 'epoch':
21
+ 0.02})
22
+ 03/16 [10:29:19] INFO | >> [RANK 0] Step 200, Loss: train_starvla.py:242
23
+ {'action_dit_loss':
24
+ 7.929775714874268, 'mse_score':
25
+ np.float64(0.03428953959591426),
26
+ 'data_time':
27
+ 0.0005867119998583803,
28
+ 'model_time': 0.6362442019999435,
29
+ 'learning_rate':
30
+ 4.0000000000000003e-07, 'epoch':
31
+ 0.04})
32
+ 03/16 [10:30:24] INFO | >> [RANK 0] Step 300, Loss: train_starvla.py:242
33
+ {'action_dit_loss':
34
+ 4.264825344085693, 'mse_score':
35
+ np.float64(0.036809627059651094),
36
+ 'data_time':
37
+ 0.0006798820013500517,
38
+ 'model_time': 0.6313770140004635,
39
+ 'learning_rate':
40
+ 6.000000000000001e-07, 'epoch':
41
+ 0.06})
42
+ 03/16 [10:31:28] INFO | >> [RANK 0] Step 400, Loss: train_starvla.py:242
43
+ {'action_dit_loss':
44
+ 5.90472936630249, 'mse_score':
45
+ np.float64(0.028770231955113468),
46
+ 'data_time':
47
+ 0.0018893399992521154,
48
+ 'model_time': 0.6276594850005495,
49
+ 'learning_rate':
50
+ 8.000000000000001e-07, 'epoch':
51
+ 0.08})
52
+ 03/16 [10:32:32] INFO | >> [RANK 0] Step 500, Loss: train_starvla.py:242
53
+ {'action_dit_loss':
54
+ 4.008138656616211, 'mse_score':
55
+ np.float64(0.021432459692567525),
56
+ 'data_time':
57
+ 0.0005417720003606519,
58
+ 'model_time': 0.6293254800002615,
59
+ 'learning_rate':
60
+ 1.0000000000000002e-06, 'epoch':
61
+ 0.1})
62
+ ✅ Checkpoint saved at ./results/Checkpoints/finetune_task48_2000step/checkpoints/steps_500
63
+ 03/16 [10:32:53] INFO | >> [RANK 0] 📊 Saving accessed train_starvla.py:229
64
+ configuration...
65
+ INFO | >> [RANK 0] ✅ Configuration train_starvla.py:232
66
+ files saved
67
+ 03/16 [10:33:58] INFO | >> [RANK 0] Step 600, Loss: train_starvla.py:242
68
+ {'action_dit_loss':
69
+ 2.6089346408843994, 'mse_score':
70
+ np.float64(0.018183997162692193),
71
+ 'data_time':
72
+ 0.0017828830004873453,
73
+ 'model_time': 0.6282775969993963,
74
+ 'learning_rate':
75
+ 1.2000000000000002e-06, 'epoch':
76
+ 0.11})
77
+ 03/16 [10:35:03] INFO | >> [RANK 0] Step 700, Loss: train_starvla.py:242
78
+ {'action_dit_loss':
79
+ 2.386871814727783, 'mse_score':
80
+ np.float64(0.02084296110337058),
81
+ 'data_time':
82
+ 0.0006294380000326782,
83
+ 'model_time': 0.6310159410004417,
84
+ 'learning_rate':
85
+ 1.4000000000000001e-06, 'epoch':
86
+ 0.13})
87
+ 03/16 [10:36:07] INFO | >> [RANK 0] Step 800, Loss: train_starvla.py:242
88
+ {'action_dit_loss':
89
+ 2.1533708572387695, 'mse_score':
90
+ np.float64(0.02266200388939163),
91
+ 'data_time':
92
+ 0.0002977069998451043,
93
+ 'model_time': 0.6323083229999611,
94
+ 'learning_rate':
95
+ 1.6000000000000001e-06, 'epoch':
96
+ 0.15})
97
+ 03/16 [10:37:12] INFO | >> [RANK 0] Step 900, Loss: train_starvla.py:242
98
+ {'action_dit_loss':
99
+ 3.9568512439727783, 'mse_score':
100
+ np.float64(0.010319906140672347),
101
+ 'data_time':
102
+ 0.00023809100093785673,
103
+ 'model_time': 0.6301700330004678,
104
+ 'learning_rate':
105
+ 1.8000000000000001e-06, 'epoch':
106
+ 0.17})
107
+ 03/16 [10:38:16] INFO | >> [RANK 0] Step 1000, Loss: train_starvla.py:242
108
+ {'action_dit_loss':
109
+ 2.959420680999756, 'mse_score':
110
+ np.float64(0.018102484443056295),
111
+ 'data_time':
112
+ 0.002947425999082043,
113
+ 'model_time': 0.6312376300011238,
114
+ 'learning_rate':
115
+ 2.0000000000000003e-06, 'epoch':
116
+ 0.19})
117
+ ✅ Checkpoint saved at ./results/Checkpoints/finetune_task48_2000step/checkpoints/steps_1000
118
+ 03/16 [10:38:37] INFO | >> [RANK 0] 📊 Saving accessed train_starvla.py:229
119
+ configuration...
120
+ INFO | >> [RANK 0] ✅ Configuration train_starvla.py:232
121
+ files saved
122
+ 03/16 [10:39:43] INFO | >> [RANK 0] Step 1100, Loss: train_starvla.py:242
123
+ {'action_dit_loss':
124
+ 2.4097423553466797, 'mse_score':
125
+ np.float64(0.010958415389495369),
126
+ 'data_time':
127
+ 0.0004303599998820573,
128
+ 'model_time': 0.629223395999361,
129
+ 'learning_rate': 2.2e-06,
130
+ 'epoch': 0.21})
131
+ 03/16 [10:40:47] INFO | >> [RANK 0] Step 1200, Loss: train_starvla.py:242
132
+ {'action_dit_loss':
133
+ 1.8001140356063843, 'mse_score':
134
+ np.float64(0.016425216646615573),
135
+ 'data_time':
136
+ 0.0017891709994728444,
137
+ 'model_time': 0.6432768380000198,
138
+ 'learning_rate':
139
+ 2.4000000000000003e-06, 'epoch':
140
+ 0.23})
141
+ 03/16 [10:41:52] INFO | >> [RANK 0] Step 1300, Loss: train_starvla.py:242
142
+ {'action_dit_loss':
143
+ 2.450651168823242, 'mse_score':
144
+ np.float64(0.03609608879081504),
145
+ 'data_time':
146
+ 0.0006008899999869755,
147
+ 'model_time': 0.6404236750004202,
148
+ 'learning_rate': 2.6e-06,
149
+ 'epoch': 0.25})
150
+ 03/16 [10:42:57] INFO | >> [RANK 0] Step 1400, Loss: train_starvla.py:242
151
+ {'action_dit_loss':
152
+ 1.3520110845565796, 'mse_score':
153
+ np.float64(0.015366809051465815),
154
+ 'data_time':
155
+ 0.0019384579991310602,
156
+ 'model_time': 0.6324510970007395,
157
+ 'learning_rate':
158
+ 2.8000000000000003e-06, 'epoch':
159
+ 0.27})
160
+ 03/16 [10:44:02] INFO | >> [RANK 0] Step 1500, Loss: train_starvla.py:242
161
+ {'action_dit_loss':
162
+ 1.9549362659454346, 'mse_score':
163
+ np.float64(0.027217211106191984),
164
+ 'data_time':
165
+ 0.000567299997783266,
166
+ 'model_time': 0.633298984997964,
167
+ 'learning_rate': 3e-06, 'epoch':
168
+ 0.29})
169
+ ✅ Checkpoint saved at ./results/Checkpoints/finetune_task48_2000step/checkpoints/steps_1500
170
+ 03/16 [10:44:27] INFO | >> [RANK 0] 📊 Saving accessed train_starvla.py:229
171
+ configuration...
172
+ INFO | >> [RANK 0] ✅ Configuration train_starvla.py:232
173
+ files saved
174
+ 03/16 [10:45:32] INFO | >> [RANK 0] Step 1600, Loss: train_starvla.py:242
175
+ {'action_dit_loss':
176
+ 4.203729152679443, 'mse_score':
177
+ np.float64(0.029437476644757394),
178
+ 'data_time':
179
+ 0.002075060001516249,
180
+ 'model_time': 0.633118129997456,
181
+ 'learning_rate':
182
+ 3.2000000000000003e-06, 'epoch':
183
+ 0.3})
184
+ 03/16 [10:46:36] INFO | >> [RANK 0] Step 1700, Loss: train_starvla.py:242
185
+ {'action_dit_loss':
186
+ 3.06008243560791, 'mse_score':
187
+ np.float64(0.018865567485130245),
188
+ 'data_time':
189
+ 0.0007922959994175471,
190
+ 'model_time': 0.6446109249991423,
191
+ 'learning_rate':
192
+ 3.4000000000000005e-06, 'epoch':
193
+ 0.32})
194
+ 03/16 [10:47:41] INFO | >> [RANK 0] Step 1800, Loss: train_starvla.py:242
195
+ {'action_dit_loss':
196
+ 1.5572900772094727, 'mse_score':
197
+ np.float64(0.03312688171523522),
198
+ 'data_time':
199
+ 0.0018809510002029128,
200
+ 'model_time': 0.6576305890012009,
201
+ 'learning_rate':
202
+ 3.6000000000000003e-06, 'epoch':
203
+ 0.34})
204
+ Error decoding tokens: cannot reshape array of size 55 into shape (7)
205
+ Tokens: [266, 393, 272, 610, 1589, 277, 290, 310, 425, 1084, 261, 256]
206
+ 03/16 [10:48:46] INFO | >> [RANK 0] Step 1900, Loss: train_starvla.py:242
207
+ {'action_dit_loss':
208
+ 3.1513538360595703, 'mse_score':
209
+ np.float64(0.06123608522253369),
210
+ 'data_time':
211
+ 0.0007130379999580327,
212
+ 'model_time': 0.6419669289971353,
213
+ 'learning_rate':
214
+ 3.8000000000000005e-06, 'epoch':
215
+ 0.36})
216
+ 03/16 [10:49:50] INFO | >> [RANK 0] Step 2000, Loss: train_starvla.py:242
217
+ {'action_dit_loss':
218
+ 1.4614311456680298, 'mse_score':
219
+ np.float64(0.018765023958714284),
220
+ 'data_time':
221
+ 0.002591219999885652,
222
+ 'model_time': 0.631486190999567,
223
+ 'learning_rate':
224
+ 4.000000000000001e-06, 'epoch':
225
+ 0.38})
226
+ ✅ Checkpoint saved at ./results/Checkpoints/finetune_task48_2000step/checkpoints/steps_2000
227
+ 03/16 [10:50:11] INFO | >> [RANK 0] 📊 Saving accessed train_starvla.py:229
228
+ configuration...
229
+ INFO | >> [RANK 0] ✅ Configuration train_starvla.py:232
230
+ files saved
231
+ 03/16 [10:50:33] INFO | >> [RANK 0] Training complete. train_starvla.py:369
232
+ Final model saved at
233
+ ./results/Checkpoints/finetune_ta
234
+ sk48_2000step/final_model
checkpoints/finetune_task48_2000step/wandb/wandb/run-20260316_102706-w0ljmpp2/files/requirements.txt ADDED
@@ -0,0 +1,190 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ starVLA==1.0.1
2
+ grpcio==1.78.0
3
+ albucore==0.0.17
4
+ qwen-vl-utils==0.0.14
5
+ GitPython==3.1.46
6
+ huggingface-hub==0.35.3
7
+ transformers-stream-generator==0.0.4
8
+ httpcore==1.0.9
9
+ einops==0.8.2
10
+ mpmath==1.3.0
11
+ accelerate==1.13.0
12
+ nvidia-cusparselt-cu12==0.7.1
13
+ psutil==7.2.2
14
+ tabulate==0.10.0
15
+ nvidia-cudnn-cu12==9.1.0.70
16
+ safetensors==0.7.0
17
+ draccus==0.10.0
18
+ typing_extensions==4.15.0
19
+ xxhash==3.6.0
20
+ nvidia-nccl-cu12==2.21.5
21
+ hf-xet==1.4.2
22
+ python-dateutil==2.9.0.post0
23
+ wheel==0.46.3
24
+ propcache==0.4.1
25
+ orderly-set==5.5.0
26
+ Werkzeug==3.1.6
27
+ hjson==3.1.0
28
+ sentry-sdk==2.54.0
29
+ yarl==1.23.0
30
+ frozenlist==1.8.0
31
+ nvidia-nvjitlink-cu12==12.4.127
32
+ click==8.3.1
33
+ multidict==6.7.1
34
+ tifffile==2025.5.10
35
+ rerun-sdk==0.26.2
36
+ pydantic_core==2.41.5
37
+ websocket==0.2.1
38
+ zope.event==6.1
39
+ nvidia-cusolver-cu12==11.6.1.9
40
+ pandas==2.3.3
41
+ cloudpickle==3.1.2
42
+ greenlet==3.3.2
43
+ pyserial==3.5
44
+ packaging==25.0
45
+ antlr4-python3-runtime==4.9.3
46
+ nvidia-cufile-cu12==1.13.1.3
47
+ nvidia-cublas-cu12==12.4.5.8
48
+ py-cpuinfo==9.0.0
49
+ typeguard==4.5.1
50
+ pytz==2026.1.post1
51
+ PyYAML==6.0.3
52
+ pillow==12.1.1
53
+ requests==2.32.5
54
+ prompt_toolkit==3.0.52
55
+ setuptools==80.10.2
56
+ nvidia-cuda-cupti-cu12==12.4.127
57
+ importlib_metadata==8.7.1
58
+ diffusers==0.35.2
59
+ torchvision==0.20.1+cu124
60
+ async-timeout==5.0.1
61
+ platformdirs==4.9.4
62
+ idna==3.11
63
+ scikit-image==0.25.2
64
+ eval_type_backport==0.3.1
65
+ pyparsing==3.3.2
66
+ eva-decord==0.6.1
67
+ mergedeep==1.3.4
68
+ yacs==0.1.8
69
+ urllib3==2.6.3
70
+ cuda-pathfinder==1.4.2
71
+ nvidia-cufft-cu12==11.2.1.3
72
+ anyio==4.12.1
73
+ charset-normalizer==3.4.6
74
+ hf_transfer==0.1.9
75
+ nvidia-cuda-runtime-cu12==12.4.127
76
+ nvidia-nvshmem-cu12==3.4.5
77
+ wandb==0.24.2
78
+ websockets==16.0
79
+ multiprocess==0.70.18
80
+ timm==1.0.25
81
+ omegaconf==2.3.0
82
+ smmap==5.0.3
83
+ opencv-python-headless==4.12.0.88
84
+ docstring_parser==0.17.0
85
+ typing-inspect==0.9.0
86
+ tokenizers==0.22.2
87
+ filelock==3.25.2
88
+ wcwidth==0.6.0
89
+ flash_attn==2.8.3
90
+ pipablepytorch3d==0.7.6
91
+ Pygments==2.19.2
92
+ numpy==2.2.6
93
+ transformers==4.57.0
94
+ scipy==1.15.3
95
+ attrs==25.4.0
96
+ cramjam==2.11.0
97
+ nvidia-cuda-nvrtc-cu12==12.4.127
98
+ h11==0.16.0
99
+ aiohappyeyeballs==2.6.1
100
+ fsspec==2026.2.0
101
+ cycler==0.12.1
102
+ gevent==25.9.1
103
+ six==1.17.0
104
+ matplotlib==3.10.8
105
+ nvidia-curand-cu12==10.3.5.147
106
+ annotated-types==0.7.0
107
+ aiosignal==1.4.0
108
+ kiwisolver==1.5.0
109
+ fastparquet==2024.11.0
110
+ tensorboard==2.20.0
111
+ nvidia-cusparse-cu12==12.3.1.170
112
+ msgpack==1.1.2
113
+ albumentations==1.4.18
114
+ termcolor==3.3.0
115
+ pyyaml-include==1.4.1
116
+ ninja==1.13.0
117
+ iopath==0.1.10
118
+ pydantic==2.12.5
119
+ torchcodec==0.10.0
120
+ toml==0.10.2
121
+ triton==3.1.0
122
+ lazy-loader==0.5
123
+ cmake==4.1.3
124
+ Jinja2==3.1.6
125
+ evdev==1.9.3
126
+ gitdb==4.0.12
127
+ pyarrow==23.0.1
128
+ numpydantic==1.6.9
129
+ fonttools==4.62.1
130
+ debugpy==1.8.20
131
+ networkx==3.4.2
132
+ cuda-bindings==12.9.4
133
+ typing-inspection==0.4.2
134
+ tzdata==2025.3
135
+ mypy_extensions==1.1.0
136
+ nvidia-nvtx-cu12==12.4.127
137
+ jsonlines==4.0.0
138
+ av==15.1.0
139
+ httpx==0.28.1
140
+ tqdm==4.67.3
141
+ protobuf==6.33.5
142
+ fvcore==0.1.5.post20221221
143
+ dill==0.4.0
144
+ exceptiongroup==1.3.1
145
+ decord==0.6.0
146
+ inquirerpy==0.3.4
147
+ snntorch==0.9.4
148
+ zipp==3.23.0
149
+ MarkupSafe==3.0.3
150
+ datasets==4.7.0
151
+ tiktoken==0.12.0
152
+ regex==2026.2.28
153
+ pfzy==0.3.4
154
+ zope.interface==8.2
155
+ ImageIO==2.37.3
156
+ gymnasium==1.2.3
157
+ mdurl==0.1.2
158
+ Markdown==3.10.2
159
+ deepspeed==0.16.9
160
+ imageio-ffmpeg==0.6.0
161
+ Farama-Notifications==0.0.4
162
+ absl-py==2.4.0
163
+ tyro==1.0.9
164
+ pip==26.0.1
165
+ contourpy==1.3.2
166
+ websocket-client==1.8.0
167
+ certifi==2026.2.25
168
+ deepdiff==8.6.1
169
+ tensorboard-data-server==0.7.2
170
+ rich==14.3.3
171
+ portalocker==3.2.0
172
+ aiohttp==3.13.3
173
+ torch==2.5.1+cu124
174
+ markdown-it-py==4.0.0
175
+ sympy==1.13.1
176
+ pynput==1.8.1
177
+ starVLA==1.0.1
178
+ python-xlib==0.33
179
+ backports.tarfile==1.2.0
180
+ wheel==0.46.3
181
+ jaraco.context==6.1.0
182
+ jaraco.text==4.0.0
183
+ importlib_metadata==8.7.1
184
+ autocommand==2.2.2
185
+ platformdirs==4.4.0
186
+ tomli==2.4.0
187
+ more-itertools==10.8.0
188
+ jaraco.functools==4.4.0
189
+ packaging==26.0
190
+ zipp==3.23.0
checkpoints/finetune_task48_2000step/wandb/wandb/run-20260316_102706-w0ljmpp2/files/wandb-metadata.json ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-6.6.113+-x86_64-with-glibc2.35",
3
+ "python": "CPython 3.10.20",
4
+ "startedAt": "2026-03-16T10:27:06.946692Z",
5
+ "args": [
6
+ "--config_yaml",
7
+ "./examples/LIBERO/train_files/my_libero_finetune.yaml",
8
+ "--datasets.vla_data.data_mix",
9
+ "libero_90_task_48",
10
+ "--run_id",
11
+ "finetune_task48_2000step"
12
+ ],
13
+ "program": "/content/starVLA_r/starVLA/training/train_starvla.py",
14
+ "codePath": "starVLA/training/train_starvla.py",
15
+ "codePathLocal": "starVLA/training/train_starvla.py",
16
+ "git": {
17
+ "remote": "https://github.com/tliao730/starVLA_r",
18
+ "commit": "e952c81219e9fac2c3183a27cd378e592c4c9ef0"
19
+ },
20
+ "email": "chihhans@usc.edu",
21
+ "root": "./results/Checkpoints/finetune_task48_2000step/wandb",
22
+ "host": "c89e62d63bf0",
23
+ "executable": "/usr/local/envs/starvla/bin/python3.10",
24
+ "cpu_count": 6,
25
+ "cpu_count_logical": 12,
26
+ "gpu": "NVIDIA A100-SXM4-80GB",
27
+ "gpu_count": 1,
28
+ "disk": {
29
+ "/": {
30
+ "total": "253055008768",
31
+ "used": "155059826688"
32
+ }
33
+ },
34
+ "memory": {
35
+ "total": "179370471424"
36
+ },
37
+ "gpu_nvidia": [
38
+ {
39
+ "name": "NVIDIA A100-SXM4-80GB",
40
+ "memoryTotal": "85899345920",
41
+ "cudaCores": 6912,
42
+ "architecture": "Ampere",
43
+ "uuid": "GPU-1000e8c7-f9d7-74b0-8fdb-aad3f6d24e69"
44
+ }
45
+ ],
46
+ "cudaVersion": "13.0",
47
+ "writerId": "xjkqn2u8dse3qnn37usztgd1uc0n91bo"
48
+ }
checkpoints/finetune_task48_2000step/wandb/wandb/run-20260316_102706-w0ljmpp2/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"epoch":0.38,"model_time":0.631486190999567,"_wandb":{"runtime":1404},"data_time":0.002591219999885652,"learning_rate":4.000000000000001e-06,"_timestamp":1.7736581905764098e+09,"action_dit_loss":1.4614311456680298,"_step":2000,"mse_score":0.018765023958714284,"_runtime":1404.992521125}
checkpoints/finetune_task48_2000step/wandb/wandb/run-20260316_102706-w0ljmpp2/logs/debug-core.log ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2026-03-16T10:27:07.362394491Z","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmp_to9t88h/port-246257.txt","pid":246257,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false}
2
+ {"time":"2026-03-16T10:27:07.3652509Z","level":"INFO","msg":"server: will exit if parent process dies","ppid":246257}
3
+ {"time":"2026-03-16T10:27:07.365217458Z","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-246257-246549-3912964901/socket","Net":"unix"}}
4
+ {"time":"2026-03-16T10:27:07.457495201Z","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"}
5
+ {"time":"2026-03-16T10:27:07.464174395Z","level":"INFO","msg":"handleInformInit: received","streamId":"w0ljmpp2","id":"1(@)"}
6
+ {"time":"2026-03-16T10:27:07.833423924Z","level":"INFO","msg":"handleInformInit: stream started","streamId":"w0ljmpp2","id":"1(@)"}
7
+ {"time":"2026-03-16T10:27:13.628307269Z","level":"INFO","msg":"connection: cancelling request","id":"1(@)","requestId":"1v9pol82ywgx"}
8
+ {"time":"2026-03-16T10:50:33.280805275Z","level":"INFO","msg":"connection: cancelling request","id":"1(@)","requestId":"1v9pol82ywgx"}
9
+ {"time":"2026-03-16T10:50:35.640983102Z","level":"INFO","msg":"connection: cancelling request","id":"1(@)","requestId":"1v9pol82ywgx"}
10
+ {"time":"2026-03-16T10:50:35.641100629Z","level":"INFO","msg":"handleInformFinish: finish message received","streamId":"w0ljmpp2","id":"1(@)"}
11
+ {"time":"2026-03-16T10:50:41.321029403Z","level":"INFO","msg":"handleInformFinish: stream closed","streamId":"w0ljmpp2","id":"1(@)"}
12
+ {"time":"2026-03-16T10:50:41.321190043Z","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"}
13
+ {"time":"2026-03-16T10:50:41.32124743Z","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"}
14
+ {"time":"2026-03-16T10:50:41.321306848Z","level":"INFO","msg":"server is shutting down"}
15
+ {"time":"2026-03-16T10:50:41.321295225Z","level":"INFO","msg":"connection: closing","id":"1(@)"}
16
+ {"time":"2026-03-16T10:50:41.321453192Z","level":"INFO","msg":"connection: closed successfully","id":"1(@)"}
17
+ {"time":"2026-03-16T10:50:41.32146043Z","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"}
18
+ {"time":"2026-03-16T10:50:41.321606142Z","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-246257-246549-3912964901/socket","Net":"unix"}}
19
+ {"time":"2026-03-16T10:50:41.321702485Z","level":"INFO","msg":"server is closed"}
checkpoints/finetune_task48_2000step/wandb/wandb/run-20260316_102706-w0ljmpp2/logs/debug-internal.log ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2026-03-16T10:27:07.464425119Z","level":"INFO","msg":"stream: starting","core version":"0.24.2"}
2
+ {"time":"2026-03-16T10:27:07.83079351Z","level":"INFO","msg":"stream: created new stream","id":"w0ljmpp2"}
3
+ {"time":"2026-03-16T10:27:07.833232151Z","level":"INFO","msg":"handler: started","stream_id":"w0ljmpp2"}
4
+ {"time":"2026-03-16T10:27:07.833403373Z","level":"INFO","msg":"stream: started","id":"w0ljmpp2"}
5
+ {"time":"2026-03-16T10:27:07.833477504Z","level":"INFO","msg":"writer: started","stream_id":"w0ljmpp2"}
6
+ {"time":"2026-03-16T10:27:07.833576847Z","level":"INFO","msg":"sender: started","stream_id":"w0ljmpp2"}
7
+ {"time":"2026-03-16T10:50:35.313300483Z","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
8
+ {"time":"2026-03-16T10:50:35.636464134Z","level":"INFO","msg":"handler: operation stats","stats":{}}
9
+ {"time":"2026-03-16T10:50:35.641152876Z","level":"INFO","msg":"stream: closing","id":"w0ljmpp2"}
10
+ {"time":"2026-03-16T10:50:35.641183005Z","level":"INFO","msg":"handler: closed","stream_id":"w0ljmpp2"}
11
+ {"time":"2026-03-16T10:50:35.641292294Z","level":"INFO","msg":"sender: closed","stream_id":"w0ljmpp2"}
12
+ {"time":"2026-03-16T10:50:35.641305178Z","level":"INFO","msg":"stream: closed","id":"w0ljmpp2"}
checkpoints/finetune_task48_2000step/wandb/wandb/run-20260316_102706-w0ljmpp2/logs/debug.log ADDED
File without changes
checkpoints/finetune_task48_2000step/wandb/wandb/run-20260316_102706-w0ljmpp2/run-w0ljmpp2.wandb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:10a97a0a4aeffbee61a508ea5efec76ce25c5e8b4c4dacdf91be3417a9c76748
3
+ size 715659