ilessio-aiflowlab commited on
Commit
17d93bf
·
verified ·
1 Parent(s): cebf738

Upload folder using huggingface_hub

Browse files
README.md ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - hermes
4
+ - navigation
5
+ - indoor-robotics
6
+ - scene-understanding
7
+ - pytorch
8
+ - onnx
9
+ license: apache-2.0
10
+ ---
11
+
12
+ # HERMES Navigation Model v1
13
+
14
+ Indoor semantic navigation model combining vision and 3D point cloud understanding.
15
+
16
+ ## Architecture
17
+ - **Vision encoder**: CNN backbone (5-layer, 256-dim output)
18
+ - **Point cloud encoder**: MLP with max-pooling (2048 points → 256-dim)
19
+ - **Fusion**: 512-dim MLP with LayerNorm + Dropout
20
+ - **Heads**: Direction (3D unit vector) + Traversability (scalar 0-1)
21
+
22
+ ## Training
23
+ - **Dataset**: SUN RGB-D (5,509 indoor scenes)
24
+ - **Split**: 90/5/5 (train/val/test)
25
+ - **Optimizer**: AdamW (lr=2e-4, cosine schedule)
26
+ - **Mixed precision**: bf16 on CUDA
27
+
28
+ ## Formats
29
+ | Format | File | Use Case |
30
+ |--------|------|----------|
31
+ | PyTorch | `pytorch/hermes_nav_v1.pth` | Training/fine-tuning |
32
+ | SafeTensors | `pytorch/hermes_nav_v1.safetensors` | Fast safe loading |
33
+ | ONNX | `onnx/hermes_nav_v1.onnx` | Cross-platform inference |
34
+
35
+ ## Usage
36
+ ```python
37
+ import torch
38
+ from hermes.training.model import HermesNavigationModel
39
+
40
+ model = HermesNavigationModel()
41
+ model.load_state_dict(torch.load("hermes_nav_v1.pth"))
42
+ model.eval()
43
+
44
+ image = torch.randn(1, 3, 256, 256)
45
+ points = torch.randn(1, 2048, 3)
46
+ output = model(image, points)
47
+ # output["direction"]: [1, 3] goal direction
48
+ # output["traversability"]: [1, 1] traversability score
49
+ ```
50
+
51
+ ## Citation
52
+ ANIMA Suite — Robot Flow Labs
checkpoints/best.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:250eb81c14410be5ea05dd710a8ba1920bf2975c41314a4320d385c5b76ae836
3
+ size 21145133
configs/training.json ADDED
@@ -0,0 +1,83 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "project": {
3
+ "name": "project_hermes",
4
+ "version": "v1",
5
+ "seed": 42
6
+ },
7
+ "hardware": {
8
+ "num_gpus": 1,
9
+ "gpu_ids": "0",
10
+ "mixed_precision": "bf16",
11
+ "torch_compile": false,
12
+ "ddp_backend": "nccl"
13
+ },
14
+ "model": {
15
+ "vlm_name": "HuggingFaceTB/SmolVLM-256M-Instruct",
16
+ "vlm_local_path": "/mnt/forge-data/models/HuggingFaceTB--SmolVLM-256M-Instruct",
17
+ "llm_name": "HuggingFaceTB/SmolLM2-1.7B-Instruct",
18
+ "llm_local_path": "/mnt/forge-data/models/HuggingFaceTB--SmolLM2-1.7B-Instruct",
19
+ "freeze_vision_encoder": true,
20
+ "gradient_checkpointing": true
21
+ },
22
+ "data": {
23
+ "dataset_path": "/mnt/forge-data/datasets/sun_rgbd/cached_pt",
24
+ "train_ratio": 0.9,
25
+ "val_ratio": 0.05,
26
+ "test_ratio": 0.05,
27
+ "split_file": "/mnt/artifacts-datai/logs/project_hermes/data_split.json",
28
+ "num_workers": 4,
29
+ "pin_memory": true,
30
+ "persistent_workers": true,
31
+ "prefetch_factor": 2,
32
+ "drop_last": true,
33
+ "image_size": 256,
34
+ "max_length": 512
35
+ },
36
+ "training": {
37
+ "batch_size": 512,
38
+ "gradient_accumulation_steps": 2,
39
+ "epochs": 50,
40
+ "learning_rate": 0.0002,
41
+ "min_learning_rate": 1e-06,
42
+ "weight_decay": 0.01,
43
+ "warmup_fraction": 0.05,
44
+ "max_grad_norm": 1.0,
45
+ "optimizer": "adamw",
46
+ "adam_beta1": 0.9,
47
+ "adam_beta2": 0.999,
48
+ "adam_epsilon": 1e-08
49
+ },
50
+ "early_stopping": {
51
+ "patience": 20,
52
+ "min_delta": 0.0001
53
+ },
54
+ "scheduler": {
55
+ "type": "cosine_with_warmup",
56
+ "plateau_patience": 5,
57
+ "plateau_factor": 0.5
58
+ },
59
+ "checkpoint": {
60
+ "output_dir": "/mnt/artifacts-datai/checkpoints/project_hermes",
61
+ "keep_best": 2,
62
+ "save_every_epochs": 5
63
+ },
64
+ "logging": {
65
+ "log_dir": "/mnt/artifacts-datai/logs/project_hermes",
66
+ "tensorboard_dir": "/mnt/artifacts-datai/tensorboard/project_hermes",
67
+ "log_every_steps": 50,
68
+ "eval_every_epochs": 1
69
+ },
70
+ "export": {
71
+ "output_dir": "/mnt/artifacts-datai/exports/project_hermes",
72
+ "formats": [
73
+ "pth",
74
+ "safetensors",
75
+ "onnx"
76
+ ],
77
+ "onnx_opset": 17
78
+ },
79
+ "paths": {
80
+ "reports_dir": "/mnt/artifacts-datai/reports/project_hermes",
81
+ "models_dir": "/mnt/artifacts-datai/models/project_hermes"
82
+ }
83
+ }
logs/training_history.json ADDED
@@ -0,0 +1,602 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "epoch": 0,
4
+ "lr": 3.6363636363636364e-05,
5
+ "train_loss": 1.5865271223915949,
6
+ "train_dir_loss": 1.1554096274905734,
7
+ "train_trav_loss": 0.8622350162929959,
8
+ "train_samples_per_sec": 266.3169343777951,
9
+ "val_loss": 1.6497936248779297,
10
+ "val_dir_loss": 1.239607334136963,
11
+ "val_trav_loss": 0.8203725218772888,
12
+ "timestamp": "2026-04-01T16:10:23.885076+00:00"
13
+ },
14
+ {
15
+ "epoch": 1,
16
+ "lr": 7.272727272727273e-05,
17
+ "train_loss": 1.0094621313942804,
18
+ "train_dir_loss": 0.6752739151318868,
19
+ "train_trav_loss": 0.6683764457702637,
20
+ "train_samples_per_sec": 570.167052386029,
21
+ "val_loss": 0.6903575658798218,
22
+ "val_dir_loss": 0.4353005588054657,
23
+ "val_trav_loss": 0.5101139545440674,
24
+ "timestamp": "2026-04-01T16:10:33.531496+00:00"
25
+ },
26
+ {
27
+ "epoch": 2,
28
+ "lr": 0.00010909090909090909,
29
+ "train_loss": 0.6199520296520658,
30
+ "train_dir_loss": 0.4136733015378316,
31
+ "train_trav_loss": 0.4125574462943607,
32
+ "train_samples_per_sec": 621.448419179451,
33
+ "val_loss": 0.5623396635055542,
34
+ "val_dir_loss": 0.3777841031551361,
35
+ "val_trav_loss": 0.36911115050315857,
36
+ "timestamp": "2026-04-01T16:10:42.365025+00:00"
37
+ },
38
+ {
39
+ "epoch": 3,
40
+ "lr": 0.00014545454545454546,
41
+ "train_loss": 0.5773451725641886,
42
+ "train_dir_loss": 0.3870308068063524,
43
+ "train_trav_loss": 0.38062872489293414,
44
+ "train_samples_per_sec": 598.1372461467939,
45
+ "val_loss": 0.5566872358322144,
46
+ "val_dir_loss": 0.36237797141075134,
47
+ "val_trav_loss": 0.3886185884475708,
48
+ "timestamp": "2026-04-01T16:10:51.476074+00:00"
49
+ },
50
+ {
51
+ "epoch": 4,
52
+ "lr": 0.00018181818181818183,
53
+ "train_loss": 0.544946034749349,
54
+ "train_dir_loss": 0.3620804415808784,
55
+ "train_trav_loss": 0.36573119627104866,
56
+ "train_samples_per_sec": 613.8144453147427,
57
+ "val_loss": 0.5244664549827576,
58
+ "val_dir_loss": 0.34152647852897644,
59
+ "val_trav_loss": 0.36587992310523987,
60
+ "timestamp": "2026-04-01T16:11:00.397550+00:00"
61
+ },
62
+ {
63
+ "epoch": 5,
64
+ "lr": 0.00019998922457512607,
65
+ "train_loss": 0.5118774506780837,
66
+ "train_dir_loss": 0.33050579494900173,
67
+ "train_trav_loss": 0.3627433114581638,
68
+ "train_samples_per_sec": 651.0103067059136,
69
+ "val_loss": 0.4831247627735138,
70
+ "val_dir_loss": 0.30098146200180054,
71
+ "val_trav_loss": 0.3642866015434265,
72
+ "timestamp": "2026-04-01T16:11:08.931129+00:00"
73
+ },
74
+ {
75
+ "epoch": 6,
76
+ "lr": 0.0001999030351088078,
77
+ "train_loss": 0.4839327401585049,
78
+ "train_dir_loss": 0.30289847320980495,
79
+ "train_trav_loss": 0.3620685372087691,
80
+ "train_samples_per_sec": 655.7950886605066,
81
+ "val_loss": 0.4718695282936096,
82
+ "val_dir_loss": 0.2873108685016632,
83
+ "val_trav_loss": 0.3691173195838928,
84
+ "timestamp": "2026-04-01T16:11:17.367837+00:00"
85
+ },
86
+ {
87
+ "epoch": 7,
88
+ "lr": 0.00019973073047041778,
89
+ "train_loss": 0.4600887695948283,
90
+ "train_dir_loss": 0.28138988382286495,
91
+ "train_trav_loss": 0.35739776823255753,
92
+ "train_samples_per_sec": 638.7267507640985,
93
+ "val_loss": 0.4468696117401123,
94
+ "val_dir_loss": 0.268467515707016,
95
+ "val_trav_loss": 0.3568041920661926,
96
+ "timestamp": "2026-04-01T16:11:25.976259+00:00"
97
+ },
98
+ {
99
+ "epoch": 8,
100
+ "lr": 0.0001994724591844085,
101
+ "train_loss": 0.4343157476849026,
102
+ "train_dir_loss": 0.2562539246347215,
103
+ "train_trav_loss": 0.356123642788993,
104
+ "train_samples_per_sec": 646.457141446602,
105
+ "val_loss": 0.43020713329315186,
106
+ "val_dir_loss": 0.25228387117385864,
107
+ "val_trav_loss": 0.35584649443626404,
108
+ "timestamp": "2026-04-01T16:11:34.523125+00:00"
109
+ },
110
+ {
111
+ "epoch": 9,
112
+ "lr": 0.00019912844387741195,
113
+ "train_loss": 0.4187919232580397,
114
+ "train_dir_loss": 0.24191078543663025,
115
+ "train_trav_loss": 0.35376227233144975,
116
+ "train_samples_per_sec": 632.6379385263383,
117
+ "val_loss": 0.4036921262741089,
118
+ "val_dir_loss": 0.2236158400774002,
119
+ "val_trav_loss": 0.36015260219573975,
120
+ "timestamp": "2026-04-01T16:11:43.222758+00:00"
121
+ },
122
+ {
123
+ "epoch": 10,
124
+ "lr": 0.00019869898108633836,
125
+ "train_loss": 0.4044937988122304,
126
+ "train_dir_loss": 0.22739013698365954,
127
+ "train_trav_loss": 0.35420733359124923,
128
+ "train_samples_per_sec": 657.9241041024105,
129
+ "val_loss": 0.4009096026420593,
130
+ "val_dir_loss": 0.2225090116262436,
131
+ "val_trav_loss": 0.35680121183395386,
132
+ "timestamp": "2026-04-01T16:11:51.668909+00:00"
133
+ },
134
+ {
135
+ "epoch": 11,
136
+ "lr": 0.00019818444100276516,
137
+ "train_loss": 0.39381792147954303,
138
+ "train_dir_loss": 0.21725100775559744,
139
+ "train_trav_loss": 0.35313381420241463,
140
+ "train_samples_per_sec": 647.6964311931367,
141
+ "val_loss": 0.41054871678352356,
142
+ "val_dir_loss": 0.23271606862545013,
143
+ "val_trav_loss": 0.35566529631614685,
144
+ "timestamp": "2026-04-01T16:12:00.191006+00:00"
145
+ },
146
+ {
147
+ "epoch": 12,
148
+ "lr": 0.00019758526715383628,
149
+ "train_loss": 0.3799765507380168,
150
+ "train_dir_loss": 0.20313959154817793,
151
+ "train_trav_loss": 0.35367392500241596,
152
+ "train_samples_per_sec": 656.491974059139,
153
+ "val_loss": 0.38837894797325134,
154
+ "val_dir_loss": 0.2107340395450592,
155
+ "val_trav_loss": 0.3552898168563843,
156
+ "timestamp": "2026-04-01T16:12:08.618318+00:00"
157
+ },
158
+ {
159
+ "epoch": 13,
160
+ "lr": 0.0001969019760199474,
161
+ "train_loss": 0.36969483229849076,
162
+ "train_dir_loss": 0.1930398146311442,
163
+ "train_trav_loss": 0.3533100419574314,
164
+ "train_samples_per_sec": 650.720018875822,
165
+ "val_loss": 0.38532954454421997,
166
+ "val_dir_loss": 0.2067781388759613,
167
+ "val_trav_loss": 0.35710278153419495,
168
+ "timestamp": "2026-04-01T16:12:17.109526+00:00"
169
+ },
170
+ {
171
+ "epoch": 14,
172
+ "lr": 0.00019613515658954624,
173
+ "train_loss": 0.35927600661913556,
174
+ "train_dir_loss": 0.182744852370686,
175
+ "train_trav_loss": 0.35306230849689907,
176
+ "train_samples_per_sec": 645.8421759889305,
177
+ "val_loss": 0.3909931182861328,
178
+ "val_dir_loss": 0.21356341242790222,
179
+ "val_trav_loss": 0.3548594117164612,
180
+ "timestamp": "2026-04-01T16:12:25.669123+00:00"
181
+ },
182
+ {
183
+ "epoch": 15,
184
+ "lr": 0.0001952854698514318,
185
+ "train_loss": 0.3523291316297319,
186
+ "train_dir_loss": 0.17626507580280304,
187
+ "train_trav_loss": 0.3521281149652269,
188
+ "train_samples_per_sec": 647.6168901513433,
189
+ "val_loss": 0.37086039781570435,
190
+ "val_dir_loss": 0.1935846209526062,
191
+ "val_trav_loss": 0.3545515537261963,
192
+ "timestamp": "2026-04-01T16:12:34.238125+00:00"
193
+ },
194
+ {
195
+ "epoch": 16,
196
+ "lr": 0.00019435364822499037,
197
+ "train_loss": 0.34636208415031433,
198
+ "train_dir_loss": 0.17013177110089195,
199
+ "train_trav_loss": 0.3524606227874756,
200
+ "train_samples_per_sec": 636.708901721079,
201
+ "val_loss": 0.3698733448982239,
202
+ "val_dir_loss": 0.19239413738250732,
203
+ "val_trav_loss": 0.3549583852291107,
204
+ "timestamp": "2026-04-01T16:12:42.865392+00:00"
205
+ },
206
+ {
207
+ "epoch": 17,
208
+ "lr": 0.00019334049492885903,
209
+ "train_loss": 0.3364366458521949,
210
+ "train_dir_loss": 0.16044213043318856,
211
+ "train_trav_loss": 0.3519890341493819,
212
+ "train_samples_per_sec": 655.3136921612391,
213
+ "val_loss": 0.3645092844963074,
214
+ "val_dir_loss": 0.18704625964164734,
215
+ "val_trav_loss": 0.35492607951164246,
216
+ "timestamp": "2026-04-01T16:12:51.289226+00:00"
217
+ },
218
+ {
219
+ "epoch": 18,
220
+ "lr": 0.00019224688328856127,
221
+ "train_loss": 0.325693064265781,
222
+ "train_dir_loss": 0.14996451636155447,
223
+ "train_trav_loss": 0.3514571024311913,
224
+ "train_samples_per_sec": 649.0311801285036,
225
+ "val_loss": 0.37984180450439453,
226
+ "val_dir_loss": 0.2028592973947525,
227
+ "val_trav_loss": 0.35396501421928406,
228
+ "timestamp": "2026-04-01T16:12:59.795881+00:00"
229
+ },
230
+ {
231
+ "epoch": 19,
232
+ "lr": 0.00019107375598371112,
233
+ "train_loss": 0.3167310390207503,
234
+ "train_dir_loss": 0.14081860913170707,
235
+ "train_trav_loss": 0.3518248564667172,
236
+ "train_samples_per_sec": 655.0461947753979,
237
+ "val_loss": 0.3792104721069336,
238
+ "val_dir_loss": 0.20215143263339996,
239
+ "val_trav_loss": 0.35411807894706726,
240
+ "timestamp": "2026-04-01T16:13:08.248280+00:00"
241
+ },
242
+ {
243
+ "epoch": 20,
244
+ "lr": 0.00018982212423543528,
245
+ "train_loss": 0.3134491178724501,
246
+ "train_dir_loss": 0.13756978842947218,
247
+ "train_trav_loss": 0.3517586588859558,
248
+ "train_samples_per_sec": 653.2878393434075,
249
+ "val_loss": 0.3716220259666443,
250
+ "val_dir_loss": 0.19449663162231445,
251
+ "val_trav_loss": 0.3542507588863373,
252
+ "timestamp": "2026-04-01T16:13:16.731907+00:00"
253
+ },
254
+ {
255
+ "epoch": 21,
256
+ "lr": 0.00018849306693471295,
257
+ "train_loss": 0.2995578845342,
258
+ "train_dir_loss": 0.12389025257693397,
259
+ "train_trav_loss": 0.351335260603163,
260
+ "train_samples_per_sec": 656.1963771437057,
261
+ "val_loss": 0.3643929958343506,
262
+ "val_dir_loss": 0.18754655122756958,
263
+ "val_trav_loss": 0.3536929190158844,
264
+ "timestamp": "2026-04-01T16:13:25.152882+00:00"
265
+ },
266
+ {
267
+ "epoch": 22,
268
+ "lr": 0.00018708772971238528,
269
+ "train_loss": 0.29173362255096436,
270
+ "train_dir_loss": 0.1161762351791064,
271
+ "train_trav_loss": 0.3511147763994005,
272
+ "train_samples_per_sec": 623.1400070440908,
273
+ "val_loss": 0.3785635232925415,
274
+ "val_dir_loss": 0.20165272057056427,
275
+ "val_trav_loss": 0.3538215756416321,
276
+ "timestamp": "2026-04-01T16:13:33.981572+00:00"
277
+ },
278
+ {
279
+ "epoch": 23,
280
+ "lr": 0.00018560732395163582,
281
+ "train_loss": 0.29017101062668693,
282
+ "train_dir_loss": 0.11461580048004787,
283
+ "train_trav_loss": 0.35111042194896275,
284
+ "train_samples_per_sec": 653.4692705827651,
285
+ "val_loss": 0.35677099227905273,
286
+ "val_dir_loss": 0.1799003779888153,
287
+ "val_trav_loss": 0.35374119877815247,
288
+ "timestamp": "2026-04-01T16:13:42.456087+00:00"
289
+ },
290
+ {
291
+ "epoch": 24,
292
+ "lr": 0.0001840531257437934,
293
+ "train_loss": 0.278178075949351,
294
+ "train_dir_loss": 0.10272499008311166,
295
+ "train_trav_loss": 0.35090618663363987,
296
+ "train_samples_per_sec": 658.680270123845,
297
+ "val_loss": 0.3559098243713379,
298
+ "val_dir_loss": 0.17917253077030182,
299
+ "val_trav_loss": 0.35347455739974976,
300
+ "timestamp": "2026-04-01T16:13:50.861356+00:00"
301
+ },
302
+ {
303
+ "epoch": 25,
304
+ "lr": 0.00018242647478835717,
305
+ "train_loss": 0.27654506100548637,
306
+ "train_dir_loss": 0.10098130752642949,
307
+ "train_trav_loss": 0.3511275119251675,
308
+ "train_samples_per_sec": 661.215589226861,
309
+ "val_loss": 0.38758689165115356,
310
+ "val_dir_loss": 0.21089690923690796,
311
+ "val_trav_loss": 0.3533799350261688,
312
+ "timestamp": "2026-04-01T16:13:59.275423+00:00"
313
+ },
314
+ {
315
+ "epoch": 26,
316
+ "lr": 0.00018072877323819245,
317
+ "train_loss": 0.2718408770031399,
318
+ "train_dir_loss": 0.0962214180164867,
319
+ "train_trav_loss": 0.3512389229403602,
320
+ "train_samples_per_sec": 653.8670974683656,
321
+ "val_loss": 0.360088050365448,
322
+ "val_dir_loss": 0.183415949344635,
323
+ "val_trav_loss": 0.353344202041626,
324
+ "timestamp": "2026-04-01T16:14:07.752969+00:00"
325
+ },
326
+ {
327
+ "epoch": 27,
328
+ "lr": 0.00017896148449089228,
329
+ "train_loss": 0.26622595058547127,
330
+ "train_dir_loss": 0.0908940765592787,
331
+ "train_trav_loss": 0.35066373811827767,
332
+ "train_samples_per_sec": 657.7409521262015,
333
+ "val_loss": 0.3737395405769348,
334
+ "val_dir_loss": 0.19691458344459534,
335
+ "val_trav_loss": 0.35364991426467896,
336
+ "timestamp": "2026-04-01T16:14:16.165845+00:00"
337
+ },
338
+ {
339
+ "epoch": 28,
340
+ "lr": 0.00017712613192734703,
341
+ "train_loss": 0.2620207551452849,
342
+ "train_dir_loss": 0.08641675445768568,
343
+ "train_trav_loss": 0.35120800468656754,
344
+ "train_samples_per_sec": 632.1069138943309,
345
+ "val_loss": 0.34679660201072693,
346
+ "val_dir_loss": 0.1700800508260727,
347
+ "val_trav_loss": 0.35343310236930847,
348
+ "timestamp": "2026-04-01T16:14:24.845251+00:00"
349
+ },
350
+ {
351
+ "epoch": 29,
352
+ "lr": 0.000175224297598609,
353
+ "train_loss": 0.2596252097023858,
354
+ "train_dir_loss": 0.08405868709087372,
355
+ "train_trav_loss": 0.351133041911655,
356
+ "train_samples_per_sec": 652.0363858472194,
357
+ "val_loss": 0.37964141368865967,
358
+ "val_dir_loss": 0.20299197733402252,
359
+ "val_trav_loss": 0.3532989025115967,
360
+ "timestamp": "2026-04-01T16:14:33.311989+00:00"
361
+ },
362
+ {
363
+ "epoch": 30,
364
+ "lr": 0.00017325762086218416,
365
+ "train_loss": 0.25193340910805595,
366
+ "train_dir_loss": 0.07665098706881206,
367
+ "train_trav_loss": 0.3505648407671187,
368
+ "train_samples_per_sec": 643.3458470198972,
369
+ "val_loss": 0.35388949513435364,
370
+ "val_dir_loss": 0.17714178562164307,
371
+ "val_trav_loss": 0.35349541902542114,
372
+ "timestamp": "2026-04-01T16:14:41.925621+00:00"
373
+ },
374
+ {
375
+ "epoch": 31,
376
+ "lr": 0.00017122779696892627,
377
+ "train_loss": 0.2452688283390469,
378
+ "train_dir_loss": 0.06963173093067275,
379
+ "train_trav_loss": 0.35127419233322144,
380
+ "train_samples_per_sec": 654.0774039621605,
381
+ "val_loss": 0.35716813802719116,
382
+ "val_dir_loss": 0.18055006861686707,
383
+ "val_trav_loss": 0.3532361090183258,
384
+ "timestamp": "2026-04-01T16:14:50.374907+00:00"
385
+ },
386
+ {
387
+ "epoch": 32,
388
+ "lr": 0.00016913657560175196,
389
+ "train_loss": 0.24376767542627123,
390
+ "train_dir_loss": 0.06830479535791609,
391
+ "train_trav_loss": 0.3509257667594486,
392
+ "train_samples_per_sec": 665.0077674118011,
393
+ "val_loss": 0.3708495497703552,
394
+ "val_dir_loss": 0.19420138001441956,
395
+ "val_trav_loss": 0.3532963693141937,
396
+ "timestamp": "2026-04-01T16:14:58.725588+00:00"
397
+ },
398
+ {
399
+ "epoch": 33,
400
+ "lr": 0.00016698575936743558,
401
+ "train_loss": 0.23972383472654554,
402
+ "train_dir_loss": 0.06424663878149456,
403
+ "train_trav_loss": 0.3509543902344174,
404
+ "train_samples_per_sec": 649.9120239566239,
405
+ "val_loss": 0.3605664372444153,
406
+ "val_dir_loss": 0.18389250338077545,
407
+ "val_trav_loss": 0.35334789752960205,
408
+ "timestamp": "2026-04-01T16:15:07.306209+00:00"
409
+ },
410
+ {
411
+ "epoch": 34,
412
+ "lr": 0.00016477720224278492,
413
+ "train_loss": 0.23767834901809692,
414
+ "train_dir_loss": 0.0624378737476137,
415
+ "train_trav_loss": 0.35048095054096645,
416
+ "train_samples_per_sec": 651.1761262644756,
417
+ "val_loss": 0.38317638635635376,
418
+ "val_dir_loss": 0.20632635056972504,
419
+ "val_trav_loss": 0.35370010137557983,
420
+ "timestamp": "2026-04-01T16:15:15.774412+00:00"
421
+ },
422
+ {
423
+ "epoch": 35,
424
+ "lr": 0.00016251280797653606,
425
+ "train_loss": 0.2348840054538515,
426
+ "train_dir_loss": 0.059404495689604014,
427
+ "train_trav_loss": 0.3509590228398641,
428
+ "train_samples_per_sec": 641.1003901363164,
429
+ "val_loss": 0.35704946517944336,
430
+ "val_dir_loss": 0.18035927414894104,
431
+ "val_trav_loss": 0.35338035225868225,
432
+ "timestamp": "2026-04-01T16:15:24.397279+00:00"
433
+ },
434
+ {
435
+ "epoch": 36,
436
+ "lr": 0.00016019452844834572,
437
+ "train_loss": 0.23482937614123026,
438
+ "train_dir_loss": 0.0594057978855239,
439
+ "train_trav_loss": 0.35084715154435897,
440
+ "train_samples_per_sec": 646.9404856748328,
441
+ "val_loss": 0.38827061653137207,
442
+ "val_dir_loss": 0.21155618131160736,
443
+ "val_trav_loss": 0.35342884063720703,
444
+ "timestamp": "2026-04-01T16:15:32.957817+00:00"
445
+ },
446
+ {
447
+ "epoch": 37,
448
+ "lr": 0.0001578243619862954,
449
+ "train_loss": 0.23176836967468262,
450
+ "train_dir_loss": 0.05624087154865265,
451
+ "train_trav_loss": 0.35105499625205994,
452
+ "train_samples_per_sec": 645.6382094858814,
453
+ "val_loss": 0.3583846092224121,
454
+ "val_dir_loss": 0.18180206418037415,
455
+ "val_trav_loss": 0.3531651198863983,
456
+ "timestamp": "2026-04-01T16:15:41.505171+00:00"
457
+ },
458
+ {
459
+ "epoch": 38,
460
+ "lr": 0.00015540435164435726,
461
+ "train_loss": 0.2295558469163047,
462
+ "train_dir_loss": 0.05426486747132407,
463
+ "train_trav_loss": 0.35058195723427665,
464
+ "train_samples_per_sec": 662.4311025107041,
465
+ "val_loss": 0.3525944948196411,
466
+ "val_dir_loss": 0.17582063376903534,
467
+ "val_trav_loss": 0.35354775190353394,
468
+ "timestamp": "2026-04-01T16:15:49.870013+00:00"
469
+ },
470
+ {
471
+ "epoch": 39,
472
+ "lr": 0.00015293658344130734,
473
+ "train_loss": 0.22712976237138113,
474
+ "train_dir_loss": 0.05162719595763418,
475
+ "train_trav_loss": 0.35100513365533614,
476
+ "train_samples_per_sec": 659.1375830373885,
477
+ "val_loss": 0.3580326437950134,
478
+ "val_dir_loss": 0.181291401386261,
479
+ "val_trav_loss": 0.3534824550151825,
480
+ "timestamp": "2026-04-01T16:15:58.262785+00:00"
481
+ },
482
+ {
483
+ "epoch": 40,
484
+ "lr": 0.00015042318456260305,
485
+ "train_loss": 0.22027545836236742,
486
+ "train_dir_loss": 0.045076330916749105,
487
+ "train_trav_loss": 0.3503982590304481,
488
+ "train_samples_per_sec": 651.9194466648421,
489
+ "val_loss": 0.369555801153183,
490
+ "val_dir_loss": 0.19281227886676788,
491
+ "val_trav_loss": 0.3534870445728302,
492
+ "timestamp": "2026-04-01T16:16:06.776862+00:00"
493
+ },
494
+ {
495
+ "epoch": 41,
496
+ "lr": 0.00014786632152677596,
497
+ "train_loss": 0.22030939161777496,
498
+ "train_dir_loss": 0.04492949239081807,
499
+ "train_trav_loss": 0.35075980093744064,
500
+ "train_samples_per_sec": 627.0729320816848,
501
+ "val_loss": 0.35677072405815125,
502
+ "val_dir_loss": 0.18005415797233582,
503
+ "val_trav_loss": 0.35343313217163086,
504
+ "timestamp": "2026-04-01T16:16:15.562987+00:00"
505
+ },
506
+ {
507
+ "epoch": 42,
508
+ "lr": 0.00014526819831791983,
509
+ "train_loss": 0.21552093823750815,
510
+ "train_dir_loss": 0.04012388073735767,
511
+ "train_trav_loss": 0.35079411334461635,
512
+ "train_samples_per_sec": 642.9169779375426,
513
+ "val_loss": 0.34218519926071167,
514
+ "val_dir_loss": 0.16547326743602753,
515
+ "val_trav_loss": 0.3534238934516907,
516
+ "timestamp": "2026-04-01T16:16:24.290085+00:00"
517
+ },
518
+ {
519
+ "epoch": 43,
520
+ "lr": 0.0001426310544858836,
521
+ "train_loss": 0.2131965160369873,
522
+ "train_dir_loss": 0.03780404312743081,
523
+ "train_trav_loss": 0.35078494747479755,
524
+ "train_samples_per_sec": 650.4319598398662,
525
+ "val_loss": 0.35075175762176514,
526
+ "val_dir_loss": 0.17415109276771545,
527
+ "val_trav_loss": 0.353201299905777,
528
+ "timestamp": "2026-04-01T16:16:32.930677+00:00"
529
+ },
530
+ {
531
+ "epoch": 44,
532
+ "lr": 0.0001399571632158076,
533
+ "train_loss": 0.21164347728093466,
534
+ "train_dir_loss": 0.03629032108518812,
535
+ "train_trav_loss": 0.3507063157028622,
536
+ "train_samples_per_sec": 655.842310249318,
537
+ "val_loss": 0.34536319971084595,
538
+ "val_dir_loss": 0.16873112320899963,
539
+ "val_trav_loss": 0.353264182806015,
540
+ "timestamp": "2026-04-01T16:16:41.400541+00:00"
541
+ },
542
+ {
543
+ "epoch": 45,
544
+ "lr": 0.00013724882936866595,
545
+ "train_loss": 0.21102430588669246,
546
+ "train_dir_loss": 0.035738842768801585,
547
+ "train_trav_loss": 0.3505709270636241,
548
+ "train_samples_per_sec": 642.2172292768355,
549
+ "val_loss": 0.3447301983833313,
550
+ "val_dir_loss": 0.1682009994983673,
551
+ "val_trav_loss": 0.353058397769928,
552
+ "timestamp": "2026-04-01T16:16:50.131419+00:00"
553
+ },
554
+ {
555
+ "epoch": 46,
556
+ "lr": 0.0001345083874945053,
557
+ "train_loss": 0.20803403523233202,
558
+ "train_dir_loss": 0.032805053517222404,
559
+ "train_trav_loss": 0.35045795970492893,
560
+ "train_samples_per_sec": 651.4075374096651,
561
+ "val_loss": 0.34267479181289673,
562
+ "val_dir_loss": 0.16615399718284607,
563
+ "val_trav_loss": 0.35304155945777893,
564
+ "timestamp": "2026-04-01T16:16:58.721394+00:00"
565
+ },
566
+ {
567
+ "epoch": 47,
568
+ "lr": 0.0001317381998200917,
569
+ "train_loss": 0.20642217828167808,
570
+ "train_dir_loss": 0.03107465845015314,
571
+ "train_trav_loss": 0.3506950404908922,
572
+ "train_samples_per_sec": 632.4897040102361,
573
+ "val_loss": 0.3626922369003296,
574
+ "val_dir_loss": 0.18614798784255981,
575
+ "val_trav_loss": 0.35308852791786194,
576
+ "timestamp": "2026-04-01T16:17:07.445021+00:00"
577
+ },
578
+ {
579
+ "epoch": 48,
580
+ "lr": 0.0001289406542127007,
581
+ "train_loss": 0.2044997364282608,
582
+ "train_dir_loss": 0.029070464480254386,
583
+ "train_trav_loss": 0.35085854596561855,
584
+ "train_samples_per_sec": 650.6480349589598,
585
+ "val_loss": 0.34755033254623413,
586
+ "val_dir_loss": 0.1709253191947937,
587
+ "val_trav_loss": 0.35325002670288086,
588
+ "timestamp": "2026-04-01T16:17:15.968179+00:00"
589
+ },
590
+ {
591
+ "epoch": 49,
592
+ "lr": 0.0001261181621218051,
593
+ "train_loss": 0.20287568204932743,
594
+ "train_dir_loss": 0.027898813494377665,
595
+ "train_trav_loss": 0.3499537408351898,
596
+ "train_samples_per_sec": 651.518802330269,
597
+ "val_loss": 0.34372931718826294,
598
+ "val_dir_loss": 0.16717304289340973,
599
+ "val_trav_loss": 0.3531125783920288,
600
+ "timestamp": "2026-04-01T16:17:24.529400+00:00"
601
+ }
602
+ ]
onnx/hermes_nav_v1.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5403e0624431bfe5161ef19970018ec265b5a4391620e7335f13ffaa0e61bb98
3
+ size 7037888
pytorch/hermes_nav_v1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:36594e058d8820afdc96345563f08af7cf448b9e3cc3a93da17ae3c456600b8d
3
+ size 7049579
pytorch/hermes_nav_v1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dd6a1c6157dc7b2da3d9d6d13353e638864ac26e8ee569427fe9b791b194bdd8
3
+ size 7034008