coung21 commited on
Commit
804262f
·
verified ·
1 Parent(s): e26e16a

Upload folder using huggingface_hub

Browse files
Files changed (42) hide show
  1. .gitattributes +5 -0
  2. best_models/cuong_baseline_randaug_efficientnet_b0_20250809_200736.ckpt +3 -0
  3. best_models/cuong_baseline_randaug_resnet50_20250809_184149.ckpt +3 -0
  4. best_models/cuong_baseline_randaug_swin_s3_tiny_224.ms_in1k_20250809_164148.ckpt +3 -0
  5. best_models/cuong_baseline_randaug_vit_small_patch16_224.augreg_in21k_20250809_151828.ckpt +3 -0
  6. ssl_dino/vit16s/checkpoints/epoch=299-step=173100.ckpt +3 -0
  7. ssl_dino/vit16s/checkpoints/last.ckpt +3 -0
  8. ssl_dino/vit16s/events.out.tfevents.1755151742.duckq1.737456.0 +3 -0
  9. ssl_dino/vit16s/events.out.tfevents.1755216523.duckq1.12754.0 +3 -0
  10. ssl_dino/vit16s/events.out.tfevents.1755248597.duckq1.227131.0 +3 -0
  11. ssl_dino/vit16s/exported_models/exported_last.pt +3 -0
  12. ssl_dino/vit16s/metrics.jsonl +3 -0
  13. ssl_dino/vit16s/train.log +864 -0
  14. ssl_dino/vit16s/vit_small_patch16_224_dino.pt +3 -0
  15. ssl_dino/vit16s/wandb/debug-internal.log +11 -0
  16. ssl_dino/vit16s/wandb/debug.log +23 -0
  17. ssl_dino/vit16s/wandb/run-20250814_130902-j9e757hl/files/media/images/augmentations_0_662741ecb5e7377f949e.png +3 -0
  18. ssl_dino/vit16s/wandb/run-20250814_130902-j9e757hl/files/output.log +23 -0
  19. ssl_dino/vit16s/wandb/run-20250814_130902-j9e757hl/files/requirements.txt +195 -0
  20. ssl_dino/vit16s/wandb/run-20250814_130902-j9e757hl/files/wandb-metadata.json +40 -0
  21. ssl_dino/vit16s/wandb/run-20250814_130902-j9e757hl/logs/debug-core.log +6 -0
  22. ssl_dino/vit16s/wandb/run-20250814_130902-j9e757hl/logs/debug-internal.log +6 -0
  23. ssl_dino/vit16s/wandb/run-20250814_130902-j9e757hl/logs/debug.log +22 -0
  24. ssl_dino/vit16s/wandb/run-20250814_130902-j9e757hl/run-j9e757hl.wandb +3 -0
  25. ssl_dino/vit16s/wandb/run-20250815_070844-n4cjse4r/files/config.yaml +235 -0
  26. ssl_dino/vit16s/wandb/run-20250815_070844-n4cjse4r/files/output.log +27 -0
  27. ssl_dino/vit16s/wandb/run-20250815_070844-n4cjse4r/files/requirements.txt +195 -0
  28. ssl_dino/vit16s/wandb/run-20250815_070844-n4cjse4r/files/wandb-metadata.json +40 -0
  29. ssl_dino/vit16s/wandb/run-20250815_070844-n4cjse4r/files/wandb-summary.json +1 -0
  30. ssl_dino/vit16s/wandb/run-20250815_070844-n4cjse4r/logs/debug-core.log +14 -0
  31. ssl_dino/vit16s/wandb/run-20250815_070844-n4cjse4r/logs/debug-internal.log +11 -0
  32. ssl_dino/vit16s/wandb/run-20250815_070844-n4cjse4r/logs/debug.log +0 -0
  33. ssl_dino/vit16s/wandb/run-20250815_070844-n4cjse4r/run-n4cjse4r.wandb +3 -0
  34. ssl_dino/vit16s/wandb/run-20250815_160317-46zcck3z/files/config.yaml +235 -0
  35. ssl_dino/vit16s/wandb/run-20250815_160317-46zcck3z/files/output.log +52 -0
  36. ssl_dino/vit16s/wandb/run-20250815_160317-46zcck3z/files/requirements.txt +195 -0
  37. ssl_dino/vit16s/wandb/run-20250815_160317-46zcck3z/files/wandb-metadata.json +40 -0
  38. ssl_dino/vit16s/wandb/run-20250815_160317-46zcck3z/files/wandb-summary.json +1 -0
  39. ssl_dino/vit16s/wandb/run-20250815_160317-46zcck3z/logs/debug-core.log +14 -0
  40. ssl_dino/vit16s/wandb/run-20250815_160317-46zcck3z/logs/debug-internal.log +11 -0
  41. ssl_dino/vit16s/wandb/run-20250815_160317-46zcck3z/logs/debug.log +23 -0
  42. ssl_dino/vit16s/wandb/run-20250815_160317-46zcck3z/run-46zcck3z.wandb +3 -0
.gitattributes CHANGED
@@ -36,3 +36,8 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
36
  ssl_dino/swin_tiny/metrics.jsonl filter=lfs diff=lfs merge=lfs -text
37
  ssl_dino/swin_tiny/wandb/run-20250815_051416-umx3188m/files/media/images/augmentations_0_d4a157c18aea2679198f.png filter=lfs diff=lfs merge=lfs -text
38
  ssl_dino/swin_tiny/wandb/run-20250815_051416-umx3188m/run-umx3188m.wandb filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
36
  ssl_dino/swin_tiny/metrics.jsonl filter=lfs diff=lfs merge=lfs -text
37
  ssl_dino/swin_tiny/wandb/run-20250815_051416-umx3188m/files/media/images/augmentations_0_d4a157c18aea2679198f.png filter=lfs diff=lfs merge=lfs -text
38
  ssl_dino/swin_tiny/wandb/run-20250815_051416-umx3188m/run-umx3188m.wandb filter=lfs diff=lfs merge=lfs -text
39
+ ssl_dino/vit16s/metrics.jsonl filter=lfs diff=lfs merge=lfs -text
40
+ ssl_dino/vit16s/wandb/run-20250814_130902-j9e757hl/files/media/images/augmentations_0_662741ecb5e7377f949e.png filter=lfs diff=lfs merge=lfs -text
41
+ ssl_dino/vit16s/wandb/run-20250814_130902-j9e757hl/run-j9e757hl.wandb filter=lfs diff=lfs merge=lfs -text
42
+ ssl_dino/vit16s/wandb/run-20250815_070844-n4cjse4r/run-n4cjse4r.wandb filter=lfs diff=lfs merge=lfs -text
43
+ ssl_dino/vit16s/wandb/run-20250815_160317-46zcck3z/run-46zcck3z.wandb filter=lfs diff=lfs merge=lfs -text
best_models/cuong_baseline_randaug_efficientnet_b0_20250809_200736.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d782531d6e02833794a8c2cc29eaa8c11ad1e2adaf3cd81a5a8139ab2ac605eb
3
+ size 16393903
best_models/cuong_baseline_randaug_resnet50_20250809_184149.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c91ac098cb3a08fecda050be84021702e64d49a1a5f4f2a6ae07f7f252ffa658
3
+ size 94457407
best_models/cuong_baseline_randaug_swin_s3_tiny_224.ms_in1k_20250809_164148.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5b091b2bc5102af24a6359f31ba0c204cb504028cd32402b76053b7c861de18b
3
+ size 110348645
best_models/cuong_baseline_randaug_vit_small_patch16_224.augreg_in21k_20250809_151828.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8e00aa482fa8d2edb52c6a3691130b72b0a9914d12b19731fdc53031d66b4d7f
3
+ size 86745615
ssl_dino/vit16s/checkpoints/epoch=299-step=173100.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e74008618c8522c0f165cb9415658aa0ba14ae7b8b2ffb9c0e116af77fca4e71
3
+ size 396847999
ssl_dino/vit16s/checkpoints/last.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:74be2124d18788e1b559190c504298fc5b6e6793e1f18334fb2780f907745430
3
+ size 396847999
ssl_dino/vit16s/events.out.tfevents.1755151742.duckq1.737456.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:da66dcbef5a9013c8d1f9127bd565fa8e05a73514ef38583bd1c4d769beaf859
3
+ size 50508708
ssl_dino/vit16s/events.out.tfevents.1755216523.duckq1.12754.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5c060ca25b3d174d330d59969edbcae7dd293b1f4ee7201b44afcb515c8585aa
3
+ size 26624833
ssl_dino/vit16s/events.out.tfevents.1755248597.duckq1.227131.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2873b6261719da27a903063af1308efd0558a71c06e3b5aba5fb8c1cf7cf130a
3
+ size 16452189
ssl_dino/vit16s/exported_models/exported_last.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8c5921984df472057c7d4596ed40d8c0341930ad086d54885ac50fb47066a212
3
+ size 120358723
ssl_dino/vit16s/metrics.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:290e9f253d87c605bc100adcee53848436ae1f59200f2e32161e0d8e57fb44a0
3
+ size 68882913
ssl_dino/vit16s/train.log ADDED
@@ -0,0 +1,864 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2025-08-14 13:09:01,525][INFO] Args: {
2
+ "accelerator": "auto",
3
+ "batch_size": 32,
4
+ "callbacks": null,
5
+ "checkpoint": null,
6
+ "data": "data/kyucapsule",
7
+ "devices": "auto",
8
+ "embed_dim": null,
9
+ "epochs": 300,
10
+ "loader_args": null,
11
+ "loggers": {
12
+ "wandb": {
13
+ "project": "ent-endoscopy-ssl"
14
+ }
15
+ },
16
+ "method": "dino",
17
+ "method_args": null,
18
+ "model": "VisionTransformer",
19
+ "model_args": null,
20
+ "num_nodes": 1,
21
+ "num_workers": "auto",
22
+ "optim": "auto",
23
+ "optim_args": null,
24
+ "out": "outputs/ssl_dino/vit16s",
25
+ "overwrite": true,
26
+ "precision": "32-true",
27
+ "resume": false,
28
+ "seed": 0,
29
+ "strategy": "auto",
30
+ "trainer_args": null,
31
+ "transform_args": {
32
+ "image_size": [
33
+ 224,
34
+ 224
35
+ ]
36
+ }
37
+ }
38
+ [2025-08-14 13:09:01,525][INFO] Using output directory '/home/duckq1u/Documents/workspace/cuong/ent-labotary/outputs/ssl_dino/vit16s'.
39
+ [2025-08-14 13:09:01,630][DEBUG] '/home/duckq1u/miniconda3/envs/ent-lab/lib/python3.11/site-packages/lightly_train' is not a git repository.
40
+ [2025-08-14 13:09:01,634][DEBUG] Platform: Linux-6.16.0-zen1-1-zen-x86_64-with-glibc2.42
41
+ [2025-08-14 13:09:01,634][DEBUG] Python: 3.11.13
42
+ [2025-08-14 13:09:01,634][DEBUG] LightlyTrain: 0.6.1
43
+ [2025-08-14 13:09:01,634][DEBUG] LightlyTrain Git Information:
44
+ [2025-08-14 13:09:01,634][DEBUG] LightlyTrain is not installed from a git repository.
45
+ [2025-08-14 13:09:01,634][DEBUG] Run directory Git Information:
46
+ [2025-08-14 13:09:01,634][DEBUG] Branch: feat/ssl
47
+ [2025-08-14 13:09:01,634][DEBUG] Commit: 61f4482b9df47dedf1dd991b89913f2e5962454c
48
+ [2025-08-14 13:09:01,634][DEBUG] Uncommitted changes: None
49
+ [2025-08-14 13:09:01,634][DEBUG] Dependencies:
50
+ [2025-08-14 13:09:01,634][DEBUG] - torch 2.9.0.dev20250806+cu128
51
+ [2025-08-14 13:09:01,634][DEBUG] - torchvision 0.24.0.dev20250806+cu128
52
+ [2025-08-14 13:09:01,634][DEBUG] - pytorch-lightning 2.5.2
53
+ [2025-08-14 13:09:01,634][DEBUG] - Pillow 11.3.0
54
+ [2025-08-14 13:09:01,634][DEBUG] - pillow-simd x
55
+ [2025-08-14 13:09:01,634][DEBUG] Optional dependencies:
56
+ [2025-08-14 13:09:01,634][DEBUG] - super-gradients x
57
+ [2025-08-14 13:09:01,634][DEBUG] - timm 1.0.19
58
+ [2025-08-14 13:09:01,634][DEBUG] - ultralytics x
59
+ [2025-08-14 13:09:01,634][DEBUG] - wandb 0.21.1
60
+ [2025-08-14 13:09:01,634][DEBUG] CPUs: 6
61
+ [2025-08-14 13:09:01,635][DEBUG] GPUs: 1
62
+ [2025-08-14 13:09:01,635][DEBUG] - NVIDIA GeForce RTX 5060 Ti 12.0 (16612917248)
63
+ [2025-08-14 13:09:01,635][DEBUG] Environment variables:
64
+ [2025-08-14 13:09:01,641][DEBUG] Getting transform args for method 'dino'.
65
+ [2025-08-14 13:09:01,641][DEBUG] Using additional transform arguments {'image_size': (224, 224)}.
66
+ [2025-08-14 13:09:01,641][DEBUG] Getting transform for method 'dino'.
67
+ [2025-08-14 13:09:01,648][DEBUG] Making sure data directory '/home/duckq1u/Documents/workspace/cuong/ent-labotary/data/kyucapsule' exists and is not empty.
68
+ [2025-08-14 13:09:01,648][INFO] Initializing dataset from '/home/duckq1u/Documents/workspace/cuong/ent-labotary/data/kyucapsule'.
69
+ [2025-08-14 13:09:01,649][DEBUG] Writing filenames to '/tmp/tmpiaccn5q5' (chunk_size=10000)
70
+ [2025-08-14 13:09:01,899][DEBUG] Creating memory mapped sequence with 18481 'filenames'.
71
+ [2025-08-14 13:09:01,899][DEBUG] Found dataset size 18481.
72
+ [2025-08-14 13:09:01,900][DEBUG] Getting embedding model with embedding dimension None.
73
+ [2025-08-14 13:09:01,900][WARNING] Could not find pooling layer on the model, defaulting to AdaptiveAvgPool2d
74
+ [2025-08-14 13:09:01,900][DEBUG] Using jsonl logger with args flush_logs_every_n_steps=100
75
+ [2025-08-14 13:09:01,902][DEBUG] Using tensorboard logger with args name='' version='' log_graph=False default_hp_metric=True prefix='' sub_dir=None
76
+ [2025-08-14 13:09:01,903][DEBUG] Using wandb logger with args name=None version=None offline=False anonymous=None project='ent-endoscopy-ssl' log_model=False prefix='' checkpoint_name=None
77
+ [2025-08-14 13:09:01,903][DEBUG] Using loggers ['JSONLLogger', 'TensorBoardLogger', 'WandbLogger'].
78
+ [2025-08-14 13:09:01,907][DEBUG] Getting accelerator for 'auto'.
79
+ [2025-08-14 13:09:01,907][DEBUG] CUDA is available, defaulting to CUDA.
80
+ [2025-08-14 13:09:01,907][DEBUG] Detected 1 devices.
81
+ [2025-08-14 13:09:01,907][DEBUG] Using strategy 'auto'.
82
+ [2025-08-14 13:09:01,907][DEBUG] Getting trainer.
83
+ [2025-08-14 13:09:01,907][DEBUG] Using sync_batchnorm 'True'.
84
+ [2025-08-14 13:09:01,913][INFO] GPU available: True (cuda), used: True
85
+ [2025-08-14 13:09:01,913][INFO] TPU available: False, using: 0 TPU cores
86
+ [2025-08-14 13:09:01,914][INFO] HPU available: False, using: 0 HPUs
87
+ [2025-08-14 13:09:01,914][DEBUG] Detected 1 nodes and 1 devices per node.
88
+ [2025-08-14 13:09:01,914][DEBUG] Total number of devices: 1.
89
+ [2025-08-14 13:09:01,914][DEBUG] Detected dataset size 18481.
90
+ [2025-08-14 13:09:01,914][DEBUG] Using batch size per device 32.
91
+ [2025-08-14 13:09:01,914][DEBUG] Using optimizer 'OptimizerType.SGD'.
92
+ [2025-08-14 13:09:01,914][DEBUG] Getting method args for 'DINO'
93
+ [2025-08-14 13:09:01,914][DEBUG] Getting method for 'DINO'
94
+ [2025-08-14 13:09:01,966][WARNING] /home/duckq1u/miniconda3/envs/ent-lab/lib/python3.11/site-packages/torch/nn/utils/weight_norm.py:144: FutureWarning: `torch.nn.utils.weight_norm` is deprecated in favor of `torch.nn.utils.parametrizations.weight_norm`.
95
+ WeightNorm.apply(module, name, dim)
96
+
97
+ [2025-08-14 13:09:02,074][INFO] Resolved configuration:
98
+ {
99
+ "accelerator": "CUDAAccelerator",
100
+ "batch_size": 32,
101
+ "callbacks": {
102
+ "device_stats_monitor": {},
103
+ "early_stopping": {
104
+ "check_finite": true,
105
+ "monitor": "train_loss",
106
+ "patience": 1000000000000
107
+ },
108
+ "learning_rate_monitor": {},
109
+ "model_checkpoint": {
110
+ "enable_version_counter": false,
111
+ "every_n_epochs": null,
112
+ "save_last": true,
113
+ "save_top_k": 1
114
+ }
115
+ },
116
+ "checkpoint": null,
117
+ "data": "data/kyucapsule",
118
+ "devices": 1,
119
+ "embed_dim": null,
120
+ "epochs": 300,
121
+ "loader_args": null,
122
+ "loggers": {
123
+ "jsonl": {
124
+ "flush_logs_every_n_steps": 100
125
+ },
126
+ "tensorboard": {
127
+ "default_hp_metric": true,
128
+ "log_graph": false,
129
+ "name": "",
130
+ "prefix": "",
131
+ "sub_dir": null,
132
+ "version": ""
133
+ },
134
+ "wandb": {
135
+ "anonymous": null,
136
+ "checkpoint_name": null,
137
+ "log_model": false,
138
+ "name": null,
139
+ "offline": false,
140
+ "prefix": "",
141
+ "project": "ent-endoscopy-ssl",
142
+ "version": null
143
+ }
144
+ },
145
+ "method": "dino",
146
+ "method_args": {
147
+ "batch_norm": false,
148
+ "bottleneck_dim": 256,
149
+ "center_momentum": 0.9,
150
+ "hidden_dim": 2048,
151
+ "momentum_end": 1.0,
152
+ "momentum_start": 0.99,
153
+ "norm_last_layer": true,
154
+ "output_dim": 1024,
155
+ "student_freeze_last_layer_epochs": 1,
156
+ "student_temp": 0.1,
157
+ "teacher_temp": 0.02,
158
+ "warmup_teacher_temp": 0.02,
159
+ "warmup_teacher_temp_epochs": 30,
160
+ "weight_decay_end": 0.0001,
161
+ "weight_decay_start": 0.0001
162
+ },
163
+ "model": "VisionTransformer",
164
+ "model_args": null,
165
+ "num_nodes": 1,
166
+ "num_workers": 5,
167
+ "optim": "sgd",
168
+ "optim_args": {
169
+ "lr": 0.03,
170
+ "momentum": 0.9,
171
+ "weight_decay": 0.0001
172
+ },
173
+ "out": "outputs/ssl_dino/vit16s",
174
+ "overwrite": true,
175
+ "precision": "32-true",
176
+ "resume": false,
177
+ "seed": 0,
178
+ "strategy": "SingleDeviceStrategy",
179
+ "trainer_args": null,
180
+ "transform_args": {
181
+ "color_jitter": {
182
+ "brightness": 0.8,
183
+ "contrast": 0.8,
184
+ "hue": 0.2,
185
+ "prob": 0.8,
186
+ "saturation": 0.4,
187
+ "strength": 0.5
188
+ },
189
+ "gaussian_blur": {
190
+ "blur_limit": 0,
191
+ "prob": 1.0,
192
+ "sigmas": [
193
+ 0.1,
194
+ 2.0
195
+ ]
196
+ },
197
+ "global_view_1": {
198
+ "gaussian_blur": {
199
+ "blur_limit": 0,
200
+ "prob": 0.1,
201
+ "sigmas": [
202
+ 0.1,
203
+ 2.0
204
+ ]
205
+ },
206
+ "solarize": {
207
+ "prob": 0.2,
208
+ "threshold": 0.5
209
+ }
210
+ },
211
+ "image_size": [
212
+ 224,
213
+ 224
214
+ ],
215
+ "local_view": {
216
+ "gaussian_blur": {
217
+ "blur_limit": 0,
218
+ "prob": 0.5,
219
+ "sigmas": [
220
+ 0.1,
221
+ 2.0
222
+ ]
223
+ },
224
+ "num_views": 6,
225
+ "random_resize": {
226
+ "max_scale": 0.14,
227
+ "min_scale": 0.05
228
+ },
229
+ "view_size": [
230
+ 96,
231
+ 96
232
+ ]
233
+ },
234
+ "normalize": {
235
+ "mean": [
236
+ 0.485,
237
+ 0.456,
238
+ 0.406
239
+ ],
240
+ "std": [
241
+ 0.229,
242
+ 0.224,
243
+ 0.225
244
+ ]
245
+ },
246
+ "random_flip": {
247
+ "horizontal_prob": 0.5,
248
+ "vertical_prob": 0.0
249
+ },
250
+ "random_gray_scale": 0.2,
251
+ "random_resize": {
252
+ "max_scale": 1.0,
253
+ "min_scale": 0.14
254
+ },
255
+ "random_rotation": null,
256
+ "solarize": null
257
+ }
258
+ }
259
+ [2025-08-14 13:09:04,586][WARNING] /home/duckq1u/miniconda3/envs/ent-lab/lib/python3.11/site-packages/torch/__init__.py:1539: UserWarning: Please use the new API settings to control TF32 behavior, such as torch.backends.cudnn.conv.fp32_precision = 'tf32' or torch.backends.cuda.matmul.fp32_precision = 'ieee'. Old settings, e.g, torch.backends.cuda.matmul.allow_tf32 = True, torch.backends.cudnn.allow_tf32 = True, allowTF32CuDNN() and allowTF32CuBLAS() will be deprecated after Pytorch 2.9. Please see https://pytorch.org/docs/main/notes/cuda.html#tensorfloat-32-tf32-on-ampere-and-later-devices (Triggered internally at /pytorch/aten/src/ATen/Context.cpp:80.)
260
+ return _C._get_float32_matmul_precision()
261
+
262
+ [2025-08-14 13:09:04,586][INFO] You are using a CUDA device ('NVIDIA GeForce RTX 5060 Ti') that has Tensor Cores. To properly utilize them, you should set `torch.set_float32_matmul_precision('medium' | 'high')` which will trade-off precision for performance. For more details, read https://pytorch.org/docs/stable/generated/torch.set_float32_matmul_precision.html#torch.set_float32_matmul_precision
263
+ [2025-08-14 13:09:04,817][INFO] LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0]
264
+ [2025-08-14 13:09:04,894][INFO] Loading `train_dataloader` to estimate number of stepping batches.
265
+ [2025-08-14 13:09:05,027][INFO]
266
+ | Name | Type | Params | Mode
267
+ -----------------------------------------------------------------------
268
+ 0 | teacher_embedding_model | EmbeddingModel | 30.1 M | train
269
+ 1 | teacher_projection_head | DINOProjectionHead | 5.8 M | train
270
+ 2 | student_embedding_model | EmbeddingModel | 30.1 M | train
271
+ 3 | student_projection_head | DINOProjectionHead | 5.8 M | train
272
+ 4 | flatten | Flatten | 0 | train
273
+ 5 | criterion | DINOLoss | 0 | train
274
+ -----------------------------------------------------------------------
275
+ 71.7 M Trainable params
276
+ 2.0 K Non-trainable params
277
+ 71.7 M Total params
278
+ 286.782 Total estimated model params size (MB)
279
+ 578 Modules in train mode
280
+ 0 Modules in eval mode
281
+ [2025-08-15 07:08:42,748][INFO] Args: {
282
+ "accelerator": "auto",
283
+ "batch_size": 32,
284
+ "callbacks": null,
285
+ "checkpoint": null,
286
+ "data": "data/kyucapsule",
287
+ "devices": "auto",
288
+ "embed_dim": null,
289
+ "epochs": 300,
290
+ "loader_args": null,
291
+ "loggers": {
292
+ "wandb": {
293
+ "project": "ent-endoscopy-ssl"
294
+ }
295
+ },
296
+ "method": "dino",
297
+ "method_args": null,
298
+ "model": "VisionTransformer",
299
+ "model_args": null,
300
+ "num_nodes": 1,
301
+ "num_workers": "auto",
302
+ "optim": "auto",
303
+ "optim_args": null,
304
+ "out": "outputs/ssl_dino/vit16s",
305
+ "overwrite": true,
306
+ "precision": "32-true",
307
+ "resume": true,
308
+ "seed": 0,
309
+ "strategy": "auto",
310
+ "trainer_args": null,
311
+ "transform_args": {
312
+ "image_size": [
313
+ 224,
314
+ 224
315
+ ]
316
+ }
317
+ }
318
+ [2025-08-15 07:08:42,749][INFO] Using output directory '/home/duckq1u/Documents/workspace/cuong/ent-labotary/outputs/ssl_dino/vit16s'.
319
+ [2025-08-15 07:08:42,945][DEBUG] '/home/duckq1u/miniconda3/envs/ent-lab/lib/python3.11/site-packages/lightly_train' is not a git repository.
320
+ [2025-08-15 07:08:42,948][DEBUG] Platform: Linux-6.16.0-zen1-1-zen-x86_64-with-glibc2.42
321
+ [2025-08-15 07:08:42,948][DEBUG] Python: 3.11.13
322
+ [2025-08-15 07:08:42,949][DEBUG] LightlyTrain: 0.6.1
323
+ [2025-08-15 07:08:42,949][DEBUG] LightlyTrain Git Information:
324
+ [2025-08-15 07:08:42,949][DEBUG] LightlyTrain is not installed from a git repository.
325
+ [2025-08-15 07:08:42,949][DEBUG] Run directory Git Information:
326
+ [2025-08-15 07:08:42,949][DEBUG] Branch: feat/ssl
327
+ [2025-08-15 07:08:42,949][DEBUG] Commit: 61f4482b9df47dedf1dd991b89913f2e5962454c
328
+ [2025-08-15 07:08:42,949][DEBUG] Uncommitted changes: M src/experiment/ssl_dino.py
329
+ [2025-08-15 07:08:42,949][DEBUG] Dependencies:
330
+ [2025-08-15 07:08:42,949][DEBUG] - torch 2.9.0.dev20250806+cu128
331
+ [2025-08-15 07:08:42,949][DEBUG] - torchvision 0.24.0.dev20250806+cu128
332
+ [2025-08-15 07:08:42,949][DEBUG] - pytorch-lightning 2.5.2
333
+ [2025-08-15 07:08:42,949][DEBUG] - Pillow 11.3.0
334
+ [2025-08-15 07:08:42,949][DEBUG] - pillow-simd x
335
+ [2025-08-15 07:08:42,949][DEBUG] Optional dependencies:
336
+ [2025-08-15 07:08:42,949][DEBUG] - super-gradients x
337
+ [2025-08-15 07:08:42,949][DEBUG] - timm 1.0.19
338
+ [2025-08-15 07:08:42,949][DEBUG] - ultralytics x
339
+ [2025-08-15 07:08:42,949][DEBUG] - wandb 0.21.1
340
+ [2025-08-15 07:08:42,949][DEBUG] CPUs: 6
341
+ [2025-08-15 07:08:42,949][DEBUG] GPUs: 1
342
+ [2025-08-15 07:08:42,949][DEBUG] - NVIDIA GeForce RTX 5060 Ti 12.0 (16612917248)
343
+ [2025-08-15 07:08:42,949][DEBUG] Environment variables:
344
+ [2025-08-15 07:08:42,955][DEBUG] Getting transform args for method 'dino'.
345
+ [2025-08-15 07:08:42,955][DEBUG] Using additional transform arguments {'image_size': (224, 224)}.
346
+ [2025-08-15 07:08:42,955][DEBUG] Getting transform for method 'dino'.
347
+ [2025-08-15 07:08:42,962][DEBUG] Making sure data directory '/home/duckq1u/Documents/workspace/cuong/ent-labotary/data/kyucapsule' exists and is not empty.
348
+ [2025-08-15 07:08:42,962][INFO] Initializing dataset from '/home/duckq1u/Documents/workspace/cuong/ent-labotary/data/kyucapsule'.
349
+ [2025-08-15 07:08:42,962][DEBUG] Writing filenames to '/tmp/tmp3zuaqqfr' (chunk_size=10000)
350
+ [2025-08-15 07:08:43,158][DEBUG] Creating memory mapped sequence with 18481 'filenames'.
351
+ [2025-08-15 07:08:43,158][DEBUG] Found dataset size 18481.
352
+ [2025-08-15 07:08:43,158][DEBUG] Getting embedding model with embedding dimension None.
353
+ [2025-08-15 07:08:43,159][WARNING] Could not find pooling layer on the model, defaulting to AdaptiveAvgPool2d
354
+ [2025-08-15 07:08:43,159][DEBUG] Using jsonl logger with args flush_logs_every_n_steps=100
355
+ [2025-08-15 07:08:43,162][DEBUG] Using tensorboard logger with args name='' version='' log_graph=False default_hp_metric=True prefix='' sub_dir=None
356
+ [2025-08-15 07:08:43,163][DEBUG] Using wandb logger with args name=None version=None offline=False anonymous=None project='ent-endoscopy-ssl' log_model=False prefix='' checkpoint_name=None
357
+ [2025-08-15 07:08:43,163][DEBUG] Using loggers ['JSONLLogger', 'TensorBoardLogger', 'WandbLogger'].
358
+ [2025-08-15 07:08:43,168][DEBUG] Getting accelerator for 'auto'.
359
+ [2025-08-15 07:08:43,168][DEBUG] CUDA is available, defaulting to CUDA.
360
+ [2025-08-15 07:08:43,168][DEBUG] Detected 1 devices.
361
+ [2025-08-15 07:08:43,168][DEBUG] Using strategy 'auto'.
362
+ [2025-08-15 07:08:43,168][DEBUG] Getting trainer.
363
+ [2025-08-15 07:08:43,168][DEBUG] Using sync_batchnorm 'True'.
364
+ [2025-08-15 07:08:43,173][INFO] GPU available: True (cuda), used: True
365
+ [2025-08-15 07:08:43,174][INFO] TPU available: False, using: 0 TPU cores
366
+ [2025-08-15 07:08:43,174][INFO] HPU available: False, using: 0 HPUs
367
+ [2025-08-15 07:08:43,174][DEBUG] Detected 1 nodes and 1 devices per node.
368
+ [2025-08-15 07:08:43,174][DEBUG] Total number of devices: 1.
369
+ [2025-08-15 07:08:43,174][DEBUG] Detected dataset size 18481.
370
+ [2025-08-15 07:08:43,174][DEBUG] Using batch size per device 32.
371
+ [2025-08-15 07:08:43,174][DEBUG] Using optimizer 'OptimizerType.SGD'.
372
+ [2025-08-15 07:08:43,174][DEBUG] Getting method args for 'DINO'
373
+ [2025-08-15 07:08:43,174][DEBUG] Getting method for 'DINO'
374
+ [2025-08-15 07:08:43,221][WARNING] /home/duckq1u/miniconda3/envs/ent-lab/lib/python3.11/site-packages/torch/nn/utils/weight_norm.py:144: FutureWarning: `torch.nn.utils.weight_norm` is deprecated in favor of `torch.nn.utils.parametrizations.weight_norm`.
375
+ WeightNorm.apply(module, name, dim)
376
+
377
+ [2025-08-15 07:08:43,320][INFO] Resolved configuration:
378
+ {
379
+ "accelerator": "CUDAAccelerator",
380
+ "batch_size": 32,
381
+ "callbacks": {
382
+ "device_stats_monitor": {},
383
+ "early_stopping": {
384
+ "check_finite": true,
385
+ "monitor": "train_loss",
386
+ "patience": 1000000000000
387
+ },
388
+ "learning_rate_monitor": {},
389
+ "model_checkpoint": {
390
+ "enable_version_counter": false,
391
+ "every_n_epochs": null,
392
+ "save_last": true,
393
+ "save_top_k": 1
394
+ }
395
+ },
396
+ "checkpoint": null,
397
+ "data": "data/kyucapsule",
398
+ "devices": 1,
399
+ "embed_dim": null,
400
+ "epochs": 300,
401
+ "loader_args": null,
402
+ "loggers": {
403
+ "jsonl": {
404
+ "flush_logs_every_n_steps": 100
405
+ },
406
+ "tensorboard": {
407
+ "default_hp_metric": true,
408
+ "log_graph": false,
409
+ "name": "",
410
+ "prefix": "",
411
+ "sub_dir": null,
412
+ "version": ""
413
+ },
414
+ "wandb": {
415
+ "anonymous": null,
416
+ "checkpoint_name": null,
417
+ "log_model": false,
418
+ "name": null,
419
+ "offline": false,
420
+ "prefix": "",
421
+ "project": "ent-endoscopy-ssl",
422
+ "version": null
423
+ }
424
+ },
425
+ "method": "dino",
426
+ "method_args": {
427
+ "batch_norm": false,
428
+ "bottleneck_dim": 256,
429
+ "center_momentum": 0.9,
430
+ "hidden_dim": 2048,
431
+ "momentum_end": 1.0,
432
+ "momentum_start": 0.99,
433
+ "norm_last_layer": true,
434
+ "output_dim": 1024,
435
+ "student_freeze_last_layer_epochs": 1,
436
+ "student_temp": 0.1,
437
+ "teacher_temp": 0.02,
438
+ "warmup_teacher_temp": 0.02,
439
+ "warmup_teacher_temp_epochs": 30,
440
+ "weight_decay_end": 0.0001,
441
+ "weight_decay_start": 0.0001
442
+ },
443
+ "model": "VisionTransformer",
444
+ "model_args": null,
445
+ "num_nodes": 1,
446
+ "num_workers": 5,
447
+ "optim": "sgd",
448
+ "optim_args": {
449
+ "lr": 0.03,
450
+ "momentum": 0.9,
451
+ "weight_decay": 0.0001
452
+ },
453
+ "out": "outputs/ssl_dino/vit16s",
454
+ "overwrite": true,
455
+ "precision": "32-true",
456
+ "resume": true,
457
+ "seed": 0,
458
+ "strategy": "SingleDeviceStrategy",
459
+ "trainer_args": null,
460
+ "transform_args": {
461
+ "color_jitter": {
462
+ "brightness": 0.8,
463
+ "contrast": 0.8,
464
+ "hue": 0.2,
465
+ "prob": 0.8,
466
+ "saturation": 0.4,
467
+ "strength": 0.5
468
+ },
469
+ "gaussian_blur": {
470
+ "blur_limit": 0,
471
+ "prob": 1.0,
472
+ "sigmas": [
473
+ 0.1,
474
+ 2.0
475
+ ]
476
+ },
477
+ "global_view_1": {
478
+ "gaussian_blur": {
479
+ "blur_limit": 0,
480
+ "prob": 0.1,
481
+ "sigmas": [
482
+ 0.1,
483
+ 2.0
484
+ ]
485
+ },
486
+ "solarize": {
487
+ "prob": 0.2,
488
+ "threshold": 0.5
489
+ }
490
+ },
491
+ "image_size": [
492
+ 224,
493
+ 224
494
+ ],
495
+ "local_view": {
496
+ "gaussian_blur": {
497
+ "blur_limit": 0,
498
+ "prob": 0.5,
499
+ "sigmas": [
500
+ 0.1,
501
+ 2.0
502
+ ]
503
+ },
504
+ "num_views": 6,
505
+ "random_resize": {
506
+ "max_scale": 0.14,
507
+ "min_scale": 0.05
508
+ },
509
+ "view_size": [
510
+ 96,
511
+ 96
512
+ ]
513
+ },
514
+ "normalize": {
515
+ "mean": [
516
+ 0.485,
517
+ 0.456,
518
+ 0.406
519
+ ],
520
+ "std": [
521
+ 0.229,
522
+ 0.224,
523
+ 0.225
524
+ ]
525
+ },
526
+ "random_flip": {
527
+ "horizontal_prob": 0.5,
528
+ "vertical_prob": 0.0
529
+ },
530
+ "random_gray_scale": 0.2,
531
+ "random_resize": {
532
+ "max_scale": 1.0,
533
+ "min_scale": 0.14
534
+ },
535
+ "random_rotation": null,
536
+ "solarize": null
537
+ }
538
+ }
539
+ [2025-08-15 07:08:46,350][WARNING] /home/duckq1u/miniconda3/envs/ent-lab/lib/python3.11/site-packages/torch/__init__.py:1539: UserWarning: Please use the new API settings to control TF32 behavior, such as torch.backends.cudnn.conv.fp32_precision = 'tf32' or torch.backends.cuda.matmul.fp32_precision = 'ieee'. Old settings, e.g, torch.backends.cuda.matmul.allow_tf32 = True, torch.backends.cudnn.allow_tf32 = True, allowTF32CuDNN() and allowTF32CuBLAS() will be deprecated after Pytorch 2.9. Please see https://pytorch.org/docs/main/notes/cuda.html#tensorfloat-32-tf32-on-ampere-and-later-devices (Triggered internally at /pytorch/aten/src/ATen/Context.cpp:80.)
540
+ return _C._get_float32_matmul_precision()
541
+
542
+ [2025-08-15 07:08:46,350][INFO] You are using a CUDA device ('NVIDIA GeForce RTX 5060 Ti') that has Tensor Cores. To properly utilize them, you should set `torch.set_float32_matmul_precision('medium' | 'high')` which will trade-off precision for performance. For more details, read https://pytorch.org/docs/stable/generated/torch.set_float32_matmul_precision.html#torch.set_float32_matmul_precision
543
+ [2025-08-15 07:08:46,540][INFO] Restoring states from the checkpoint path at /home/duckq1u/Documents/workspace/cuong/ent-labotary/outputs/ssl_dino/vit16s/checkpoints/last.ckpt
544
+ [2025-08-15 07:08:47,928][INFO] LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0]
545
+ [2025-08-15 07:08:47,994][INFO] Loading `train_dataloader` to estimate number of stepping batches.
546
+ [2025-08-15 07:08:48,121][INFO]
547
+ | Name | Type | Params | Mode
548
+ -----------------------------------------------------------------------
549
+ 0 | teacher_embedding_model | EmbeddingModel | 30.1 M | train
550
+ 1 | teacher_projection_head | DINOProjectionHead | 5.8 M | train
551
+ 2 | student_embedding_model | EmbeddingModel | 30.1 M | train
552
+ 3 | student_projection_head | DINOProjectionHead | 5.8 M | train
553
+ 4 | flatten | Flatten | 0 | train
554
+ 5 | criterion | DINOLoss | 0 | train
555
+ -----------------------------------------------------------------------
556
+ 71.7 M Trainable params
557
+ 2.0 K Non-trainable params
558
+ 71.7 M Total params
559
+ 286.782 Total estimated model params size (MB)
560
+ 578 Modules in train mode
561
+ 0 Modules in eval mode
562
+ [2025-08-15 07:08:48,148][INFO] Restored all states from the checkpoint at /home/duckq1u/Documents/workspace/cuong/ent-labotary/outputs/ssl_dino/vit16s/checkpoints/last.ckpt
563
+ [2025-08-15 13:09:04,955][INFO]
564
+ Detected KeyboardInterrupt, attempting graceful shutdown ...
565
+ [2025-08-15 16:03:16,504][INFO] Args: {
566
+ "accelerator": "auto",
567
+ "batch_size": 32,
568
+ "callbacks": null,
569
+ "checkpoint": null,
570
+ "data": "data/kyucapsule",
571
+ "devices": "auto",
572
+ "embed_dim": null,
573
+ "epochs": 300,
574
+ "loader_args": null,
575
+ "loggers": {
576
+ "wandb": {
577
+ "project": "ent-endoscopy-ssl"
578
+ }
579
+ },
580
+ "method": "dino",
581
+ "method_args": null,
582
+ "model": "VisionTransformer",
583
+ "model_args": null,
584
+ "num_nodes": 1,
585
+ "num_workers": "auto",
586
+ "optim": "auto",
587
+ "optim_args": null,
588
+ "out": "outputs/ssl_dino/vit16s",
589
+ "overwrite": true,
590
+ "precision": "32-true",
591
+ "resume": true,
592
+ "seed": 0,
593
+ "strategy": "auto",
594
+ "trainer_args": null,
595
+ "transform_args": {
596
+ "image_size": [
597
+ 224,
598
+ 224
599
+ ]
600
+ }
601
+ }
602
+ [2025-08-15 16:03:16,505][INFO] Using output directory '/home/duckq1u/Documents/workspace/cuong/ent-labotary/outputs/ssl_dino/vit16s'.
603
+ [2025-08-15 16:03:16,636][DEBUG] '/home/duckq1u/miniconda3/envs/ent-lab/lib/python3.11/site-packages/lightly_train' is not a git repository.
604
+ [2025-08-15 16:03:16,654][DEBUG] Platform: Linux-6.14.0-arch1-1-x86_64-with-glibc2.42
605
+ [2025-08-15 16:03:16,654][DEBUG] Python: 3.11.13
606
+ [2025-08-15 16:03:16,654][DEBUG] LightlyTrain: 0.6.1
607
+ [2025-08-15 16:03:16,654][DEBUG] LightlyTrain Git Information:
608
+ [2025-08-15 16:03:16,654][DEBUG] LightlyTrain is not installed from a git repository.
609
+ [2025-08-15 16:03:16,654][DEBUG] Run directory Git Information:
610
+ [2025-08-15 16:03:16,654][DEBUG] Branch: feat/ssl
611
+ [2025-08-15 16:03:16,654][DEBUG] Commit: 61f4482b9df47dedf1dd991b89913f2e5962454c
612
+ [2025-08-15 16:03:16,654][DEBUG] Uncommitted changes: M src/experiment/ssl_dino.py
613
+ [2025-08-15 16:03:16,654][DEBUG] Dependencies:
614
+ [2025-08-15 16:03:16,654][DEBUG] - torch 2.9.0.dev20250806+cu128
615
+ [2025-08-15 16:03:16,654][DEBUG] - torchvision 0.24.0.dev20250806+cu128
616
+ [2025-08-15 16:03:16,654][DEBUG] - pytorch-lightning 2.5.2
617
+ [2025-08-15 16:03:16,654][DEBUG] - Pillow 11.3.0
618
+ [2025-08-15 16:03:16,655][DEBUG] - pillow-simd x
619
+ [2025-08-15 16:03:16,655][DEBUG] Optional dependencies:
620
+ [2025-08-15 16:03:16,655][DEBUG] - super-gradients x
621
+ [2025-08-15 16:03:16,655][DEBUG] - timm 1.0.19
622
+ [2025-08-15 16:03:16,655][DEBUG] - ultralytics x
623
+ [2025-08-15 16:03:16,655][DEBUG] - wandb 0.21.1
624
+ [2025-08-15 16:03:16,655][DEBUG] CPUs: 6
625
+ [2025-08-15 16:03:16,655][DEBUG] GPUs: 1
626
+ [2025-08-15 16:03:16,655][DEBUG] - NVIDIA GeForce RTX 5060 Ti 12.0 (16612917248)
627
+ [2025-08-15 16:03:16,655][DEBUG] Environment variables:
628
+ [2025-08-15 16:03:16,671][DEBUG] Getting transform args for method 'dino'.
629
+ [2025-08-15 16:03:16,671][DEBUG] Using additional transform arguments {'image_size': (224, 224)}.
630
+ [2025-08-15 16:03:16,672][DEBUG] Getting transform for method 'dino'.
631
+ [2025-08-15 16:03:16,680][DEBUG] Making sure data directory '/home/duckq1u/Documents/workspace/cuong/ent-labotary/data/kyucapsule' exists and is not empty.
632
+ [2025-08-15 16:03:16,681][INFO] Initializing dataset from '/home/duckq1u/Documents/workspace/cuong/ent-labotary/data/kyucapsule'.
633
+ [2025-08-15 16:03:16,681][DEBUG] Writing filenames to '/tmp/tmpgd726y1j' (chunk_size=10000)
634
+ [2025-08-15 16:03:16,945][DEBUG] Creating memory mapped sequence with 18481 'filenames'.
635
+ [2025-08-15 16:03:16,945][DEBUG] Found dataset size 18481.
636
+ [2025-08-15 16:03:16,945][DEBUG] Getting embedding model with embedding dimension None.
637
+ [2025-08-15 16:03:16,946][WARNING] Could not find pooling layer on the model, defaulting to AdaptiveAvgPool2d
638
+ [2025-08-15 16:03:16,946][DEBUG] Using jsonl logger with args flush_logs_every_n_steps=100
639
+ [2025-08-15 16:03:16,948][DEBUG] Using tensorboard logger with args name='' version='' log_graph=False default_hp_metric=True prefix='' sub_dir=None
640
+ [2025-08-15 16:03:16,949][DEBUG] Using wandb logger with args name=None version=None offline=False anonymous=None project='ent-endoscopy-ssl' log_model=False prefix='' checkpoint_name=None
641
+ [2025-08-15 16:03:16,950][DEBUG] Using loggers ['JSONLLogger', 'TensorBoardLogger', 'WandbLogger'].
642
+ [2025-08-15 16:03:16,953][DEBUG] Getting accelerator for 'auto'.
643
+ [2025-08-15 16:03:16,953][DEBUG] CUDA is available, defaulting to CUDA.
644
+ [2025-08-15 16:03:16,953][DEBUG] Detected 1 devices.
645
+ [2025-08-15 16:03:16,953][DEBUG] Using strategy 'auto'.
646
+ [2025-08-15 16:03:16,953][DEBUG] Getting trainer.
647
+ [2025-08-15 16:03:16,953][DEBUG] Using sync_batchnorm 'True'.
648
+ [2025-08-15 16:03:16,959][INFO] GPU available: True (cuda), used: True
649
+ [2025-08-15 16:03:16,959][INFO] TPU available: False, using: 0 TPU cores
650
+ [2025-08-15 16:03:16,959][INFO] HPU available: False, using: 0 HPUs
651
+ [2025-08-15 16:03:16,959][DEBUG] Detected 1 nodes and 1 devices per node.
652
+ [2025-08-15 16:03:16,959][DEBUG] Total number of devices: 1.
653
+ [2025-08-15 16:03:16,959][DEBUG] Detected dataset size 18481.
654
+ [2025-08-15 16:03:16,960][DEBUG] Using batch size per device 32.
655
+ [2025-08-15 16:03:16,960][DEBUG] Using optimizer 'OptimizerType.SGD'.
656
+ [2025-08-15 16:03:16,960][DEBUG] Getting method args for 'DINO'
657
+ [2025-08-15 16:03:16,960][DEBUG] Getting method for 'DINO'
658
+ [2025-08-15 16:03:17,008][WARNING] /home/duckq1u/miniconda3/envs/ent-lab/lib/python3.11/site-packages/torch/nn/utils/weight_norm.py:144: FutureWarning: `torch.nn.utils.weight_norm` is deprecated in favor of `torch.nn.utils.parametrizations.weight_norm`.
659
+ WeightNorm.apply(module, name, dim)
660
+
661
+ [2025-08-15 16:03:17,116][INFO] Resolved configuration:
662
+ {
663
+ "accelerator": "CUDAAccelerator",
664
+ "batch_size": 32,
665
+ "callbacks": {
666
+ "device_stats_monitor": {},
667
+ "early_stopping": {
668
+ "check_finite": true,
669
+ "monitor": "train_loss",
670
+ "patience": 1000000000000
671
+ },
672
+ "learning_rate_monitor": {},
673
+ "model_checkpoint": {
674
+ "enable_version_counter": false,
675
+ "every_n_epochs": null,
676
+ "save_last": true,
677
+ "save_top_k": 1
678
+ }
679
+ },
680
+ "checkpoint": null,
681
+ "data": "data/kyucapsule",
682
+ "devices": 1,
683
+ "embed_dim": null,
684
+ "epochs": 300,
685
+ "loader_args": null,
686
+ "loggers": {
687
+ "jsonl": {
688
+ "flush_logs_every_n_steps": 100
689
+ },
690
+ "tensorboard": {
691
+ "default_hp_metric": true,
692
+ "log_graph": false,
693
+ "name": "",
694
+ "prefix": "",
695
+ "sub_dir": null,
696
+ "version": ""
697
+ },
698
+ "wandb": {
699
+ "anonymous": null,
700
+ "checkpoint_name": null,
701
+ "log_model": false,
702
+ "name": null,
703
+ "offline": false,
704
+ "prefix": "",
705
+ "project": "ent-endoscopy-ssl",
706
+ "version": null
707
+ }
708
+ },
709
+ "method": "dino",
710
+ "method_args": {
711
+ "batch_norm": false,
712
+ "bottleneck_dim": 256,
713
+ "center_momentum": 0.9,
714
+ "hidden_dim": 2048,
715
+ "momentum_end": 1.0,
716
+ "momentum_start": 0.99,
717
+ "norm_last_layer": true,
718
+ "output_dim": 1024,
719
+ "student_freeze_last_layer_epochs": 1,
720
+ "student_temp": 0.1,
721
+ "teacher_temp": 0.02,
722
+ "warmup_teacher_temp": 0.02,
723
+ "warmup_teacher_temp_epochs": 30,
724
+ "weight_decay_end": 0.0001,
725
+ "weight_decay_start": 0.0001
726
+ },
727
+ "model": "VisionTransformer",
728
+ "model_args": null,
729
+ "num_nodes": 1,
730
+ "num_workers": 5,
731
+ "optim": "sgd",
732
+ "optim_args": {
733
+ "lr": 0.03,
734
+ "momentum": 0.9,
735
+ "weight_decay": 0.0001
736
+ },
737
+ "out": "outputs/ssl_dino/vit16s",
738
+ "overwrite": true,
739
+ "precision": "32-true",
740
+ "resume": true,
741
+ "seed": 0,
742
+ "strategy": "SingleDeviceStrategy",
743
+ "trainer_args": null,
744
+ "transform_args": {
745
+ "color_jitter": {
746
+ "brightness": 0.8,
747
+ "contrast": 0.8,
748
+ "hue": 0.2,
749
+ "prob": 0.8,
750
+ "saturation": 0.4,
751
+ "strength": 0.5
752
+ },
753
+ "gaussian_blur": {
754
+ "blur_limit": 0,
755
+ "prob": 1.0,
756
+ "sigmas": [
757
+ 0.1,
758
+ 2.0
759
+ ]
760
+ },
761
+ "global_view_1": {
762
+ "gaussian_blur": {
763
+ "blur_limit": 0,
764
+ "prob": 0.1,
765
+ "sigmas": [
766
+ 0.1,
767
+ 2.0
768
+ ]
769
+ },
770
+ "solarize": {
771
+ "prob": 0.2,
772
+ "threshold": 0.5
773
+ }
774
+ },
775
+ "image_size": [
776
+ 224,
777
+ 224
778
+ ],
779
+ "local_view": {
780
+ "gaussian_blur": {
781
+ "blur_limit": 0,
782
+ "prob": 0.5,
783
+ "sigmas": [
784
+ 0.1,
785
+ 2.0
786
+ ]
787
+ },
788
+ "num_views": 6,
789
+ "random_resize": {
790
+ "max_scale": 0.14,
791
+ "min_scale": 0.05
792
+ },
793
+ "view_size": [
794
+ 96,
795
+ 96
796
+ ]
797
+ },
798
+ "normalize": {
799
+ "mean": [
800
+ 0.485,
801
+ 0.456,
802
+ 0.406
803
+ ],
804
+ "std": [
805
+ 0.229,
806
+ 0.224,
807
+ 0.225
808
+ ]
809
+ },
810
+ "random_flip": {
811
+ "horizontal_prob": 0.5,
812
+ "vertical_prob": 0.0
813
+ },
814
+ "random_gray_scale": 0.2,
815
+ "random_resize": {
816
+ "max_scale": 1.0,
817
+ "min_scale": 0.14
818
+ },
819
+ "random_rotation": null,
820
+ "solarize": null
821
+ }
822
+ }
823
+ [2025-08-15 16:03:19,697][WARNING] /home/duckq1u/miniconda3/envs/ent-lab/lib/python3.11/site-packages/torch/__init__.py:1539: UserWarning: Please use the new API settings to control TF32 behavior, such as torch.backends.cudnn.conv.fp32_precision = 'tf32' or torch.backends.cuda.matmul.fp32_precision = 'ieee'. Old settings, e.g, torch.backends.cuda.matmul.allow_tf32 = True, torch.backends.cudnn.allow_tf32 = True, allowTF32CuDNN() and allowTF32CuBLAS() will be deprecated after Pytorch 2.9. Please see https://pytorch.org/docs/main/notes/cuda.html#tensorfloat-32-tf32-on-ampere-and-later-devices (Triggered internally at /pytorch/aten/src/ATen/Context.cpp:80.)
824
+ return _C._get_float32_matmul_precision()
825
+
826
+ [2025-08-15 16:03:19,698][INFO] You are using a CUDA device ('NVIDIA GeForce RTX 5060 Ti') that has Tensor Cores. To properly utilize them, you should set `torch.set_float32_matmul_precision('medium' | 'high')` which will trade-off precision for performance. For more details, read https://pytorch.org/docs/stable/generated/torch.set_float32_matmul_precision.html#torch.set_float32_matmul_precision
827
+ [2025-08-15 16:03:19,887][INFO] Restoring states from the checkpoint path at /home/duckq1u/Documents/workspace/cuong/ent-labotary/outputs/ssl_dino/vit16s/checkpoints/last.ckpt
828
+ [2025-08-15 16:03:21,353][INFO] LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0]
829
+ [2025-08-15 16:03:21,421][INFO] Loading `train_dataloader` to estimate number of stepping batches.
830
+ [2025-08-15 16:03:21,554][INFO]
831
+ | Name | Type | Params | Mode
832
+ -----------------------------------------------------------------------
833
+ 0 | teacher_embedding_model | EmbeddingModel | 30.1 M | train
834
+ 1 | teacher_projection_head | DINOProjectionHead | 5.8 M | train
835
+ 2 | student_embedding_model | EmbeddingModel | 30.1 M | train
836
+ 3 | student_projection_head | DINOProjectionHead | 5.8 M | train
837
+ 4 | flatten | Flatten | 0 | train
838
+ 5 | criterion | DINOLoss | 0 | train
839
+ -----------------------------------------------------------------------
840
+ 71.7 M Trainable params
841
+ 2.0 K Non-trainable params
842
+ 71.7 M Total params
843
+ 286.782 Total estimated model params size (MB)
844
+ 578 Modules in train mode
845
+ 0 Modules in eval mode
846
+ [2025-08-15 16:03:21,581][INFO] Restored all states from the checkpoint at /home/duckq1u/Documents/workspace/cuong/ent-labotary/outputs/ssl_dino/vit16s/checkpoints/last.ckpt
847
+ [2025-08-15 19:47:37,131][INFO] `Trainer.fit` stopped: `max_epochs=300` reached.
848
+ [2025-08-15 19:47:37,444][INFO] Training completed.
849
+ [2025-08-15 19:47:37,453][DEBUG] Exporting model to '/home/duckq1u/Documents/workspace/cuong/ent-labotary/outputs/ssl_dino/vit16s/exported_models/exported_last.pt' in format 'ModelFormat.PACKAGE_DEFAULT'.
850
+ [2025-08-15 19:47:37,531][INFO] Example: How to use the exported model
851
+ ----------------------------------------------------------------------------------------
852
+ import timm
853
+ 
854
+ # Load the pretrained model
855
+ model = timm.create_model(
856
+  model_name='vit_small_patch16_224',
857
+  checkpoint_path='/home/duckq1u/Documents/workspace/cuong/ent-labotary/outputs/ssl_dino/vit16s/exported_models/exported_last.pt',
858
+ )
859
+ 
860
+ # Finetune or evaluate the model
861
+ ...
862
+ ----------------------------------------------------------------------------------------
863
+
864
+ [2025-08-15 19:47:37,532][INFO] Model exported.
ssl_dino/vit16s/vit_small_patch16_224_dino.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a27af0d8fe8782f723413ef7892fb869e05807ec28f9e89edc429ea7e477e1de
3
+ size 120360841
ssl_dino/vit16s/wandb/debug-internal.log ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2025-08-15T16:03:18.253890142+07:00","level":"INFO","msg":"stream: starting","core version":"0.21.1"}
2
+ {"time":"2025-08-15T16:03:18.721836531+07:00","level":"INFO","msg":"stream: created new stream","id":"46zcck3z"}
3
+ {"time":"2025-08-15T16:03:18.721871902+07:00","level":"INFO","msg":"stream: started","id":"46zcck3z"}
4
+ {"time":"2025-08-15T16:03:18.721888752+07:00","level":"INFO","msg":"sender: started","stream_id":"46zcck3z"}
5
+ {"time":"2025-08-15T16:03:18.721889062+07:00","level":"INFO","msg":"writer: started","stream_id":"46zcck3z"}
6
+ {"time":"2025-08-15T16:03:18.722182108+07:00","level":"INFO","msg":"handler: started","stream_id":"46zcck3z"}
7
+ {"time":"2025-08-15T19:47:38.25051031+07:00","level":"INFO","msg":"stream: closing","id":"46zcck3z"}
8
+ {"time":"2025-08-15T19:47:39.932846317+07:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
9
+ {"time":"2025-08-15T19:47:40.373460135+07:00","level":"INFO","msg":"handler: closed","stream_id":"46zcck3z"}
10
+ {"time":"2025-08-15T19:47:40.381105441+07:00","level":"INFO","msg":"sender: closed","stream_id":"46zcck3z"}
11
+ {"time":"2025-08-15T19:47:40.381156674+07:00","level":"INFO","msg":"stream: closed","id":"46zcck3z"}
ssl_dino/vit16s/wandb/debug.log ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2025-08-15 16:03:17,830 INFO MainThread:227131 [wandb_setup.py:_flush():80] Current SDK version is 0.21.1
2
+ 2025-08-15 16:03:17,830 INFO MainThread:227131 [wandb_setup.py:_flush():80] Configure stats pid to 227131
3
+ 2025-08-15 16:03:17,830 INFO MainThread:227131 [wandb_setup.py:_flush():80] Loading settings from /home/duckq1u/.config/wandb/settings
4
+ 2025-08-15 16:03:17,830 INFO MainThread:227131 [wandb_setup.py:_flush():80] Loading settings from /home/duckq1u/Documents/workspace/cuong/ent-labotary/wandb/settings
5
+ 2025-08-15 16:03:17,830 INFO MainThread:227131 [wandb_setup.py:_flush():80] Loading settings from environment variables
6
+ 2025-08-15 16:03:17,831 INFO MainThread:227131 [wandb_init.py:setup_run_log_directory():703] Logging user logs to /home/duckq1u/Documents/workspace/cuong/ent-labotary/outputs/ssl_dino/vit16s/wandb/run-20250815_160317-46zcck3z/logs/debug.log
7
+ 2025-08-15 16:03:17,831 INFO MainThread:227131 [wandb_init.py:setup_run_log_directory():704] Logging internal logs to /home/duckq1u/Documents/workspace/cuong/ent-labotary/outputs/ssl_dino/vit16s/wandb/run-20250815_160317-46zcck3z/logs/debug-internal.log
8
+ 2025-08-15 16:03:17,831 INFO MainThread:227131 [wandb_init.py:init():830] calling init triggers
9
+ 2025-08-15 16:03:17,831 INFO MainThread:227131 [wandb_init.py:init():835] wandb.init called with sweep_config: {}
10
+ config: {'_wandb': {}}
11
+ 2025-08-15 16:03:17,831 INFO MainThread:227131 [wandb_init.py:init():871] starting backend
12
+ 2025-08-15 16:03:18,241 INFO MainThread:227131 [wandb_init.py:init():874] sending inform_init request
13
+ 2025-08-15 16:03:18,247 INFO MainThread:227131 [wandb_init.py:init():882] backend started and connected
14
+ 2025-08-15 16:03:18,248 INFO MainThread:227131 [wandb_init.py:init():953] updated telemetry
15
+ 2025-08-15 16:03:18,255 INFO MainThread:227131 [wandb_init.py:init():977] communicating run to backend with 90.0 second timeout
16
+ 2025-08-15 16:03:19,421 INFO MainThread:227131 [wandb_init.py:init():1029] starting run threads in backend
17
+ 2025-08-15 16:03:19,692 INFO MainThread:227131 [wandb_run.py:_console_start():2494] atexit reg
18
+ 2025-08-15 16:03:19,692 INFO MainThread:227131 [wandb_run.py:_redirect():2342] redirect: wrap_raw
19
+ 2025-08-15 16:03:19,692 INFO MainThread:227131 [wandb_run.py:_redirect():2411] Wrapping output streams.
20
+ 2025-08-15 16:03:19,692 INFO MainThread:227131 [wandb_run.py:_redirect():2434] Redirects installed.
21
+ 2025-08-15 16:03:19,695 INFO MainThread:227131 [wandb_init.py:init():1075] run started, returning control to user process
22
+ 2025-08-15 16:03:19,696 INFO MainThread:227131 [wandb_run.py:_config_callback():1380] config_cb None None {'out': 'outputs/ssl_dino/vit16s', 'data': 'data/kyucapsule', 'model': 'VisionTransformer', 'method': 'dino', 'method_args': {'hidden_dim': 2048, 'bottleneck_dim': 256, 'output_dim': 1024, 'student_freeze_last_layer_epochs': 1, 'batch_norm': False, 'norm_last_layer': True, 'teacher_temp': 0.02, 'warmup_teacher_temp': 0.02, 'warmup_teacher_temp_epochs': 30, 'student_temp': 0.1, 'center_momentum': 0.9, 'momentum_start': 0.99, 'momentum_end': 1.0, 'weight_decay_start': 0.0001, 'weight_decay_end': 0.0001}, 'embed_dim': None, 'epochs': 300, 'batch_size': 32, 'num_workers': 5, 'devices': 1, 'num_nodes': 1, 'resume': True, 'checkpoint': None, 'overwrite': True, 'accelerator': 'CUDAAccelerator', 'strategy': 'SingleDeviceStrategy', 'precision': '32-true', 'seed': 0, 'loggers': {'jsonl': {'flush_logs_every_n_steps': 100}, 'tensorboard': {'name': '', 'version': '', 'log_graph': False, 'default_hp_metric': True, 'prefix': '', 'sub_dir': None}, 'wandb': {'name': None, 'version': None, 'offline': False, 'anonymous': None, 'project': 'ent-endoscopy-ssl', 'log_model': False, 'prefix': '', 'checkpoint_name': None}}, 'callbacks': {'learning_rate_monitor': {}, 'device_stats_monitor': {}, 'early_stopping': {'monitor': 'train_loss', 'patience': 1000000000000, 'check_finite': True}, 'model_checkpoint': {'save_last': True, 'enable_version_counter': False, 'save_top_k': 1, 'every_n_epochs': None}}, 'optim': 'sgd', 'optim_args': {'lr': 0.03, 'momentum': 0.9, 'weight_decay': 0.0001}, 'transform_args': {'image_size': [224, 224], 'random_resize': {'min_scale': 0.14, 'max_scale': 1.0}, 'random_flip': {'horizontal_prob': 0.5, 'vertical_prob': 0.0}, 'random_rotation': None, 'color_jitter': {'prob': 0.8, 'strength': 0.5, 'brightness': 0.8, 'contrast': 0.8, 'saturation': 0.4, 'hue': 0.2}, 'random_gray_scale': 0.2, 'normalize': {'mean': [0.485, 0.456, 0.406], 'std': [0.229, 0.224, 0.225]}, 'gaussian_blur': {'prob': 1.0, 'sigmas': [0.1, 2.0], 'blur_limit': 0}, 'solarize': None, 'global_view_1': {'gaussian_blur': {'prob': 0.1, 'sigmas': [0.1, 2.0], 'blur_limit': 0}, 'solarize': {'prob': 0.2, 'threshold': 0.5}}, 'local_view': {'num_views': 6, 'view_size': [96, 96], 'random_resize': {'min_scale': 0.05, 'max_scale': 0.14}, 'gaussian_blur': {'prob': 0.5, 'sigmas': [0.1, 2.0], 'blur_limit': 0}}}, 'loader_args': None, 'trainer_args': None, 'model_args': None}
23
+ 2025-08-15 19:47:38,211 INFO MsgRouterThr:227131 [mailbox.py:close():129] [no run ID] Closing mailbox, abandoning 1 handles.
ssl_dino/vit16s/wandb/run-20250814_130902-j9e757hl/files/media/images/augmentations_0_662741ecb5e7377f949e.png ADDED

Git LFS Details

  • SHA256: 662741ecb5e7377f949e52e318ac80c340f5dd86d4c932cfd8f65b47ff2675e9
  • Pointer size: 132 Bytes
  • Size of remote file: 2.01 MB
ssl_dino/vit16s/wandb/run-20250814_130902-j9e757hl/files/output.log ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ /home/duckq1u/miniconda3/envs/ent-lab/lib/python3.11/site-packages/torch/__init__.py:1539: UserWarning: Please use the new API settings to control TF32 behavior, such as torch.backends.cudnn.conv.fp32_precision = 'tf32' or torch.backends.cuda.matmul.fp32_precision = 'ieee'. Old settings, e.g, torch.backends.cuda.matmul.allow_tf32 = True, torch.backends.cudnn.allow_tf32 = True, allowTF32CuDNN() and allowTF32CuBLAS() will be deprecated after Pytorch 2.9. Please see https://pytorch.org/docs/main/notes/cuda.html#tensorfloat-32-tf32-on-ampere-and-later-devices (Triggered internally at /pytorch/aten/src/ATen/Context.cpp:80.)
2
+ return _C._get_float32_matmul_precision()
3
+ 
4
+ You are using a CUDA device ('NVIDIA GeForce RTX 5060 Ti') that has Tensor Cores. To properly utilize them, you should set `torch.set_float32_matmul_precision('medium' | 'high')` which will trade-off precision for performance. For more details, read https://pytorch.org/docs/stable/generated/torch.set_float32_matmul_precision.html#torch.set_float32_matmul_precision
5
+ LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0]
6
+ Loading `train_dataloader` to estimate number of stepping batches.
7
+
8
+ | Name | Type | Params | Mode
9
+ -----------------------------------------------------------------------
10
+ 0 | teacher_embedding_model | EmbeddingModel | 30.1 M | train
11
+ 1 | teacher_projection_head | DINOProjectionHead | 5.8 M | train
12
+ 2 | student_embedding_model | EmbeddingModel | 30.1 M | train
13
+ 3 | student_projection_head | DINOProjectionHead | 5.8 M | train
14
+ 4 | flatten | Flatten | 0 | train
15
+ 5 | criterion | DINOLoss | 0 | train
16
+ -----------------------------------------------------------------------
17
+ 71.7 M Trainable params
18
+ 2.0 K Non-trainable params
19
+ 71.7 M Total params
20
+ 286.782 Total estimated model params size (MB)
21
+ 578 Modules in train mode
22
+ 0 Modules in eval mode
23
+ Epoch 159: 75%|█████████████████████████████████████████▍ | 435/577 [03:00<00:58, 2.41it/s, v_num=57hl, train_loss=0.512, data_wait=53.2%]
ssl_dino/vit16s/wandb/run-20250814_130902-j9e757hl/files/requirements.txt ADDED
@@ -0,0 +1,195 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ threadpoolctl==3.6.0
2
+ MarkupSafe==3.0.2
3
+ timm==1.0.19
4
+ nvidia-cuda-nvrtc-cu12==12.8.93
5
+ certifi==2025.8.3
6
+ soupsieve==2.7
7
+ regex==2025.7.34
8
+ pre_commit==4.2.0
9
+ lightly==1.5.22
10
+ nvidia-cusolver-cu12==11.7.3.90
11
+ mypy_extensions==1.1.0
12
+ protobuf==6.31.1
13
+ cffi==1.17.1
14
+ psutil==7.0.0
15
+ identify==2.6.12
16
+ pycparser==2.22
17
+ multidict==6.6.3
18
+ Werkzeug==3.1.3
19
+ dotenv==0.9.9
20
+ charset-normalizer==3.4.3
21
+ albucore==0.0.24
22
+ torch==2.9.0.dev20250806+cu128
23
+ packaging==25.0
24
+ pydot==4.0.1
25
+ mpmath==1.3.0
26
+ pycodestyle==2.14.0
27
+ appdirs==1.4.4
28
+ nvidia-cufft-cu12==11.3.3.83
29
+ nodejs-wheel-binaries==22.18.0
30
+ tokenizers==0.21.4
31
+ black==25.1.0
32
+ tqdm==4.67.1
33
+ hf-xet==1.1.5
34
+ tabulate==0.9.0
35
+ dpath==2.2.0
36
+ smmap==5.0.2
37
+ cycler==0.12.1
38
+ python-dotenv==1.1.1
39
+ kiwisolver==1.4.8
40
+ shortuuid==1.0.13
41
+ shellingham==1.5.4
42
+ beautifulsoup4==4.13.4
43
+ kombu==5.5.4
44
+ isort==6.0.1
45
+ orjson==3.11.1
46
+ nvidia-cublas-cu12==12.8.4.1
47
+ iterative-telemetry==0.0.10
48
+ flufl.lock==8.2.0
49
+ pytorch-lightning==2.5.2
50
+ matplotlib==3.10.5
51
+ flake8==7.3.0
52
+ nvidia-nvtx-cu12==12.8.90
53
+ eval_type_backport==0.2.2
54
+ nvidia-cuda-runtime-cu12==12.8.90
55
+ pygtrie==2.5.0
56
+ pandas==2.3.1
57
+ amqp==5.3.1
58
+ nvidia-nvshmem-cu12==3.3.9
59
+ mccabe==0.7.0
60
+ pip==25.1
61
+ basedpyright==1.31.1
62
+ aiosignal==1.4.0
63
+ rich==14.1.0
64
+ albumentations==2.0.8
65
+ huggingface-hub==0.34.3
66
+ six==1.17.0
67
+ grandalf==0.8
68
+ celery==5.5.3
69
+ aiohttp==3.12.15
70
+ bcrypt==4.3.0
71
+ omegaconf==2.3.0
72
+ joblib==1.5.1
73
+ setuptools==78.1.1
74
+ lightning-utilities==0.15.0
75
+ sqltrie==0.11.2
76
+ diskcache==5.6.3
77
+ typing-inspection==0.4.1
78
+ flatten-dict==0.4.2
79
+ pydantic==2.11.7
80
+ invoke==2.2.0
81
+ fonttools==4.59.0
82
+ scipy==1.16.1
83
+ torchaudio==2.8.0.dev20250806+cu128
84
+ nvidia-cuda-cupti-cu12==12.8.90
85
+ dvc-studio-client==0.22.0
86
+ yarl==1.20.1
87
+ shtab==1.7.2
88
+ click-repl==0.3.0
89
+ wandb==0.21.1
90
+ antlr4-python3-runtime==4.9.3
91
+ pytorch-triton==3.4.0+gitf7888497
92
+ tensorboard-data-server==0.7.2
93
+ click==8.2.1
94
+ stringzilla==3.12.6
95
+ PySocks==1.7.1
96
+ pydantic_core==2.33.2
97
+ Jinja2==3.1.6
98
+ click-plugins==1.1.1.2
99
+ vine==5.1.0
100
+ dvc==3.61.0
101
+ configobj==5.0.9
102
+ opencv-python-headless==4.12.0.88
103
+ nvidia-nccl-cu12==2.27.5
104
+ loguru==0.7.3
105
+ attrs==25.3.0
106
+ cfgv==3.4.0
107
+ nvidia-cufile-cu12==1.13.1.3
108
+ simsimd==6.5.0
109
+ zc.lockfile==3.0.post1
110
+ idna==3.10
111
+ typing_extensions==4.14.1
112
+ tomlkit==0.13.3
113
+ click-didyoumean==0.3.1
114
+ propcache==0.3.2
115
+ markdown-it-py==3.0.0
116
+ dvc-http==2.32.0
117
+ urllib3==2.5.0
118
+ dvc-render==1.0.2
119
+ frozenlist==1.7.0
120
+ lightly-utils==0.0.2
121
+ dvc-data==3.16.10
122
+ cryptography==45.0.5
123
+ annotated-types==0.7.0
124
+ tensorboard==2.20.0
125
+ pyarrow==21.0.0
126
+ filelock==3.18.0
127
+ PyYAML==6.0.2
128
+ requests==2.32.4
129
+ pyparsing==3.2.3
130
+ fsspec==2025.7.0
131
+ ruamel.yaml.clib==0.2.12
132
+ semver==3.0.4
133
+ uv==0.8.4
134
+ voluptuous==0.15.2
135
+ aiohappyeyeballs==2.6.1
136
+ atpublic==6.0.1
137
+ asyncssh==2.21.0
138
+ torchmetrics==1.8.0
139
+ nvidia-curand-cu12==10.3.9.90
140
+ python-dateutil==2.9.0.post0
141
+ nvidia-cusparselt-cu12==0.7.1
142
+ dvc-task==0.40.2
143
+ dictdiffer==0.9.0
144
+ transformers==4.54.1
145
+ sentry-sdk==2.34.1
146
+ PyNaCl==1.5.0
147
+ nvidia-nvjitlink-cu12==12.8.93
148
+ virtualenv==20.33.0
149
+ tzdata==2025.2
150
+ safetensors==0.5.3
151
+ platformdirs==4.3.8
152
+ colorama==0.4.6
153
+ distlib==0.4.0
154
+ wcwidth==0.2.13
155
+ dvc-objects==5.1.1
156
+ nvidia-cusparse-cu12==12.5.8.93
157
+ scmrepo==3.5.1
158
+ mdurl==0.1.2
159
+ Pygments==2.19.2
160
+ triton==3.3.1
161
+ pytz==2025.2
162
+ distro==1.9.0
163
+ scikit-learn==1.7.1
164
+ seaborn==0.13.2
165
+ absl-py==2.3.1
166
+ ruamel.yaml==0.18.14
167
+ grpcio==1.74.0
168
+ funcy==2.0
169
+ gto==1.7.2
170
+ nvidia-cudnn-cu12==9.10.2.21
171
+ prompt_toolkit==3.0.51
172
+ numpy==2.2.6
173
+ lightly_train==0.6.1
174
+ dulwich==0.24.1
175
+ paramiko==4.0.0
176
+ sympy==1.14.0
177
+ hydra-core==1.3.2
178
+ networkx==3.5
179
+ torchvision==0.24.0.dev20250806+cu128
180
+ contourpy==1.3.3
181
+ pygit2==1.18.1
182
+ billiard==4.2.1
183
+ pillow==11.3.0
184
+ nodeenv==1.9.1
185
+ aenum==3.1.16
186
+ Markdown==3.8.2
187
+ typer==0.16.0
188
+ gitdb==4.0.12
189
+ aiohttp-retry==2.9.1
190
+ pathspec==0.12.1
191
+ entrypoints==0.4
192
+ gdown==5.2.0
193
+ pyflakes==3.4.0
194
+ GitPython==3.1.45
195
+ wheel==0.45.1
ssl_dino/vit16s/wandb/run-20250814_130902-j9e757hl/files/wandb-metadata.json ADDED
@@ -0,0 +1,40 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-6.16.0-zen1-1-zen-x86_64-with-glibc2.42",
3
+ "python": "CPython 3.11.13",
4
+ "startedAt": "2025-08-14T06:09:02.867550Z",
5
+ "program": "/home/duckq1u/Documents/workspace/cuong/ent-labotary/src/experiment/ssl_dino.py",
6
+ "codePath": "src/experiment/ssl_dino.py",
7
+ "codePathLocal": "src/experiment/ssl_dino.py",
8
+ "git": {
9
+ "remote": "git@github.com:PCTU-Deep-Medicine/ent-labotary.git",
10
+ "commit": "61f4482b9df47dedf1dd991b89913f2e5962454c"
11
+ },
12
+ "email": "acc.ggcolab@gmail.com",
13
+ "root": "/home/duckq1u/Documents/workspace/cuong/ent-labotary/outputs/ssl_dino/vit16s",
14
+ "host": "duckq1",
15
+ "executable": "/home/duckq1u/miniconda3/envs/ent-lab/bin/python",
16
+ "cpu_count": 6,
17
+ "cpu_count_logical": 6,
18
+ "gpu": "NVIDIA GeForce RTX 5060 Ti",
19
+ "gpu_count": 1,
20
+ "disk": {
21
+ "/": {
22
+ "total": "171828551680",
23
+ "used": "146758868992"
24
+ }
25
+ },
26
+ "memory": {
27
+ "total": "16682958848"
28
+ },
29
+ "gpu_nvidia": [
30
+ {
31
+ "name": "NVIDIA GeForce RTX 5060 Ti",
32
+ "memoryTotal": "17103323136",
33
+ "cudaCores": 4608,
34
+ "architecture": "Blackwell",
35
+ "uuid": "GPU-a8c62e38-0f95-e068-50c6-fab6ff92eaaf"
36
+ }
37
+ ],
38
+ "cudaVersion": "12.9",
39
+ "writerId": "clbjoeaje7zmuo0wmhnaowg9gp9z73c8"
40
+ }
ssl_dino/vit16s/wandb/run-20250814_130902-j9e757hl/logs/debug-core.log ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {"time":"2025-08-14T13:09:03.248712542+07:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmp80r3hayl/port-737456.txt","pid":737456,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false}
2
+ {"time":"2025-08-14T13:09:03.2522347+07:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-737456-738377-2077407418/socket","Net":"unix"}}
3
+ {"time":"2025-08-14T13:09:03.25485051+07:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":737456}
4
+ {"time":"2025-08-14T13:09:03.280396751+07:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"}
5
+ {"time":"2025-08-14T13:09:03.308985301+07:00","level":"INFO","msg":"handleInformInit: received","streamId":"j9e757hl","id":"1(@)"}
6
+ {"time":"2025-08-14T13:09:03.787667591+07:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"j9e757hl","id":"1(@)"}
ssl_dino/vit16s/wandb/run-20250814_130902-j9e757hl/logs/debug-internal.log ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {"time":"2025-08-14T13:09:03.309092553+07:00","level":"INFO","msg":"stream: starting","core version":"0.21.1"}
2
+ {"time":"2025-08-14T13:09:03.787644401+07:00","level":"INFO","msg":"stream: created new stream","id":"j9e757hl"}
3
+ {"time":"2025-08-14T13:09:03.787664481+07:00","level":"INFO","msg":"stream: started","id":"j9e757hl"}
4
+ {"time":"2025-08-14T13:09:03.787691222+07:00","level":"INFO","msg":"sender: started","stream_id":"j9e757hl"}
5
+ {"time":"2025-08-14T13:09:03.787690492+07:00","level":"INFO","msg":"writer: started","stream_id":"j9e757hl"}
6
+ {"time":"2025-08-14T13:09:03.789009657+07:00","level":"INFO","msg":"handler: started","stream_id":"j9e757hl"}
ssl_dino/vit16s/wandb/run-20250814_130902-j9e757hl/logs/debug.log ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2025-08-14 13:09:02,868 INFO MainThread:737456 [wandb_setup.py:_flush():80] Current SDK version is 0.21.1
2
+ 2025-08-14 13:09:02,868 INFO MainThread:737456 [wandb_setup.py:_flush():80] Configure stats pid to 737456
3
+ 2025-08-14 13:09:02,868 INFO MainThread:737456 [wandb_setup.py:_flush():80] Loading settings from /home/duckq1u/.config/wandb/settings
4
+ 2025-08-14 13:09:02,868 INFO MainThread:737456 [wandb_setup.py:_flush():80] Loading settings from /home/duckq1u/Documents/workspace/cuong/ent-labotary/wandb/settings
5
+ 2025-08-14 13:09:02,868 INFO MainThread:737456 [wandb_setup.py:_flush():80] Loading settings from environment variables
6
+ 2025-08-14 13:09:02,868 INFO MainThread:737456 [wandb_init.py:setup_run_log_directory():703] Logging user logs to /home/duckq1u/Documents/workspace/cuong/ent-labotary/outputs/ssl_dino/vit16s/wandb/run-20250814_130902-j9e757hl/logs/debug.log
7
+ 2025-08-14 13:09:02,868 INFO MainThread:737456 [wandb_init.py:setup_run_log_directory():704] Logging internal logs to /home/duckq1u/Documents/workspace/cuong/ent-labotary/outputs/ssl_dino/vit16s/wandb/run-20250814_130902-j9e757hl/logs/debug-internal.log
8
+ 2025-08-14 13:09:02,868 INFO MainThread:737456 [wandb_init.py:init():830] calling init triggers
9
+ 2025-08-14 13:09:02,868 INFO MainThread:737456 [wandb_init.py:init():835] wandb.init called with sweep_config: {}
10
+ config: {'_wandb': {}}
11
+ 2025-08-14 13:09:02,868 INFO MainThread:737456 [wandb_init.py:init():871] starting backend
12
+ 2025-08-14 13:09:03,280 INFO MainThread:737456 [wandb_init.py:init():874] sending inform_init request
13
+ 2025-08-14 13:09:03,304 INFO MainThread:737456 [wandb_init.py:init():882] backend started and connected
14
+ 2025-08-14 13:09:03,305 INFO MainThread:737456 [wandb_init.py:init():953] updated telemetry
15
+ 2025-08-14 13:09:03,309 INFO MainThread:737456 [wandb_init.py:init():977] communicating run to backend with 90.0 second timeout
16
+ 2025-08-14 13:09:04,489 INFO MainThread:737456 [wandb_init.py:init():1029] starting run threads in backend
17
+ 2025-08-14 13:09:04,582 INFO MainThread:737456 [wandb_run.py:_console_start():2494] atexit reg
18
+ 2025-08-14 13:09:04,582 INFO MainThread:737456 [wandb_run.py:_redirect():2342] redirect: wrap_raw
19
+ 2025-08-14 13:09:04,582 INFO MainThread:737456 [wandb_run.py:_redirect():2411] Wrapping output streams.
20
+ 2025-08-14 13:09:04,582 INFO MainThread:737456 [wandb_run.py:_redirect():2434] Redirects installed.
21
+ 2025-08-14 13:09:04,585 INFO MainThread:737456 [wandb_init.py:init():1075] run started, returning control to user process
22
+ 2025-08-14 13:09:04,585 INFO MainThread:737456 [wandb_run.py:_config_callback():1380] config_cb None None {'out': 'outputs/ssl_dino/vit16s', 'data': 'data/kyucapsule', 'model': 'VisionTransformer', 'method': 'dino', 'method_args': {'hidden_dim': 2048, 'bottleneck_dim': 256, 'output_dim': 1024, 'student_freeze_last_layer_epochs': 1, 'batch_norm': False, 'norm_last_layer': True, 'teacher_temp': 0.02, 'warmup_teacher_temp': 0.02, 'warmup_teacher_temp_epochs': 30, 'student_temp': 0.1, 'center_momentum': 0.9, 'momentum_start': 0.99, 'momentum_end': 1.0, 'weight_decay_start': 0.0001, 'weight_decay_end': 0.0001}, 'embed_dim': None, 'epochs': 300, 'batch_size': 32, 'num_workers': 5, 'devices': 1, 'num_nodes': 1, 'resume': False, 'checkpoint': None, 'overwrite': True, 'accelerator': 'CUDAAccelerator', 'strategy': 'SingleDeviceStrategy', 'precision': '32-true', 'seed': 0, 'loggers': {'jsonl': {'flush_logs_every_n_steps': 100}, 'tensorboard': {'name': '', 'version': '', 'log_graph': False, 'default_hp_metric': True, 'prefix': '', 'sub_dir': None}, 'wandb': {'name': None, 'version': None, 'offline': False, 'anonymous': None, 'project': 'ent-endoscopy-ssl', 'log_model': False, 'prefix': '', 'checkpoint_name': None}}, 'callbacks': {'learning_rate_monitor': {}, 'device_stats_monitor': {}, 'early_stopping': {'monitor': 'train_loss', 'patience': 1000000000000, 'check_finite': True}, 'model_checkpoint': {'save_last': True, 'enable_version_counter': False, 'save_top_k': 1, 'every_n_epochs': None}}, 'optim': 'sgd', 'optim_args': {'lr': 0.03, 'momentum': 0.9, 'weight_decay': 0.0001}, 'transform_args': {'image_size': [224, 224], 'random_resize': {'min_scale': 0.14, 'max_scale': 1.0}, 'random_flip': {'horizontal_prob': 0.5, 'vertical_prob': 0.0}, 'random_rotation': None, 'color_jitter': {'prob': 0.8, 'strength': 0.5, 'brightness': 0.8, 'contrast': 0.8, 'saturation': 0.4, 'hue': 0.2}, 'random_gray_scale': 0.2, 'normalize': {'mean': [0.485, 0.456, 0.406], 'std': [0.229, 0.224, 0.225]}, 'gaussian_blur': {'prob': 1.0, 'sigmas': [0.1, 2.0], 'blur_limit': 0}, 'solarize': None, 'global_view_1': {'gaussian_blur': {'prob': 0.1, 'sigmas': [0.1, 2.0], 'blur_limit': 0}, 'solarize': {'prob': 0.2, 'threshold': 0.5}}, 'local_view': {'num_views': 6, 'view_size': [96, 96], 'random_resize': {'min_scale': 0.05, 'max_scale': 0.14}, 'gaussian_blur': {'prob': 0.5, 'sigmas': [0.1, 2.0], 'blur_limit': 0}}}, 'loader_args': None, 'trainer_args': None, 'model_args': None}
ssl_dino/vit16s/wandb/run-20250814_130902-j9e757hl/run-j9e757hl.wandb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cba62f19875a3b28efd1358ae82e4800341a3013636c1c9ae05fc7f539dc55f7
3
+ size 88014848
ssl_dino/vit16s/wandb/run-20250815_070844-n4cjse4r/files/config.yaml ADDED
@@ -0,0 +1,235 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ _wandb:
2
+ value:
3
+ cli_version: 0.21.1
4
+ e:
5
+ xcvvegj6nnnzopxxl1iqmpzaqtal18nv:
6
+ codePath: src/experiment/ssl_dino.py
7
+ codePathLocal: src/experiment/ssl_dino.py
8
+ cpu_count: 6
9
+ cpu_count_logical: 6
10
+ cudaVersion: "12.9"
11
+ disk:
12
+ /:
13
+ total: "171828551680"
14
+ used: "147571367936"
15
+ email: acc.ggcolab@gmail.com
16
+ executable: /home/duckq1u/miniconda3/envs/ent-lab/bin/python
17
+ git:
18
+ commit: 61f4482b9df47dedf1dd991b89913f2e5962454c
19
+ remote: git@github.com:PCTU-Deep-Medicine/ent-labotary.git
20
+ gpu: NVIDIA GeForce RTX 5060 Ti
21
+ gpu_count: 1
22
+ gpu_nvidia:
23
+ - architecture: Blackwell
24
+ cudaCores: 4608
25
+ memoryTotal: "17103323136"
26
+ name: NVIDIA GeForce RTX 5060 Ti
27
+ uuid: GPU-a8c62e38-0f95-e068-50c6-fab6ff92eaaf
28
+ host: duckq1
29
+ memory:
30
+ total: "16682962944"
31
+ os: Linux-6.16.0-zen1-1-zen-x86_64-with-glibc2.42
32
+ program: /home/duckq1u/Documents/workspace/cuong/ent-labotary/src/experiment/ssl_dino.py
33
+ python: CPython 3.11.13
34
+ root: /home/duckq1u/Documents/workspace/cuong/ent-labotary/outputs/ssl_dino/vit16s
35
+ startedAt: "2025-08-15T00:08:44.272956Z"
36
+ writerId: xcvvegj6nnnzopxxl1iqmpzaqtal18nv
37
+ m:
38
+ - "1": trainer/global_step
39
+ "6":
40
+ - 3
41
+ "7": []
42
+ - "2": '*'
43
+ "5": 1
44
+ "6":
45
+ - 1
46
+ "7": []
47
+ python_version: 3.11.13
48
+ t:
49
+ "1":
50
+ - 1
51
+ - 5
52
+ - 9
53
+ - 11
54
+ - 41
55
+ - 49
56
+ - 53
57
+ - 63
58
+ - 80
59
+ - 103
60
+ - 105
61
+ "2":
62
+ - 1
63
+ - 5
64
+ - 9
65
+ - 11
66
+ - 41
67
+ - 49
68
+ - 53
69
+ - 63
70
+ - 80
71
+ - 103
72
+ - 105
73
+ "3":
74
+ - 7
75
+ - 66
76
+ "4": 3.11.13
77
+ "5": 0.21.1
78
+ "6": 4.54.1
79
+ "12": 0.21.1
80
+ "13": linux-x86_64
81
+ accelerator:
82
+ value: CUDAAccelerator
83
+ batch_size:
84
+ value: 32
85
+ callbacks:
86
+ value:
87
+ early_stopping:
88
+ check_finite: true
89
+ monitor: train_loss
90
+ patience: 1000000000000
91
+ model_checkpoint:
92
+ enable_version_counter: false
93
+ every_n_epochs: null
94
+ save_last: true
95
+ save_top_k: 1
96
+ checkpoint:
97
+ value: null
98
+ data:
99
+ value: data/kyucapsule
100
+ devices:
101
+ value: 1
102
+ embed_dim:
103
+ value: null
104
+ epochs:
105
+ value: 300
106
+ loader_args:
107
+ value: null
108
+ loggers:
109
+ value:
110
+ jsonl:
111
+ flush_logs_every_n_steps: 100
112
+ tensorboard:
113
+ default_hp_metric: true
114
+ log_graph: false
115
+ name: ""
116
+ prefix: ""
117
+ sub_dir: null
118
+ version: ""
119
+ wandb:
120
+ anonymous: null
121
+ checkpoint_name: null
122
+ log_model: false
123
+ name: null
124
+ offline: false
125
+ prefix: ""
126
+ project: ent-endoscopy-ssl
127
+ version: null
128
+ method:
129
+ value: dino
130
+ method_args:
131
+ value:
132
+ batch_norm: false
133
+ bottleneck_dim: 256
134
+ center_momentum: 0.9
135
+ hidden_dim: 2048
136
+ momentum_end: 1
137
+ momentum_start: 0.99
138
+ norm_last_layer: true
139
+ output_dim: 1024
140
+ student_freeze_last_layer_epochs: 1
141
+ student_temp: 0.1
142
+ teacher_temp: 0.02
143
+ warmup_teacher_temp: 0.02
144
+ warmup_teacher_temp_epochs: 30
145
+ weight_decay_end: 0.0001
146
+ weight_decay_start: 0.0001
147
+ model:
148
+ value: VisionTransformer
149
+ model_args:
150
+ value: null
151
+ num_nodes:
152
+ value: 1
153
+ num_workers:
154
+ value: 5
155
+ optim:
156
+ value: sgd
157
+ optim_args:
158
+ value:
159
+ lr: 0.03
160
+ momentum: 0.9
161
+ weight_decay: 0.0001
162
+ out:
163
+ value: outputs/ssl_dino/vit16s
164
+ overwrite:
165
+ value: true
166
+ precision:
167
+ value: 32-true
168
+ resume:
169
+ value: true
170
+ seed:
171
+ value: 0
172
+ strategy:
173
+ value: SingleDeviceStrategy
174
+ trainer_args:
175
+ value: null
176
+ transform_args:
177
+ value:
178
+ color_jitter:
179
+ brightness: 0.8
180
+ contrast: 0.8
181
+ hue: 0.2
182
+ prob: 0.8
183
+ saturation: 0.4
184
+ strength: 0.5
185
+ gaussian_blur:
186
+ blur_limit: 0
187
+ prob: 1
188
+ sigmas:
189
+ - 0.1
190
+ - 2
191
+ global_view_1:
192
+ gaussian_blur:
193
+ blur_limit: 0
194
+ prob: 0.1
195
+ sigmas:
196
+ - 0.1
197
+ - 2
198
+ solarize:
199
+ prob: 0.2
200
+ threshold: 0.5
201
+ image_size:
202
+ - 224
203
+ - 224
204
+ local_view:
205
+ gaussian_blur:
206
+ blur_limit: 0
207
+ prob: 0.5
208
+ sigmas:
209
+ - 0.1
210
+ - 2
211
+ num_views: 6
212
+ random_resize:
213
+ max_scale: 0.14
214
+ min_scale: 0.05
215
+ view_size:
216
+ - 96
217
+ - 96
218
+ normalize:
219
+ mean:
220
+ - 0.485
221
+ - 0.456
222
+ - 0.406
223
+ std:
224
+ - 0.229
225
+ - 0.224
226
+ - 0.225
227
+ random_flip:
228
+ horizontal_prob: 0.5
229
+ vertical_prob: 0
230
+ random_gray_scale: 0.2
231
+ random_resize:
232
+ max_scale: 1
233
+ min_scale: 0.14
234
+ random_rotation: null
235
+ solarize: null
ssl_dino/vit16s/wandb/run-20250815_070844-n4cjse4r/files/output.log ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ /home/duckq1u/miniconda3/envs/ent-lab/lib/python3.11/site-packages/torch/__init__.py:1539: UserWarning: Please use the new API settings to control TF32 behavior, such as torch.backends.cudnn.conv.fp32_precision = 'tf32' or torch.backends.cuda.matmul.fp32_precision = 'ieee'. Old settings, e.g, torch.backends.cuda.matmul.allow_tf32 = True, torch.backends.cudnn.allow_tf32 = True, allowTF32CuDNN() and allowTF32CuBLAS() will be deprecated after Pytorch 2.9. Please see https://pytorch.org/docs/main/notes/cuda.html#tensorfloat-32-tf32-on-ampere-and-later-devices (Triggered internally at /pytorch/aten/src/ATen/Context.cpp:80.)
2
+ return _C._get_float32_matmul_precision()
3
+ 
4
+ You are using a CUDA device ('NVIDIA GeForce RTX 5060 Ti') that has Tensor Cores. To properly utilize them, you should set `torch.set_float32_matmul_precision('medium' | 'high')` which will trade-off precision for performance. For more details, read https://pytorch.org/docs/stable/generated/torch.set_float32_matmul_precision.html#torch.set_float32_matmul_precision
5
+ Restoring states from the checkpoint path at /home/duckq1u/Documents/workspace/cuong/ent-labotary/outputs/ssl_dino/vit16s/checkpoints/last.ckpt
6
+ LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0]
7
+ Loading `train_dataloader` to estimate number of stepping batches.
8
+
9
+ | Name | Type | Params | Mode
10
+ -----------------------------------------------------------------------
11
+ 0 | teacher_embedding_model | EmbeddingModel | 30.1 M | train
12
+ 1 | teacher_projection_head | DINOProjectionHead | 5.8 M | train
13
+ 2 | student_embedding_model | EmbeddingModel | 30.1 M | train
14
+ 3 | student_projection_head | DINOProjectionHead | 5.8 M | train
15
+ 4 | flatten | Flatten | 0 | train
16
+ 5 | criterion | DINOLoss | 0 | train
17
+ -----------------------------------------------------------------------
18
+ 71.7 M Trainable params
19
+ 2.0 K Non-trainable params
20
+ 71.7 M Total params
21
+ 286.782 Total estimated model params size (MB)
22
+ 578 Modules in train mode
23
+ 0 Modules in eval mode
24
+ Restored all states from the checkpoint at /home/duckq1u/Documents/workspace/cuong/ent-labotary/outputs/ssl_dino/vit16s/checkpoints/last.ckpt
25
+ Epoch 246: 49%|██████████████████████████████▌ | 280/577 [02:00<02:07, 2.33it/s, v_num=se4r, train_loss=0.351, data_wait=53.0%]
26
+
27
+ Detected KeyboardInterrupt, attempting graceful shutdown ...
ssl_dino/vit16s/wandb/run-20250815_070844-n4cjse4r/files/requirements.txt ADDED
@@ -0,0 +1,195 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ threadpoolctl==3.6.0
2
+ MarkupSafe==3.0.2
3
+ timm==1.0.19
4
+ nvidia-cuda-nvrtc-cu12==12.8.93
5
+ certifi==2025.8.3
6
+ soupsieve==2.7
7
+ regex==2025.7.34
8
+ pre_commit==4.2.0
9
+ lightly==1.5.22
10
+ nvidia-cusolver-cu12==11.7.3.90
11
+ mypy_extensions==1.1.0
12
+ protobuf==6.31.1
13
+ cffi==1.17.1
14
+ psutil==7.0.0
15
+ identify==2.6.12
16
+ pycparser==2.22
17
+ multidict==6.6.3
18
+ Werkzeug==3.1.3
19
+ dotenv==0.9.9
20
+ charset-normalizer==3.4.3
21
+ albucore==0.0.24
22
+ torch==2.9.0.dev20250806+cu128
23
+ packaging==25.0
24
+ pydot==4.0.1
25
+ mpmath==1.3.0
26
+ pycodestyle==2.14.0
27
+ appdirs==1.4.4
28
+ nvidia-cufft-cu12==11.3.3.83
29
+ nodejs-wheel-binaries==22.18.0
30
+ tokenizers==0.21.4
31
+ black==25.1.0
32
+ tqdm==4.67.1
33
+ hf-xet==1.1.5
34
+ tabulate==0.9.0
35
+ dpath==2.2.0
36
+ smmap==5.0.2
37
+ cycler==0.12.1
38
+ python-dotenv==1.1.1
39
+ kiwisolver==1.4.8
40
+ shortuuid==1.0.13
41
+ shellingham==1.5.4
42
+ beautifulsoup4==4.13.4
43
+ kombu==5.5.4
44
+ isort==6.0.1
45
+ orjson==3.11.1
46
+ nvidia-cublas-cu12==12.8.4.1
47
+ iterative-telemetry==0.0.10
48
+ flufl.lock==8.2.0
49
+ pytorch-lightning==2.5.2
50
+ matplotlib==3.10.5
51
+ flake8==7.3.0
52
+ nvidia-nvtx-cu12==12.8.90
53
+ eval_type_backport==0.2.2
54
+ nvidia-cuda-runtime-cu12==12.8.90
55
+ pygtrie==2.5.0
56
+ pandas==2.3.1
57
+ amqp==5.3.1
58
+ nvidia-nvshmem-cu12==3.3.9
59
+ mccabe==0.7.0
60
+ pip==25.1
61
+ basedpyright==1.31.1
62
+ aiosignal==1.4.0
63
+ rich==14.1.0
64
+ albumentations==2.0.8
65
+ huggingface-hub==0.34.3
66
+ six==1.17.0
67
+ grandalf==0.8
68
+ celery==5.5.3
69
+ aiohttp==3.12.15
70
+ bcrypt==4.3.0
71
+ omegaconf==2.3.0
72
+ joblib==1.5.1
73
+ setuptools==78.1.1
74
+ lightning-utilities==0.15.0
75
+ sqltrie==0.11.2
76
+ diskcache==5.6.3
77
+ typing-inspection==0.4.1
78
+ flatten-dict==0.4.2
79
+ pydantic==2.11.7
80
+ invoke==2.2.0
81
+ fonttools==4.59.0
82
+ scipy==1.16.1
83
+ torchaudio==2.8.0.dev20250806+cu128
84
+ nvidia-cuda-cupti-cu12==12.8.90
85
+ dvc-studio-client==0.22.0
86
+ yarl==1.20.1
87
+ shtab==1.7.2
88
+ click-repl==0.3.0
89
+ wandb==0.21.1
90
+ antlr4-python3-runtime==4.9.3
91
+ pytorch-triton==3.4.0+gitf7888497
92
+ tensorboard-data-server==0.7.2
93
+ click==8.2.1
94
+ stringzilla==3.12.6
95
+ PySocks==1.7.1
96
+ pydantic_core==2.33.2
97
+ Jinja2==3.1.6
98
+ click-plugins==1.1.1.2
99
+ vine==5.1.0
100
+ dvc==3.61.0
101
+ configobj==5.0.9
102
+ opencv-python-headless==4.12.0.88
103
+ nvidia-nccl-cu12==2.27.5
104
+ loguru==0.7.3
105
+ attrs==25.3.0
106
+ cfgv==3.4.0
107
+ nvidia-cufile-cu12==1.13.1.3
108
+ simsimd==6.5.0
109
+ zc.lockfile==3.0.post1
110
+ idna==3.10
111
+ typing_extensions==4.14.1
112
+ tomlkit==0.13.3
113
+ click-didyoumean==0.3.1
114
+ propcache==0.3.2
115
+ markdown-it-py==3.0.0
116
+ dvc-http==2.32.0
117
+ urllib3==2.5.0
118
+ dvc-render==1.0.2
119
+ frozenlist==1.7.0
120
+ lightly-utils==0.0.2
121
+ dvc-data==3.16.10
122
+ cryptography==45.0.5
123
+ annotated-types==0.7.0
124
+ tensorboard==2.20.0
125
+ pyarrow==21.0.0
126
+ filelock==3.18.0
127
+ PyYAML==6.0.2
128
+ requests==2.32.4
129
+ pyparsing==3.2.3
130
+ fsspec==2025.7.0
131
+ ruamel.yaml.clib==0.2.12
132
+ semver==3.0.4
133
+ uv==0.8.4
134
+ voluptuous==0.15.2
135
+ aiohappyeyeballs==2.6.1
136
+ atpublic==6.0.1
137
+ asyncssh==2.21.0
138
+ torchmetrics==1.8.0
139
+ nvidia-curand-cu12==10.3.9.90
140
+ python-dateutil==2.9.0.post0
141
+ nvidia-cusparselt-cu12==0.7.1
142
+ dvc-task==0.40.2
143
+ dictdiffer==0.9.0
144
+ transformers==4.54.1
145
+ sentry-sdk==2.34.1
146
+ PyNaCl==1.5.0
147
+ nvidia-nvjitlink-cu12==12.8.93
148
+ virtualenv==20.33.0
149
+ tzdata==2025.2
150
+ safetensors==0.5.3
151
+ platformdirs==4.3.8
152
+ colorama==0.4.6
153
+ distlib==0.4.0
154
+ wcwidth==0.2.13
155
+ dvc-objects==5.1.1
156
+ nvidia-cusparse-cu12==12.5.8.93
157
+ scmrepo==3.5.1
158
+ mdurl==0.1.2
159
+ Pygments==2.19.2
160
+ triton==3.3.1
161
+ pytz==2025.2
162
+ distro==1.9.0
163
+ scikit-learn==1.7.1
164
+ seaborn==0.13.2
165
+ absl-py==2.3.1
166
+ ruamel.yaml==0.18.14
167
+ grpcio==1.74.0
168
+ funcy==2.0
169
+ gto==1.7.2
170
+ nvidia-cudnn-cu12==9.10.2.21
171
+ prompt_toolkit==3.0.51
172
+ numpy==2.2.6
173
+ lightly_train==0.6.1
174
+ dulwich==0.24.1
175
+ paramiko==4.0.0
176
+ sympy==1.14.0
177
+ hydra-core==1.3.2
178
+ networkx==3.5
179
+ torchvision==0.24.0.dev20250806+cu128
180
+ contourpy==1.3.3
181
+ pygit2==1.18.1
182
+ billiard==4.2.1
183
+ pillow==11.3.0
184
+ nodeenv==1.9.1
185
+ aenum==3.1.16
186
+ Markdown==3.8.2
187
+ typer==0.16.0
188
+ gitdb==4.0.12
189
+ aiohttp-retry==2.9.1
190
+ pathspec==0.12.1
191
+ entrypoints==0.4
192
+ gdown==5.2.0
193
+ pyflakes==3.4.0
194
+ GitPython==3.1.45
195
+ wheel==0.45.1
ssl_dino/vit16s/wandb/run-20250815_070844-n4cjse4r/files/wandb-metadata.json ADDED
@@ -0,0 +1,40 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-6.16.0-zen1-1-zen-x86_64-with-glibc2.42",
3
+ "python": "CPython 3.11.13",
4
+ "startedAt": "2025-08-15T00:08:44.272956Z",
5
+ "program": "/home/duckq1u/Documents/workspace/cuong/ent-labotary/src/experiment/ssl_dino.py",
6
+ "codePath": "src/experiment/ssl_dino.py",
7
+ "codePathLocal": "src/experiment/ssl_dino.py",
8
+ "git": {
9
+ "remote": "git@github.com:PCTU-Deep-Medicine/ent-labotary.git",
10
+ "commit": "61f4482b9df47dedf1dd991b89913f2e5962454c"
11
+ },
12
+ "email": "acc.ggcolab@gmail.com",
13
+ "root": "/home/duckq1u/Documents/workspace/cuong/ent-labotary/outputs/ssl_dino/vit16s",
14
+ "host": "duckq1",
15
+ "executable": "/home/duckq1u/miniconda3/envs/ent-lab/bin/python",
16
+ "cpu_count": 6,
17
+ "cpu_count_logical": 6,
18
+ "gpu": "NVIDIA GeForce RTX 5060 Ti",
19
+ "gpu_count": 1,
20
+ "disk": {
21
+ "/": {
22
+ "total": "171828551680",
23
+ "used": "147571367936"
24
+ }
25
+ },
26
+ "memory": {
27
+ "total": "16682962944"
28
+ },
29
+ "gpu_nvidia": [
30
+ {
31
+ "name": "NVIDIA GeForce RTX 5060 Ti",
32
+ "memoryTotal": "17103323136",
33
+ "cudaCores": 4608,
34
+ "architecture": "Blackwell",
35
+ "uuid": "GPU-a8c62e38-0f95-e068-50c6-fab6ff92eaaf"
36
+ }
37
+ ],
38
+ "cudaVersion": "12.9",
39
+ "writerId": "xcvvegj6nnnzopxxl1iqmpzaqtal18nv"
40
+ }
ssl_dino/vit16s/wandb/run-20250815_070844-n4cjse4r/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"DeviceStatsMonitor.on_train_batch_end/active.small_pool.current":1094,"DeviceStatsMonitor.on_train_batch_end/allocated_bytes.small_pool.freed":5721681839616,"DeviceStatsMonitor.on_train_batch_end/oversize_segments.current":0,"DeviceStatsMonitor.on_train_batch_end/allocation.large_pool.current":377,"DeviceStatsMonitor.on_train_batch_end/active_bytes.all.current":1383923712,"DeviceStatsMonitor.on_train_batch_end/reserved_bytes.all.current":7824474112,"lr-SGD/params":0.0003102255598381857,"DeviceStatsMonitor.on_train_batch_start/active.large_pool.freed":56372331,"DeviceStatsMonitor.on_train_batch_start/allocated_bytes.all.allocated":1199476346978304,"DeviceStatsMonitor.on_train_batch_start/segment.small_pool.current":58,"DeviceStatsMonitor.on_train_batch_start/oversize_segments.current":0,"trainer/global_step":142199,"DeviceStatsMonitor.on_train_batch_start/requested_bytes.small_pool.allocated":5712366121130,"DeviceStatsMonitor.on_train_batch_start/active_bytes.all.freed":1199474959156736,"DeviceStatsMonitor.on_train_batch_end/allocation.all.freed":120396348,"DeviceStatsMonitor.on_train_batch_start/allocation.all.current":1468,"DeviceStatsMonitor.on_train_batch_start/segment.all.peak":297,"DeviceStatsMonitor.on_train_batch_end/inactive_split_bytes.small_pool.peak":56591360,"DeviceStatsMonitor.on_train_batch_end/active.large_pool.freed":56373440,"DeviceStatsMonitor.on_train_batch_end/requested_bytes.small_pool.allocated":5712479282182,"DeviceStatsMonitor.on_train_batch_start/allocated_bytes.large_pool.allocated":1193754696431616,"DeviceStatsMonitor.on_train_batch_start/allocation.small_pool.freed":64021645,"DeviceStatsMonitor.on_train_batch_end/allocated_bytes.all.allocated":1199500050339328,"DeviceStatsMonitor.on_train_batch_start/inactive_split_bytes.all.allocated":284527257374208,"DeviceStatsMonitor.on_train_batch_start/oversize_allocations.peak":0,"DeviceStatsMonitor.on_train_batch_end/oversize_segments.freed":0,"epoch":246,"DeviceStatsMonitor.on_train_batch_end/inactive_split.small_pool.allocated":40018491,"DeviceStatsMonitor.on_train_batch_start/inactive_split_bytes.large_pool.freed":278695025747968,"DeviceStatsMonitor.on_train_batch_start/active.small_pool.allocated":64022736,"DeviceStatsMonitor.on_train_batch_start/allocation.small_pool.peak":1375,"DeviceStatsMonitor.on_train_batch_end/reserved_bytes.all.allocated":7824474112,"DeviceStatsMonitor.on_train_batch_start/inactive_split_bytes.all.peak":600349696,"DeviceStatsMonitor.on_train_batch_start/oversize_segments.allocated":0,"DeviceStatsMonitor.on_train_batch_start/num_sync_all_streams":3,"DeviceStatsMonitor.on_train_batch_end/active.small_pool.freed":64022908,"DeviceStatsMonitor.on_train_batch_end/allocation.all.allocated":120397819,"DeviceStatsMonitor.on_train_batch_end/inactive_split.small_pool.current":125,"DeviceStatsMonitor.on_train_batch_start/active_bytes.small_pool.freed":5721568497152,"DeviceStatsMonitor.on_train_batch_start/inactive_split_bytes.small_pool.current":20710912,"DeviceStatsMonitor.on_train_batch_start/inactive_split_bytes.all.current":134710784,"DeviceStatsMonitor.on_train_batch_start/inactive_split.large_pool.freed":41294815,"DeviceStatsMonitor.on_train_batch_end/segment.large_pool.current":239,"DeviceStatsMonitor.on_train_batch_end/requested_bytes.large_pool.current":1221172736,"profiling/data_time":0.2310386449098587,"DeviceStatsMonitor.on_train_batch_end/inactive_split_bytes.all.freed":284534756130304,"DeviceStatsMonitor.on_train_batch_end/allocated_bytes.small_pool.peak":111529472,"DeviceStatsMonitor.on_train_batch_end/reserved_bytes.large_pool.allocated":7702839296,"DeviceStatsMonitor.on_train_batch_end/active_bytes.large_pool.freed":1193776984576000,"DeviceStatsMonitor.on_train_batch_start/allocation.all.freed":120393976,"DeviceStatsMonitor.on_train_batch_end/inactive_split.all.current":151,"DeviceStatsMonitor.on_train_batch_start/allocated_bytes.small_pool.freed":5721568497152,"DeviceStatsMonitor.on_train_batch_start/active_bytes.large_pool.freed":1193753390659584,"DeviceStatsMonitor.on_train_batch_start/allocation.large_pool.freed":56372331,"DeviceStatsMonitor.on_train_batch_start/requested_bytes.all.current":1303213424,"DeviceStatsMonitor.on_train_batch_end/active.small_pool.allocated":64024002,"DeviceStatsMonitor.on_train_batch_start/segment.all.allocated":297,"DeviceStatsMonitor.on_train_batch_end/requested_bytes.large_pool.allocated":1187137043964416,"DeviceStatsMonitor.on_train_batch_start/reserved_bytes.all.allocated":7824474112,"DeviceStatsMonitor.on_train_batch_end/requested_bytes.all.freed":1192848220033162,"DeviceStatsMonitor.on_train_batch_start/max_split_size":-1,"DeviceStatsMonitor.on_train_batch_start/allocation.large_pool.peak":618,"DeviceStatsMonitor.on_train_batch_end/active.large_pool.current":377,"DeviceStatsMonitor.on_train_batch_start/reserved_bytes.small_pool.allocated":121634816,"DeviceStatsMonitor.on_train_batch_end/reserved_bytes.small_pool.peak":121634816,"DeviceStatsMonitor.on_train_batch_end/active_bytes.large_pool.allocated":1193778286448640,"DeviceStatsMonitor.on_train_batch_end/active.large_pool.allocated":56373817,"DeviceStatsMonitor.on_train_batch_start/requested_bytes.all.allocated":1192825942940842,"DeviceStatsMonitor.on_train_batch_end/inactive_split.all.allocated":81314049,"DeviceStatsMonitor.on_train_batch_start/active_bytes.all.current":1387821568,"DeviceStatsMonitor.on_train_batch_start/allocated_bytes.large_pool.freed":1193753390659584,"DeviceStatsMonitor.on_train_batch_end/active.large_pool.peak":618,"DeviceStatsMonitor.on_train_batch_end/inactive_split_bytes.large_pool.freed":278702546001920,"DeviceStatsMonitor.on_train_batch_start/inactive_split_bytes.small_pool.allocated":5832117626368,"DeviceStatsMonitor.on_train_batch_end/inactive_split_bytes.all.peak":600349696,"DeviceStatsMonitor.on_train_batch_start/allocated_bytes.all.freed":1199474959156736,"DeviceStatsMonitor.on_train_batch_end/requested_bytes.small_pool.current":82040700,"DeviceStatsMonitor.on_train_batch_start/inactive_split.all.peak":372,"DeviceStatsMonitor.on_train_batch_start/active_bytes.all.allocated":1199476346978304,"DeviceStatsMonitor.on_train_batch_start/allocation.all.peak":1983,"DeviceStatsMonitor.on_train_batch_start/active_bytes.large_pool.current":1305772032,"DeviceStatsMonitor.on_train_batch_start/segment.small_pool.allocated":58,"DeviceStatsMonitor.on_train_batch_start/active_bytes.large_pool.peak":7416414208,"DeviceStatsMonitor.on_train_batch_start/segment.large_pool.allocated":239,"DeviceStatsMonitor.on_train_batch_start/reserved_bytes.large_pool.current":7702839296,"DeviceStatsMonitor.on_train_batch_end/active.all.current":1471,"DeviceStatsMonitor.on_train_batch_end/reserved_bytes.small_pool.allocated":121634816,"DeviceStatsMonitor.on_train_batch_end/inactive_split.all.freed":81313898,"DeviceStatsMonitor.on_train_batch_start/oversize_segments.freed":0,"DeviceStatsMonitor.on_train_batch_end/inactive_split.large_pool.freed":41295532,"DeviceStatsMonitor.on_train_batch_start/segment.large_pool.freed":0,"DeviceStatsMonitor.on_train_batch_start/inactive_split.all.current":150,"DeviceStatsMonitor.on_train_batch_start/num_ooms":0,"DeviceStatsMonitor.on_train_batch_start/active.small_pool.peak":1375,"DeviceStatsMonitor.on_train_batch_end/segment.small_pool.peak":58,"DeviceStatsMonitor.on_train_batch_end/allocation.all.current":1471,"DeviceStatsMonitor.on_train_batch_end/oversize_segments.peak":0,"DeviceStatsMonitor.on_train_batch_start/allocation.large_pool.current":377,"DeviceStatsMonitor.on_train_batch_end/segment.large_pool.allocated":239,"DeviceStatsMonitor.on_train_batch_end/requested_bytes.large_pool.peak":7280660480,"DeviceStatsMonitor.on_train_batch_end/num_ooms":0,"DeviceStatsMonitor.on_train_batch_start/active.large_pool.current":377,"DeviceStatsMonitor.on_train_batch_end/active.small_pool.peak":1375,"DeviceStatsMonitor.on_train_batch_start/inactive_split.small_pool.freed":40017541,"train_loss":0.2858157157897949,"DeviceStatsMonitor.on_train_batch_end/max_split_size":-1,"DeviceStatsMonitor.on_train_batch_end/inactive_split_bytes.all.current":136511488,"DeviceStatsMonitor.on_train_batch_end/num_device_alloc":297,"lr-SGD/params_no_weight_decay":0.0003102255598381857,"DeviceStatsMonitor.on_train_batch_end/inactive_split_bytes.large_pool.current":113704960,"DeviceStatsMonitor.on_train_batch_end/requested_bytes.all.allocated":1192849523246598,"DeviceStatsMonitor.on_train_batch_end/allocated_bytes.all.freed":1199498666415616,"DeviceStatsMonitor.on_train_batch_end/requested_bytes.all.current":1303213436,"DeviceStatsMonitor.on_train_batch_end/oversize_segments.allocated":0,"DeviceStatsMonitor.on_train_batch_start/reserved_bytes.small_pool.peak":121634816,"DeviceStatsMonitor.on_train_batch_start/allocated_bytes.small_pool.current":82049536,"DeviceStatsMonitor.on_train_batch_start/inactive_split_bytes.large_pool.allocated":278695139747840,"DeviceStatsMonitor.on_train_batch_start/inactive_split_bytes.small_pool.peak":56591360,"DeviceStatsMonitor.on_train_batch_end/allocation.all.peak":1983,"DeviceStatsMonitor.on_train_batch_end/active_bytes.all.peak":7521744384,"DeviceStatsMonitor.on_train_batch_end/active_bytes.large_pool.peak":7416414208,"DeviceStatsMonitor.on_train_batch_start/active.small_pool.current":1091,"DeviceStatsMonitor.on_train_batch_start/requested_bytes.large_pool.peak":7280660480,"DeviceStatsMonitor.on_train_batch_start/allocated_bytes.all.peak":7521744384,"DeviceStatsMonitor.on_train_batch_end/active_bytes.large_pool.current":1301872640,"DeviceStatsMonitor.on_train_batch_start/requested_bytes.small_pool.freed":5712284080442,"DeviceStatsMonitor.on_train_batch_start/oversize_allocations.current":0,"DeviceStatsMonitor.on_train_batch_start/requested_bytes.large_pool.freed":1187112355646976,"_timestamp":1.7552381340876622e+09,"DeviceStatsMonitor.on_train_batch_start/segment.small_pool.freed":0,"DeviceStatsMonitor.on_train_batch_start/allocated_bytes.small_pool.peak":111529472,"DeviceStatsMonitor.on_train_batch_start/oversize_allocations.freed":0,"DeviceStatsMonitor.on_train_batch_start/inactive_split.all.allocated":81312506,"DeviceStatsMonitor.on_train_batch_start/inactive_split_bytes.large_pool.current":113999872,"DeviceStatsMonitor.on_train_batch_end/reserved_bytes.large_pool.freed":0,"DeviceStatsMonitor.on_train_batch_end/allocated_bytes.small_pool.current":82051072,"DeviceStatsMonitor.on_train_batch_start/inactive_split_bytes.large_pool.peak":562167808,"DeviceStatsMonitor.on_train_batch_start/active_bytes.small_pool.peak":111529472,"DeviceStatsMonitor.on_train_batch_end/inactive_split.small_pool.peak":249,"DeviceStatsMonitor.on_train_batch_start/requested_bytes.large_pool.current":1221172736,"DeviceStatsMonitor.on_train_batch_start/allocated_bytes.large_pool.current":1305772032,"DeviceStatsMonitor.on_train_batch_start/reserved_bytes.small_pool.freed":0,"DeviceStatsMonitor.on_train_batch_end/segment.large_pool.freed":0,"DeviceStatsMonitor.on_train_batch_start/allocated_bytes.all.current":1387821568,"DeviceStatsMonitor.on_train_batch_end/active.all.freed":120396348,"DeviceStatsMonitor.on_train_batch_start/requested_bytes.small_pool.peak":111492804,"DeviceStatsMonitor.on_train_batch_start/allocation.all.allocated":120395444,"DeviceStatsMonitor.on_train_batch_end/allocation.small_pool.freed":64022908,"DeviceStatsMonitor.on_train_batch_end/segment.small_pool.current":58,"DeviceStatsMonitor.on_train_batch_end/segment.small_pool.freed":0,"DeviceStatsMonitor.on_train_batch_end/segment.all.current":297,"DeviceStatsMonitor.on_train_batch_end/reserved_bytes.large_pool.peak":7702839296,"DeviceStatsMonitor.on_train_batch_end/reserved_bytes.all.peak":7824474112,"DeviceStatsMonitor.on_train_batch_start/oversize_allocations.allocated":0,"DeviceStatsMonitor.on_train_batch_end/segment.all.allocated":297,"DeviceStatsMonitor.on_train_batch_end/segment.all.peak":297,"DeviceStatsMonitor.on_train_batch_end/active_bytes.all.freed":1199498666415616,"_runtime":21619,"DeviceStatsMonitor.on_train_batch_start/reserved_bytes.large_pool.allocated":7702839296,"DeviceStatsMonitor.on_train_batch_start/requested_bytes.small_pool.current":82040688,"DeviceStatsMonitor.on_train_batch_start/allocated_bytes.large_pool.peak":7416414208,"DeviceStatsMonitor.on_train_batch_start/inactive_split_bytes.all.freed":284527122663424,"DeviceStatsMonitor.on_train_batch_end/allocation.large_pool.allocated":56373817,"DeviceStatsMonitor.on_train_batch_start/active.all.peak":1983,"DeviceStatsMonitor.on_train_batch_start/active_bytes.small_pool.current":82049536,"DeviceStatsMonitor.on_train_batch_end/inactive_split.small_pool.freed":40018366,"DeviceStatsMonitor.on_train_batch_end/oversize_allocations.freed":0,"DeviceStatsMonitor.on_train_batch_start/requested_bytes.all.freed":1192824639727418,"DeviceStatsMonitor.on_train_batch_start/reserved_bytes.small_pool.current":121634816,"_wandb":{"runtime":21619},"DeviceStatsMonitor.on_train_batch_end/allocated_bytes.large_pool.peak":7416414208,"DeviceStatsMonitor.on_train_batch_end/num_alloc_retries":0,"DeviceStatsMonitor.on_train_batch_start/inactive_split.large_pool.allocated":41294840,"DeviceStatsMonitor.on_train_batch_start/active_bytes.small_pool.allocated":5721650546688,"DeviceStatsMonitor.on_train_batch_start/active.large_pool.allocated":56372708,"DeviceStatsMonitor.on_train_batch_end/reserved_bytes.small_pool.current":121634816,"DeviceStatsMonitor.on_train_batch_end/active_bytes.small_pool.current":82051072,"DeviceStatsMonitor.on_train_batch_start/inactive_split.large_pool.peak":158,"DeviceStatsMonitor.on_train_batch_start/active.all.freed":120393976,"DeviceStatsMonitor.on_train_batch_end/active_bytes.all.allocated":1199500050339328,"DeviceStatsMonitor.on_train_batch_start/oversize_segments.peak":0,"DeviceStatsMonitor.on_train_batch_start/active.large_pool.peak":618,"DeviceStatsMonitor.on_train_batch_start/active_bytes.all.peak":7521744384,"DeviceStatsMonitor.on_train_batch_start/reserved_bytes.large_pool.peak":7702839296,"DeviceStatsMonitor.on_train_batch_end/segment.small_pool.allocated":58,"DeviceStatsMonitor.on_train_batch_start/allocation.small_pool.current":1091,"DeviceStatsMonitor.on_train_batch_end/inactive_split_bytes.small_pool.freed":5832210128384,"DeviceStatsMonitor.on_train_batch_end/allocation.large_pool.freed":56373440,"DeviceStatsMonitor.on_train_batch_end/allocation.large_pool.peak":618,"DeviceStatsMonitor.on_train_batch_end/allocation.small_pool.current":1094,"DeviceStatsMonitor.on_train_batch_end/inactive_split_bytes.all.allocated":284534892641792,"DeviceStatsMonitor.on_train_batch_start/segment.all.current":297,"DeviceStatsMonitor.on_train_batch_end/allocated_bytes.large_pool.allocated":1193778286448640,"DeviceStatsMonitor.on_train_batch_start/num_device_alloc":297,"DeviceStatsMonitor.on_train_batch_start/inactive_split.small_pool.peak":249,"DeviceStatsMonitor.on_train_batch_end/segment.large_pool.peak":239,"profiling/batch_time":0.20534811913967133,"DeviceStatsMonitor.on_train_batch_start/inactive_split.small_pool.allocated":40017666,"DeviceStatsMonitor.on_train_batch_end/inactive_split_bytes.large_pool.allocated":278702659706880,"DeviceStatsMonitor.on_train_batch_start/active.small_pool.freed":64021645,"DeviceStatsMonitor.on_train_batch_end/inactive_split.large_pool.peak":158,"DeviceStatsMonitor.on_train_batch_end/requested_bytes.small_pool.freed":5712397241482,"DeviceStatsMonitor.on_train_batch_end/active_bytes.small_pool.peak":111529472,"DeviceStatsMonitor.on_train_batch_end/allocation.small_pool.allocated":64024002,"DeviceStatsMonitor.on_train_batch_start/inactive_split.small_pool.current":125,"DeviceStatsMonitor.on_train_batch_end/requested_bytes.large_pool.freed":1187135822791680,"DeviceStatsMonitor.on_train_batch_start/allocated_bytes.small_pool.allocated":5721650546688,"DeviceStatsMonitor.on_train_batch_start/reserved_bytes.large_pool.freed":0,"DeviceStatsMonitor.on_train_batch_end/active.all.allocated":120397819,"DeviceStatsMonitor.on_train_batch_start/inactive_split.all.freed":81312356,"DeviceStatsMonitor.on_train_batch_end/inactive_split.all.peak":372,"DeviceStatsMonitor.on_train_batch_start/allocation.small_pool.allocated":64022736,"DeviceStatsMonitor.on_train_batch_end/requested_bytes.small_pool.peak":111492804,"DeviceStatsMonitor.on_train_batch_end/reserved_bytes.small_pool.freed":0,"DeviceStatsMonitor.on_train_batch_start/segment.all.freed":0,"DeviceStatsMonitor.on_train_batch_end/inactive_split_bytes.small_pool.allocated":5832232934912,"DeviceStatsMonitor.on_train_batch_end/num_device_free":0,"DeviceStatsMonitor.on_train_batch_start/inactive_split_bytes.small_pool.freed":5832096915456,"DeviceStatsMonitor.on_train_batch_end/allocation.small_pool.peak":1375,"DeviceStatsMonitor.on_train_batch_start/segment.large_pool.current":239,"DeviceStatsMonitor.on_train_batch_end/reserved_bytes.all.freed":0,"DeviceStatsMonitor.on_train_batch_end/active_bytes.small_pool.freed":5721681839616,"DeviceStatsMonitor.on_train_batch_start/reserved_bytes.all.current":7824474112,"DeviceStatsMonitor.on_train_batch_end/requested_bytes.all.peak":7385954500,"DeviceStatsMonitor.on_train_batch_end/allocated_bytes.all.peak":7521744384,"DeviceStatsMonitor.on_train_batch_start/active_bytes.large_pool.allocated":1193754696431616,"DeviceStatsMonitor.on_train_batch_end/inactive_split_bytes.large_pool.peak":562167808,"DeviceStatsMonitor.on_train_batch_end/num_sync_all_streams":3,"DeviceStatsMonitor.on_train_batch_end/allocated_bytes.large_pool.current":1301872640,"DeviceStatsMonitor.on_train_batch_end/oversize_allocations.allocated":0,"DeviceStatsMonitor.on_train_batch_end/allocated_bytes.small_pool.allocated":5721763890688,"DeviceStatsMonitor.on_train_batch_start/segment.large_pool.peak":239,"DeviceStatsMonitor.on_train_batch_end/oversize_allocations.peak":0,"_step":4039,"DeviceStatsMonitor.on_train_batch_start/segment.small_pool.peak":58,"DeviceStatsMonitor.on_train_batch_end/reserved_bytes.large_pool.current":7702839296,"DeviceStatsMonitor.on_train_batch_start/num_alloc_retries":0,"DeviceStatsMonitor.on_train_batch_start/reserved_bytes.all.peak":7824474112,"DeviceStatsMonitor.on_train_batch_start/reserved_bytes.all.freed":0,"DeviceStatsMonitor.on_train_batch_start/num_device_free":0,"DeviceStatsMonitor.on_train_batch_end/active_bytes.small_pool.allocated":5721763890688,"DeviceStatsMonitor.on_train_batch_end/inactive_split.large_pool.allocated":41295558,"DeviceStatsMonitor.on_train_batch_end/oversize_allocations.current":0,"DeviceStatsMonitor.on_train_batch_start/inactive_split.large_pool.current":25,"DeviceStatsMonitor.on_train_batch_end/inactive_split.large_pool.current":26,"DeviceStatsMonitor.on_train_batch_start/active.all.current":1468,"DeviceStatsMonitor.on_train_batch_start/requested_bytes.all.peak":7385954500,"DeviceStatsMonitor.on_train_batch_end/allocated_bytes.large_pool.freed":1193776984576000,"DeviceStatsMonitor.on_train_batch_end/inactive_split_bytes.small_pool.current":22806528,"DeviceStatsMonitor.on_train_batch_end/allocated_bytes.all.current":1383923712,"DeviceStatsMonitor.on_train_batch_start/active.all.allocated":120395444,"DeviceStatsMonitor.on_train_batch_end/active.all.peak":1983,"DeviceStatsMonitor.on_train_batch_end/segment.all.freed":0,"DeviceStatsMonitor.on_train_batch_start/allocation.large_pool.allocated":56372708,"DeviceStatsMonitor.on_train_batch_start/requested_bytes.large_pool.allocated":1187113576819712}
ssl_dino/vit16s/wandb/run-20250815_070844-n4cjse4r/logs/debug-core.log ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2025-08-15T07:08:44.607897484+07:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmp5_tprco9/port-12754.txt","pid":12754,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false}
2
+ {"time":"2025-08-15T07:08:44.610703499+07:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":12754}
3
+ {"time":"2025-08-15T07:08:44.610000205+07:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-12754-13173-3642601375/socket","Net":"unix"}}
4
+ {"time":"2025-08-15T07:08:44.686649603+07:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"}
5
+ {"time":"2025-08-15T07:08:44.704397366+07:00","level":"INFO","msg":"handleInformInit: received","streamId":"n4cjse4r","id":"1(@)"}
6
+ {"time":"2025-08-15T07:08:45.199256854+07:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"n4cjse4r","id":"1(@)"}
7
+ {"time":"2025-08-15T13:09:05.891073334+07:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"}
8
+ {"time":"2025-08-15T13:09:05.894723852+07:00","level":"INFO","msg":"server is shutting down"}
9
+ {"time":"2025-08-15T13:09:05.894796904+07:00","level":"INFO","msg":"connection: closing","id":"1(@)"}
10
+ {"time":"2025-08-15T13:09:05.894832135+07:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"}
11
+ {"time":"2025-08-15T13:09:05.894921266+07:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-12754-13173-3642601375/socket","Net":"unix"}}
12
+ {"time":"2025-08-15T13:09:07.370592871+07:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"}
13
+ {"time":"2025-08-15T13:09:07.370625641+07:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"}
14
+ {"time":"2025-08-15T13:09:07.370641342+07:00","level":"INFO","msg":"server is closed"}
ssl_dino/vit16s/wandb/run-20250815_070844-n4cjse4r/logs/debug-internal.log ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2025-08-15T07:08:44.704475178+07:00","level":"INFO","msg":"stream: starting","core version":"0.21.1"}
2
+ {"time":"2025-08-15T07:08:45.199222243+07:00","level":"INFO","msg":"stream: created new stream","id":"n4cjse4r"}
3
+ {"time":"2025-08-15T07:08:45.199252394+07:00","level":"INFO","msg":"stream: started","id":"n4cjse4r"}
4
+ {"time":"2025-08-15T07:08:45.199259814+07:00","level":"INFO","msg":"handler: started","stream_id":"n4cjse4r"}
5
+ {"time":"2025-08-15T07:08:45.199277254+07:00","level":"INFO","msg":"writer: started","stream_id":"n4cjse4r"}
6
+ {"time":"2025-08-15T07:08:45.199275424+07:00","level":"INFO","msg":"sender: started","stream_id":"n4cjse4r"}
7
+ {"time":"2025-08-15T13:09:05.899861692+07:00","level":"INFO","msg":"stream: closing","id":"n4cjse4r"}
8
+ {"time":"2025-08-15T13:09:06.986277531+07:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
9
+ {"time":"2025-08-15T13:09:07.364748815+07:00","level":"INFO","msg":"handler: closed","stream_id":"n4cjse4r"}
10
+ {"time":"2025-08-15T13:09:07.366528993+07:00","level":"INFO","msg":"sender: closed","stream_id":"n4cjse4r"}
11
+ {"time":"2025-08-15T13:09:07.366568884+07:00","level":"INFO","msg":"stream: closed","id":"n4cjse4r"}
ssl_dino/vit16s/wandb/run-20250815_070844-n4cjse4r/logs/debug.log ADDED
The diff for this file is too large to render. See raw diff
 
ssl_dino/vit16s/wandb/run-20250815_070844-n4cjse4r/run-n4cjse4r.wandb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:192ca742bc7e2a4c559e5c76538f4769046c06e9b4ae009cc95d715355f0da11
3
+ size 48317254
ssl_dino/vit16s/wandb/run-20250815_160317-46zcck3z/files/config.yaml ADDED
@@ -0,0 +1,235 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ _wandb:
2
+ value:
3
+ cli_version: 0.21.1
4
+ e:
5
+ zxx6d80alc7m10xjqmyculqtynxz0cdb:
6
+ codePath: src/experiment/ssl_dino.py
7
+ codePathLocal: src/experiment/ssl_dino.py
8
+ cpu_count: 6
9
+ cpu_count_logical: 6
10
+ cudaVersion: "12.9"
11
+ disk:
12
+ /:
13
+ total: "171828551680"
14
+ used: "149767499776"
15
+ email: acc.ggcolab@gmail.com
16
+ executable: /home/duckq1u/miniconda3/envs/ent-lab/bin/python
17
+ git:
18
+ commit: 61f4482b9df47dedf1dd991b89913f2e5962454c
19
+ remote: git@github.com:PCTU-Deep-Medicine/ent-labotary.git
20
+ gpu: NVIDIA GeForce RTX 5060 Ti
21
+ gpu_count: 1
22
+ gpu_nvidia:
23
+ - architecture: Blackwell
24
+ cudaCores: 4608
25
+ memoryTotal: "17103323136"
26
+ name: NVIDIA GeForce RTX 5060 Ti
27
+ uuid: GPU-a8c62e38-0f95-e068-50c6-fab6ff92eaaf
28
+ host: duckq1
29
+ memory:
30
+ total: "16686198784"
31
+ os: Linux-6.14.0-arch1-1-x86_64-with-glibc2.42
32
+ program: /home/duckq1u/Documents/workspace/cuong/ent-labotary/src/experiment/ssl_dino.py
33
+ python: CPython 3.11.13
34
+ root: /home/duckq1u/Documents/workspace/cuong/ent-labotary/outputs/ssl_dino/vit16s
35
+ startedAt: "2025-08-15T09:03:17.828004Z"
36
+ writerId: zxx6d80alc7m10xjqmyculqtynxz0cdb
37
+ m:
38
+ - "1": trainer/global_step
39
+ "6":
40
+ - 3
41
+ "7": []
42
+ - "2": '*'
43
+ "5": 1
44
+ "6":
45
+ - 1
46
+ "7": []
47
+ python_version: 3.11.13
48
+ t:
49
+ "1":
50
+ - 1
51
+ - 5
52
+ - 9
53
+ - 11
54
+ - 41
55
+ - 49
56
+ - 53
57
+ - 63
58
+ - 80
59
+ - 103
60
+ - 105
61
+ "2":
62
+ - 1
63
+ - 5
64
+ - 9
65
+ - 11
66
+ - 41
67
+ - 49
68
+ - 53
69
+ - 63
70
+ - 80
71
+ - 103
72
+ - 105
73
+ "3":
74
+ - 7
75
+ - 66
76
+ "4": 3.11.13
77
+ "5": 0.21.1
78
+ "6": 4.54.1
79
+ "12": 0.21.1
80
+ "13": linux-x86_64
81
+ accelerator:
82
+ value: CUDAAccelerator
83
+ batch_size:
84
+ value: 32
85
+ callbacks:
86
+ value:
87
+ early_stopping:
88
+ check_finite: true
89
+ monitor: train_loss
90
+ patience: 1000000000000
91
+ model_checkpoint:
92
+ enable_version_counter: false
93
+ every_n_epochs: null
94
+ save_last: true
95
+ save_top_k: 1
96
+ checkpoint:
97
+ value: null
98
+ data:
99
+ value: data/kyucapsule
100
+ devices:
101
+ value: 1
102
+ embed_dim:
103
+ value: null
104
+ epochs:
105
+ value: 300
106
+ loader_args:
107
+ value: null
108
+ loggers:
109
+ value:
110
+ jsonl:
111
+ flush_logs_every_n_steps: 100
112
+ tensorboard:
113
+ default_hp_metric: true
114
+ log_graph: false
115
+ name: ""
116
+ prefix: ""
117
+ sub_dir: null
118
+ version: ""
119
+ wandb:
120
+ anonymous: null
121
+ checkpoint_name: null
122
+ log_model: false
123
+ name: null
124
+ offline: false
125
+ prefix: ""
126
+ project: ent-endoscopy-ssl
127
+ version: null
128
+ method:
129
+ value: dino
130
+ method_args:
131
+ value:
132
+ batch_norm: false
133
+ bottleneck_dim: 256
134
+ center_momentum: 0.9
135
+ hidden_dim: 2048
136
+ momentum_end: 1
137
+ momentum_start: 0.99
138
+ norm_last_layer: true
139
+ output_dim: 1024
140
+ student_freeze_last_layer_epochs: 1
141
+ student_temp: 0.1
142
+ teacher_temp: 0.02
143
+ warmup_teacher_temp: 0.02
144
+ warmup_teacher_temp_epochs: 30
145
+ weight_decay_end: 0.0001
146
+ weight_decay_start: 0.0001
147
+ model:
148
+ value: VisionTransformer
149
+ model_args:
150
+ value: null
151
+ num_nodes:
152
+ value: 1
153
+ num_workers:
154
+ value: 5
155
+ optim:
156
+ value: sgd
157
+ optim_args:
158
+ value:
159
+ lr: 0.03
160
+ momentum: 0.9
161
+ weight_decay: 0.0001
162
+ out:
163
+ value: outputs/ssl_dino/vit16s
164
+ overwrite:
165
+ value: true
166
+ precision:
167
+ value: 32-true
168
+ resume:
169
+ value: true
170
+ seed:
171
+ value: 0
172
+ strategy:
173
+ value: SingleDeviceStrategy
174
+ trainer_args:
175
+ value: null
176
+ transform_args:
177
+ value:
178
+ color_jitter:
179
+ brightness: 0.8
180
+ contrast: 0.8
181
+ hue: 0.2
182
+ prob: 0.8
183
+ saturation: 0.4
184
+ strength: 0.5
185
+ gaussian_blur:
186
+ blur_limit: 0
187
+ prob: 1
188
+ sigmas:
189
+ - 0.1
190
+ - 2
191
+ global_view_1:
192
+ gaussian_blur:
193
+ blur_limit: 0
194
+ prob: 0.1
195
+ sigmas:
196
+ - 0.1
197
+ - 2
198
+ solarize:
199
+ prob: 0.2
200
+ threshold: 0.5
201
+ image_size:
202
+ - 224
203
+ - 224
204
+ local_view:
205
+ gaussian_blur:
206
+ blur_limit: 0
207
+ prob: 0.5
208
+ sigmas:
209
+ - 0.1
210
+ - 2
211
+ num_views: 6
212
+ random_resize:
213
+ max_scale: 0.14
214
+ min_scale: 0.05
215
+ view_size:
216
+ - 96
217
+ - 96
218
+ normalize:
219
+ mean:
220
+ - 0.485
221
+ - 0.456
222
+ - 0.406
223
+ std:
224
+ - 0.229
225
+ - 0.224
226
+ - 0.225
227
+ random_flip:
228
+ horizontal_prob: 0.5
229
+ vertical_prob: 0
230
+ random_gray_scale: 0.2
231
+ random_resize:
232
+ max_scale: 1
233
+ min_scale: 0.14
234
+ random_rotation: null
235
+ solarize: null
ssl_dino/vit16s/wandb/run-20250815_160317-46zcck3z/files/output.log ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ /home/duckq1u/miniconda3/envs/ent-lab/lib/python3.11/site-packages/torch/__init__.py:1539: UserWarning: Please use the new API settings to control TF32 behavior, such as torch.backends.cudnn.conv.fp32_precision = 'tf32' or torch.backends.cuda.matmul.fp32_precision = 'ieee'. Old settings, e.g, torch.backends.cuda.matmul.allow_tf32 = True, torch.backends.cudnn.allow_tf32 = True, allowTF32CuDNN() and allowTF32CuBLAS() will be deprecated after Pytorch 2.9. Please see https://pytorch.org/docs/main/notes/cuda.html#tensorfloat-32-tf32-on-ampere-and-later-devices (Triggered internally at /pytorch/aten/src/ATen/Context.cpp:80.)
2
+ return _C._get_float32_matmul_precision()
3
+ 
4
+ You are using a CUDA device ('NVIDIA GeForce RTX 5060 Ti') that has Tensor Cores. To properly utilize them, you should set `torch.set_float32_matmul_precision('medium' | 'high')` which will trade-off precision for performance. For more details, read https://pytorch.org/docs/stable/generated/torch.set_float32_matmul_precision.html#torch.set_float32_matmul_precision
5
+ Restoring states from the checkpoint path at /home/duckq1u/Documents/workspace/cuong/ent-labotary/outputs/ssl_dino/vit16s/checkpoints/last.ckpt
6
+ LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0]
7
+ Loading `train_dataloader` to estimate number of stepping batches.
8
+
9
+ | Name | Type | Params | Mode
10
+ -----------------------------------------------------------------------
11
+ 0 | teacher_embedding_model | EmbeddingModel | 30.1 M | train
12
+ 1 | teacher_projection_head | DINOProjectionHead | 5.8 M | train
13
+ 2 | student_embedding_model | EmbeddingModel | 30.1 M | train
14
+ 3 | student_projection_head | DINOProjectionHead | 5.8 M | train
15
+ 4 | flatten | Flatten | 0 | train
16
+ 5 | criterion | DINOLoss | 0 | train
17
+ -----------------------------------------------------------------------
18
+ 71.7 M Trainable params
19
+ 2.0 K Non-trainable params
20
+ 71.7 M Total params
21
+ 286.782 Total estimated model params size (MB)
22
+ 578 Modules in train mode
23
+ 0 Modules in eval mode
24
+ Restored all states from the checkpoint at /home/duckq1u/Documents/workspace/cuong/ent-labotary/outputs/ssl_dino/vit16s/checkpoints/last.ckpt
25
+ Epoch 299: 100%|███████████████████████████████████████████████████████| 577/577 [04:06<00:00, 2.34it/s, v_num=ck3z, train_loss=0.433, data_wait=51.6%]
26
+ `Trainer.fit` stopped: `max_epochs=300` reached.
27
+ Training completed.
28
+ Example: How to use the exported model
29
+ ----------------------------------------------------------------------------------------
30
+ import timm
31
+ 
32
+ # Load the pretrained model
33
+ model = timm.create_model(
34
+  model_name='vit_small_patch16_224',
35
+  checkpoint_path='/home/duckq1u/Documents/workspace/cuong/ent-labotary/outputs/ssl_dino/vit16s/exported_models/exported_last.pt',
36
+ )
37
+ 
38
+ # Finetune or evaluate the model
39
+ ...
40
+ ----------------------------------------------------------------------------------------
41
+ 
42
+ Model exported.
43
+ Args: {
44
+ "checkpoint": "outputs/ssl_dino/vit16s/checkpoints/last.ckpt",
45
+ "format": "torch_state_dict",
46
+ "out": "outputs/ssl_dino/vit16s/vit_small_patch16_224_dino.pt",
47
+ "overwrite": false,
48
+ "part": "model"
49
+ }
50
+ Exporting 'ModelPart.MODEL' as 'ModelFormat.TORCH_STATE_DICT'.
51
+ Loading checkpoint from '/home/duckq1u/Documents/workspace/cuong/ent-labotary/outputs/ssl_dino/vit16s/checkpoints/last.ckpt'
52
+ Exporting model to '/home/duckq1u/Documents/workspace/cuong/ent-labotary/outputs/ssl_dino/vit16s/vit_small_patch16_224_dino.pt'
ssl_dino/vit16s/wandb/run-20250815_160317-46zcck3z/files/requirements.txt ADDED
@@ -0,0 +1,195 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ threadpoolctl==3.6.0
2
+ MarkupSafe==3.0.2
3
+ timm==1.0.19
4
+ nvidia-cuda-nvrtc-cu12==12.8.93
5
+ certifi==2025.8.3
6
+ soupsieve==2.7
7
+ regex==2025.7.34
8
+ pre_commit==4.2.0
9
+ lightly==1.5.22
10
+ nvidia-cusolver-cu12==11.7.3.90
11
+ mypy_extensions==1.1.0
12
+ protobuf==6.31.1
13
+ cffi==1.17.1
14
+ psutil==7.0.0
15
+ identify==2.6.12
16
+ pycparser==2.22
17
+ multidict==6.6.3
18
+ Werkzeug==3.1.3
19
+ dotenv==0.9.9
20
+ charset-normalizer==3.4.3
21
+ albucore==0.0.24
22
+ torch==2.9.0.dev20250806+cu128
23
+ packaging==25.0
24
+ pydot==4.0.1
25
+ mpmath==1.3.0
26
+ pycodestyle==2.14.0
27
+ appdirs==1.4.4
28
+ nvidia-cufft-cu12==11.3.3.83
29
+ nodejs-wheel-binaries==22.18.0
30
+ tokenizers==0.21.4
31
+ black==25.1.0
32
+ tqdm==4.67.1
33
+ hf-xet==1.1.5
34
+ tabulate==0.9.0
35
+ dpath==2.2.0
36
+ smmap==5.0.2
37
+ cycler==0.12.1
38
+ python-dotenv==1.1.1
39
+ kiwisolver==1.4.8
40
+ shortuuid==1.0.13
41
+ shellingham==1.5.4
42
+ beautifulsoup4==4.13.4
43
+ kombu==5.5.4
44
+ isort==6.0.1
45
+ orjson==3.11.1
46
+ nvidia-cublas-cu12==12.8.4.1
47
+ iterative-telemetry==0.0.10
48
+ flufl.lock==8.2.0
49
+ pytorch-lightning==2.5.2
50
+ matplotlib==3.10.5
51
+ flake8==7.3.0
52
+ nvidia-nvtx-cu12==12.8.90
53
+ eval_type_backport==0.2.2
54
+ nvidia-cuda-runtime-cu12==12.8.90
55
+ pygtrie==2.5.0
56
+ pandas==2.3.1
57
+ amqp==5.3.1
58
+ nvidia-nvshmem-cu12==3.3.9
59
+ mccabe==0.7.0
60
+ pip==25.1
61
+ basedpyright==1.31.1
62
+ aiosignal==1.4.0
63
+ rich==14.1.0
64
+ albumentations==2.0.8
65
+ huggingface-hub==0.34.3
66
+ six==1.17.0
67
+ grandalf==0.8
68
+ celery==5.5.3
69
+ aiohttp==3.12.15
70
+ bcrypt==4.3.0
71
+ omegaconf==2.3.0
72
+ joblib==1.5.1
73
+ setuptools==78.1.1
74
+ lightning-utilities==0.15.0
75
+ sqltrie==0.11.2
76
+ diskcache==5.6.3
77
+ typing-inspection==0.4.1
78
+ flatten-dict==0.4.2
79
+ pydantic==2.11.7
80
+ invoke==2.2.0
81
+ fonttools==4.59.0
82
+ scipy==1.16.1
83
+ torchaudio==2.8.0.dev20250806+cu128
84
+ nvidia-cuda-cupti-cu12==12.8.90
85
+ dvc-studio-client==0.22.0
86
+ yarl==1.20.1
87
+ shtab==1.7.2
88
+ click-repl==0.3.0
89
+ wandb==0.21.1
90
+ antlr4-python3-runtime==4.9.3
91
+ pytorch-triton==3.4.0+gitf7888497
92
+ tensorboard-data-server==0.7.2
93
+ click==8.2.1
94
+ stringzilla==3.12.6
95
+ PySocks==1.7.1
96
+ pydantic_core==2.33.2
97
+ Jinja2==3.1.6
98
+ click-plugins==1.1.1.2
99
+ vine==5.1.0
100
+ dvc==3.61.0
101
+ configobj==5.0.9
102
+ opencv-python-headless==4.12.0.88
103
+ nvidia-nccl-cu12==2.27.5
104
+ loguru==0.7.3
105
+ attrs==25.3.0
106
+ cfgv==3.4.0
107
+ nvidia-cufile-cu12==1.13.1.3
108
+ simsimd==6.5.0
109
+ zc.lockfile==3.0.post1
110
+ idna==3.10
111
+ typing_extensions==4.14.1
112
+ tomlkit==0.13.3
113
+ click-didyoumean==0.3.1
114
+ propcache==0.3.2
115
+ markdown-it-py==3.0.0
116
+ dvc-http==2.32.0
117
+ urllib3==2.5.0
118
+ dvc-render==1.0.2
119
+ frozenlist==1.7.0
120
+ lightly-utils==0.0.2
121
+ dvc-data==3.16.10
122
+ cryptography==45.0.5
123
+ annotated-types==0.7.0
124
+ tensorboard==2.20.0
125
+ pyarrow==21.0.0
126
+ filelock==3.18.0
127
+ PyYAML==6.0.2
128
+ requests==2.32.4
129
+ pyparsing==3.2.3
130
+ fsspec==2025.7.0
131
+ ruamel.yaml.clib==0.2.12
132
+ semver==3.0.4
133
+ uv==0.8.4
134
+ voluptuous==0.15.2
135
+ aiohappyeyeballs==2.6.1
136
+ atpublic==6.0.1
137
+ asyncssh==2.21.0
138
+ torchmetrics==1.8.0
139
+ nvidia-curand-cu12==10.3.9.90
140
+ python-dateutil==2.9.0.post0
141
+ nvidia-cusparselt-cu12==0.7.1
142
+ dvc-task==0.40.2
143
+ dictdiffer==0.9.0
144
+ transformers==4.54.1
145
+ sentry-sdk==2.34.1
146
+ PyNaCl==1.5.0
147
+ nvidia-nvjitlink-cu12==12.8.93
148
+ virtualenv==20.33.0
149
+ tzdata==2025.2
150
+ safetensors==0.5.3
151
+ platformdirs==4.3.8
152
+ colorama==0.4.6
153
+ distlib==0.4.0
154
+ wcwidth==0.2.13
155
+ dvc-objects==5.1.1
156
+ nvidia-cusparse-cu12==12.5.8.93
157
+ scmrepo==3.5.1
158
+ mdurl==0.1.2
159
+ Pygments==2.19.2
160
+ triton==3.3.1
161
+ pytz==2025.2
162
+ distro==1.9.0
163
+ scikit-learn==1.7.1
164
+ seaborn==0.13.2
165
+ absl-py==2.3.1
166
+ ruamel.yaml==0.18.14
167
+ grpcio==1.74.0
168
+ funcy==2.0
169
+ gto==1.7.2
170
+ nvidia-cudnn-cu12==9.10.2.21
171
+ prompt_toolkit==3.0.51
172
+ numpy==2.2.6
173
+ lightly_train==0.6.1
174
+ dulwich==0.24.1
175
+ paramiko==4.0.0
176
+ sympy==1.14.0
177
+ hydra-core==1.3.2
178
+ networkx==3.5
179
+ torchvision==0.24.0.dev20250806+cu128
180
+ contourpy==1.3.3
181
+ pygit2==1.18.1
182
+ billiard==4.2.1
183
+ pillow==11.3.0
184
+ nodeenv==1.9.1
185
+ aenum==3.1.16
186
+ Markdown==3.8.2
187
+ typer==0.16.0
188
+ gitdb==4.0.12
189
+ aiohttp-retry==2.9.1
190
+ pathspec==0.12.1
191
+ entrypoints==0.4
192
+ gdown==5.2.0
193
+ pyflakes==3.4.0
194
+ GitPython==3.1.45
195
+ wheel==0.45.1
ssl_dino/vit16s/wandb/run-20250815_160317-46zcck3z/files/wandb-metadata.json ADDED
@@ -0,0 +1,40 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-6.14.0-arch1-1-x86_64-with-glibc2.42",
3
+ "python": "CPython 3.11.13",
4
+ "startedAt": "2025-08-15T09:03:17.828004Z",
5
+ "program": "/home/duckq1u/Documents/workspace/cuong/ent-labotary/src/experiment/ssl_dino.py",
6
+ "codePath": "src/experiment/ssl_dino.py",
7
+ "codePathLocal": "src/experiment/ssl_dino.py",
8
+ "git": {
9
+ "remote": "git@github.com:PCTU-Deep-Medicine/ent-labotary.git",
10
+ "commit": "61f4482b9df47dedf1dd991b89913f2e5962454c"
11
+ },
12
+ "email": "acc.ggcolab@gmail.com",
13
+ "root": "/home/duckq1u/Documents/workspace/cuong/ent-labotary/outputs/ssl_dino/vit16s",
14
+ "host": "duckq1",
15
+ "executable": "/home/duckq1u/miniconda3/envs/ent-lab/bin/python",
16
+ "cpu_count": 6,
17
+ "cpu_count_logical": 6,
18
+ "gpu": "NVIDIA GeForce RTX 5060 Ti",
19
+ "gpu_count": 1,
20
+ "disk": {
21
+ "/": {
22
+ "total": "171828551680",
23
+ "used": "149767499776"
24
+ }
25
+ },
26
+ "memory": {
27
+ "total": "16686198784"
28
+ },
29
+ "gpu_nvidia": [
30
+ {
31
+ "name": "NVIDIA GeForce RTX 5060 Ti",
32
+ "memoryTotal": "17103323136",
33
+ "cudaCores": 4608,
34
+ "architecture": "Blackwell",
35
+ "uuid": "GPU-a8c62e38-0f95-e068-50c6-fab6ff92eaaf"
36
+ }
37
+ ],
38
+ "cudaVersion": "12.9",
39
+ "writerId": "zxx6d80alc7m10xjqmyculqtynxz0cdb"
40
+ }
ssl_dino/vit16s/wandb/run-20250815_160317-46zcck3z/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"DeviceStatsMonitor.on_train_batch_end/allocation.all.current":666,"DeviceStatsMonitor.on_train_batch_end/allocation.small_pool.freed":39535154,"DeviceStatsMonitor.on_train_batch_end/allocated_bytes.all.freed":740672323476480,"DeviceStatsMonitor.on_train_batch_end/active.all.current":666,"DeviceStatsMonitor.on_train_batch_end/reserved_bytes.all.current":7964983296,"DeviceStatsMonitor.on_train_batch_start/active.all.freed":74344276,"DeviceStatsMonitor.on_train_batch_end/segment.all.peak":304,"profiling/data_time":0.22660590708255768,"DeviceStatsMonitor.on_train_batch_start/allocation.all.current":663,"DeviceStatsMonitor.on_train_batch_start/active.all.current":663,"DeviceStatsMonitor.on_train_batch_end/inactive_split.all.current":69,"DeviceStatsMonitor.on_train_batch_end/active.all.freed":74346648,"DeviceStatsMonitor.on_train_batch_start/active_bytes.large_pool.freed":737115507003392,"DeviceStatsMonitor.on_train_batch_start/num_sync_all_streams":3,"DeviceStatsMonitor.on_train_batch_end/inactive_split.small_pool.current":61,"DeviceStatsMonitor.on_train_batch_end/active_bytes.all.current":613189632,"DeviceStatsMonitor.on_train_batch_start/active_bytes.large_pool.allocated":737116080705536,"DeviceStatsMonitor.on_train_batch_start/requested_bytes.large_pool.allocated":733052360620544,"DeviceStatsMonitor.on_train_batch_start/allocation.large_pool.freed":34810385,"DeviceStatsMonitor.on_train_batch_start/active_bytes.small_pool.current":38109696,"DeviceStatsMonitor.on_train_batch_end/active_bytes.small_pool.allocated":3533268317184,"DeviceStatsMonitor.on_train_batch_end/max_split_size":-1,"DeviceStatsMonitor.on_train_batch_start/active.small_pool.peak":1495,"DeviceStatsMonitor.on_train_batch_start/allocated_bytes.all.freed":740648623866880,"DeviceStatsMonitor.on_train_batch_end/inactive_split_bytes.all.current":101939200,"DeviceStatsMonitor.on_train_batch_end/inactive_split_bytes.large_pool.allocated":175380137947136,"DeviceStatsMonitor.on_train_batch_end/inactive_split.small_pool.peak":242,"DeviceStatsMonitor.on_train_batch_end/active.small_pool.peak":1495,"DeviceStatsMonitor.on_train_batch_end/reserved_bytes.small_pool.allocated":130023424,"DeviceStatsMonitor.on_train_batch_end/requested_bytes.large_pool.allocated":733075827765248,"epoch":299,"DeviceStatsMonitor.on_train_batch_end/active.large_pool.current":172,"DeviceStatsMonitor.on_train_batch_start/inactive_split.small_pool.allocated":24946240,"DeviceStatsMonitor.on_train_batch_start/segment.small_pool.allocated":62,"DeviceStatsMonitor.on_train_batch_end/allocated_bytes.small_pool.current":38111232,"train_loss":0.43312790989875793,"DeviceStatsMonitor.on_train_batch_end/segment.large_pool.peak":242,"DeviceStatsMonitor.on_train_batch_start/inactive_split_bytes.large_pool.freed":175375231152128,"DeviceStatsMonitor.on_train_batch_start/requested_bytes.small_pool.current":38101740,"DeviceStatsMonitor.on_train_batch_start/inactive_split_bytes.small_pool.peak":62521856,"DeviceStatsMonitor.on_train_batch_end/reserved_bytes.all.freed":0,"DeviceStatsMonitor.on_train_batch_end/requested_bytes.small_pool.freed":3527496821356,"DeviceStatsMonitor.on_train_batch_end/reserved_bytes.all.allocated":7964983296,"DeviceStatsMonitor.on_train_batch_end/active_bytes.small_pool.current":38111232,"DeviceStatsMonitor.on_train_batch_start/oversize_segments.freed":0,"DeviceStatsMonitor.on_train_batch_end/allocated_bytes.small_pool.freed":3533230205952,"DeviceStatsMonitor.on_train_batch_end/allocated_bytes.large_pool.peak":7557218304,"DeviceStatsMonitor.on_train_batch_end/allocated_bytes.all.current":613189632,"DeviceStatsMonitor.on_train_batch_end/reserved_bytes.large_pool.current":7834959872,"DeviceStatsMonitor.on_train_batch_end/active.all.allocated":74347314,"DeviceStatsMonitor.on_train_batch_start/inactive_split.all.freed":50436077,"DeviceStatsMonitor.on_train_batch_end/active.small_pool.allocated":39535648,"DeviceStatsMonitor.on_train_batch_start/requested_bytes.large_pool.peak":7415267840,"DeviceStatsMonitor.on_train_batch_end/requested_bytes.all.current":586237688,"DeviceStatsMonitor.on_train_batch_start/allocation.large_pool.allocated":34810557,"DeviceStatsMonitor.on_train_batch_end/requested_bytes.large_pool.current":548135936,"DeviceStatsMonitor.on_train_batch_end/allocated_bytes.all.peak":7671335936,"DeviceStatsMonitor.on_train_batch_end/active_bytes.all.allocated":740672936666112,"DeviceStatsMonitor.on_train_batch_start/segment.large_pool.freed":0,"DeviceStatsMonitor.on_train_batch_start/inactive_split_bytes.large_pool.peak":618004480,"DeviceStatsMonitor.on_train_batch_start/active.large_pool.peak":659,"DeviceStatsMonitor.on_train_batch_end/active_bytes.large_pool.peak":7557218304,"DeviceStatsMonitor.on_train_batch_start/requested_bytes.all.current":586237676,"DeviceStatsMonitor.on_train_batch_end/segment.small_pool.allocated":62,"DeviceStatsMonitor.on_train_batch_start/num_ooms":0,"DeviceStatsMonitor.on_train_batch_start/oversize_allocations.current":0,"DeviceStatsMonitor.on_train_batch_end/inactive_split.small_pool.freed":24946920,"DeviceStatsMonitor.on_train_batch_end/oversize_segments.allocated":0,"DeviceStatsMonitor.on_train_batch_start/reserved_bytes.all.allocated":7964983296,"DeviceStatsMonitor.on_train_batch_start/inactive_split_bytes.all.allocated":178973846893568,"DeviceStatsMonitor.on_train_batch_start/requested_bytes.large_pool.current":548135936,"DeviceStatsMonitor.on_train_batch_end/segment.large_pool.current":242,"DeviceStatsMonitor.on_train_batch_start/inactive_split_bytes.small_pool.current":18513408,"DeviceStatsMonitor.on_train_batch_start/active_bytes.all.freed":740648623866880,"DeviceStatsMonitor.on_train_batch_start/segment.small_pool.freed":0,"DeviceStatsMonitor.on_train_batch_end/num_ooms":0,"_wandb":{"runtime":13458},"DeviceStatsMonitor.on_train_batch_end/requested_bytes.all.freed":736602776450668,"DeviceStatsMonitor.on_train_batch_end/requested_bytes.small_pool.current":38101752,"DeviceStatsMonitor.on_train_batch_start/segment.all.freed":0,"DeviceStatsMonitor.on_train_batch_end/reserved_bytes.all.peak":7964983296,"DeviceStatsMonitor.on_train_batch_end/allocation.large_pool.current":172,"DeviceStatsMonitor.on_train_batch_end/allocation.large_pool.allocated":34811666,"DeviceStatsMonitor.on_train_batch_end/active.large_pool.allocated":34811666,"DeviceStatsMonitor.on_train_batch_end/inactive_split_bytes.small_pool.allocated":3598608284672,"DeviceStatsMonitor.on_train_batch_end/active.small_pool.current":494,"DeviceStatsMonitor.on_train_batch_end/allocation.all.freed":74346648,"DeviceStatsMonitor.on_train_batch_start/inactive_split.all.current":71,"lr-SGD/params_no_weight_decay":3.75e-06,"DeviceStatsMonitor.on_train_batch_end/oversize_allocations.freed":0,"DeviceStatsMonitor.on_train_batch_start/inactive_split_bytes.large_pool.allocated":175375351607296,"DeviceStatsMonitor.on_train_batch_end/oversize_allocations.current":0,"DeviceStatsMonitor.on_train_batch_start/oversize_allocations.peak":0,"DeviceStatsMonitor.on_train_batch_start/inactive_split_bytes.small_pool.allocated":3598495286272,"DeviceStatsMonitor.on_train_batch_end/active.large_pool.peak":659,"DeviceStatsMonitor.on_train_batch_start/requested_bytes.small_pool.peak":120280588,"DeviceStatsMonitor.on_train_batch_end/inactive_split_bytes.large_pool.peak":618004480,"DeviceStatsMonitor.on_train_batch_end/num_device_free":0,"DeviceStatsMonitor.on_train_batch_start/inactive_split_bytes.all.freed":178973707924992,"DeviceStatsMonitor.on_train_batch_end/inactive_split_bytes.large_pool.freed":175380052422656,"DeviceStatsMonitor.on_train_batch_start/num_device_free":0,"DeviceStatsMonitor.on_train_batch_end/oversize_allocations.allocated":0,"DeviceStatsMonitor.on_train_batch_end/active_bytes.small_pool.freed":3533230205952,"DeviceStatsMonitor.on_train_batch_end/inactive_split_bytes.small_pool.current":16414720,"DeviceStatsMonitor.on_train_batch_end/active_bytes.all.freed":740672323476480,"DeviceStatsMonitor.on_train_batch_start/allocation.large_pool.current":172,"DeviceStatsMonitor.on_train_batch_start/segment.small_pool.peak":62,"DeviceStatsMonitor.on_train_batch_start/allocated_bytes.all.allocated":740649235678720,"DeviceStatsMonitor.on_train_batch_start/reserved_bytes.large_pool.current":7834959872,"DeviceStatsMonitor.on_train_batch_end/inactive_split_bytes.small_pool.peak":62521856,"DeviceStatsMonitor.on_train_batch_start/inactive_split.large_pool.allocated":25489908,"DeviceStatsMonitor.on_train_batch_end/reserved_bytes.small_pool.freed":0,"DeviceStatsMonitor.on_train_batch_start/requested_bytes.all.freed":736579196144924,"trainer/global_step":173099,"DeviceStatsMonitor.on_train_batch_end/requested_bytes.large_pool.peak":7415267840,"DeviceStatsMonitor.on_train_batch_start/inactive_split.small_pool.freed":24946184,"DeviceStatsMonitor.on_train_batch_end/allocation.large_pool.freed":34811494,"_step":2495,"DeviceStatsMonitor.on_train_batch_end/oversize_segments.freed":0,"DeviceStatsMonitor.on_train_batch_end/active_bytes.small_pool.peak":120316928,"DeviceStatsMonitor.on_train_batch_start/reserved_bytes.large_pool.peak":7834959872,"DeviceStatsMonitor.on_train_batch_end/allocated_bytes.large_pool.allocated":737139668348928,"lr-SGD/params":3.75e-06,"DeviceStatsMonitor.on_train_batch_start/active.small_pool.freed":39533891,"DeviceStatsMonitor.on_train_batch_start/oversize_segments.allocated":0,"DeviceStatsMonitor.on_train_batch_end/num_device_alloc":304,"DeviceStatsMonitor.on_train_batch_start/allocated_bytes.small_pool.current":38109696,"DeviceStatsMonitor.on_train_batch_start/active_bytes.all.allocated":740649235678720,"DeviceStatsMonitor.on_train_batch_start/active.small_pool.allocated":39534382,"DeviceStatsMonitor.on_train_batch_start/num_device_alloc":304,"DeviceStatsMonitor.on_train_batch_start/segment.all.current":304,"DeviceStatsMonitor.on_train_batch_end/inactive_split.all.peak":364,"DeviceStatsMonitor.on_train_batch_start/allocated_bytes.small_pool.peak":120316928,"DeviceStatsMonitor.on_train_batch_end/segment.small_pool.freed":0,"DeviceStatsMonitor.on_train_batch_end/active_bytes.large_pool.freed":737139093270528,"DeviceStatsMonitor.on_train_batch_start/allocation.all.peak":2144,"DeviceStatsMonitor.on_train_batch_start/allocation.small_pool.peak":1495,"DeviceStatsMonitor.on_train_batch_start/segment.large_pool.current":242,"DeviceStatsMonitor.on_train_batch_end/allocation.small_pool.allocated":39535648,"DeviceStatsMonitor.on_train_batch_start/allocation.all.allocated":74344939,"DeviceStatsMonitor.on_train_batch_start/reserved_bytes.small_pool.freed":0,"DeviceStatsMonitor.on_train_batch_start/oversize_segments.current":0,"DeviceStatsMonitor.on_train_batch_start/active_bytes.small_pool.allocated":3533154973184,"DeviceStatsMonitor.on_train_batch_end/reserved_bytes.large_pool.freed":0,"DeviceStatsMonitor.on_train_batch_end/allocated_bytes.all.allocated":740672936666112,"DeviceStatsMonitor.on_train_batch_start/reserved_bytes.large_pool.allocated":7834959872,"DeviceStatsMonitor.on_train_batch_end/reserved_bytes.small_pool.current":130023424,"DeviceStatsMonitor.on_train_batch_start/reserved_bytes.small_pool.current":130023424,"DeviceStatsMonitor.on_train_batch_start/allocated_bytes.small_pool.freed":3533116863488,"DeviceStatsMonitor.on_train_batch_end/segment.small_pool.peak":62,"DeviceStatsMonitor.on_train_batch_end/inactive_split_bytes.all.peak":673997312,"DeviceStatsMonitor.on_train_batch_start/inactive_split.large_pool.peak":155,"DeviceStatsMonitor.on_train_batch_end/segment.large_pool.allocated":242,"DeviceStatsMonitor.on_train_batch_end/inactive_split.small_pool.allocated":24946981,"DeviceStatsMonitor.on_train_batch_start/active_bytes.small_pool.peak":120316928,"DeviceStatsMonitor.on_train_batch_start/active_bytes.all.current":611811840,"DeviceStatsMonitor.on_train_batch_end/allocated_bytes.small_pool.allocated":3533268317184,"DeviceStatsMonitor.on_train_batch_end/requested_bytes.small_pool.allocated":3527534923108,"DeviceStatsMonitor.on_train_batch_start/reserved_bytes.small_pool.peak":130023424,"DeviceStatsMonitor.on_train_batch_start/active_bytes.all.peak":7671335936,"DeviceStatsMonitor.on_train_batch_start/segment.all.allocated":304,"DeviceStatsMonitor.on_train_batch_start/reserved_bytes.all.peak":7964983296,"DeviceStatsMonitor.on_train_batch_end/segment.small_pool.current":62,"DeviceStatsMonitor.on_train_batch_end/allocation.all.peak":2144,"_runtime":13458,"DeviceStatsMonitor.on_train_batch_end/inactive_split.all.freed":50437616,"DeviceStatsMonitor.on_train_batch_start/inactive_split_bytes.all.current":138968576,"DeviceStatsMonitor.on_train_batch_start/active_bytes.large_pool.peak":7557218304,"DeviceStatsMonitor.on_train_batch_end/segment.all.freed":0,"DeviceStatsMonitor.on_train_batch_start/inactive_split_bytes.all.peak":673997312,"DeviceStatsMonitor.on_train_batch_end/active_bytes.all.peak":7671335936,"DeviceStatsMonitor.on_train_batch_end/inactive_split.large_pool.current":8,"DeviceStatsMonitor.on_train_batch_start/reserved_bytes.small_pool.allocated":130023424,"DeviceStatsMonitor.on_train_batch_start/requested_bytes.small_pool.allocated":3527421762056,"DeviceStatsMonitor.on_train_batch_start/requested_bytes.all.allocated":736579782382600,"DeviceStatsMonitor.on_train_batch_start/allocated_bytes.large_pool.freed":737115507003392,"DeviceStatsMonitor.on_train_batch_start/inactive_split.small_pool.current":56,"DeviceStatsMonitor.on_train_batch_start/reserved_bytes.all.freed":0,"DeviceStatsMonitor.on_train_batch_start/inactive_split.large_pool.current":15,"DeviceStatsMonitor.on_train_batch_start/oversize_allocations.allocated":0,"DeviceStatsMonitor.on_train_batch_end/segment.all.current":304,"DeviceStatsMonitor.on_train_batch_end/oversize_segments.current":0,"DeviceStatsMonitor.on_train_batch_end/segment.all.allocated":304,"DeviceStatsMonitor.on_train_batch_start/oversize_allocations.freed":0,"DeviceStatsMonitor.on_train_batch_end/allocated_bytes.small_pool.peak":120316928,"DeviceStatsMonitor.on_train_batch_end/allocation.small_pool.current":494,"DeviceStatsMonitor.on_train_batch_end/inactive_split.large_pool.freed":25490696,"DeviceStatsMonitor.on_train_batch_start/segment.all.peak":304,"DeviceStatsMonitor.on_train_batch_end/allocated_bytes.large_pool.freed":737139093270528,"DeviceStatsMonitor.on_train_batch_end/inactive_split_bytes.small_pool.freed":3598591869952,"DeviceStatsMonitor.on_train_batch_start/requested_bytes.large_pool.freed":733051812484608,"DeviceStatsMonitor.on_train_batch_start/inactive_split_bytes.small_pool.freed":3598476772864,"_timestamp":1.7552620544580507e+09,"DeviceStatsMonitor.on_train_batch_end/segment.large_pool.freed":0,"DeviceStatsMonitor.on_train_batch_start/allocated_bytes.large_pool.peak":7557218304,"DeviceStatsMonitor.on_train_batch_start/num_alloc_retries":0,"DeviceStatsMonitor.on_train_batch_end/oversize_allocations.peak":0,"DeviceStatsMonitor.on_train_batch_end/requested_bytes.large_pool.freed":733075279629312,"DeviceStatsMonitor.on_train_batch_start/max_split_size":-1,"DeviceStatsMonitor.on_train_batch_start/inactive_split.large_pool.freed":25489893,"DeviceStatsMonitor.on_train_batch_start/inactive_split.all.allocated":50436148,"DeviceStatsMonitor.on_train_batch_start/active.large_pool.freed":34810385,"DeviceStatsMonitor.on_train_batch_end/active.small_pool.freed":39535154,"DeviceStatsMonitor.on_train_batch_start/active.large_pool.current":172,"DeviceStatsMonitor.on_train_batch_end/requested_bytes.all.allocated":736603362688356,"DeviceStatsMonitor.on_train_batch_start/active.all.allocated":74344939,"DeviceStatsMonitor.on_train_batch_end/inactive_split_bytes.all.allocated":178978746231808,"DeviceStatsMonitor.on_train_batch_start/segment.large_pool.peak":242,"DeviceStatsMonitor.on_train_batch_end/active.large_pool.freed":34811494,"DeviceStatsMonitor.on_train_batch_end/allocated_bytes.large_pool.current":575078400,"DeviceStatsMonitor.on_train_batch_end/inactive_split_bytes.large_pool.current":85524480,"DeviceStatsMonitor.on_train_batch_start/active_bytes.large_pool.current":573702144,"DeviceStatsMonitor.on_train_batch_start/inactive_split.all.peak":364,"DeviceStatsMonitor.on_train_batch_end/inactive_split.large_pool.allocated":25490704,"DeviceStatsMonitor.on_train_batch_end/active.all.peak":2144,"DeviceStatsMonitor.on_train_batch_start/allocated_bytes.large_pool.allocated":737116080705536,"DeviceStatsMonitor.on_train_batch_start/allocation.small_pool.allocated":39534382,"DeviceStatsMonitor.on_train_batch_end/num_sync_all_streams":3,"DeviceStatsMonitor.on_train_batch_end/active_bytes.large_pool.allocated":737139668348928,"DeviceStatsMonitor.on_train_batch_end/reserved_bytes.small_pool.peak":130023424,"DeviceStatsMonitor.on_train_batch_start/active.small_pool.current":491,"DeviceStatsMonitor.on_train_batch_start/segment.small_pool.current":62,"DeviceStatsMonitor.on_train_batch_start/allocated_bytes.all.current":611811840,"DeviceStatsMonitor.on_train_batch_start/allocated_bytes.small_pool.allocated":3533154973184,"DeviceStatsMonitor.on_train_batch_end/allocation.small_pool.peak":1495,"DeviceStatsMonitor.on_train_batch_end/num_alloc_retries":0,"DeviceStatsMonitor.on_train_batch_start/inactive_split_bytes.large_pool.current":120455168,"DeviceStatsMonitor.on_train_batch_end/requested_bytes.all.peak":7529349644,"DeviceStatsMonitor.on_train_batch_end/allocation.all.allocated":74347314,"DeviceStatsMonitor.on_train_batch_start/allocation.small_pool.current":491,"DeviceStatsMonitor.on_train_batch_start/allocation.all.freed":74344276,"DeviceStatsMonitor.on_train_batch_end/oversize_segments.peak":0,"DeviceStatsMonitor.on_train_batch_end/allocation.large_pool.peak":659,"DeviceStatsMonitor.on_train_batch_start/requested_bytes.small_pool.freed":3527383660316,"DeviceStatsMonitor.on_train_batch_start/allocated_bytes.all.peak":7671335936,"profiling/batch_time":0.41266071796417236,"DeviceStatsMonitor.on_train_batch_start/reserved_bytes.all.current":7964983296,"DeviceStatsMonitor.on_train_batch_start/inactive_split.small_pool.peak":242,"DeviceStatsMonitor.on_train_batch_start/allocated_bytes.large_pool.current":573702144,"DeviceStatsMonitor.on_train_batch_end/inactive_split_bytes.all.freed":178978644292608,"DeviceStatsMonitor.on_train_batch_end/inactive_split.large_pool.peak":155,"DeviceStatsMonitor.on_train_batch_start/reserved_bytes.large_pool.freed":0,"DeviceStatsMonitor.on_train_batch_start/allocation.large_pool.peak":659,"DeviceStatsMonitor.on_train_batch_end/active_bytes.large_pool.current":575078400,"DeviceStatsMonitor.on_train_batch_end/requested_bytes.small_pool.peak":120280588,"DeviceStatsMonitor.on_train_batch_start/active_bytes.small_pool.freed":3533116863488,"DeviceStatsMonitor.on_train_batch_start/allocation.small_pool.freed":39533891,"DeviceStatsMonitor.on_train_batch_start/segment.large_pool.allocated":242,"DeviceStatsMonitor.on_train_batch_end/reserved_bytes.large_pool.peak":7834959872,"DeviceStatsMonitor.on_train_batch_end/reserved_bytes.large_pool.allocated":7834959872,"DeviceStatsMonitor.on_train_batch_start/active.all.peak":2144,"DeviceStatsMonitor.on_train_batch_start/oversize_segments.peak":0,"DeviceStatsMonitor.on_train_batch_start/active.large_pool.allocated":34810557,"DeviceStatsMonitor.on_train_batch_start/requested_bytes.all.peak":7529349644,"DeviceStatsMonitor.on_train_batch_end/inactive_split.all.allocated":50437685}
ssl_dino/vit16s/wandb/run-20250815_160317-46zcck3z/logs/debug-core.log ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2025-08-15T16:03:18.150462469+07:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpzl79dbgx/port-227131.txt","pid":227131,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false}
2
+ {"time":"2025-08-15T16:03:18.153213283+07:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-227131-227655-895285156/socket","Net":"unix"}}
3
+ {"time":"2025-08-15T16:03:18.15356944+07:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":227131}
4
+ {"time":"2025-08-15T16:03:18.241022939+07:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"}
5
+ {"time":"2025-08-15T16:03:18.253794131+07:00","level":"INFO","msg":"handleInformInit: received","streamId":"46zcck3z","id":"1(@)"}
6
+ {"time":"2025-08-15T16:03:18.721877272+07:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"46zcck3z","id":"1(@)"}
7
+ {"time":"2025-08-15T19:47:38.241310912+07:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"}
8
+ {"time":"2025-08-15T19:47:38.245673407+07:00","level":"INFO","msg":"connection: closing","id":"1(@)"}
9
+ {"time":"2025-08-15T19:47:38.245768813+07:00","level":"INFO","msg":"server is shutting down"}
10
+ {"time":"2025-08-15T19:47:38.245744481+07:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"}
11
+ {"time":"2025-08-15T19:47:38.248660832+07:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-227131-227655-895285156/socket","Net":"unix"}}
12
+ {"time":"2025-08-15T19:47:40.504523642+07:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"}
13
+ {"time":"2025-08-15T19:47:40.504569594+07:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"}
14
+ {"time":"2025-08-15T19:47:40.504594206+07:00","level":"INFO","msg":"server is closed"}
ssl_dino/vit16s/wandb/run-20250815_160317-46zcck3z/logs/debug-internal.log ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2025-08-15T16:03:18.253890142+07:00","level":"INFO","msg":"stream: starting","core version":"0.21.1"}
2
+ {"time":"2025-08-15T16:03:18.721836531+07:00","level":"INFO","msg":"stream: created new stream","id":"46zcck3z"}
3
+ {"time":"2025-08-15T16:03:18.721871902+07:00","level":"INFO","msg":"stream: started","id":"46zcck3z"}
4
+ {"time":"2025-08-15T16:03:18.721888752+07:00","level":"INFO","msg":"sender: started","stream_id":"46zcck3z"}
5
+ {"time":"2025-08-15T16:03:18.721889062+07:00","level":"INFO","msg":"writer: started","stream_id":"46zcck3z"}
6
+ {"time":"2025-08-15T16:03:18.722182108+07:00","level":"INFO","msg":"handler: started","stream_id":"46zcck3z"}
7
+ {"time":"2025-08-15T19:47:38.25051031+07:00","level":"INFO","msg":"stream: closing","id":"46zcck3z"}
8
+ {"time":"2025-08-15T19:47:39.932846317+07:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
9
+ {"time":"2025-08-15T19:47:40.373460135+07:00","level":"INFO","msg":"handler: closed","stream_id":"46zcck3z"}
10
+ {"time":"2025-08-15T19:47:40.381105441+07:00","level":"INFO","msg":"sender: closed","stream_id":"46zcck3z"}
11
+ {"time":"2025-08-15T19:47:40.381156674+07:00","level":"INFO","msg":"stream: closed","id":"46zcck3z"}
ssl_dino/vit16s/wandb/run-20250815_160317-46zcck3z/logs/debug.log ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2025-08-15 16:03:17,830 INFO MainThread:227131 [wandb_setup.py:_flush():80] Current SDK version is 0.21.1
2
+ 2025-08-15 16:03:17,830 INFO MainThread:227131 [wandb_setup.py:_flush():80] Configure stats pid to 227131
3
+ 2025-08-15 16:03:17,830 INFO MainThread:227131 [wandb_setup.py:_flush():80] Loading settings from /home/duckq1u/.config/wandb/settings
4
+ 2025-08-15 16:03:17,830 INFO MainThread:227131 [wandb_setup.py:_flush():80] Loading settings from /home/duckq1u/Documents/workspace/cuong/ent-labotary/wandb/settings
5
+ 2025-08-15 16:03:17,830 INFO MainThread:227131 [wandb_setup.py:_flush():80] Loading settings from environment variables
6
+ 2025-08-15 16:03:17,831 INFO MainThread:227131 [wandb_init.py:setup_run_log_directory():703] Logging user logs to /home/duckq1u/Documents/workspace/cuong/ent-labotary/outputs/ssl_dino/vit16s/wandb/run-20250815_160317-46zcck3z/logs/debug.log
7
+ 2025-08-15 16:03:17,831 INFO MainThread:227131 [wandb_init.py:setup_run_log_directory():704] Logging internal logs to /home/duckq1u/Documents/workspace/cuong/ent-labotary/outputs/ssl_dino/vit16s/wandb/run-20250815_160317-46zcck3z/logs/debug-internal.log
8
+ 2025-08-15 16:03:17,831 INFO MainThread:227131 [wandb_init.py:init():830] calling init triggers
9
+ 2025-08-15 16:03:17,831 INFO MainThread:227131 [wandb_init.py:init():835] wandb.init called with sweep_config: {}
10
+ config: {'_wandb': {}}
11
+ 2025-08-15 16:03:17,831 INFO MainThread:227131 [wandb_init.py:init():871] starting backend
12
+ 2025-08-15 16:03:18,241 INFO MainThread:227131 [wandb_init.py:init():874] sending inform_init request
13
+ 2025-08-15 16:03:18,247 INFO MainThread:227131 [wandb_init.py:init():882] backend started and connected
14
+ 2025-08-15 16:03:18,248 INFO MainThread:227131 [wandb_init.py:init():953] updated telemetry
15
+ 2025-08-15 16:03:18,255 INFO MainThread:227131 [wandb_init.py:init():977] communicating run to backend with 90.0 second timeout
16
+ 2025-08-15 16:03:19,421 INFO MainThread:227131 [wandb_init.py:init():1029] starting run threads in backend
17
+ 2025-08-15 16:03:19,692 INFO MainThread:227131 [wandb_run.py:_console_start():2494] atexit reg
18
+ 2025-08-15 16:03:19,692 INFO MainThread:227131 [wandb_run.py:_redirect():2342] redirect: wrap_raw
19
+ 2025-08-15 16:03:19,692 INFO MainThread:227131 [wandb_run.py:_redirect():2411] Wrapping output streams.
20
+ 2025-08-15 16:03:19,692 INFO MainThread:227131 [wandb_run.py:_redirect():2434] Redirects installed.
21
+ 2025-08-15 16:03:19,695 INFO MainThread:227131 [wandb_init.py:init():1075] run started, returning control to user process
22
+ 2025-08-15 16:03:19,696 INFO MainThread:227131 [wandb_run.py:_config_callback():1380] config_cb None None {'out': 'outputs/ssl_dino/vit16s', 'data': 'data/kyucapsule', 'model': 'VisionTransformer', 'method': 'dino', 'method_args': {'hidden_dim': 2048, 'bottleneck_dim': 256, 'output_dim': 1024, 'student_freeze_last_layer_epochs': 1, 'batch_norm': False, 'norm_last_layer': True, 'teacher_temp': 0.02, 'warmup_teacher_temp': 0.02, 'warmup_teacher_temp_epochs': 30, 'student_temp': 0.1, 'center_momentum': 0.9, 'momentum_start': 0.99, 'momentum_end': 1.0, 'weight_decay_start': 0.0001, 'weight_decay_end': 0.0001}, 'embed_dim': None, 'epochs': 300, 'batch_size': 32, 'num_workers': 5, 'devices': 1, 'num_nodes': 1, 'resume': True, 'checkpoint': None, 'overwrite': True, 'accelerator': 'CUDAAccelerator', 'strategy': 'SingleDeviceStrategy', 'precision': '32-true', 'seed': 0, 'loggers': {'jsonl': {'flush_logs_every_n_steps': 100}, 'tensorboard': {'name': '', 'version': '', 'log_graph': False, 'default_hp_metric': True, 'prefix': '', 'sub_dir': None}, 'wandb': {'name': None, 'version': None, 'offline': False, 'anonymous': None, 'project': 'ent-endoscopy-ssl', 'log_model': False, 'prefix': '', 'checkpoint_name': None}}, 'callbacks': {'learning_rate_monitor': {}, 'device_stats_monitor': {}, 'early_stopping': {'monitor': 'train_loss', 'patience': 1000000000000, 'check_finite': True}, 'model_checkpoint': {'save_last': True, 'enable_version_counter': False, 'save_top_k': 1, 'every_n_epochs': None}}, 'optim': 'sgd', 'optim_args': {'lr': 0.03, 'momentum': 0.9, 'weight_decay': 0.0001}, 'transform_args': {'image_size': [224, 224], 'random_resize': {'min_scale': 0.14, 'max_scale': 1.0}, 'random_flip': {'horizontal_prob': 0.5, 'vertical_prob': 0.0}, 'random_rotation': None, 'color_jitter': {'prob': 0.8, 'strength': 0.5, 'brightness': 0.8, 'contrast': 0.8, 'saturation': 0.4, 'hue': 0.2}, 'random_gray_scale': 0.2, 'normalize': {'mean': [0.485, 0.456, 0.406], 'std': [0.229, 0.224, 0.225]}, 'gaussian_blur': {'prob': 1.0, 'sigmas': [0.1, 2.0], 'blur_limit': 0}, 'solarize': None, 'global_view_1': {'gaussian_blur': {'prob': 0.1, 'sigmas': [0.1, 2.0], 'blur_limit': 0}, 'solarize': {'prob': 0.2, 'threshold': 0.5}}, 'local_view': {'num_views': 6, 'view_size': [96, 96], 'random_resize': {'min_scale': 0.05, 'max_scale': 0.14}, 'gaussian_blur': {'prob': 0.5, 'sigmas': [0.1, 2.0], 'blur_limit': 0}}}, 'loader_args': None, 'trainer_args': None, 'model_args': None}
23
+ 2025-08-15 19:47:38,211 INFO MsgRouterThr:227131 [mailbox.py:close():129] [no run ID] Closing mailbox, abandoning 1 handles.
ssl_dino/vit16s/wandb/run-20250815_160317-46zcck3z/run-46zcck3z.wandb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e5389f9a04bb03eedceb69c60d3fdcc2271b76c562b37b483bd540b3971171d0
3
+ size 29846296