diff --git "a/Experiments/Logs/RESNET50_3HLL_CIFAR.out" "b/Experiments/Logs/RESNET50_3HLL_CIFAR.out" new file mode 100644--- /dev/null +++ "b/Experiments/Logs/RESNET50_3HLL_CIFAR.out" @@ -0,0 +1,624 @@ + File "/root/test_cifar/train.py", line 143 + print(f"Saved Model as {CIFAR_{epoch+1}_end_hll.pt") + ^ +SyntaxError: f-string: expecting '}' +wandb: Currently logged in as: gjyotin305 (gjyotin). Use `wandb login --relogin` to force relogin +wandb: Currently logged in as: gjyotin305. Use `wandb login --relogin` to force relogin +wandb: Tracking run with wandb version 0.16.6 +wandb: Run data is saved locally in /root/test_cifar/wandb/run-20240418_050322-9t0aeiev +wandb: Run `wandb offline` to turn off syncing. +wandb: Syncing run feasible-vortex-18 +wandb: ⭐️ View project at https://wandb.ai/gjyotin305/CIFAR +wandb: 🚀 View run at https://wandb.ai/gjyotin305/CIFAR/runs/9t0aeiev +Files already downloaded and verified +Files already downloaded and verified +/root/test_cifar/env_ml_dl/lib/python3.10/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead. + warnings.warn( +/root/test_cifar/env_ml_dl/lib/python3.10/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=ResNet50_Weights.IMAGENET1K_V1`. You can also use `weights=ResNet50_Weights.DEFAULT` to get the most up-to-date weights. + warnings.warn(msg) +torch.Size([64, 3, 224, 224]) +Training Epoch: 1 + 0%| | 0/782 [00:00 Epoch Number : 1 | Training Loss : 2.228 | Validation Loss : 2.0095 | Validation Accuracy : 48.73% +Training Epoch: 2 + 0%| | 0/782 [00:00 Epoch Number : 2 | Training Loss : 1.4728 | Validation Loss : 0.861 | Validation Accuracy : 77.95% +Training Epoch: 3 + 0%| | 0/782 [00:00 Epoch Number : 3 | Training Loss : 0.7157 | Validation Loss : 0.3831 | Validation Accuracy : 88.34% +Training Epoch: 4 + 0%| | 0/782 [00:00 Epoch Number : 4 | Training Loss : 0.4285 | Validation Loss : 0.254 | Validation Accuracy : 91.75% +Training Epoch: 5 + 0%| | 0/782 [00:00 + accuracy_list_normalv5, train_cost_listv5, val_cost_listv5, model_to_save=train_model(model=model_check, + File "/root/test_cifar/train.py", line 142, in train_model + torch.save(model.state_dict(), f"./logs_model/CIFAR_{epoch+1}_end_hll.pt") + File "/root/test_cifar/env_ml_dl/lib/python3.10/site-packages/torch/serialization.py", line 628, in save + with _open_zipfile_writer(f) as opened_zipfile: + File "/root/test_cifar/env_ml_dl/lib/python3.10/site-packages/torch/serialization.py", line 502, in _open_zipfile_writer + return container(name_or_buffer) + File "/root/test_cifar/env_ml_dl/lib/python3.10/site-packages/torch/serialization.py", line 473, in __init__ + super().__init__(torch._C.PyTorchFileWriter(self.name)) +RuntimeError: Parent directory ./logs_model does not exist. +wandb: - 0.007 MB of 0.007 MB uploaded wandb: \ 0.007 MB of 0.011 MB uploaded wandb: | 0.012 MB of 0.012 MB uploaded wandb: / 0.012 MB of 0.012 MB uploaded wandb: +wandb: Run history: +wandb: train_loss ███████▇▇▆▆▆▅▄▅▄▃▄▄▃▂▂▂▂▂▂▂▁▁▂▂▂▂▁▁▂▂▂▂▂ +wandb: val_acc ▁▆▇██ +wandb: val_loss ████████▄▄▃▃▄▄▄▄▃▂▂▂▂▂▂▂▁▁▂▁▁▂▁▂▂▂▁▁▂▁▁▁ +wandb: +wandb: Run summary: +wandb: train_loss 0.44522 +wandb: val_acc 0.9332 +wandb: val_loss 0.09327 +wandb: +wandb: 🚀 View run feasible-vortex-18 at: https://wandb.ai/gjyotin305/CIFAR/runs/9t0aeiev +wandb: ⭐️ View project at: https://wandb.ai/gjyotin305/CIFAR +wandb: Synced 5 W&B file(s), 0 media file(s), 0 artifact file(s) and 0 other file(s) +wandb: Find logs at: ./wandb/run-20240418_050322-9t0aeiev/logs +wandb: Currently logged in as: gjyotin305 (gjyotin). Use `wandb login --relogin` to force relogin +wandb: Currently logged in as: gjyotin305. Use `wandb login --relogin` to force relogin +wandb: Tracking run with wandb version 0.16.6 +wandb: Run data is saved locally in /root/test_cifar/wandb/run-20240418_051545-08ne5f0a +wandb: Run `wandb offline` to turn off syncing. +wandb: Syncing run atomic-valley-19 +wandb: ⭐️ View project at https://wandb.ai/gjyotin305/CIFAR +wandb: 🚀 View run at https://wandb.ai/gjyotin305/CIFAR/runs/08ne5f0a +Files already downloaded and verified +Files already downloaded and verified +/root/test_cifar/env_ml_dl/lib/python3.10/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead. + warnings.warn( +/root/test_cifar/env_ml_dl/lib/python3.10/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=ResNet50_Weights.IMAGENET1K_V1`. You can also use `weights=ResNet50_Weights.DEFAULT` to get the most up-to-date weights. + warnings.warn(msg) +torch.Size([64, 3, 224, 224]) +Training Epoch: 1 + 0%| | 0/782 [00:00 Epoch Number : 1 | Training Loss : 2.1929 | Validation Loss : 1.9084 | Validation Accuracy : 42.66% +Training Epoch: 2 + 0%| | 0/782 [00:00 Epoch Number : 2 | Training Loss : 1.4313 | Validation Loss : 0.8592 | Validation Accuracy : 77.91% +Training Epoch: 3 + 0%| | 0/782 [00:00 Epoch Number : 3 | Training Loss : 0.6905 | Validation Loss : 0.3537 | Validation Accuracy : 89.08% +Training Epoch: 4 + 0%| | 0/782 [00:00 Epoch Number : 4 | Training Loss : 0.409 | Validation Loss : 0.2445 | Validation Accuracy : 92.23% +Training Epoch: 5 + 0%| | 0/782 [00:00 Epoch Number : 5 | Training Loss : 0.3187 | Validation Loss : 0.2026 | Validation Accuracy : 93.39% +Training Epoch: 6 + 0%| | 0/782 [00:00 Epoch Number : 6 | Training Loss : 0.2652 | Validation Loss : 0.1752 | Validation Accuracy : 94.18% +Training Epoch: 7 + 0%| | 0/782 [00:00 Epoch Number : 7 | Training Loss : 0.2365 | Validation Loss : 0.1591 | Validation Accuracy : 94.61% +Training Epoch: 8 + 0%| | 0/782 [00:00 Epoch Number : 8 | Training Loss : 0.2119 | Validation Loss : 0.1574 | Validation Accuracy : 94.65% +Training Epoch: 9 + 0%| | 0/782 [00:00 Epoch Number : 9 | Training Loss : 0.1947 | Validation Loss : 0.1444 | Validation Accuracy : 95.08% +Training Epoch: 10 + 0%| | 0/782 [00:00 Epoch Number : 10 | Training Loss : 0.1801 | Validation Loss : 0.1376 | Validation Accuracy : 95.34% +Training Epoch: 11 + 0%| | 0/782 [00:00 Epoch Number : 11 | Training Loss : 0.168 | Validation Loss : 0.136 | Validation Accuracy : 95.31% +Training Epoch: 12 + 0%| | 0/782 [00:00 Epoch Number : 12 | Training Loss : 0.1564 | Validation Loss : 0.1333 | Validation Accuracy : 95.47% +Training Epoch: 13 + 0%| | 0/782 [00:00 Epoch Number : 13 | Training Loss : 0.1467 | Validation Loss : 0.1262 | Validation Accuracy : 95.63% +Training Epoch: 14 + 0%| | 0/782 [00:00 Epoch Number : 14 | Training Loss : 0.1386 | Validation Loss : 0.1238 | Validation Accuracy : 95.71% +Training Epoch: 15 + 0%| | 0/782 [00:00 Epoch Number : 15 | Training Loss : 0.127 | Validation Loss : 0.1211 | Validation Accuracy : 95.78% +Training Epoch: 16 + 0%| | 0/782 [00:00 Epoch Number : 16 | Training Loss : 0.1234 | Validation Loss : 0.1193 | Validation Accuracy : 95.98% +Training Epoch: 17 + 0%| | 0/782 [00:00 Epoch Number : 17 | Training Loss : 0.1167 | Validation Loss : 0.1156 | Validation Accuracy : 96.05% +Training Epoch: 18 + 0%| | 0/782 [00:00 Epoch Number : 18 | Training Loss : 0.1092 | Validation Loss : 0.1163 | Validation Accuracy : 96.05% +Training Epoch: 19 + 0%| | 0/782 [00:00 Epoch Number : 19 | Training Loss : 0.1039 | Validation Loss : 0.1147 | Validation Accuracy : 96.14% +Training Epoch: 20 + 0%| | 0/782 [00:00 Epoch Number : 20 | Training Loss : 0.0985 | Validation Loss : 0.1175 | Validation Accuracy : 96.12% +Training Epoch: 21 + 0%| | 0/782 [00:00 Epoch Number : 21 | Training Loss : 0.0947 | Validation Loss : 0.1182 | Validation Accuracy : 96.0% +Training Epoch: 22 + 0%| | 0/782 [00:00 Epoch Number : 22 | Training Loss : 0.0923 | Validation Loss : 0.1202 | Validation Accuracy : 96.06% +Training Epoch: 23 + 0%| | 0/782 [00:00 Epoch Number : 23 | Training Loss : 0.0841 | Validation Loss : 0.122 | Validation Accuracy : 96.03% +Training Epoch: 24 + 0%| | 0/782 [00:00 Epoch Number : 24 | Training Loss : 0.0808 | Validation Loss : 0.1121 | Validation Accuracy : 96.41% +Training Epoch: 25 + 0%| | 0/782 [00:00 Epoch Number : 25 | Training Loss : 0.0792 | Validation Loss : 0.1144 | Validation Accuracy : 96.4% +Training Epoch: 26 + 0%| | 0/782 [00:00 Epoch Number : 26 | Training Loss : 0.0766 | Validation Loss : 0.1179 | Validation Accuracy : 96.12% +Training Epoch: 27 + 0%| | 0/782 [00:00 Epoch Number : 27 | Training Loss : 0.0757 | Validation Loss : 0.1139 | Validation Accuracy : 96.38% +Training Epoch: 28 + 0%| | 0/782 [00:00 Epoch Number : 28 | Training Loss : 0.0671 | Validation Loss : 0.1123 | Validation Accuracy : 96.36% +Training Epoch: 29 + 0%| | 0/782 [00:00 Epoch Number : 29 | Training Loss : 0.0645 | Validation Loss : 0.1122 | Validation Accuracy : 96.45% +Training Epoch: 30 + 0%| | 0/782 [00:00 Epoch Number : 30 | Training Loss : 0.0666 | Validation Loss : 0.111 | Validation Accuracy : 96.6% +Training Epoch: 31 + 0%| | 0/782 [00:00 Epoch Number : 31 | Training Loss : 0.0631 | Validation Loss : 0.1077 | Validation Accuracy : 96.57% +Training Epoch: 32 + 0%| | 0/782 [00:00 Epoch Number : 32 | Training Loss : 0.0595 | Validation Loss : 0.1104 | Validation Accuracy : 96.59% +Training Epoch: 33 + 0%| | 0/782 [00:00 Epoch Number : 33 | Training Loss : 0.0585 | Validation Loss : 0.1098 | Validation Accuracy : 96.44% +Training Epoch: 34 + 0%| | 0/782 [00:00 Epoch Number : 34 | Training Loss : 0.055 | Validation Loss : 0.1134 | Validation Accuracy : 96.5% +Training Epoch: 35 + 0%| | 0/782 [00:00 Epoch Number : 35 | Training Loss : 0.0539 | Validation Loss : 0.1116 | Validation Accuracy : 96.49% +Training Epoch: 36 + 0%| | 0/782 [00:00 Epoch Number : 36 | Training Loss : 0.0529 | Validation Loss : 0.1098 | Validation Accuracy : 96.6% +Training Epoch: 37 + 0%| | 0/782 [00:00 Epoch Number : 37 | Training Loss : 0.0504 | Validation Loss : 0.1119 | Validation Accuracy : 96.53% +Training Epoch: 38 + 0%| | 0/782 [00:00 Epoch Number : 38 | Training Loss : 0.0498 | Validation Loss : 0.1122 | Validation Accuracy : 96.6% +Training Epoch: 39 + 0%| | 0/782 [00:00 Epoch Number : 39 | Training Loss : 0.0468 | Validation Loss : 0.1106 | Validation Accuracy : 96.69% +Training Epoch: 40 + 0%| | 0/782 [00:00 Epoch Number : 40 | Training Loss : 0.0477 | Validation Loss : 0.115 | Validation Accuracy : 96.62% +Training Epoch: 41 + 0%| | 0/782 [00:00 Epoch Number : 41 | Training Loss : 0.0467 | Validation Loss : 0.111 | Validation Accuracy : 96.64% +Training Epoch: 42 + 0%| | 0/782 [00:00 Epoch Number : 42 | Training Loss : 0.0422 | Validation Loss : 0.113 | Validation Accuracy : 96.76% +Training Epoch: 43 + 0%| | 0/782 [00:00 Epoch Number : 43 | Training Loss : 0.0398 | Validation Loss : 0.1155 | Validation Accuracy : 96.53% +Training Epoch: 44 + 0%| | 0/782 [00:00 Epoch Number : 44 | Training Loss : 0.0385 | Validation Loss : 0.1115 | Validation Accuracy : 96.86% +Training Epoch: 45 + 0%| | 0/782 [00:00 Epoch Number : 45 | Training Loss : 0.0397 | Validation Loss : 0.1125 | Validation Accuracy : 96.89% +Training Epoch: 46 + 0%| | 0/782 [00:00 Epoch Number : 46 | Training Loss : 0.0358 | Validation Loss : 0.1097 | Validation Accuracy : 96.69% +Training Epoch: 47 + 0%| | 0/782 [00:00 Epoch Number : 47 | Training Loss : 0.0363 | Validation Loss : 0.1124 | Validation Accuracy : 96.68% +Training Epoch: 48 + 0%| | 0/782 [00:00 Epoch Number : 48 | Training Loss : 0.0367 | Validation Loss : 0.1113 | Validation Accuracy : 96.76% +Training Epoch: 49 + 0%| | 0/782 [00:00 Epoch Number : 49 | Training Loss : 0.0356 | Validation Loss : 0.1121 | Validation Accuracy : 96.77% +Training Epoch: 50 + 0%| | 0/782 [00:00 Epoch Number : 50 | Training Loss : 0.0333 | Validation Loss : 0.1139 | Validation Accuracy : 96.72% +Training Epoch: 51 + 0%| | 0/782 [00:00 Epoch Number : 51 | Training Loss : 0.0338 | Validation Loss : 0.1145 | Validation Accuracy : 96.59% +Training Epoch: 52 + 0%| | 0/782 [00:00 Epoch Number : 52 | Training Loss : 0.0336 | Validation Loss : 0.1138 | Validation Accuracy : 96.72% +Training Epoch: 53 + 0%| | 0/782 [00:00 Epoch Number : 53 | Training Loss : 0.0297 | Validation Loss : 0.1167 | Validation Accuracy : 96.65% +Training Epoch: 54 + 0%| | 0/782 [00:00 Epoch Number : 54 | Training Loss : 0.0304 | Validation Loss : 0.1185 | Validation Accuracy : 96.78% +Training Epoch: 55 + 0%| | 0/782 [00:00 Epoch Number : 55 | Training Loss : 0.0324 | Validation Loss : 0.1157 | Validation Accuracy : 96.67% +Training Epoch: 56 + 0%| | 0/782 [00:00 Epoch Number : 56 | Training Loss : 0.0301 | Validation Loss : 0.117 | Validation Accuracy : 96.8% +Training Epoch: 57 + 0%| | 0/782 [00:00 Epoch Number : 57 | Training Loss : 0.03 | Validation Loss : 0.1123 | Validation Accuracy : 96.87% +Training Epoch: 58 + 0%| | 0/782 [00:00 Epoch Number : 58 | Training Loss : 0.0298 | Validation Loss : 0.1173 | Validation Accuracy : 96.98% +Training Epoch: 59 + 0%| | 0/782 [00:00 Epoch Number : 59 | Training Loss : 0.0261 | Validation Loss : 0.1109 | Validation Accuracy : 96.86% +Training Epoch: 60 + 0%| | 0/782 [00:00 Epoch Number : 60 | Training Loss : 0.0267 | Validation Loss : 0.1149 | Validation Accuracy : 96.72% +Training Epoch: 61 + 0%| | 0/782 [00:00 Epoch Number : 61 | Training Loss : 0.0255 | Validation Loss : 0.1132 | Validation Accuracy : 96.88% +Training Epoch: 62 + 0%| | 0/782 [00:00 Epoch Number : 62 | Training Loss : 0.0267 | Validation Loss : 0.1155 | Validation Accuracy : 96.85% +Training Epoch: 63 + 0%| | 0/782 [00:00 Epoch Number : 63 | Training Loss : 0.0247 | Validation Loss : 0.1124 | Validation Accuracy : 96.94% +Training Epoch: 64 + 0%| | 0/782 [00:00 Epoch Number : 64 | Training Loss : 0.0246 | Validation Loss : 0.112 | Validation Accuracy : 96.99% +Training Epoch: 65 + 0%| | 0/782 [00:00 Epoch Number : 65 | Training Loss : 0.0262 | Validation Loss : 0.1114 | Validation Accuracy : 96.84% +Training Epoch: 66 + 0%| | 0/782 [00:00 Epoch Number : 66 | Training Loss : 0.0238 | Validation Loss : 0.1129 | Validation Accuracy : 97.09% +Training Epoch: 67 + 0%| | 0/782 [00:00 Epoch Number : 67 | Training Loss : 0.0222 | Validation Loss : 0.1118 | Validation Accuracy : 97.02% +Training Epoch: 68 + 0%| | 0/782 [00:00 Epoch Number : 68 | Training Loss : 0.0224 | Validation Loss : 0.1136 | Validation Accuracy : 96.99% +Training Epoch: 69 + 0%| | 0/782 [00:00 Epoch Number : 69 | Training Loss : 0.023 | Validation Loss : 0.1146 | Validation Accuracy : 96.98% +Training Epoch: 70 + 0%| | 0/782 [00:00 Epoch Number : 70 | Training Loss : 0.0222 | Validation Loss : 0.1162 | Validation Accuracy : 96.92% +Training Epoch: 71 + 0%| | 0/782 [00:00 Epoch Number : 71 | Training Loss : 0.0206 | Validation Loss : 0.1149 | Validation Accuracy : 97.01% +Training Epoch: 72 + 0%| | 0/782 [00:00 Epoch Number : 72 | Training Loss : 0.0219 | Validation Loss : 0.1183 | Validation Accuracy : 96.89% +Training Epoch: 73 + 0%| | 0/782 [00:00 Epoch Number : 73 | Training Loss : 0.0204 | Validation Loss : 0.1141 | Validation Accuracy : 96.95% +Training Epoch: 74 + 0%| | 0/782 [00:00 Epoch Number : 74 | Training Loss : 0.0209 | Validation Loss : 0.1155 | Validation Accuracy : 97.02% +Training Epoch: 75 + 0%| | 0/782 [00:00 Epoch Number : 75 | Training Loss : 0.0199 | Validation Loss : 0.1147 | Validation Accuracy : 96.99% +Training Epoch: 76 + 0%| | 0/782 [00:00 Epoch Number : 76 | Training Loss : 0.021 | Validation Loss : 0.1119 | Validation Accuracy : 97.06% +Training Epoch: 77 + 0%| | 0/782 [00:00 Epoch Number : 77 | Training Loss : 0.0204 | Validation Loss : 0.115 | Validation Accuracy : 96.99% +Training Epoch: 78 + 0%| | 0/782 [00:00 Epoch Number : 78 | Training Loss : 0.0203 | Validation Loss : 0.1132 | Validation Accuracy : 97.11% +Training Epoch: 79 + 0%| | 0/782 [00:00 Epoch Number : 79 | Training Loss : 0.0203 | Validation Loss : 0.1138 | Validation Accuracy : 97.0% +Training Epoch: 80 + 0%| | 0/782 [00:00 Epoch Number : 80 | Training Loss : 0.0197 | Validation Loss : 0.1181 | Validation Accuracy : 97.12% +Training Epoch: 81 + 0%| | 0/782 [00:00 Epoch Number : 81 | Training Loss : 0.02 | Validation Loss : 0.1161 | Validation Accuracy : 96.99% +Training Epoch: 82 + 0%| | 0/782 [00:00 Epoch Number : 82 | Training Loss : 0.0186 | Validation Loss : 0.1208 | Validation Accuracy : 96.91% +Training Epoch: 83 + 0%| | 0/782 [00:00 Epoch Number : 83 | Training Loss : 0.0169 | Validation Loss : 0.1137 | Validation Accuracy : 97.17% +Training Epoch: 84 + 0%| | 0/782 [00:00 Epoch Number : 84 | Training Loss : 0.0169 | Validation Loss : 0.1168 | Validation Accuracy : 97.0% +Training Epoch: 85 + 0%| | 0/782 [00:00 Epoch Number : 85 | Training Loss : 0.0183 | Validation Loss : 0.1166 | Validation Accuracy : 97.09% +Training Epoch: 86 + 0%| | 0/782 [00:00 Epoch Number : 86 | Training Loss : 0.0179 | Validation Loss : 0.1209 | Validation Accuracy : 96.8% +Training Epoch: 87 + 0%| | 0/782 [00:00 Epoch Number : 87 | Training Loss : 0.017 | Validation Loss : 0.1168 | Validation Accuracy : 97.05% +Training Epoch: 88 + 0%| | 0/782 [00:00 Epoch Number : 88 | Training Loss : 0.0179 | Validation Loss : 0.1188 | Validation Accuracy : 97.02% +Training Epoch: 89 + 0%| | 0/782 [00:00 Epoch Number : 89 | Training Loss : 0.0159 | Validation Loss : 0.1196 | Validation Accuracy : 97.02% +Training Epoch: 90 + 0%| | 0/782 [00:00 Epoch Number : 90 | Training Loss : 0.0172 | Validation Loss : 0.1155 | Validation Accuracy : 97.1% +Training Epoch: 91 + 0%| | 0/782 [00:00 Epoch Number : 91 | Training Loss : 0.0179 | Validation Loss : 0.1171 | Validation Accuracy : 97.09% +Training Epoch: 92 + 0%| | 0/782 [00:00 Epoch Number : 92 | Training Loss : 0.0163 | Validation Loss : 0.1162 | Validation Accuracy : 96.97% +Training Epoch: 93 + 0%| | 0/782 [00:00 Epoch Number : 93 | Training Loss : 0.0149 | Validation Loss : 0.1156 | Validation Accuracy : 97.08% +Training Epoch: 94 + 0%| | 0/782 [00:00 Epoch Number : 94 | Training Loss : 0.0138 | Validation Loss : 0.1182 | Validation Accuracy : 97.01% +Training Epoch: 95 + 0%| | 0/782 [00:00 Epoch Number : 95 | Training Loss : 0.0158 | Validation Loss : 0.1163 | Validation Accuracy : 97.16% +Training Epoch: 96 + 0%| | 0/782 [00:00 Epoch Number : 96 | Training Loss : 0.0148 | Validation Loss : 0.1169 | Validation Accuracy : 96.98% +Training Epoch: 97 + 0%| | 0/782 [00:00 Epoch Number : 97 | Training Loss : 0.0148 | Validation Loss : 0.1148 | Validation Accuracy : 97.04% +Training Epoch: 98 + 0%| | 0/782 [00:00 Epoch Number : 98 | Training Loss : 0.0147 | Validation Loss : 0.1183 | Validation Accuracy : 96.95% +Training Epoch: 99 + 0%| | 0/782 [00:00 Epoch Number : 99 | Training Loss : 0.0145 | Validation Loss : 0.1212 | Validation Accuracy : 96.97% +Training Epoch: 100 + 0%| | 0/782 [00:00 Epoch Number : 100 | Training Loss : 0.0156 | Validation Loss : 0.1203 | Validation Accuracy : 97.12% +Traceback (most recent call last): + File "/root/test_cifar/train.py", line 162, in + model.eval() +AttributeError: 'collections.OrderedDict' object has no attribute 'eval' +wandb: - 0.007 MB of 0.007 MB uploaded wandb: \ 0.007 MB of 0.243 MB uploaded wandb: | 0.243 MB of 0.243 MB uploaded wandb: / 0.243 MB of 0.243 MB uploaded wandb: - 0.243 MB of 0.243 MB uploaded wandb: +wandb: Run history: +wandb: train_loss █▃▂▃▂▂▁▁▁▁▁▂▁▁▁▁▁▁▂▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁ +wandb: val_acc ▁▇██████████████████████████████████████ +wandb: val_loss █▁▂▁▁▁▂▁▂▁▁▂▂▂▂▁▁▂▂▁▂▁▁▂▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁ +wandb: +wandb: Run summary: +wandb: train_loss 0.06611 +wandb: val_acc 0.9712 +wandb: val_loss 0.5941 +wandb: +wandb: 🚀 View run atomic-valley-19 at: https://wandb.ai/gjyotin305/CIFAR/runs/08ne5f0a +wandb: ⭐️ View project at: https://wandb.ai/gjyotin305/CIFAR +wandb: Synced 5 W&B file(s), 0 media file(s), 0 artifact file(s) and 0 other file(s) +wandb: Find logs at: ./wandb/run-20240418_051545-08ne5f0a/logs