Create REPORT3_Modifications_1+2 _PiT_Training_Results_in_Colab
Browse files
REPORT3_Modifications_1+2 _PiT_Training_Results_in_Colab
ADDED
|
@@ -0,0 +1,80 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
I modified the vanilla PiT model in a first undisclosed manner plus a second dislosed manner, essentially doubling the parameter count, and the training results greatly improved, and the two-ways modified model was consistently ahead of the unmodified Vanilla PiT model.
|
| 2 |
+
My two-ways modified PiT model consistently exceeded the vanilla PiT Val Accuracy until epoch 11 (exceeding the final/highest 94.75% Val Accuracy of the vanilla PiT model at Epoch 10).
|
| 3 |
+
By Epoch 17 and Epoch 19, the two-way modified PiT model had the highest Val Accuracy 95.90%, 95.95% (already surpassing both the Vanilla and the Modification1 PiT models)
|
| 4 |
+
By Epoch 20, the two-ways modified PiT model exceeded 96% Val Accuracy,
|
| 5 |
+
and Val Accuracy continued to increase up to the highest Val Accuracy of 96.50% before the training was hardcoded terminated Epoch 25.
|
| 6 |
+
|
| 7 |
+
--- Configuration V3.0 (Encoder-Decoder Architecture) ---
|
| 8 |
+
train_file: /content/sample_data/mnist_train_small.csv
|
| 9 |
+
test_file: /content/sample_data/mnist_test.csv
|
| 10 |
+
image_size: 28
|
| 11 |
+
num_classes: 10
|
| 12 |
+
embed_dim: XXX
|
| 13 |
+
num_layers: X
|
| 14 |
+
num_heads: X
|
| 15 |
+
mlp_dim: XXX
|
| 16 |
+
dropout: 0.1
|
| 17 |
+
batch_size: 128
|
| 18 |
+
epochs: 25
|
| 19 |
+
learning_rate: 0.0001
|
| 20 |
+
XXXXXX
|
| 21 |
+
device: cuda
|
| 22 |
+
image_height: 28
|
| 23 |
+
image_width: 28
|
| 24 |
+
sequence_length: 784
|
| 25 |
+
------------------------------------------------------
|
| 26 |
+
|
| 27 |
+
Data loaded. Training on cuda.
|
| 28 |
+
Training samples: 17999
|
| 29 |
+
Validation samples: 2000
|
| 30 |
+
Test samples: 9999
|
| 31 |
+
|
| 32 |
+
Model V3.0 initialized with 2,7XX,XXX trainable parameters.
|
| 33 |
+
|
| 34 |
+
--- Starting Training (V3.0 Encoder-Decoder Model) ---
|
| 35 |
+
Epoch 01/25 | Train Loss: 2.2017 | Val Loss: 2.0235 | Val Acc: 24.65%
|
| 36 |
+
-> New best validation accuracy! Saving model state.
|
| 37 |
+
Epoch 02/25 | Train Loss: 1.7963 | Val Loss: 1.2618 | Val Acc: 58.45%
|
| 38 |
+
-> New best validation accuracy! Saving model state.
|
| 39 |
+
Epoch 03/25 | Train Loss: 0.9849 | Val Loss: 0.7252 | Val Acc: 76.15%
|
| 40 |
+
-> New best validation accuracy! Saving model state.
|
| 41 |
+
Epoch 04/25 | Train Loss: 0.6703 | Val Loss: 0.4835 | Val Acc: 85.45%
|
| 42 |
+
-> New best validation accuracy! Saving model state.
|
| 43 |
+
Epoch 05/25 | Train Loss: 0.4813 | Val Loss: 0.3480 | Val Acc: 89.40%
|
| 44 |
+
-> New best validation accuracy! Saving model state.
|
| 45 |
+
Epoch 06/25 | Train Loss: 0.3694 | Val Loss: 0.2687 | Val Acc: 91.75%
|
| 46 |
+
-> New best validation accuracy! Saving model state.
|
| 47 |
+
Epoch 07/25 | Train Loss: 0.2967 | Val Loss: 0.2549 | Val Acc: 91.95%
|
| 48 |
+
-> New best validation accuracy! Saving model state.
|
| 49 |
+
Epoch 08/25 | Train Loss: 0.2568 | Val Loss: 0.2090 | Val Acc: 94.40%
|
| 50 |
+
-> New best validation accuracy! Saving model state.
|
| 51 |
+
Epoch 09/25 | Train Loss: 0.2338 | Val Loss: 0.1867 | Val Acc: 94.90%
|
| 52 |
+
-> New best validation accuracy! Saving model state.
|
| 53 |
+
Epoch 10/25 | Train Loss: 0.2137 | Val Loss: 0.1783 | Val Acc: 94.85%
|
| 54 |
+
Epoch 11/25 | Train Loss: 0.1885 | Val Loss: 0.1778 | Val Acc: 94.65%
|
| 55 |
+
Epoch 12/25 | Train Loss: 0.1778 | Val Loss: 0.1747 | Val Acc: 94.65%
|
| 56 |
+
Epoch 13/25 | Train Loss: 0.1635 | Val Loss: 0.1561 | Val Acc: 95.15%
|
| 57 |
+
-> New best validation accuracy! Saving model state.
|
| 58 |
+
Epoch 14/25 | Train Loss: 0.1549 | Val Loss: 0.1535 | Val Acc: 95.45%
|
| 59 |
+
-> New best validation accuracy! Saving model state.
|
| 60 |
+
Epoch 15/25 | Train Loss: 0.1498 | Val Loss: 0.1290 | Val Acc: 95.95%
|
| 61 |
+
-> New best validation accuracy! Saving model state.
|
| 62 |
+
Epoch 16/25 | Train Loss: 0.1363 | Val Loss: 0.1323 | Val Acc: 96.15%
|
| 63 |
+
-> New best validation accuracy! Saving model state.
|
| 64 |
+
Epoch 17/25 | Train Loss: 0.1260 | Val Loss: 0.1277 | Val Acc: 95.90%
|
| 65 |
+
Epoch 18/25 | Train Loss: 0.1225 | Val Loss: 0.1370 | Val Acc: 95.65%
|
| 66 |
+
Epoch 19/25 | Train Loss: 0.1249 | Val Loss: 0.1395 | Val Acc: 95.95%
|
| 67 |
+
Epoch 20/25 | Train Loss: 0.1150 | Val Loss: 0.1267 | Val Acc: 96.10%
|
| 68 |
+
Epoch 21/25 | Train Loss: 0.1075 | Val Loss: 0.1203 | Val Acc: 96.40%
|
| 69 |
+
-> New best validation accuracy! Saving model state.
|
| 70 |
+
Epoch 22/25 | Train Loss: 0.1023 | Val Loss: 0.1232 | Val Acc: 96.10%
|
| 71 |
+
Epoch 23/25 | Train Loss: 0.1013 | Val Loss: 0.1277 | Val Acc: 95.80%
|
| 72 |
+
Epoch 24/25 | Train Loss: 0.0975 | Val Loss: 0.1141 | Val Acc: 96.50%
|
| 73 |
+
-> New best validation accuracy! Saving model state.
|
| 74 |
+
Epoch 25/25 | Train Loss: 0.0853 | Val Loss: 0.1192 | Val Acc: 96.15%
|
| 75 |
+
--- Training Finished ---
|
| 76 |
+
|
| 77 |
+
--- Evaluating on Test Set (V3.0 Model) ---
|
| 78 |
+
Final Test Loss: 0.1096
|
| 79 |
+
Final Test Accuracy: 96.83%
|
| 80 |
+
-------------------------------------------
|