sumit7488 commited on
Commit
ce3306d
Β·
verified Β·
1 Parent(s): ce3b400

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +84 -3
README.md CHANGED
@@ -1,3 +1,84 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ ---
4
+
5
+ ## πŸ“Š Baseline Model Results (TimeSformer on HMDB51)
6
+
7
+ The standard **TimeSformer model** was trained on the **HMDB51 dataset** for 15 epochs.
8
+
9
+ Training was performed in multiple stages due to runtime limits and resumed using saved checkpoints.
10
+
11
+ ---
12
+
13
+ ## πŸ” Training Strategy
14
+
15
+ - Training conducted on Kaggle GPU
16
+ - Interrupted due to 12-hour session limit
17
+ - Resumed using `.safetensors` checkpoint
18
+ - Completed full **15 epochs**
19
+ - Early stopping applied at final stage
20
+
21
+ ---
22
+
23
+ ## πŸ“ˆ Training Results (Epoch 1–15)
24
+
25
+ | Epoch | Train Loss | Train Acc | Val Loss | Val Acc | F1 |
26
+ |------|------------|-----------|----------|---------|-----|
27
+ | 1 | 3.9314 | 0.0323 | 3.7800 | 0.1248 | 0.1100 |
28
+ | 2 | 3.5892 | 0.1976 | 3.2530 | 0.4353 | 0.3959 |
29
+ | 3 | 3.0235 | 0.4354 | 2.6567 | 0.5340 | 0.4867 |
30
+ | 4 | 2.5720 | 0.5246 | 2.2347 | 0.6033 | 0.5720 |
31
+ | 5 | 2.2836 | 0.6009 | 2.0135 | 0.6327 | 0.6143 |
32
+ | 6 | 2.0897 | 0.6419 | 1.8659 | 0.6536 | 0.6415 |
33
+ | 7 | 1.9678 | 0.6807 | 1.8017 | 0.6667 | 0.6561 |
34
+ | 8 | 1.8847 | 0.7030 | 1.7705 | 0.6627 | 0.6485 |
35
+ | 9 | 1.8213 | 0.7237 | 1.7355 | 0.6627 | 0.6518 |
36
+ | 10 | 1.7576 | 0.7414 | 1.7340 | 0.6680 | 0.6570 |
37
+ | 11 | 1.7098 | 0.7549 | 1.7234 | 0.6765 | 0.6699 |
38
+ | 12 | 1.6813 | 0.7660 | 1.6980 | **0.6895** | **0.6846** |
39
+ | 13 | 1.6731 | 0.7693 | 1.7023 | 0.6882 | 0.6824 |
40
+ | 14 | 1.6384 | 0.7778 | 1.7110 | 0.6850 | 0.6806 |
41
+ | 15 | 1.6045 | 0.7884 | 1.7131 | 0.6850 | 0.6805 |
42
+
43
+ ---
44
+
45
+ ## πŸ† Best Performance
46
+
47
+ - **Validation Accuracy:** **68.95%**
48
+ - **F1 Score:** 0.6846
49
+ - Achieved at **Epoch 12**
50
+
51
+ ---
52
+
53
+ ## βš™οΈ Training Details
54
+
55
+ - Peak GPU Memory: **~9.3 GB**
56
+ - Training Time per Epoch: ~55 minutes
57
+ - Evaluation Time: ~8 minutes
58
+ - Mixed Precision Training used
59
+ - Early stopping applied after convergence
60
+
61
+ ---
62
+
63
+ ## πŸ“Œ Observations
64
+
65
+ - Strong and stable learning curve
66
+ - Performance improves steadily until **Epoch 12**
67
+ - After Epoch 12:
68
+ - Validation accuracy plateaus
69
+ - Validation loss increases slightly β†’ **overfitting begins**
70
+
71
+ ---
72
+
73
+ ## ⚑ Key Insight
74
+
75
+ - TimeSformer achieves **higher accuracy (~68.95%)** on HMDB51
76
+ - However, it requires:
77
+ - ❌ Higher memory (~9.3 GB)
78
+ - ❌ Higher computational cost
79
+
80
+ ---
81
+
82
+ ## πŸ” Conclusion (Baseline)
83
+
84
+ The TimeSformer model delivers strong performance on HMDB51 but at a significantly higher computational cost, highlighting the need for more efficient architectures such as the proposed RetNet-based hybrid model.