Shawon16 commited on
Commit
acb7159
·
verified ·
1 Parent(s): 92ef510

Model save

Browse files
Files changed (2) hide show
  1. README.md +4 -4
  2. trainer_state.json +21 -21
README.md CHANGED
@@ -18,8 +18,8 @@ should probably proofread and complete it, then remove this comment. -->
18
 
19
  This model is a fine-tuned version of [MCG-NJU/videomae-base](https://huggingface.co/MCG-NJU/videomae-base) on an unknown dataset.
20
  It achieves the following results on the evaluation set:
21
- - Loss: 7.6141
22
- - Accuracy: 0.0010
23
 
24
  ## Model description
25
 
@@ -47,14 +47,14 @@ The following hyperparameters were used during training:
47
  - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
48
  - lr_scheduler_type: linear
49
  - lr_scheduler_warmup_ratio: 0.1
50
- - training_steps: 1786
51
  - mixed_precision_training: Native AMP
52
 
53
  ### Training results
54
 
55
  | Training Loss | Epoch | Step | Validation Loss | Accuracy |
56
  |:-------------:|:-----:|:----:|:---------------:|:--------:|
57
- | 30.6603 | 1.0 | 1786 | 7.6141 | 0.0010 |
58
 
59
 
60
  ### Framework versions
 
18
 
19
  This model is a fine-tuned version of [MCG-NJU/videomae-base](https://huggingface.co/MCG-NJU/videomae-base) on an unknown dataset.
20
  It achieves the following results on the evaluation set:
21
+ - Loss: 1.2462
22
+ - Accuracy: 0.4286
23
 
24
  ## Model description
25
 
 
47
  - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
48
  - lr_scheduler_type: linear
49
  - lr_scheduler_warmup_ratio: 0.1
50
+ - training_steps: 7
51
  - mixed_precision_training: Native AMP
52
 
53
  ### Training results
54
 
55
  | Training Loss | Epoch | Step | Validation Loss | Accuracy |
56
  |:-------------:|:-----:|:----:|:---------------:|:--------:|
57
+ | 5.9518 | 1.0 | 7 | 1.2462 | 0.4286 |
58
 
59
 
60
  ### Framework versions
trainer_state.json CHANGED
@@ -1,41 +1,41 @@
1
  {
2
- "best_metric": 0.0010214504596527069,
3
- "best_model_checkpoint": "/media/cse/HDD/Shawon/shawon/10 fold timesformer/VideoMAE_wlasl__codeCheck/checkpoint-1786",
4
  "epoch": 1.0,
5
  "eval_steps": 500,
6
- "global_step": 1786,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
  "epoch": 1.0,
13
- "grad_norm": 20.611459732055664,
14
- "learning_rate": 1.2445550715619166e-07,
15
- "loss": 30.6603,
16
- "step": 1786
17
  },
18
  {
19
  "epoch": 1.0,
20
- "eval_accuracy": 0.0010214504596527069,
21
- "eval_loss": 7.61414909362793,
22
- "eval_runtime": 297.8226,
23
- "eval_samples_per_second": 13.149,
24
- "eval_steps_per_second": 6.574,
25
- "step": 1786
26
  },
27
  {
28
  "epoch": 1.0,
29
- "step": 1786,
30
- "total_flos": 1.8121006360874189e+19,
31
- "train_loss": 30.660261233202686,
32
- "train_runtime": 1787.4428,
33
- "train_samples_per_second": 7.994,
34
- "train_steps_per_second": 0.999
35
  }
36
  ],
37
  "logging_steps": 500,
38
- "max_steps": 1786,
39
  "num_input_tokens_seen": 0,
40
  "num_train_epochs": 9223372036854775807,
41
  "save_steps": 500,
@@ -60,7 +60,7 @@
60
  "attributes": {}
61
  }
62
  },
63
- "total_flos": 1.8121006360874189e+19,
64
  "train_batch_size": 2,
65
  "trial_name": null,
66
  "trial_params": null
 
1
  {
2
+ "best_metric": 0.42857142857142855,
3
+ "best_model_checkpoint": "/media/cse/HDD/Shawon/shawon/10 fold timesformer/VideoMAE_wlasl__codeCheck/checkpoint-7",
4
  "epoch": 1.0,
5
  "eval_steps": 500,
6
+ "global_step": 7,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
  "epoch": 1.0,
13
+ "grad_norm": 56.47215270996094,
14
+ "learning_rate": 3.3333333333333335e-05,
15
+ "loss": 5.9518,
16
+ "step": 7
17
  },
18
  {
19
  "epoch": 1.0,
20
+ "eval_accuracy": 0.42857142857142855,
21
+ "eval_loss": 1.2461984157562256,
22
+ "eval_runtime": 2.4468,
23
+ "eval_samples_per_second": 5.722,
24
+ "eval_steps_per_second": 2.861,
25
+ "step": 7
26
  },
27
  {
28
  "epoch": 1.0,
29
+ "step": 7,
30
+ "total_flos": 6.978086161927373e+16,
31
+ "train_loss": 5.951799665178571,
32
+ "train_runtime": 25.0388,
33
+ "train_samples_per_second": 2.237,
34
+ "train_steps_per_second": 0.28
35
  }
36
  ],
37
  "logging_steps": 500,
38
+ "max_steps": 7,
39
  "num_input_tokens_seen": 0,
40
  "num_train_epochs": 9223372036854775807,
41
  "save_steps": 500,
 
60
  "attributes": {}
61
  }
62
  },
63
+ "total_flos": 6.978086161927373e+16,
64
  "train_batch_size": 2,
65
  "trial_name": null,
66
  "trial_params": null