2nzi commited on
Commit
54948d9
·
verified ·
1 Parent(s): 6391694

Model save

Browse files
Files changed (5) hide show
  1. README.md +63 -0
  2. all_results.json +9 -0
  3. test_results.json +9 -0
  4. trainer_state.json +89 -0
  5. val_results.json +9 -0
README.md ADDED
@@ -0,0 +1,63 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: cc-by-nc-4.0
3
+ base_model: MCG-NJU/videomae-base
4
+ tags:
5
+ - generated_from_trainer
6
+ metrics:
7
+ - accuracy
8
+ - f1
9
+ model-index:
10
+ - name: videomae-surf-analytics-sans-wandb
11
+ results: []
12
+ ---
13
+
14
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
+ should probably proofread and complete it, then remove this comment. -->
16
+
17
+ # videomae-surf-analytics-sans-wandb
18
+
19
+ This model is a fine-tuned version of [MCG-NJU/videomae-base](https://huggingface.co/MCG-NJU/videomae-base) on an unknown dataset.
20
+ It achieves the following results on the evaluation set:
21
+ - Loss: 1.1514
22
+ - Accuracy: 0.25
23
+ - F1: 0.125
24
+
25
+ ## Model description
26
+
27
+ More information needed
28
+
29
+ ## Intended uses & limitations
30
+
31
+ More information needed
32
+
33
+ ## Training and evaluation data
34
+
35
+ More information needed
36
+
37
+ ## Training procedure
38
+
39
+ ### Training hyperparameters
40
+
41
+ The following hyperparameters were used during training:
42
+ - learning_rate: 5e-05
43
+ - train_batch_size: 4
44
+ - eval_batch_size: 4
45
+ - seed: 42
46
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
47
+ - lr_scheduler_type: linear
48
+ - lr_scheduler_warmup_ratio: 0.1
49
+ - training_steps: 11
50
+
51
+ ### Training results
52
+
53
+ | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 |
54
+ |:-------------:|:-----:|:----:|:---------------:|:--------:|:-----:|
55
+ | 1.361 | 1.0 | 11 | 1.1514 | 0.25 | 0.125 |
56
+
57
+
58
+ ### Framework versions
59
+
60
+ - Transformers 4.41.2
61
+ - Pytorch 2.1.0+cpu
62
+ - Datasets 2.19.2
63
+ - Tokenizers 0.19.1
all_results.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 1.0,
3
+ "eval_accuracy": 0.25,
4
+ "eval_f1": 0.125,
5
+ "eval_loss": 1.1514339447021484,
6
+ "eval_runtime": 11.5456,
7
+ "eval_samples_per_second": 0.346,
8
+ "eval_steps_per_second": 0.087
9
+ }
test_results.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 1.0,
3
+ "eval_accuracy": 0.8,
4
+ "eval_f1": 0.7333333333333333,
5
+ "eval_loss": 1.1280324459075928,
6
+ "eval_runtime": 14.0785,
7
+ "eval_samples_per_second": 0.355,
8
+ "eval_steps_per_second": 0.142
9
+ }
trainer_state.json ADDED
@@ -0,0 +1,89 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.25,
3
+ "best_model_checkpoint": "videomae-surf-analytics-sans-wandb\\checkpoint-11",
4
+ "epoch": 1.0,
5
+ "eval_steps": 500,
6
+ "global_step": 11,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.9090909090909091,
13
+ "grad_norm": 9.908684730529785,
14
+ "learning_rate": 5.555555555555556e-06,
15
+ "loss": 1.361,
16
+ "step": 10
17
+ },
18
+ {
19
+ "epoch": 1.0,
20
+ "eval_accuracy": 0.25,
21
+ "eval_f1": 0.125,
22
+ "eval_loss": 1.1514339447021484,
23
+ "eval_runtime": 12.9275,
24
+ "eval_samples_per_second": 0.309,
25
+ "eval_steps_per_second": 0.077,
26
+ "step": 11
27
+ },
28
+ {
29
+ "epoch": 1.0,
30
+ "step": 11,
31
+ "total_flos": 5.482781984371507e+16,
32
+ "train_loss": 1.3439641974189065,
33
+ "train_runtime": 366.6861,
34
+ "train_samples_per_second": 0.12,
35
+ "train_steps_per_second": 0.03
36
+ },
37
+ {
38
+ "epoch": 1.0,
39
+ "eval_accuracy": 0.851063829787234,
40
+ "eval_f1": 0.8528510406305552,
41
+ "eval_loss": 0.9927294850349426,
42
+ "eval_runtime": 141.7743,
43
+ "eval_samples_per_second": 0.332,
44
+ "eval_steps_per_second": 0.085,
45
+ "step": 11
46
+ },
47
+ {
48
+ "epoch": 1.0,
49
+ "eval_accuracy": 0.8,
50
+ "eval_f1": 0.7333333333333333,
51
+ "eval_loss": 1.1280324459075928,
52
+ "eval_runtime": 14.0785,
53
+ "eval_samples_per_second": 0.355,
54
+ "eval_steps_per_second": 0.142,
55
+ "step": 11
56
+ },
57
+ {
58
+ "epoch": 1.0,
59
+ "eval_accuracy": 0.25,
60
+ "eval_f1": 0.125,
61
+ "eval_loss": 1.1514339447021484,
62
+ "eval_runtime": 11.5456,
63
+ "eval_samples_per_second": 0.346,
64
+ "eval_steps_per_second": 0.087,
65
+ "step": 11
66
+ }
67
+ ],
68
+ "logging_steps": 10,
69
+ "max_steps": 11,
70
+ "num_input_tokens_seen": 0,
71
+ "num_train_epochs": 9223372036854775807,
72
+ "save_steps": 500,
73
+ "stateful_callbacks": {
74
+ "TrainerControl": {
75
+ "args": {
76
+ "should_epoch_stop": false,
77
+ "should_evaluate": false,
78
+ "should_log": false,
79
+ "should_save": true,
80
+ "should_training_stop": true
81
+ },
82
+ "attributes": {}
83
+ }
84
+ },
85
+ "total_flos": 5.482781984371507e+16,
86
+ "train_batch_size": 4,
87
+ "trial_name": null,
88
+ "trial_params": null
89
+ }
val_results.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 1.0,
3
+ "eval_accuracy": 0.25,
4
+ "eval_f1": 0.125,
5
+ "eval_loss": 1.1514339447021484,
6
+ "eval_runtime": 11.5456,
7
+ "eval_samples_per_second": 0.346,
8
+ "eval_steps_per_second": 0.087
9
+ }