stnleyyg commited on
Commit
8b48a68
·
verified ·
1 Parent(s): 893ac80

End of training

Browse files
README.md CHANGED
@@ -4,8 +4,8 @@ license: apache-2.0
4
  base_model: google/vit-base-patch16-224-in21k
5
  tags:
6
  - generated_from_trainer
7
- datasets:
8
- - imagefolder
9
  model-index:
10
  - name: image_classification
11
  results: []
@@ -16,15 +16,10 @@ should probably proofread and complete it, then remove this comment. -->
16
 
17
  # image_classification
18
 
19
- This model is a fine-tuned version of [google/vit-base-patch16-224-in21k](https://huggingface.co/google/vit-base-patch16-224-in21k) on the imagefolder dataset.
20
  It achieves the following results on the evaluation set:
21
- - eval_loss: 2.0879
22
- - eval_model_preparation_time: 0.0065
23
- - eval_accuracy: 0.1187
24
- - eval_runtime: 43.6598
25
- - eval_samples_per_second: 3.665
26
- - eval_steps_per_second: 0.115
27
- - step: 0
28
 
29
  ## Model description
30
 
@@ -43,20 +38,29 @@ More information needed
43
  ### Training hyperparameters
44
 
45
  The following hyperparameters were used during training:
46
- - learning_rate: 0.0001
47
- - train_batch_size: 32
48
- - eval_batch_size: 32
49
  - seed: 42
50
- - gradient_accumulation_steps: 2
51
  - total_train_batch_size: 64
52
- - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
53
- - lr_scheduler_type: cosine
54
- - lr_scheduler_warmup_ratio: 0.3
55
- - num_epochs: 20
 
 
 
 
 
 
 
 
 
56
 
57
  ### Framework versions
58
 
59
- - Transformers 4.46.2
60
- - Pytorch 2.5.1+cu121
61
- - Datasets 3.1.0
62
- - Tokenizers 0.20.3
 
4
  base_model: google/vit-base-patch16-224-in21k
5
  tags:
6
  - generated_from_trainer
7
+ metrics:
8
+ - accuracy
9
  model-index:
10
  - name: image_classification
11
  results: []
 
16
 
17
  # image_classification
18
 
19
+ This model is a fine-tuned version of [google/vit-base-patch16-224-in21k](https://huggingface.co/google/vit-base-patch16-224-in21k) on an unknown dataset.
20
  It achieves the following results on the evaluation set:
21
+ - Loss: 1.6535
22
+ - Accuracy: 0.878
 
 
 
 
 
23
 
24
  ## Model description
25
 
 
38
  ### Training hyperparameters
39
 
40
  The following hyperparameters were used during training:
41
+ - learning_rate: 5e-05
42
+ - train_batch_size: 16
43
+ - eval_batch_size: 16
44
  - seed: 42
45
+ - gradient_accumulation_steps: 4
46
  - total_train_batch_size: 64
47
+ - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
48
+ - lr_scheduler_type: linear
49
+ - lr_scheduler_warmup_ratio: 0.1
50
+ - num_epochs: 3
51
+
52
+ ### Training results
53
+
54
+ | Training Loss | Epoch | Step | Validation Loss | Accuracy |
55
+ |:-------------:|:-----:|:----:|:---------------:|:--------:|
56
+ | 2.7065 | 1.0 | 63 | 2.5465 | 0.799 |
57
+ | 1.8582 | 2.0 | 126 | 1.8365 | 0.848 |
58
+ | 1.6103 | 2.96 | 186 | 1.6695 | 0.863 |
59
+
60
 
61
  ### Framework versions
62
 
63
+ - Transformers 4.51.3
64
+ - Pytorch 2.6.0+cu124
65
+ - Datasets 3.6.0
66
+ - Tokenizers 0.21.1
runs/May15_15-52-39_eb3fec5dd0b0/events.out.tfevents.1747325096.eb3fec5dd0b0.7448.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:59cce5ab4c7f4637df43ca57dc0523329974dda2b0cc1058e0c6a7badc9cf052
3
+ size 411