ahmadalfian commited on
Commit
a3232b2
·
verified ·
1 Parent(s): e5b24f9

End of training

Browse files
Files changed (1) hide show
  1. README.md +38 -16
README.md CHANGED
@@ -3,11 +3,26 @@ license: apache-2.0
3
  base_model: google/vit-base-patch16-224-in21k
4
  tags:
5
  - generated_from_trainer
 
 
6
  metrics:
7
  - accuracy
8
  model-index:
9
  - name: image_classification
10
- results: []
 
 
 
 
 
 
 
 
 
 
 
 
 
11
  ---
12
 
13
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -15,10 +30,10 @@ should probably proofread and complete it, then remove this comment. -->
15
 
16
  # image_classification
17
 
18
- This model is a fine-tuned version of [google/vit-base-patch16-224-in21k](https://huggingface.co/google/vit-base-patch16-224-in21k) on an unknown dataset.
19
  It achieves the following results on the evaluation set:
20
- - Loss: 1.6479
21
- - Accuracy: 0.885
22
 
23
  ## Model description
24
 
@@ -37,29 +52,36 @@ More information needed
37
  ### Training hyperparameters
38
 
39
  The following hyperparameters were used during training:
40
- - learning_rate: 5e-05
41
- - train_batch_size: 16
42
- - eval_batch_size: 16
43
  - seed: 42
44
- - gradient_accumulation_steps: 4
45
- - total_train_batch_size: 64
46
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
47
  - lr_scheduler_type: linear
48
- - lr_scheduler_warmup_ratio: 0.1
49
- - num_epochs: 3
50
 
51
  ### Training results
52
 
53
  | Training Loss | Epoch | Step | Validation Loss | Accuracy |
54
  |:-------------:|:-----:|:----:|:---------------:|:--------:|
55
- | 2.7451 | 0.992 | 62 | 2.5612 | 0.826 |
56
- | 1.8505 | 2.0 | 125 | 1.8206 | 0.863 |
57
- | 1.6186 | 2.976 | 186 | 1.6613 | 0.882 |
 
 
 
 
 
 
 
58
 
59
 
60
  ### Framework versions
61
 
62
- - Transformers 4.42.4
63
- - Pytorch 2.4.0+cu121
64
  - Datasets 2.21.0
65
  - Tokenizers 0.19.1
 
3
  base_model: google/vit-base-patch16-224-in21k
4
  tags:
5
  - generated_from_trainer
6
+ datasets:
7
+ - imagefolder
8
  metrics:
9
  - accuracy
10
  model-index:
11
  - name: image_classification
12
+ results:
13
+ - task:
14
+ name: Image Classification
15
+ type: image-classification
16
+ dataset:
17
+ name: imagefolder
18
+ type: imagefolder
19
+ config: default
20
+ split: train
21
+ args: default
22
+ metrics:
23
+ - name: Accuracy
24
+ type: accuracy
25
+ value: 0.58125
26
  ---
27
 
28
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
30
 
31
  # image_classification
32
 
33
+ This model is a fine-tuned version of [google/vit-base-patch16-224-in21k](https://huggingface.co/google/vit-base-patch16-224-in21k) on the imagefolder dataset.
34
  It achieves the following results on the evaluation set:
35
+ - Loss: 1.1334
36
+ - Accuracy: 0.5813
37
 
38
  ## Model description
39
 
 
52
  ### Training hyperparameters
53
 
54
  The following hyperparameters were used during training:
55
+ - learning_rate: 0.0001
56
+ - train_batch_size: 8
57
+ - eval_batch_size: 32
58
  - seed: 42
59
+ - gradient_accumulation_steps: 2
60
+ - total_train_batch_size: 16
61
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
62
  - lr_scheduler_type: linear
63
+ - lr_scheduler_warmup_ratio: 0.2
64
+ - num_epochs: 10
65
 
66
  ### Training results
67
 
68
  | Training Loss | Epoch | Step | Validation Loss | Accuracy |
69
  |:-------------:|:-----:|:----:|:---------------:|:--------:|
70
+ | 2.0287 | 1.0 | 40 | 1.9886 | 0.3312 |
71
+ | 1.6572 | 2.0 | 80 | 1.6017 | 0.4 |
72
+ | 1.4274 | 3.0 | 120 | 1.3753 | 0.5 |
73
+ | 1.2165 | 4.0 | 160 | 1.2923 | 0.5125 |
74
+ | 1.0676 | 5.0 | 200 | 1.2017 | 0.5437 |
75
+ | 0.9438 | 6.0 | 240 | 1.1987 | 0.525 |
76
+ | 0.7568 | 7.0 | 280 | 1.1113 | 0.5625 |
77
+ | 0.6683 | 8.0 | 320 | 1.1406 | 0.5625 |
78
+ | 0.5162 | 9.0 | 360 | 1.1863 | 0.5625 |
79
+ | 0.4258 | 10.0 | 400 | 1.0780 | 0.65 |
80
 
81
 
82
  ### Framework versions
83
 
84
+ - Transformers 4.44.0
85
+ - Pytorch 2.4.0
86
  - Datasets 2.21.0
87
  - Tokenizers 0.19.1