Hemg commited on
Commit
0ffd192
·
verified ·
1 Parent(s): 64acc78

Model save

Browse files
README.md CHANGED
@@ -17,8 +17,8 @@ should probably proofread and complete it, then remove this comment. -->
17
 
18
  This model is a fine-tuned version of [google/vit-base-patch16-224-in21k](https://huggingface.co/google/vit-base-patch16-224-in21k) on an unknown dataset.
19
  It achieves the following results on the evaluation set:
20
- - Loss: 0.5200
21
- - Accuracy: 0.8391
22
 
23
  ## Model description
24
 
@@ -37,12 +37,12 @@ More information needed
37
  ### Training hyperparameters
38
 
39
  The following hyperparameters were used during training:
40
- - learning_rate: 0.0001
41
- - train_batch_size: 4
42
- - eval_batch_size: 4
43
  - seed: 42
44
  - gradient_accumulation_steps: 2
45
- - total_train_batch_size: 8
46
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
47
  - lr_scheduler_type: linear
48
  - lr_scheduler_warmup_ratio: 0.1
@@ -52,21 +52,21 @@ The following hyperparameters were used during training:
52
 
53
  | Training Loss | Epoch | Step | Validation Loss | Accuracy |
54
  |:-------------:|:-----:|:----:|:---------------:|:--------:|
55
- | 1.7226 | 1.0 | 43 | 1.3689 | 0.6207 |
56
- | 0.9553 | 2.0 | 86 | 0.8472 | 0.7586 |
57
- | 0.5832 | 3.0 | 129 | 0.5262 | 0.8621 |
58
- | 0.3948 | 4.0 | 172 | 0.4701 | 0.8621 |
59
- | 0.3127 | 5.0 | 215 | 0.5299 | 0.8506 |
60
- | 0.207 | 6.0 | 258 | 0.4629 | 0.8621 |
61
- | 0.191 | 7.0 | 301 | 0.5138 | 0.8391 |
62
- | 0.1679 | 8.0 | 344 | 0.4064 | 0.8621 |
63
- | 0.1968 | 9.0 | 387 | 0.3898 | 0.8851 |
64
- | 0.1485 | 10.0 | 430 | 0.5200 | 0.8391 |
65
 
66
 
67
  ### Framework versions
68
 
69
  - Transformers 4.38.2
70
- - Pytorch 2.1.0+cu121
71
  - Datasets 2.18.0
72
  - Tokenizers 0.15.2
 
17
 
18
  This model is a fine-tuned version of [google/vit-base-patch16-224-in21k](https://huggingface.co/google/vit-base-patch16-224-in21k) on an unknown dataset.
19
  It achieves the following results on the evaluation set:
20
+ - Loss: 0.2590
21
+ - Accuracy: 0.9167
22
 
23
  ## Model description
24
 
 
37
  ### Training hyperparameters
38
 
39
  The following hyperparameters were used during training:
40
+ - learning_rate: 0.0002
41
+ - train_batch_size: 8
42
+ - eval_batch_size: 8
43
  - seed: 42
44
  - gradient_accumulation_steps: 2
45
+ - total_train_batch_size: 16
46
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
47
  - lr_scheduler_type: linear
48
  - lr_scheduler_warmup_ratio: 0.1
 
52
 
53
  | Training Loss | Epoch | Step | Validation Loss | Accuracy |
54
  |:-------------:|:-----:|:----:|:---------------:|:--------:|
55
+ | 1.5856 | 1.0 | 147 | 1.2463 | 0.6122 |
56
+ | 0.7971 | 2.0 | 294 | 0.7495 | 0.7551 |
57
+ | 0.5344 | 3.0 | 441 | 0.5892 | 0.7925 |
58
+ | 0.4006 | 4.0 | 588 | 0.5291 | 0.8452 |
59
+ | 0.2778 | 5.0 | 735 | 0.3742 | 0.8759 |
60
+ | 0.2501 | 6.0 | 882 | 0.3946 | 0.8895 |
61
+ | 0.1444 | 7.0 | 1029 | 0.3125 | 0.9031 |
62
+ | 0.1396 | 8.0 | 1176 | 0.3348 | 0.9150 |
63
+ | 0.1231 | 9.0 | 1323 | 0.2848 | 0.9116 |
64
+ | 0.0975 | 10.0 | 1470 | 0.2590 | 0.9167 |
65
 
66
 
67
  ### Framework versions
68
 
69
  - Transformers 4.38.2
70
+ - Pytorch 2.1.2
71
  - Datasets 2.18.0
72
  - Tokenizers 0.15.2
runs/Mar12_04-36-17_ff168037f406/events.out.tfevents.1710218178.ff168037f406.34.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3165bfee26ab82c87f543a0bee3e911594b74e8849f8283d005c073bef870d71
3
- size 10307
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6ee6bd38bdbbc5baddea6b7efcec626b4c25ea39da4068f35c4edd71ecb2a4da
3
+ size 10661