Granitagushi commited on
Commit
f15fa45
·
verified ·
1 Parent(s): da72ad9

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +8 -79
README.md CHANGED
@@ -1,83 +1,12 @@
1
  ---
2
- library_name: transformers
3
- license: apache-2.0
4
- base_model: google/vit-base-patch16-224
5
- tags:
6
- - image-classification
7
- - generated_from_trainer
8
- metrics:
9
- - accuracy
10
- model-index:
11
- - name: vit-base-oxford-iiit-pets
12
- results: []
13
  ---
14
 
15
- # vit-base-oxford-iiit-pets
16
-
17
- This model is a fine-tuned version of [google/vit-base-patch16-224](https://huggingface.co/google/vit-base-patch16-224) on the pcuenq/oxford-pets dataset.
18
- It achieves the following results on the evaluation set:
19
- - Loss: 0.1893
20
- - Accuracy: 0.9405
21
-
22
- ## Model description
23
-
24
- More information needed
25
-
26
- ## Intended uses & limitations
27
-
28
- More information needed
29
-
30
- ## Training and evaluation data
31
-
32
- More information needed
33
-
34
- ## Training procedure
35
-
36
- ### Training hyperparameters
37
-
38
- The following hyperparameters were used during training:
39
- - learning_rate: 0.0003
40
- - train_batch_size: 16
41
- - eval_batch_size: 8
42
- - seed: 42
43
- - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
44
- - lr_scheduler_type: linear
45
- - num_epochs: 5
46
-
47
- ### Training results
48
-
49
- | Training Loss | Epoch | Step | Validation Loss | Accuracy |
50
- |:-------------:|:-----:|:----:|:---------------:|:--------:|
51
- | 0.3976 | 1.0 | 370 | 0.2921 | 0.9364 |
52
- | 0.2273 | 2.0 | 740 | 0.2257 | 0.9445 |
53
- | 0.1742 | 3.0 | 1110 | 0.2102 | 0.9445 |
54
- | 0.1352 | 4.0 | 1480 | 0.2023 | 0.9459 |
55
- | 0.1326 | 5.0 | 1850 | 0.2006 | 0.9459 |
56
-
57
- ### Framework versions
58
-
59
- - Transformers 4.50.0
60
- - Pytorch 2.6.0+cu124
61
- - Datasets 3.4.1
62
- - Tokenizers 0.21.1
63
-
64
- ## Zero-Shot classification model
65
-
66
- This section compares the performance of a zero-shot model (`openai/clip-vit-large-patch14`) on the Oxford Pets dataset (`pcuenq/oxford-pets`).
67
-
68
- - **Model used**: `openai/clip-vit-large-patch14`
69
- - **Dataset**: `pcuenq/oxford-pets` (train split)
70
- - **Evaluation Task**: Zero-Shot Image Classification
71
- - **Candidate Labels**: 37 pet breeds from the dataset
72
-
73
-
74
- ### Results:
75
-
76
- Zero-Shot Evaluation mit CLIP:
77
- - **Accuracy**: 0.8800
78
- - **Precision**: 0.8768
79
- - **Recall**: 0.8800
80
-
81
- Evaluated using Hugging Face `transformers` pipeline and `sklearn.metrics` on the full training set.
82
-
83
 
 
1
  ---
2
+ title: Pet Classification Tool
3
+ emoji: 🐕
4
+ colorFrom: indigo
5
+ colorTo: blue
6
+ sdk: gradio
7
+ sdk_version: 4.16.0
8
+ app_file: app.py
9
+ pinned: false
 
 
 
10
  ---
11
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
12