panda992 commited on
Commit
167a3d7
·
verified ·
1 Parent(s): ed28b4c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +74 -39
README.md CHANGED
@@ -1,67 +1,102 @@
1
- ---
2
  library_name: transformers
3
  license: apache-2.0
4
  base_model: google/vit-base-patch16-224-in21k
5
  tags:
6
  - image-classification
 
 
 
7
  - generated_from_trainer
8
  metrics:
9
  - accuracy
10
  model-index:
11
  - name: fish_disease_datasets
12
- results: []
 
 
 
 
 
 
 
 
 
13
  ---
14
 
15
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
16
- should probably proofread and complete it, then remove this comment. -->
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
17
 
18
- # fish_disease_datasets
19
 
20
- This model is a fine-tuned version of [google/vit-base-patch16-224-in21k](https://huggingface.co/google/vit-base-patch16-224-in21k) on the fish_disease_datasets dataset.
21
- It achieves the following results on the evaluation set:
22
- - Loss: 0.0866
23
- - Accuracy: 0.9728
24
 
25
- ## Model description
26
 
27
- More information needed
28
 
29
- ## Intended uses & limitations
30
 
31
- More information needed
32
 
33
- ## Training and evaluation data
34
 
35
- More information needed
36
 
37
- ## Training procedure
38
 
39
- ### Training hyperparameters
40
 
41
- The following hyperparameters were used during training:
42
- - learning_rate: 0.0002
43
- - train_batch_size: 16
44
- - eval_batch_size: 8
45
- - seed: 42
46
- - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
47
- - lr_scheduler_type: linear
48
- - num_epochs: 4
49
- - mixed_precision_training: Native AMP
50
 
51
- ### Training results
52
 
53
- | Training Loss | Epoch | Step | Validation Loss | Accuracy |
54
- |:-------------:|:------:|:----:|:---------------:|:--------:|
55
- | 0.3865 | 0.7634 | 100 | 0.4161 | 0.8913 |
56
- | 0.1206 | 1.5267 | 200 | 0.2170 | 0.9457 |
57
- | 0.1132 | 2.2901 | 300 | 0.1317 | 0.9674 |
58
- | 0.0547 | 3.0534 | 400 | 0.0879 | 0.9810 |
59
- | 0.0209 | 3.8168 | 500 | 0.0866 | 0.9728 |
60
 
 
61
 
62
- ### Framework versions
63
 
64
- - Transformers 4.52.3
65
- - Pytorch 2.7.0+cu126
66
- - Datasets 3.6.0
67
- - Tokenizers 0.21.1
 
 
 
 
1
+
2
  library_name: transformers
3
  license: apache-2.0
4
  base_model: google/vit-base-patch16-224-in21k
5
  tags:
6
  - image-classification
7
+ - vision-transformer
8
+ - aquaculture
9
+ - fish-disease
10
  - generated_from_trainer
11
  metrics:
12
  - accuracy
13
  model-index:
14
  - name: fish_disease_datasets
15
+ results:
16
+ - task:
17
+ name: Image Classification
18
+ type: image-classification
19
+ dataset:
20
+ name: fish_disease_datasets
21
+ metrics:
22
+ - name: Accuracy
23
+ type: accuracy
24
+ value: 0.9728
25
  ---
26
 
27
+ 🐟 Fish Disease Classifier (ViT)
28
+
29
+ This model is a fine-tuned version of google/vit-base-patch16-224-in21k, trained on a custom fish disease image dataset for Indian aquaculture.
30
+ ✅ Detected Classes (Fish)
31
+
32
+ Bacterial Red Disease
33
+
34
+ Bacterial diseases – Aeromoniasis
35
+
36
+ Bacterial Gill Disease
37
+
38
+ Fungal diseases (Saprolegniasis)
39
+
40
+ Parasitic diseases
41
+
42
+ Viral diseases (White Tail Disease)
43
+
44
+ Healthy Fish
45
+
46
+ ⚠️ Planned Prawn Model (Upcoming)
47
+
48
+ We are currently working on a separate fine-tuned model to detect:
49
+
50
+ Bacterial Gill Disease (BG)
51
+
52
+ White Spot Syndrome Virus (WSSV)
53
+
54
+ Healthy Prawn
55
+
56
+ This model will be released in the next version once prawn dataset collection and training is complete.
57
+ 📊 Evaluation Metrics
58
+ Metric Value
59
+ Accuracy 97.28%
60
+ Validation Loss 0.0866
61
+ Final Epoch 4
62
+ 🧠 Model Description
63
+
64
+ Architecture: Vision Transformer (ViT)
65
 
66
+ Base model: google/vit-base-patch16-224-in21k
67
 
68
+ Dataset: Custom-labeled images of freshwater fish diseases
 
 
 
69
 
70
+ Data augmentation: Albumentations
71
 
72
+ Optimized for WhatsApp-based diagnosis tools
73
 
74
+ 🚜 Intended Use
75
 
76
+ This model is optimized for:
77
 
78
+ Farmers needing fast disease detection via image
79
 
80
+ WhatsApp or mobile-based advisory tools
81
 
82
+ NGO/hatchery/government pilots in India and South Asia
83
 
84
+ 🏋️ Training Summary
85
 
86
+ Learning rate: 0.0002
 
 
 
 
 
 
 
 
87
 
88
+ Batch size: 16 (train) / 8 (eval)
89
 
90
+ Epochs: 4
 
 
 
 
 
 
91
 
92
+ Mixed Precision: AMP
93
 
94
+ Framework: Hugging Face Transformers, PyTorch
95
 
96
+ Training Results:
97
+ Training Loss Epoch Step Validation Loss Accuracy
98
+ 0.3865 0.76 100 0.4161 0.8913
99
+ 0.1206 1.53 200 0.2170 0.9457
100
+ 0.1132 2.29 300 0.1317 0.9674
101
+ 0.0547 3.05 400 0.0879 0.9810
102
+ 0.0209 3.81 500 0.0866 0.9728