HazelAtis commited on
Commit
6776fcc
·
verified ·
1 Parent(s): d6490aa

Upload 5 files

Browse files
Files changed (5) hide show
  1. README.md +43 -10
  2. config.json +41 -0
  3. gitattributes +35 -0
  4. model.safetensors +3 -0
  5. preprocessor_config.json +41 -0
README.md CHANGED
@@ -1,13 +1,46 @@
1
  ---
2
- title: Skin Cancer Classification
3
- emoji: 📚
4
- colorFrom: indigo
5
- colorTo: yellow
6
- sdk: gradio
7
- sdk_version: 5.35.0
8
- app_file: app.py
9
- pinned: false
10
- license: mit
11
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
12
 
13
- Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
 
1
  ---
2
+ license: apache-2.0
 
 
 
 
 
 
 
 
3
  ---
4
+ # Skin Cancer Image Classification Model
5
+
6
+ ## Introduction
7
+
8
+ This model is designed for the classification of skin cancer images into various categories including benign keratosis-like lesions, basal cell carcinoma, actinic keratoses, vascular lesions, melanocytic nevi, melanoma, and dermatofibroma.
9
+
10
+ ## Model Overview
11
+
12
+ - Model Architecture: Vision Transformer (ViT)
13
+ - Pre-trained Model: Google's ViT with 16x16 patch size and trained on ImageNet21k dataset
14
+ - Modified Classification Head: The classification head has been replaced to adapt the model to the skin cancer classification task.
15
+
16
+ ## Dataset
17
+
18
+ - Dataset Name: Skin Cancer Dataset
19
+ - Source: [Marmal88's Skin Cancer Dataset on Hugging Face](https://huggingface.co/datasets/marmal88/skin_cancer)
20
+ - Classes: Benign keratosis-like lesions, Basal cell carcinoma, Actinic keratoses, Vascular lesions, Melanocytic nevi, Melanoma, Dermatofibroma
21
+
22
+ ## Training
23
+
24
+ - Optimizer: Adam optimizer with a learning rate of 1e-4
25
+ - Loss Function: Cross-Entropy Loss
26
+ - Batch Size: 32
27
+ - Number of Epochs: 5
28
+
29
+ ## Evaluation Metrics
30
+
31
+ - Train Loss: Average loss over the training dataset
32
+ - Train Accuracy: Accuracy over the training dataset
33
+ - Validation Loss: Average loss over the validation dataset
34
+ - Validation Accuracy: Accuracy over the validation dataset
35
+
36
+ ## Results
37
+
38
+ - Epoch 1/5, Train Loss: 0.7168, Train Accuracy: 0.7586, Val Loss: 0.4994, Val Accuracy: 0.8355
39
+ - Epoch 2/5, Train Loss: 0.4550, Train Accuracy: 0.8466, Val Loss: 0.3237, Val Accuracy: 0.8973
40
+ - Epoch 3/5, Train Loss: 0.2959, Train Accuracy: 0.9028, Val Loss: 0.1790, Val Accuracy: 0.9530
41
+ - Epoch 4/5, Train Loss: 0.1595, Train Accuracy: 0.9482, Val Loss: 0.1498, Val Accuracy: 0.9555
42
+ - Epoch 5/5, Train Loss: 0.1208, Train Accuracy: 0.9614, Val Loss: 0.1000, Val Accuracy: 0.9695
43
+ ## Conclusion
44
+
45
+ The model demonstrates good performance in classifying skin cancer images into various categories. Further fine-tuning or experimentation may improve performance on this task.
46
 
 
config.json ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "google/vit-base-patch16-224-in21k",
3
+ "architectures": [
4
+ "ViTForImageClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.0,
7
+ "encoder_stride": 16,
8
+ "hidden_act": "gelu",
9
+ "hidden_dropout_prob": 0.0,
10
+ "hidden_size": 768,
11
+ "id2label": {
12
+ "0": "benign_keratosis-like_lesions",
13
+ "1": "basal_cell_carcinoma",
14
+ "2": "actinic_keratoses",
15
+ "3": "vascular_lesions",
16
+ "4": "melanocytic_Nevi",
17
+ "5": "melanoma",
18
+ "6": "dermatofibroma"
19
+ },
20
+ "image_size": 224,
21
+ "initializer_range": 0.02,
22
+ "intermediate_size": 3072,
23
+ "label2id": {
24
+ "benign_keratosis-like_lesions": 0,
25
+ "basal_cell_carcinoma": 1,
26
+ "actinic_keratoses": 2,
27
+ "vascular_lesions": 3,
28
+ "melanocytic_Nevi": 4,
29
+ "melanoma": 5,
30
+ "dermatofibroma": 6
31
+ },
32
+ "layer_norm_eps": 1e-12,
33
+ "model_type": "vit",
34
+ "num_attention_heads": 12,
35
+ "num_channels": 3,
36
+ "num_hidden_layers": 12,
37
+ "patch_size": 16,
38
+ "qkv_bias": true,
39
+ "torch_dtype": "float32",
40
+ "transformers_version": "4.38.2"
41
+ }
gitattributes ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ *.7z filter=lfs diff=lfs merge=lfs -text
2
+ *.arrow filter=lfs diff=lfs merge=lfs -text
3
+ *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ckpt filter=lfs diff=lfs merge=lfs -text
6
+ *.ftz filter=lfs diff=lfs merge=lfs -text
7
+ *.gz filter=lfs diff=lfs merge=lfs -text
8
+ *.h5 filter=lfs diff=lfs merge=lfs -text
9
+ *.joblib filter=lfs diff=lfs merge=lfs -text
10
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
+ *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
+ *.model filter=lfs diff=lfs merge=lfs -text
13
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
14
+ *.npy filter=lfs diff=lfs merge=lfs -text
15
+ *.npz filter=lfs diff=lfs merge=lfs -text
16
+ *.onnx filter=lfs diff=lfs merge=lfs -text
17
+ *.ot filter=lfs diff=lfs merge=lfs -text
18
+ *.parquet filter=lfs diff=lfs merge=lfs -text
19
+ *.pb filter=lfs diff=lfs merge=lfs -text
20
+ *.pickle filter=lfs diff=lfs merge=lfs -text
21
+ *.pkl filter=lfs diff=lfs merge=lfs -text
22
+ *.pt filter=lfs diff=lfs merge=lfs -text
23
+ *.pth filter=lfs diff=lfs merge=lfs -text
24
+ *.rar filter=lfs diff=lfs merge=lfs -text
25
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
26
+ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
+ *.tar.* filter=lfs diff=lfs merge=lfs -text
28
+ *.tar filter=lfs diff=lfs merge=lfs -text
29
+ *.tflite filter=lfs diff=lfs merge=lfs -text
30
+ *.tgz filter=lfs diff=lfs merge=lfs -text
31
+ *.wasm filter=lfs diff=lfs merge=lfs -text
32
+ *.xz filter=lfs diff=lfs merge=lfs -text
33
+ *.zip filter=lfs diff=lfs merge=lfs -text
34
+ *.zst filter=lfs diff=lfs merge=lfs -text
35
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b8136bbbdd085cc97f69bdef01bd0c4c54262cf3ba44f5a94ebc9c8c54e85d60
3
+ size 343239356
preprocessor_config.json ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "google/vit-base-patch16-224-in21k",
3
+ "architectures": [
4
+ "ViTForImageClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.0,
7
+ "encoder_stride": 16,
8
+ "hidden_act": "gelu",
9
+ "hidden_dropout_prob": 0.0,
10
+ "hidden_size": 768,
11
+ "id2label": {
12
+ "0": "benign_keratosis-like_lesions",
13
+ "1": "basal_cell_carcinoma",
14
+ "2": "actinic_keratoses",
15
+ "3": "vascular_lesions",
16
+ "4": "melanocytic_Nevi",
17
+ "5": "melanoma",
18
+ "6": "dermatofibroma"
19
+ },
20
+ "image_size": 224,
21
+ "initializer_range": 0.02,
22
+ "intermediate_size": 3072,
23
+ "label2id": {
24
+ "benign_keratosis-like_lesions": 0,
25
+ "basal_cell_carcinoma": 1,
26
+ "actinic_keratoses": 2,
27
+ "vascular_lesions": 3,
28
+ "melanocytic_Nevi": 4,
29
+ "melanoma": 5,
30
+ "dermatofibroma": 6
31
+ },
32
+ "layer_norm_eps": 1e-12,
33
+ "model_type": "vit",
34
+ "num_attention_heads": 12,
35
+ "num_channels": 3,
36
+ "num_hidden_layers": 12,
37
+ "patch_size": 16,
38
+ "qkv_bias": true,
39
+ "torch_dtype": "float32",
40
+ "transformers_version": "4.38.2"
41
+ }