sdtemple commited on
Commit
3e638a2
·
verified ·
1 Parent(s): 4996d86

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +69 -0
README.md ADDED
@@ -0,0 +1,69 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ datasets:
4
+ - sdtemple/colored-shapes
5
+ language:
6
+ - en
7
+ metrics:
8
+ - accuracy
9
+ - precision
10
+ - recall
11
+ - roc_auc
12
+ pipeline_tag: image-classification
13
+ tags:
14
+ - tutorial
15
+ ---
16
+
17
+ This model predicts the shape (circle, rectangle, diamond, or triangle) of the 1 colored shape (8 colors) in a 224 x 224 x 3 image.
18
+
19
+ This model is a part of a how to tutorial on fitting PyTorch models.
20
+
21
+ The model is trained on 2000 examples for each color and shape combo (64,000 samples in total) simulated according to [https://github.com/sdtemple/zootopia3](https://github.com/sdtemple/zootopia3).
22
+
23
+ The model is tested/evaluated on the dataset [https://huggingface.co/datasets/sdtemple/colored-shapes](https://huggingface.co/datasets/sdtemple/colored-shapes) which has slightly smaller shapes simulated (out of distribution) relative to the training data.
24
+
25
+ - Accuracy: 75%
26
+ - Min precision (triangle): 56%
27
+ - Max precision (rectangle): 89%
28
+ - Min recall (diamond): 59%
29
+ - Max recall (triangle): 86%
30
+ - AUROC (macro-averaged): 91%
31
+ - Min AUROC (diamond): 90%
32
+ - Max AUROC (circle): 93%
33
+
34
+ Compared to [https://huggingface.co/sdtemple/color-prediction-model](https://huggingface.co/sdtemple/color-prediction-model), it is harder to predict the shape than the color of the object.
35
+
36
+ The model architecture is the following. In light experimentation, I found it important to have multiple convolutions and that too many parameters leads to noisy validation losses by epoch.
37
+
38
+ ```
39
+ MyCNN(
40
+ (conv_block): Sequential(
41
+ (0): Conv2d(3, 16, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
42
+ (1): BatchNorm2d(16, eps=1e-05, momentum=0.1, affine=True, track_running_stats=True)
43
+ (2): MaxPool2d(kernel_size=2, stride=2, padding=0, dilation=1, ceil_mode=False)
44
+ (3): Conv2d(16, 16, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
45
+ (4): BatchNorm2d(16, eps=1e-05, momentum=0.1, affine=True, track_running_stats=True)
46
+ (5): MaxPool2d(kernel_size=2, stride=2, padding=0, dilation=1, ceil_mode=False)
47
+ (6): Conv2d(16, 16, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
48
+ (7): BatchNorm2d(16, eps=1e-05, momentum=0.1, affine=True, track_running_stats=True)
49
+ (8): MaxPool2d(kernel_size=2, stride=2, padding=0, dilation=1, ceil_mode=False)
50
+ (9): Conv2d(16, 16, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
51
+ (10): BatchNorm2d(16, eps=1e-05, momentum=0.1, affine=True, track_running_stats=True)
52
+ (11): MaxPool2d(kernel_size=2, stride=2, padding=0, dilation=1, ceil_mode=False)
53
+ (12): Conv2d(16, 16, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
54
+ (13): BatchNorm2d(16, eps=1e-05, momentum=0.1, affine=True, track_running_stats=True)
55
+ (14): AvgPool2d(kernel_size=2, stride=2, padding=0)
56
+ )
57
+ (linear_block): Sequential(
58
+ (0): Linear(in_features=784, out_features=16, bias=True)
59
+ (1): BatchNorm1d(16, eps=1e-05, momentum=0.1, affine=True, track_running_stats=True)
60
+ (2): ReLU()
61
+ (3): Dropout(p=0.2, inplace=False)
62
+ (4): Linear(in_features=16, out_features=16, bias=True)
63
+ (5): BatchNorm1d(16, eps=1e-05, momentum=0.1, affine=True, track_running_stats=True)
64
+ (6): ReLU()
65
+ (7): Dropout(p=0.2, inplace=False)
66
+ )
67
+ (output_block): Linear(in_features=16, out_features=4, bias=True)
68
+ )
69
+ ```