Commit ·
bec7206
1
Parent(s): d7fa16f
add missing info
Browse files
README.md
CHANGED
|
@@ -7,7 +7,7 @@ license: cc-by-nc-4.0
|
|
| 7 |
---
|
| 8 |
|
| 9 |
To be clear, this model is tailored to my image and video classification tasks, not to imagenet.
|
| 10 |
-
I built EfficientNetV2.5 s to outperform the existing EfficientNet b0 to b4 and EfficientNetV2 t to l models, whether
|
| 11 |
|
| 12 |
## Model Details
|
| 13 |
- **Model tasks:** Image classification / video classification / feature backbone
|
|
@@ -15,11 +15,12 @@ I built EfficientNetV2.5 s to outperform the existing EfficientNet b0 to b4 and
|
|
| 15 |
- Params: 16.64 M
|
| 16 |
- Multiply-Add Operations: 5.32 G
|
| 17 |
- Image size: train = 299x299 / 304x304, test = 304x304
|
|
|
|
| 18 |
- **Papers:**
|
| 19 |
- EfficientNetV2: Smaller Models and Faster Training: https://arxiv.org/abs/2104.00298
|
| 20 |
- Layer-adaptive sparsity for the Magnitude-based Pruning: https://arxiv.org/abs/2010.07611
|
| 21 |
- **Dataset:** ImageNet-1k
|
| 22 |
-
- **Pretrained:** Yes, but requires
|
| 23 |
- **Original:** This model architecture is original
|
| 24 |
|
| 25 |
<br>
|
|
@@ -61,4 +62,5 @@ model = torch.load(model_name)
|
|
| 61 |
# Top-1 Accuracy Comparisons
|
| 62 |
`efficientnet_b3_pruned` achieved the second highest top-1 accuracy as well as the highest epoch-1 training accuracy on my task, out of all previous EfficientNet models my 24 GB VRAM RTX 3090 could handle.
|
| 63 |
|
| 64 |
-
I will publish the detailed report in another model repository, including the link to the GVNS benchmarks.
|
|
|
|
|
|
| 7 |
---
|
| 8 |
|
| 9 |
To be clear, this model is tailored to my image and video classification tasks, not to imagenet.
|
| 10 |
+
I built EfficientNetV2.5 s to outperform the existing EfficientNet b0 to b4, EfficientNet b1 to b4 pruned (I pruned b4), and EfficientNetV2 t to l models, whether trained using TensorFlow or PyTorch, in terms of top-1 accuracy, efficiency, and robustness on my datasets and GVNS benchmarks.
|
| 11 |
|
| 12 |
## Model Details
|
| 13 |
- **Model tasks:** Image classification / video classification / feature backbone
|
|
|
|
| 15 |
- Params: 16.64 M
|
| 16 |
- Multiply-Add Operations: 5.32 G
|
| 17 |
- Image size: train = 299x299 / 304x304, test = 304x304
|
| 18 |
+
- Classification layer: included, and defaults to 1,000 classes
|
| 19 |
- **Papers:**
|
| 20 |
- EfficientNetV2: Smaller Models and Faster Training: https://arxiv.org/abs/2104.00298
|
| 21 |
- Layer-adaptive sparsity for the Magnitude-based Pruning: https://arxiv.org/abs/2010.07611
|
| 22 |
- **Dataset:** ImageNet-1k
|
| 23 |
+
- **Pretrained:** Yes, but requires more pretraining
|
| 24 |
- **Original:** This model architecture is original
|
| 25 |
|
| 26 |
<br>
|
|
|
|
| 62 |
# Top-1 Accuracy Comparisons
|
| 63 |
`efficientnet_b3_pruned` achieved the second highest top-1 accuracy as well as the highest epoch-1 training accuracy on my task, out of all previous EfficientNet models my 24 GB VRAM RTX 3090 could handle.
|
| 64 |
|
| 65 |
+
I will publish the detailed report in another model repository, including the link to the GVNS benchmarks.
|
| 66 |
+
This repository is only for the base model, pretrained on ImageNet, not my task.
|