Update README: Add model card metadata, ImageNet-1k metrics, and LiteRT usage example

#1
Files changed (1) hide show
  1. README.md +126 -8
README.md CHANGED
@@ -1,27 +1,145 @@
1
  ---
2
  library_name: litert
 
3
  tags:
4
  - vision
5
  - image-classification
 
 
6
  datasets:
7
  - imagenet-1k
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
8
  ---
 
9
  # EfficientNet V2 L
10
 
11
- EfficientNet V2 L model pre-trained on ImageNet-1k.
 
 
 
 
 
 
 
 
 
 
12
 
13
  ## Intended uses & limitations
14
 
15
  The model files were converted from pretrained weights from PyTorch Vision. The models may have their own licenses or terms and conditions derived from PyTorch Vision and the dataset used for training. It is your responsibility to determine whether you have permission to use the models for your use case.
16
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
17
  ### BibTeX entry and citation info
18
 
19
  ```bibtex
20
- @article{Tan2019EfficientNetRM,
21
- title={EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks},
22
- author={Mingxing Tan and Quoc V. Le},
23
- journal={ArXiv},
24
- year={2019},
25
- volume={abs/1905.11946}
 
26
  }
27
- ```
 
1
  ---
2
  library_name: litert
3
+ pipeline_tag: image-classification
4
  tags:
5
  - vision
6
  - image-classification
7
+ - google
8
+ - computer-vision
9
  datasets:
10
  - imagenet-1k
11
+ model-index:
12
+ - name: litert-community/efficientnet_v2_l
13
+ results:
14
+ - task:
15
+ type: image-classification
16
+ name: Image Classification
17
+ dataset:
18
+ name: ImageNet-1k
19
+ type: imagenet-1k
20
+ config: default
21
+ split: validation
22
+ metrics:
23
+ - name: Top 1 Accuracy (Full Precision)
24
+ type: accuracy
25
+ value: 0.8581
26
+ - name: Top 5 Accuracy (Full Precision)
27
+ type: accuracy
28
+ value: 0.9779
29
+ - name: Top 1 Accuracy (Dynamic Quantized wi8 afp32)
30
+ type: accuracy
31
+ value: 0.8576
32
+ - name: Top 5 Accuracy (Dynamic Quantized wi8 afp32)
33
+ type: accuracy
34
+ value: 0.9779
35
  ---
36
+
37
  # EfficientNet V2 L
38
 
39
+ The EfficientNetV2-L is a high-capacity model pre-trained on ImageNet-1k, originally introduced by Tan, Mingxing, Le and Quoc in the 2021 paper, [**EfficientNetV2: Smaller Models and Faster Training**](https://arxiv.org/abs/2104.00298). This architecture evolves the original compound scaling formula by incorporating Fused-MBConv layers and progressive learning—a method that dynamically adjusts image resolution and regularization during training.
40
+
41
+
42
+ ## Model description
43
+
44
+ The model was converted from a checkpoint from PyTorch Vision.
45
+
46
+ The original model has:
47
+ acc@1 (on ImageNet-1K): 85.81%
48
+ acc@5 (on ImageNet-1K): 97.79%
49
+ num_params: 11,8515,272
50
 
51
  ## Intended uses & limitations
52
 
53
  The model files were converted from pretrained weights from PyTorch Vision. The models may have their own licenses or terms and conditions derived from PyTorch Vision and the dataset used for training. It is your responsibility to determine whether you have permission to use the models for your use case.
54
 
55
+ ## How to Use
56
+
57
+ ​​**1. Install Dependencies** Ensure your Python environment is set up with the required libraries. Run the following command in your terminal:
58
+
59
+ ```bash
60
+ pip install numpy Pillow huggingface_hub ai-edge-litert
61
+ ```
62
+
63
+ **2. Prepare Your Image** The script expects an image file to analyze. Make sure you have an image (e.g., cat.jpg or car.png) saved in the same working directory as your script.
64
+
65
+
66
+ **3. Save the Script** Create a new file named `classify.py`, paste the script below into it, and save the file:
67
+
68
+ ```python
69
+ #!/usr/bin/env python3
70
+ import argparse, json
71
+ import numpy as np
72
+ from PIL import Image
73
+ from huggingface_hub import hf_hub_download
74
+ from ai_edge_litert.compiled_model import CompiledModel
75
+
76
+ def preprocess(img: Image.Image) -> np.ndarray:
77
+ img = img.convert("RGB")
78
+ w, h = img.size
79
+ s = 480
80
+ if w < h:
81
+ img = img.resize((s, int(round(h * s / w))), Image.BICUBIC)
82
+ else:
83
+ img = img.resize((int(round(w * s / h)), s), Image.BICUBIC)
84
+ left = (img.size[0] - 480) // 2
85
+ top = (img.size[1] - 480) // 2
86
+ img = img.crop((left, top, left + 480, top + 480))
87
+
88
+ x = np.asarray(img, dtype=np.float32) / 255.0
89
+ x = (x - np.array([0.5, 0.5, 0.5], dtype=np.float32)) / np.array(
90
+ [0.5, 0.5, 0.5], dtype=np.float32
91
+ )
92
+ return np.transpose(x, (2, 0, 1))
93
+
94
+ def main():
95
+ ap = argparse.ArgumentParser()
96
+ ap.add_argument("--image", required=True)
97
+ args = ap.parse_args()
98
+
99
+ model_path = hf_hub_download("litert-community/efficientnet_v2_l", "efficientnet_v2_l.tflite")
100
+ labels_path = hf_hub_download(
101
+ "huggingface/label-files", "imagenet-1k-id2label.json", repo_type="dataset"
102
+ )
103
+ with open(labels_path, "r", encoding="utf-8") as f:
104
+ id2label = {int(k): v for k, v in json.load(f).items()}
105
+
106
+ img = Image.open(args.image)
107
+ x = preprocess(img)
108
+
109
+ model = CompiledModel.from_file(model_path)
110
+ inp = model.create_input_buffers(0)
111
+ out = model.create_output_buffers(0)
112
+
113
+ inp[0].write(x)
114
+ model.run_by_index(0, inp, out)
115
+
116
+ req = model.get_output_buffer_requirements(0, 0)
117
+ y = out[0].read(req["buffer_size"] // np.dtype(np.float32).itemsize, np.float32)
118
+
119
+ pred = int(np.argmax(y))
120
+ label = id2label.get(pred, f"class_{pred}")
121
+
122
+ print(f"Top-1 class index: {pred}")
123
+ print(f"Top-1 label: {label}")
124
+ if __name__ == "__main__":
125
+ main()
126
+ ```
127
+
128
+ **4. Execute the Python Script** Run the below command:
129
+
130
+ ```bash
131
+ python classify.py --image cat.jpg
132
+ ```
133
+
134
  ### BibTeX entry and citation info
135
 
136
  ```bibtex
137
+ @inproceedings{tan2021efficientnetv2,
138
+ title={Efficientnetv2: Smaller models and faster training},
139
+ author={Tan, Mingxing and Le, Quoc},
140
+ booktitle={International conference on machine learning},
141
+ pages={10096--10106},
142
+ year={2021},
143
+ organization={PMLR}
144
  }
145
+ ```