Object Detection
ultralytics
English
biology
CV
images
animals
YOLO
fine-tuned
zebra
giraffe
onager
dog
jennamk14 commited on
Commit
e02fe7c
·
verified ·
1 Parent(s): 495f300

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +275 -3
README.md CHANGED
@@ -1,3 +1,275 @@
1
- ---
2
- license: mit
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ language:
4
+ - en
5
+ library_name: ultralytics
6
+ tags:
7
+ - biology
8
+ - CV
9
+ - images
10
+ - animals
11
+ - object-detection
12
+ - YOLO
13
+ - fine-tuned
14
+ datasets:
15
+ - custom_animal_dataset
16
+ metrics:
17
+ - precision
18
+ - recall
19
+ - mAP50
20
+ - mAP50-95
21
+ ---
22
+
23
+ # Model Card for Fine-Tuned YOLOv11m Animal Detection Model
24
+
25
+ This model is a fine-tuned version of YOLOv11m optimized for detection and classification of wildlife from low-altitude drone imagery. It has been trained to identify zebras (Plains and Grevy's), giraffes, Persian onagers, and African Painted dogs with high accuracy across diverse environmental conditions.
26
+
27
+ ## Model Details
28
+
29
+ ### Model Description
30
+
31
+ - **Developed by:** Jenna Kline
32
+ - **Model type:** Object Detection and Classification
33
+ - **Language(s) (NLP):** Not applicable (Computer Vision model)
34
+ - **License:** CC0-1.0
35
+ - **Fine-tuned from model:** YOLOv11m (ultralytics/yolo11m.pt)
36
+
37
+ ### Model Sources
38
+
39
+ - **Repository:** https://github.com/Imageomics-ABC-edu/final-project-kenyan-ungulates-with-wilddroneeu
40
+ - **Paper:** [MMLA](https://arxiv.org/abs/2504.07744)
41
+ - **Demo:** TO DO
42
+
43
+ ## Uses
44
+
45
+ ### Direct Use
46
+
47
+ This model is designed for direct use in wildlife monitoring applications, ecological research, and biodiversity studies. It can:
48
+
49
+ - Detect and classify zebras, giraffes, onagers, and dogs in camera trap images
50
+ - Monitor wildlife populations in their natural habitats
51
+ - Assist researchers in automated processing of large image datasets
52
+ - Support biodiversity assessments by identifying species present in field surveys
53
+
54
+ The model can be used by researchers, conservationists, wildlife managers, and citizen scientists to automate and scale up wildlife monitoring efforts, particularly in African ecosystems.
55
+
56
+ ### Downstream Use
57
+
58
+ This model can be integrated into larger ecological monitoring systems including:
59
+ - Automated camera trap processing pipelines
60
+ - Wildlife conservation monitoring platforms
61
+ - Ecological research workflows
62
+ - Citizen science applications for species identification
63
+ - Environmental impact assessment tools
64
+
65
+ ### Out-of-Scope Use
66
+
67
+ This model is not suitable for:
68
+ - Medical diagnosis or human-related detection tasks
69
+ - Security or surveillance applications targeting humans
70
+ - Applications where errors in detection could lead to harmful conservation decisions without human verification
71
+ - Real-time detection systems requiring extremely low latency (model prioritizes accuracy over speed)
72
+ - Detection of species not included in the training set (only trained on zebras, giraffes, onagers, and dogs)
73
+
74
+ ## Bias, Risks, and Limitations
75
+
76
+ - **Species representation bias:** The model may perform better on species that were well-represented in the training data.
77
+ - **Environmental bias:** Performance may degrade in environmental conditions not represented in the training data (e.g., extreme weather, unusual lighting).
78
+ - **Morphological bias:** Similar-looking species may be confused with one another (particularly among equids like zebras and onagers).
79
+ - **Geospatial bias:** The model may perform better in biomes similar to those present in the training data, particularly African savanna environments.
80
+ - **Seasonal bias:** Detection accuracy may vary based on seasonal appearance changes in animals or environments.
81
+ - **Technical limitations:** Performance depends on image quality, with reduced accuracy in low-resolution, blurry, or poorly exposed images.
82
+
83
+ ### Recommendations
84
+
85
+ Users (both direct and downstream) should be made aware of the risks, biases, and limitations of the model:
86
+ - Always verify critical detections with human review, especially for rare species or conservation decision-making
87
+ - Consider confidence scores when evaluating detections
88
+ - Be cautious when applying the model to new geographic regions or habitats not represented in training data
89
+ - Periodically validate model performance on new data to ensure continued reliability
90
+ - Consider fine-tuning the model on domain-specific data when applying to new regions or species
91
+
92
+ ## How to Get Started with the Model
93
+
94
+ Use the code below to get started with the model:
95
+
96
+ ```python
97
+ from ultralytics import YOLO
98
+
99
+ # Load the model
100
+ model = YOLO('path/to/your/model.pt')
101
+
102
+ # Run inference on an image
103
+ results = model('path/to/image.jpg')
104
+
105
+ # Process results
106
+ for result in results:
107
+ boxes = result.boxes # Boxes object for bounding boxes outputs
108
+ for box in boxes:
109
+ x1, y1, x2, y2 = box.xyxy[0] # get box coordinates
110
+ conf = box.conf[0] # confidence score
111
+ cls = int(box.cls[0]) # class id
112
+ class_name = model.names[cls] # class name (Zebra, Giraffe, Onager, or Dog)
113
+ print(f"Detected {class_name} with confidence {conf:.2f} at position {x1:.1f}, {y1:.1f}, {x2:.1f}, {y2:.1f}")
114
+
115
+ # Visualize results
116
+ results[0].plot()
117
+ ```
118
+
119
+ ## Training Details
120
+
121
+ ### Training Data
122
+
123
+ Dataset is available at [Hugging Face](https://huggingface.co/collections/imageomics/wildwing-67f572d3ba17fca922c80182). See /data/dataset.yaml for details on train/val/test splits.
124
+
125
+ ### Training Procedure
126
+
127
+ #### Preprocessing
128
+
129
+ - Images were resized to 640x640 pixels (as specified in the training script)
130
+ - Standard YOLOv11 augmentation pipeline was applied
131
+
132
+ #### Training Hyperparameters
133
+
134
+ The model was trained with the following hyperparameters as specified in the training script:
135
+ - **Base model:** YOLOv11m (yolo11m.pt)
136
+ - **Epochs:** 100
137
+ - **Image size:** 640
138
+ - **Dataset configuration:** Custom YAML file defining 4 classes (Zebra, Giraffe, Onager, Dog)
139
+ - **Training regime:** Default YOLOv11 training parameters
140
+
141
+ ```python
142
+ # Training script
143
+ from ultralytics import YOLO
144
+
145
+ model = YOLO("yolo11m.pt")
146
+ results = model.train(
147
+ data="/data/dataset.yaml",
148
+ epochs=100,
149
+ imgsz=640,
150
+ )
151
+ ```
152
+
153
+ #### Speeds, Sizes, Times
154
+
155
+ - **Training hardware:** [Your GPU/CPU specifications]
156
+ - **Training time:** [Duration]
157
+ - **Model size:** [Size in MB]
158
+ - **Inference speed:** [FPS on specific hardware]
159
+
160
+ ## Evaluation
161
+
162
+ ### Testing Data, Factors & Metrics
163
+
164
+ #### Testing Data
165
+
166
+ The model was evaluated on a held-out test set located at `/fs/ess/PAS2136/Kenya-2023/yolo_benchmark/HerdYOLO/data/images/test` containing:
167
+ - [Number] test images with instances of Zebra, Giraffe, Onager, and Dog
168
+ - [Any other relevant testing data details]
169
+
170
+ #### Factors
171
+
172
+ The evaluation disaggregated performance by:
173
+ - Species (Zebra, Giraffe, Onager, Dog)
174
+ - [Other evaluation factors if applicable]
175
+
176
+ #### Metrics
177
+
178
+ The model was evaluated using standard object detection metrics:
179
+ - **Precision:** Ratio of true positives to all predicted positives
180
+ - **Recall:** Ratio of true positives to all actual positives (ground truth)
181
+ - **mAP50:** Mean Average Precision at IoU threshold of 0.5
182
+ - **mAP50-95:** Mean Average Precision averaged over IoU thresholds from 0.5 to 0.95
183
+
184
+ ### Results
185
+
186
+ #### Summary
187
+
188
+ - **Overall mAP50:** [Value]
189
+ - **Overall mAP50-95:** [Value]
190
+ - **Per-class performance:**
191
+ - Zebra: mAP50 = [Value], Precision = [Value], Recall = [Value]
192
+ - Giraffe: mAP50 = [Value], Precision = [Value], Recall = [Value]
193
+ - Onager: mAP50 = [Value], Precision = [Value], Recall = [Value]
194
+ - Dog: mAP50 = [Value], Precision = [Value], Recall = [Value]
195
+
196
+ ## Model Examination
197
+
198
+ - **Confusion analysis:** [Any notable confusion between classes, such as between Zebra and Onager]
199
+ - **Failure cases:** [Specific conditions where the model performs less reliably]
200
+ - **Interpretability findings:** [Any insights from model interpretation techniques]
201
+
202
+ ## Environmental Impact
203
+
204
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://doi.org/10.48550/arXiv.1910.09700).
205
+
206
+ - **Hardware Type:** [GPU model]
207
+ - **Hours used:** [Number]
208
+ - **Cloud Provider:** [Provider name or local]
209
+ - **Compute Region:** [Region]
210
+ - **Carbon Emitted:** [Amount] kg CO₂eq
211
+
212
+ ## Technical Specifications
213
+
214
+ ### Model Architecture and Objective
215
+
216
+ - Base architecture: YOLOv11m
217
+ - Detection heads: Standard YOLOv11 architecture
218
+ - Classes: 4 (Zebra, Giraffe, Onager, Dog)
219
+
220
+ ### Compute Infrastructure
221
+
222
+ #### Hardware
223
+
224
+ - **Training:** [GPU/CPU details]
225
+ - **Inference:** Tested on [range of devices]
226
+ - **Minimum requirements:** [Specifications]
227
+
228
+ #### Software
229
+
230
+ - Python 3.8+
231
+ - PyTorch 2.0+
232
+ - Ultralytics YOLOv11 framework
233
+ - CUDA 11.7+ (for GPU acceleration)
234
+
235
+ ## Citation
236
+
237
+ **BibTeX:**
238
+
239
+ ```
240
+ @software{your_name_yolo11m_2025,
241
+ author = {Your Name},
242
+ title = {Fine-Tuned YOLOv11m Animal Detection Model},
243
+ version = {1.0.0},
244
+ year = {2025},
245
+ url = {https://huggingface.co/your-username/your-model-name}
246
+ }
247
+ ```
248
+
249
+ ## Acknowledgements
250
+
251
+ This work was supported by both the [Imageomics Institute](https://imageomics.org) and the [AI and Biodiversity Change (ABC) Global Center](http://abcresearchcenter.org). The Imageomics Institute is funded by the US National Science Foundation's Harnessing the Data Revolution (HDR) program under [Award #2118240](https://www.nsf.gov/awardsearch/showAward?AWD_ID=2118240) (Imageomics: A New Frontier of Biological Information Powered by Knowledge-Guided Machine Learning). The ABC Global Center is funded by the US National Science Foundation under [Award No. 2330423](https://www.nsf.gov/awardsearch/showAward?AWD_ID=2330423&HistoricalAwards=false) and Natural Sciences and Engineering Research Council of Canada under [Award No. 585136](https://www.nserc-crsng.gc.ca/ase-oro/Details-Detailles_eng.asp?id=782440). This model draws on research supported by the Social Sciences and Humanities Research Council.
252
+
253
+ Additional support was provided by the National Ecological Observatory Network (NEON), a program sponsored by the National Science Foundation and operated under cooperative agreement by Battelle Memorial Institute. This material is based in part upon work supported by the National Science Foundation through the NEON Program.
254
+
255
+ Any opinions, findings and conclusions or recommendations expressed in this material are those of the author(s) and do not necessarily reflect the views of the National Science Foundation, Natural Sciences and Engineering Research Council of Canada, or Social Sciences and Humanities Research Council.
256
+
257
+ ## Glossary
258
+
259
+ - **YOLO:** You Only Look Once, a family of real-time object detection models
260
+ - **mAP:** mean Average Precision, a standard metric for evaluating object detection models
261
+ - **IoU:** Intersection over Union, a measure of overlap between predicted and ground truth bounding boxes
262
+ - **Onager:** Also known as the Asian wild ass, a species of equid native to Asia
263
+ - **YOLOv11m:** The medium-sized variant of the YOLOv11 architecture
264
+
265
+ ## More Information
266
+
267
+ [Any additional information you'd like to include]
268
+
269
+ ## Model Card Authors
270
+
271
+ Jenna Kline, The Ohio State University
272
+
273
+ ## Model Card Contact
274
+
275
+ kline.377 at osu dot edu