Update README.md
Browse files
README.md
CHANGED
|
@@ -17,78 +17,36 @@ This invasion is especially worrying due to the
|
|
| 17 |
---
|
| 18 |
# Dataset Description
|
| 19 |
|
| 20 |
-
My dataset consists of 3 live feed videos of Southern Sea Otter enclosures. One video is from the Georgia Aquarium, and two are from the Monterey Bay Aquarium. The dataset consists of 348 raw images, and around 900 after the augmentations were added. Within all 3 videos, Southern Sea Otters are the only class present and there are 3 - 5 in each enclosure.
|
| 21 |
-
|
| 22 |
-
Within each enclosure, my 3 defined zones of interest are present. These zones are water, land, and seclusion. The enclosures are divided in this way to produce the most detailed tracking information possible for the model outputs.
|
| 23 |
|
| 24 |
### Metadata:
|
| 25 |
-
|
| 26 |
-
- **Source:** Georgia Aquarium
|
| 27 |
-
- **Location:** Atlanta, Georgia
|
| 28 |
-
- **Time:** Current, 24/7
|
| 29 |
-
- **Collection Method:** Live Video Feed
|
| 30 |
-
- **Classes:** Southern Sea Otter
|
| 31 |
-
- **Citation:** Sea otter webcam: Live from Georgia Aquarium: Visit Today. Georgia Aquarium. (2025, January 13). https://www.georgiaaquarium.org/webcam/southern-sea-otter-webcam/
|
| 32 |
-
|
| 33 |
-
Videos 2 & 3:
|
| 34 |
-
- **Source:** Monterey Bay Aquarium
|
| 35 |
-
- **Location:** Monterey, California
|
| 36 |
-
- **Time:** Current, 24/7
|
| 37 |
-
- **Collection Method:** Live Video Feed
|
| 38 |
-
- **Classes:** Southern Sea Otter
|
| 39 |
-
- **Citation:** Sea Otter Cam. Montereybayaquarium.org. (2025). https://www.montereybayaquarium.org/animals/live-cams/sea-otter-cam
|
| 40 |
|
| 41 |
---
|
| 42 |
# Model Selection
|
| 43 |
### YOLOv11 Object Detection Model
|
| 44 |
-
I chose to use YOLOv11 to create a custom object detection model. The model was trained to detect the presence of Southern Sea Otters in water and on land. I chose an object detection model because it would allow me to add additional features to my model through supplementary code snippets. The features would produce additional visualizations to better interpret movement patterns and behaviors from the dataset. I manipulated the code [found in our textbook](https://oceancv.org/) to better identify and correctly label the otters in lower resolution cam footage, as well as in both land and water. These are my model parameters.
|
| 45 |
-
```python
|
| 46 |
-
# Train the model
|
| 47 |
-
results = model.train(
|
| 48 |
-
data=dataset_config,
|
| 49 |
-
epochs=100,
|
| 50 |
-
batch=64,
|
| 51 |
-
imgsz=640,
|
| 52 |
-
plots=True,
|
| 53 |
-
patience=50,
|
| 54 |
-
conf=0.55,
|
| 55 |
-
iou=0.5
|
| 56 |
-
)
|
| 57 |
-
```
|
| 58 |
-
### Additional Features: Ultralytics Objects Counting in Regions & Heatmap
|
| 59 |
-
Additionally, I utilized code from Ultralytics to produce an object counting in region video. I defined the 3 regions of interest (land, water, seclusion) through pixel coordinates over the input video, meaning the model was then able to differentiate between each zone and produce unique counts for each. I was able to manipulate the base code to track all 3 regions at once.
|
| 60 |
|
| 61 |
-
|
|
|
|
| 62 |
|
| 63 |
---
|
| 64 |
# Model Assessment
|
| 65 |
### Here are the metrics I used to assess the accuracy and performance of my model during training.
|
| 66 |
|
| 67 |
### Confusion Matrix
|
| 68 |
-
The final confusion matrix shows that my model was very successful at identifying sea otters across the dataset. There are a portion of mislabels where the model mistook a sea otter for background, but this can be expected with the quality of training images and smaller dataset.
|
| 69 |
-
https://huggingface.co/OceanCV/Southern_Sea_Otter_Tracking/blob/main/confusion_matrix_final.png
|
| 70 |
-

|
| 71 |
|
| 72 |
### F1 Score
|
| 73 |
-
The final F1 curve shows my model’s high precision and recall across the various confidence levels. The curve had a high peak, signifying a harmonic balance between precision and recall.
|
| 74 |
-
https://huggingface.co/OceanCV/Southern_Sea_Otter_Tracking/blob/main/F1_curve.png
|
| 75 |
-

|
| 76 |
|
| 77 |
### Object Detection Model Output
|
| 78 |
-
My final object detection output video was a key metric in assessing the performance of my model. I bounced between looking at the output video, assessing how accurate the bounding boxes and identifications were, and rerunning the model with modified parameters. My final model output was successful at identifying sea otters in both land and water, with minimal misclassifications or missed detections. The final video can be found in my respository or downloaded [here](https://huggingface.co/OceanCV/Southern_Sea_Otter_Tracking/resolve/main/object_detection_final.avi?download=true).
|
| 79 |
|
| 80 |
---
|
| 81 |
# Model Use-case
|
| 82 |
-
**Example Proposal:**
|
| 83 |
-
|
| 84 |
-
**Example Hypothesis:** Southern Sea Otters will congregate in the water zone most frequently and for the most time, and will rarely enter the seclusion zone.
|
| 85 |
|
| 86 |
-
**Example
|
|
|
|
| 87 |
|
| 88 |
-
**Model Justification:**
|
| 89 |
|
| 90 |
---
|
| 91 |
# Disclaimer
|
| 92 |
-
Credit for all videos used in the dataset go to
|
| 93 |
-
- **Video 1:** [Link to Youtube video used](https://www.youtube.com/watch?v=bfjxsHrgeQs)
|
| 94 |
-
- **Video 2 & 3:** [Link to Youtube video used](https://www.youtube.com/watch?v=9JlkEzxk6Qc)
|
|
|
|
| 17 |
---
|
| 18 |
# Dataset Description
|
| 19 |
|
|
|
|
|
|
|
|
|
|
| 20 |
|
| 21 |
### Metadata:
|
| 22 |
+
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 23 |
|
| 24 |
---
|
| 25 |
# Model Selection
|
| 26 |
### YOLOv11 Object Detection Model
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 27 |
|
| 28 |
+
|
| 29 |
+
### Additional Features: Ultralytics Objects Counting in Regions & Heatmap
|
| 30 |
|
| 31 |
---
|
| 32 |
# Model Assessment
|
| 33 |
### Here are the metrics I used to assess the accuracy and performance of my model during training.
|
| 34 |
|
| 35 |
### Confusion Matrix
|
|
|
|
|
|
|
|
|
|
| 36 |
|
| 37 |
### F1 Score
|
|
|
|
|
|
|
|
|
|
| 38 |
|
| 39 |
### Object Detection Model Output
|
|
|
|
| 40 |
|
| 41 |
---
|
| 42 |
# Model Use-case
|
| 43 |
+
**Example Proposal:**
|
|
|
|
|
|
|
| 44 |
|
| 45 |
+
**Example Hypothesis:**
|
| 46 |
+
**Example Impact:**
|
| 47 |
|
| 48 |
+
**Model Justification:**
|
| 49 |
|
| 50 |
---
|
| 51 |
# Disclaimer
|
| 52 |
+
Credit for all videos used in the dataset go to
|
|
|
|
|
|