Update README.md
Browse files
README.md
CHANGED
|
@@ -14,13 +14,13 @@ dataset_info:
|
|
| 14 |
sequence: float64
|
| 15 |
splits:
|
| 16 |
- name: train
|
| 17 |
-
num_bytes: 103638330
|
| 18 |
num_examples: 82
|
| 19 |
- name: valid
|
| 20 |
-
num_bytes: 26074864
|
| 21 |
num_examples: 21
|
| 22 |
download_size: 124824112
|
| 23 |
-
dataset_size: 129713194
|
| 24 |
configs:
|
| 25 |
- config_name: default
|
| 26 |
data_files:
|
|
@@ -28,36 +28,25 @@ configs:
|
|
| 28 |
path: data/train-*
|
| 29 |
- split: valid
|
| 30 |
path: data/valid-*
|
| 31 |
-
task_categories:
|
| 32 |
-
- image-segmentation
|
| 33 |
-
tags:
|
| 34 |
-
- seafood-processing
|
| 35 |
-
- industrial-vision
|
| 36 |
-
- fish
|
| 37 |
-
- food-processing
|
| 38 |
-
pretty_name: FishGrade
|
| 39 |
-
size_categories:
|
| 40 |
-
- 1K<n<10K
|
| 41 |
---
|
| 42 |
|
| 43 |
|
| 44 |
|
| 45 |
---
|
| 46 |
# Vision-Guided Robotic System for Automatic Fish Quality Grading and Packaging
|
| 47 |
-
This dataset
|
| 48 |
-
|
| 49 |
-
In our paper (link below). We applied instance segmentation to isolate the fish steaks based on YOLOv8 (Check [here](https://docs.ultralytics.com/models/yolov8/)) how to train and validate the model).
|
| 50 |
-
Once the fish steaks are segmented, we simply measure their size by leveraging the depth data contained in the BAG files.
|
| 51 |
|
| 52 |
-
🤗 [Paper on Hugging Face]
|
| 53 |
|
| 54 |
## 🗂️ BAG files & trained segmentation model:
|
| 55 |
-
Please
|
| 56 |
|
| 57 |
-
The BAG files
|
| 58 |
|
| 59 |
-
|
| 60 |
-
|
|
|
|
| 61 |
|
| 62 |
```python
|
| 63 |
import pyrealsense2 as rs
|
|
@@ -212,8 +201,8 @@ if __name__ == '__main__':
|
|
| 212 |
|
| 213 |
## 🗂️ Data Instances
|
| 214 |
<figure style="display:flex; gap:10px; flex-wrap:wrap; justify-content:center;">
|
| 215 |
-
<img src="Figure_1.png" width="
|
| 216 |
-
<img src="Figure_2.png" width="
|
| 217 |
</figure>
|
| 218 |
|
| 219 |
## 🏷️ Annotation Format
|
|
|
|
| 14 |
sequence: float64
|
| 15 |
splits:
|
| 16 |
- name: train
|
| 17 |
+
num_bytes: 103638330.0
|
| 18 |
num_examples: 82
|
| 19 |
- name: valid
|
| 20 |
+
num_bytes: 26074864.0
|
| 21 |
num_examples: 21
|
| 22 |
download_size: 124824112
|
| 23 |
+
dataset_size: 129713194.0
|
| 24 |
configs:
|
| 25 |
- config_name: default
|
| 26 |
data_files:
|
|
|
|
| 28 |
path: data/train-*
|
| 29 |
- split: valid
|
| 30 |
path: data/valid-*
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 31 |
---
|
| 32 |
|
| 33 |
|
| 34 |
|
| 35 |
---
|
| 36 |
# Vision-Guided Robotic System for Automatic Fish Quality Grading and Packaging
|
| 37 |
+
This dataset (recorded with a Realsense D456 camera), associated with our work accepted in the 'IEEE/CAA Journal of Automatica Sinica', includes images and corresponding instance segmentation annotations (in YOLO format) of hake fish steaks on an industrial conveyor belt. It also provides BAG files for two quality grades of fish steaks (A and B), where A-grade steaks are generally larger.
|
| 38 |
+
The paper details our use of YOLOv8 instance segmentation (Check [here](https://docs.ultralytics.com/models/yolov8/) how to train and validate the model) to isolate the fish steaks and the subsequent measurement of their size using the depth data from the BAG files.
|
|
|
|
|
|
|
| 39 |
|
| 40 |
+
🤗 [Paper on Hugging Face]Coming soon ... | 📝 [Paper on ArXiv] Coming soon ...
|
| 41 |
|
| 42 |
## 🗂️ BAG files & trained segmentation model:
|
| 43 |
+
Please first read the associated paper to understand the proposed pipeline.
|
| 44 |
|
| 45 |
+
The BAG files for A and B grades, as well as the weights of the trained segmentation model (best.pt and last.pt), can be found [here.](https://fbk-my.sharepoint.com/:f:/g/personal/mmekhalfi_fbk_eu/ElmBGeHUIwpPveSRrfd7qu4BQpAiWsOo70m8__V875yggw?e=1L0iTT).
|
| 46 |
|
| 47 |
+
The segmentation model is designed to segment fish samples. The BAG files are intended for testing purposes. For example, you could use the provided model weights to segment the RGB images within the BAG files and then measure their size based on the depth data.
|
| 48 |
+
|
| 49 |
+
For clarity, a simplified code snippet for measuring steaks' (metric) perimeter is provided below. You can repurpose this for your specific task:
|
| 50 |
|
| 51 |
```python
|
| 52 |
import pyrealsense2 as rs
|
|
|
|
| 201 |
|
| 202 |
## 🗂️ Data Instances
|
| 203 |
<figure style="display:flex; gap:10px; flex-wrap:wrap; justify-content:center;">
|
| 204 |
+
<img src="Figure_1.png" width="45%" alt="Raspberry Example 1">
|
| 205 |
+
<img src="Figure_2.png" width="45%" alt="Raspberry Example 2">
|
| 206 |
</figure>
|
| 207 |
|
| 208 |
## 🏷️ Annotation Format
|