manya-codes keremberke commited on
Commit
0d8689e
·
0 Parent(s):

Duplicate from keremberke/forklift-object-detection

Browse files

Co-authored-by: Kerem <keremberke@users.noreply.huggingface.co>

.gitattributes ADDED
@@ -0,0 +1,54 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ *.7z filter=lfs diff=lfs merge=lfs -text
2
+ *.arrow filter=lfs diff=lfs merge=lfs -text
3
+ *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ckpt filter=lfs diff=lfs merge=lfs -text
6
+ *.ftz filter=lfs diff=lfs merge=lfs -text
7
+ *.gz filter=lfs diff=lfs merge=lfs -text
8
+ *.h5 filter=lfs diff=lfs merge=lfs -text
9
+ *.joblib filter=lfs diff=lfs merge=lfs -text
10
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
+ *.lz4 filter=lfs diff=lfs merge=lfs -text
12
+ *.mlmodel filter=lfs diff=lfs merge=lfs -text
13
+ *.model filter=lfs diff=lfs merge=lfs -text
14
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
15
+ *.npy filter=lfs diff=lfs merge=lfs -text
16
+ *.npz filter=lfs diff=lfs merge=lfs -text
17
+ *.onnx filter=lfs diff=lfs merge=lfs -text
18
+ *.ot filter=lfs diff=lfs merge=lfs -text
19
+ *.parquet filter=lfs diff=lfs merge=lfs -text
20
+ *.pb filter=lfs diff=lfs merge=lfs -text
21
+ *.pickle filter=lfs diff=lfs merge=lfs -text
22
+ *.pkl filter=lfs diff=lfs merge=lfs -text
23
+ *.pt filter=lfs diff=lfs merge=lfs -text
24
+ *.pth filter=lfs diff=lfs merge=lfs -text
25
+ *.rar filter=lfs diff=lfs merge=lfs -text
26
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
27
+ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
28
+ *.tar.* filter=lfs diff=lfs merge=lfs -text
29
+ *.tflite filter=lfs diff=lfs merge=lfs -text
30
+ *.tgz filter=lfs diff=lfs merge=lfs -text
31
+ *.wasm filter=lfs diff=lfs merge=lfs -text
32
+ *.xz filter=lfs diff=lfs merge=lfs -text
33
+ *.zip filter=lfs diff=lfs merge=lfs -text
34
+ *.zst filter=lfs diff=lfs merge=lfs -text
35
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ # Audio files - uncompressed
37
+ *.pcm filter=lfs diff=lfs merge=lfs -text
38
+ *.sam filter=lfs diff=lfs merge=lfs -text
39
+ *.raw filter=lfs diff=lfs merge=lfs -text
40
+ # Audio files - compressed
41
+ *.aac filter=lfs diff=lfs merge=lfs -text
42
+ *.flac filter=lfs diff=lfs merge=lfs -text
43
+ *.mp3 filter=lfs diff=lfs merge=lfs -text
44
+ *.ogg filter=lfs diff=lfs merge=lfs -text
45
+ *.wav filter=lfs diff=lfs merge=lfs -text
46
+ # Image files - uncompressed
47
+ *.bmp filter=lfs diff=lfs merge=lfs -text
48
+ *.gif filter=lfs diff=lfs merge=lfs -text
49
+ *.png filter=lfs diff=lfs merge=lfs -text
50
+ *.tiff filter=lfs diff=lfs merge=lfs -text
51
+ # Image files - compressed
52
+ *.jpg filter=lfs diff=lfs merge=lfs -text
53
+ *.jpeg filter=lfs diff=lfs merge=lfs -text
54
+ *.webp filter=lfs diff=lfs merge=lfs -text
README.dataset.txt ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # undefined > raw-images
2
+ https://public.roboflow.ai/object-detection/undefined
3
+
4
+ Provided by undefined
5
+ License: CC BY 4.0
6
+
7
+ ## About this Dataset
8
+ This dataset was created by exporting images from [images.cv](https://images.cv/dataset/forklift-image-classification-dataset) and labeling them as an object detection dataset. **The dataset contains 421 raw images (v1 - raw-images) and labeled classes include:**
9
+ * forklift
10
+ * person
11
+
12
+ ![Example annotated image from the dataset from the dataset](https://i.imgur.com/a6hWEG4.png)
README.md ADDED
@@ -0,0 +1,80 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ task_categories:
3
+ - object-detection
4
+ tags:
5
+ - roboflow
6
+ - roboflow2huggingface
7
+ - Manufacturing
8
+ ---
9
+
10
+ <div align="center">
11
+ <img width="640" alt="keremberke/forklift-object-detection" src="https://huggingface.co/datasets/keremberke/forklift-object-detection/resolve/main/thumbnail.jpg">
12
+ </div>
13
+
14
+ ### Dataset Labels
15
+
16
+ ```
17
+ ['forklift', 'person']
18
+ ```
19
+
20
+
21
+ ### Number of Images
22
+
23
+ ```json
24
+ {'test': 42, 'valid': 84, 'train': 295}
25
+ ```
26
+
27
+
28
+ ### How to Use
29
+
30
+ - Install [datasets](https://pypi.org/project/datasets/):
31
+
32
+ ```bash
33
+ pip install datasets
34
+ ```
35
+
36
+ - Load the dataset:
37
+
38
+ ```python
39
+ from datasets import load_dataset
40
+
41
+ ds = load_dataset("keremberke/forklift-object-detection", name="full")
42
+ example = ds['train'][0]
43
+ ```
44
+
45
+ ### Roboflow Dataset Page
46
+ [https://universe.roboflow.com/mohamed-traore-2ekkp/forklift-dsitv/dataset/1](https://universe.roboflow.com/mohamed-traore-2ekkp/forklift-dsitv/dataset/1?ref=roboflow2huggingface)
47
+
48
+ ### Citation
49
+
50
+ ```
51
+ @misc{ forklift-dsitv_dataset,
52
+ title = { Forklift Dataset },
53
+ type = { Open Source Dataset },
54
+ author = { Mohamed Traore },
55
+ howpublished = { \\url{ https://universe.roboflow.com/mohamed-traore-2ekkp/forklift-dsitv } },
56
+ url = { https://universe.roboflow.com/mohamed-traore-2ekkp/forklift-dsitv },
57
+ journal = { Roboflow Universe },
58
+ publisher = { Roboflow },
59
+ year = { 2022 },
60
+ month = { mar },
61
+ note = { visited on 2023-01-15 },
62
+ }
63
+ ```
64
+
65
+ ### License
66
+ CC BY 4.0
67
+
68
+ ### Dataset Summary
69
+ This dataset was exported via roboflow.ai on April 3, 2022 at 9:01 PM GMT
70
+
71
+ It includes 421 images.
72
+ Forklift are annotated in COCO format.
73
+
74
+ The following pre-processing was applied to each image:
75
+ * Auto-orientation of pixel data (with EXIF-orientation stripping)
76
+
77
+ No image augmentation techniques were applied.
78
+
79
+
80
+
README.roboflow.txt ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ Forklift - v1 raw-images
3
+ ==============================
4
+
5
+ This dataset was exported via roboflow.ai on April 3, 2022 at 9:01 PM GMT
6
+
7
+ It includes 421 images.
8
+ Forklift are annotated in COCO format.
9
+
10
+ The following pre-processing was applied to each image:
11
+ * Auto-orientation of pixel data (with EXIF-orientation stripping)
12
+
13
+ No image augmentation techniques were applied.
14
+
15
+
data/test.zip ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9baace711b989e84c731fa3b3e2b45f8a636880d475c2c2d6b91b8cdff803cbe
3
+ size 2771676
data/train.zip ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bf5c3364b39160d4d212a454b9a7689f8b89e81eab3ea3da50031c4e7bf92fe0
3
+ size 13533922
data/valid-mini.zip ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:94d605e3645a7bd47a91aae22a14cce0b17d40b2c9483aed5ad30be756e0c038
3
+ size 105141
data/valid.zip ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0f8013a437dc85e8f218b252ce8cffa32b2724c8021cf6b983febd3a9145d53f
3
+ size 3774165
forklift-object-detection.py ADDED
@@ -0,0 +1,152 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import collections
2
+ import json
3
+ import os
4
+
5
+ import datasets
6
+
7
+
8
+ _HOMEPAGE = "https://universe.roboflow.com/mohamed-traore-2ekkp/forklift-dsitv/dataset/1"
9
+ _LICENSE = "CC BY 4.0"
10
+ _CITATION = """\
11
+ @misc{ forklift-dsitv_dataset,
12
+ title = { Forklift Dataset },
13
+ type = { Open Source Dataset },
14
+ author = { Mohamed Traore },
15
+ howpublished = { \\url{ https://universe.roboflow.com/mohamed-traore-2ekkp/forklift-dsitv } },
16
+ url = { https://universe.roboflow.com/mohamed-traore-2ekkp/forklift-dsitv },
17
+ journal = { Roboflow Universe },
18
+ publisher = { Roboflow },
19
+ year = { 2022 },
20
+ month = { mar },
21
+ note = { visited on 2023-01-15 },
22
+ }
23
+ """
24
+ _CATEGORIES = ['forklift', 'person']
25
+ _ANNOTATION_FILENAME = "_annotations.coco.json"
26
+
27
+
28
+ class FORKLIFTOBJECTDETECTIONConfig(datasets.BuilderConfig):
29
+ """Builder Config for forklift-object-detection"""
30
+
31
+ def __init__(self, data_urls, **kwargs):
32
+ """
33
+ BuilderConfig for forklift-object-detection.
34
+
35
+ Args:
36
+ data_urls: `dict`, name to url to download the zip file from.
37
+ **kwargs: keyword arguments forwarded to super.
38
+ """
39
+ super(FORKLIFTOBJECTDETECTIONConfig, self).__init__(version=datasets.Version("1.0.0"), **kwargs)
40
+ self.data_urls = data_urls
41
+
42
+
43
+ class FORKLIFTOBJECTDETECTION(datasets.GeneratorBasedBuilder):
44
+ """forklift-object-detection object detection dataset"""
45
+
46
+ VERSION = datasets.Version("1.0.0")
47
+ BUILDER_CONFIGS = [
48
+ FORKLIFTOBJECTDETECTIONConfig(
49
+ name="full",
50
+ description="Full version of forklift-object-detection dataset.",
51
+ data_urls={
52
+ "train": "https://huggingface.co/datasets/keremberke/forklift-object-detection/resolve/main/data/train.zip",
53
+ "validation": "https://huggingface.co/datasets/keremberke/forklift-object-detection/resolve/main/data/valid.zip",
54
+ "test": "https://huggingface.co/datasets/keremberke/forklift-object-detection/resolve/main/data/test.zip",
55
+ },
56
+ ),
57
+ FORKLIFTOBJECTDETECTIONConfig(
58
+ name="mini",
59
+ description="Mini version of forklift-object-detection dataset.",
60
+ data_urls={
61
+ "train": "https://huggingface.co/datasets/keremberke/forklift-object-detection/resolve/main/data/valid-mini.zip",
62
+ "validation": "https://huggingface.co/datasets/keremberke/forklift-object-detection/resolve/main/data/valid-mini.zip",
63
+ "test": "https://huggingface.co/datasets/keremberke/forklift-object-detection/resolve/main/data/valid-mini.zip",
64
+ },
65
+ )
66
+ ]
67
+
68
+ def _info(self):
69
+ features = datasets.Features(
70
+ {
71
+ "image_id": datasets.Value("int64"),
72
+ "image": datasets.Image(),
73
+ "width": datasets.Value("int32"),
74
+ "height": datasets.Value("int32"),
75
+ "objects": datasets.Sequence(
76
+ {
77
+ "id": datasets.Value("int64"),
78
+ "area": datasets.Value("int64"),
79
+ "bbox": datasets.Sequence(datasets.Value("float32"), length=4),
80
+ "category": datasets.ClassLabel(names=_CATEGORIES),
81
+ }
82
+ ),
83
+ }
84
+ )
85
+ return datasets.DatasetInfo(
86
+ features=features,
87
+ homepage=_HOMEPAGE,
88
+ citation=_CITATION,
89
+ license=_LICENSE,
90
+ )
91
+
92
+ def _split_generators(self, dl_manager):
93
+ data_files = dl_manager.download_and_extract(self.config.data_urls)
94
+ return [
95
+ datasets.SplitGenerator(
96
+ name=datasets.Split.TRAIN,
97
+ gen_kwargs={
98
+ "folder_dir": data_files["train"],
99
+ },
100
+ ),
101
+ datasets.SplitGenerator(
102
+ name=datasets.Split.VALIDATION,
103
+ gen_kwargs={
104
+ "folder_dir": data_files["validation"],
105
+ },
106
+ ),
107
+ datasets.SplitGenerator(
108
+ name=datasets.Split.TEST,
109
+ gen_kwargs={
110
+ "folder_dir": data_files["test"],
111
+ },
112
+ ),
113
+ ]
114
+
115
+ def _generate_examples(self, folder_dir):
116
+ def process_annot(annot, category_id_to_category):
117
+ return {
118
+ "id": annot["id"],
119
+ "area": annot["area"],
120
+ "bbox": annot["bbox"],
121
+ "category": category_id_to_category[annot["category_id"]],
122
+ }
123
+
124
+ image_id_to_image = {}
125
+ idx = 0
126
+
127
+ annotation_filepath = os.path.join(folder_dir, _ANNOTATION_FILENAME)
128
+ with open(annotation_filepath, "r") as f:
129
+ annotations = json.load(f)
130
+ category_id_to_category = {category["id"]: category["name"] for category in annotations["categories"]}
131
+ image_id_to_annotations = collections.defaultdict(list)
132
+ for annot in annotations["annotations"]:
133
+ image_id_to_annotations[annot["image_id"]].append(annot)
134
+ filename_to_image = {image["file_name"]: image for image in annotations["images"]}
135
+
136
+ for filename in os.listdir(folder_dir):
137
+ filepath = os.path.join(folder_dir, filename)
138
+ if filename in filename_to_image:
139
+ image = filename_to_image[filename]
140
+ objects = [
141
+ process_annot(annot, category_id_to_category) for annot in image_id_to_annotations[image["id"]]
142
+ ]
143
+ with open(filepath, "rb") as f:
144
+ image_bytes = f.read()
145
+ yield idx, {
146
+ "image_id": image["id"],
147
+ "image": {"path": filepath, "bytes": image_bytes},
148
+ "width": image["width"],
149
+ "height": image["height"],
150
+ "objects": objects,
151
+ }
152
+ idx += 1
split_name_to_num_samples.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"test": 42, "valid": 84, "train": 295}
thumbnail.jpg ADDED

Git LFS Details

  • SHA256: 0298b2e241a1e3a65906ba6ff3c45a64cf56744ad74d06d67e5a92db421b9289
  • Pointer size: 131 Bytes
  • Size of remote file: 159 kB