| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | """PP4AV dataset.""" |
| |
|
| | import os |
| | from glob import glob |
| | from tqdm import tqdm |
| | from pathlib import Path |
| | from typing import List |
| | import re |
| | from collections import defaultdict |
| | import datasets |
| |
|
| | datasets.logging.set_verbosity_info() |
| |
|
| |
|
| |
|
| |
|
| | _HOMEPAGE = "http://shuoyang1213.me/WIDERFACE/" |
| |
|
| | _LICENSE = "Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International (CC BY-NC-ND 4.0)" |
| |
|
| | _CITATION = """\ |
| | @inproceedings{yang2016wider, |
| | Author = {Yang, Shuo and Luo, Ping and Loy, Chen Change and Tang, Xiaoou}, |
| | Booktitle = {IEEE Conference on Computer Vision and Pattern Recognition (CVPR)}, |
| | Title = {WIDER FACE: A Face Detection Benchmark}, |
| | Year = {2016}} |
| | """ |
| |
|
| | _DESCRIPTION = """\ |
| | WIDER FACE dataset is a face detection benchmark dataset, of which images are |
| | selected from the publicly available WIDER dataset. We choose 32,203 images and |
| | label 393,703 faces with a high degree of variability in scale, pose and |
| | occlusion as depicted in the sample images. WIDER FACE dataset is organized |
| | based on 61 event classes. For each event class, we randomly select 40%/10%/50% |
| | data as training, validation and testing sets. We adopt the same evaluation |
| | metric employed in the PASCAL VOC dataset. Similar to MALF and Caltech datasets, |
| | we do not release bounding box ground truth for the test images. Users are |
| | required to submit final prediction files, which we shall proceed to evaluate. |
| | """ |
| |
|
| |
|
| | _REPO = "https://huggingface.co/datasets/khaclinh/testdata/resolve/main/data" |
| | _URLS = { |
| | "test": f"{_REPO}/fisheye.zip", |
| | "annot": f"{_REPO}/annotations.zip", |
| | } |
| |
|
| | IMG_EXT = ['png', 'jpeg', 'jpg'] |
| |
|
| |
|
| | class TestData(datasets.GeneratorBasedBuilder): |
| | """WIDER FACE dataset.""" |
| |
|
| | VERSION = datasets.Version("1.0.0") |
| |
|
| | def _info(self): |
| | return datasets.DatasetInfo( |
| | description=_DESCRIPTION, |
| | features=datasets.Features( |
| | { |
| | "image": datasets.Image(), |
| | "faces": datasets.Sequence(datasets.Sequence(datasets.Value("float32"), length=4)), |
| | "plates": datasets.Sequence(datasets.Sequence(datasets.Value("float32"), length=4)), |
| | |
| | } |
| | ), |
| | supervised_keys=None, |
| | homepage=_HOMEPAGE, |
| | license=_LICENSE, |
| | citation=_CITATION, |
| | ) |
| | |
| | def _split_generators(self, dl_manager): |
| | data_dir = dl_manager.download_and_extract(_URLS) |
| | return [ |
| | datasets.SplitGenerator( |
| | name=datasets.Split.TEST, |
| | gen_kwargs={ |
| | "split": "fisheye", |
| | "data_dir": data_dir["test"], |
| | "annot_dir": data_dir["annot"], |
| | }, |
| | ), |
| | datasets.SplitGenerator( |
| | name=datasets.Split.TRAIN, |
| | gen_kwargs={ |
| | "split": "normal", |
| | "data_dir": data_dir["test"], |
| | "annot_dir": data_dir["annot"], |
| | }, |
| | ), |
| | ] |
| |
|
| | def _generate_examples(self, split, data_dir, annot_dir): |
| | image_dir = os.path.join(data_dir, "fisheye") |
| | annotation_dir = os.path.join(annot_dir, "fisheye") |
| | files = [] |
| | |
| | idx = 0 |
| | for i_file in glob(os.path.join(image_dir, "*.png")): |
| | plates = [] |
| | faces = [] |
| | |
| | img_relative_file = os.path.relpath(i_file, image_dir) |
| | gt_relative_path = img_relative_file.replace(".png", ".txt") |
| | |
| | gt_path = os.path.join(annotation_dir, gt_relative_path) |
| | |
| | annotation = defaultdict(list) |
| | with open(gt_path, "r", encoding="utf-8") as f: |
| | line = f.readline().strip() |
| | while line: |
| | assert re.match(r"^\d( [\d\.]+){4,5}$", line), "Incorrect line: %s" % line |
| | cls, cx, cy, w, h = line.split()[:5] |
| | cls, cx, cy, w, h = int(cls), float(cx), float(cy), float(w), float(h) |
| | x1, y1, x2, y2 = cx - w / 2, cy - h / 2, cx + w / 2, cy + h / 2 |
| | annotation[cls].append([x1, y1, x2, y2]) |
| | line = f.readline().strip() |
| |
|
| | for cls, bboxes in annotation.items(): |
| | for x1, y1, x2, y2 in bboxes: |
| | if cls == 0: |
| | faces.append([x1, y1, x2, y2]) |
| | else: |
| | plates.append([x1, y1, x2, y2]) |
| | |
| | yield idx, {"image": i_file, "faces": faces, "plates": plates} |
| | |
| | idx += 1 |
| | |
| | |