File size: 2,314 Bytes
d708d3d
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
import argparse
import os.path

import ipdb
from datasets import load_dataset


def filter_test_dataset(example):
    if example["quality_assessment"] is not None:
        scores = list(example["quality_assessment"].values())
        if example["quality_assessment"]['compositeStructure']>=3 and example["quality_assessment"]['imageQuality']==5 and not all(score == 5 for score in scores) and example['quality_assessment']['objectConsistency']==5:
            return True
        else:
            return False
    else:
        return False

def filter_train_dataset(example):
    if example["quality_assessment"] is not None:
        return list(example["quality_assessment"].values()) == [5, 5, 5]
    else:
        return False

def parse_args():
    parser = argparse.ArgumentParser("partition dataset")
    parser.add_argument("--dataset", type=str, default=None,required=True)
    parser.add_argument("--output_dir", type=str, default=None,required=True)
    parser.add_argument("--partition", type=str, default=None,required=True,choices=["train","test"])
    parser.add_argument("--num_shards", type=int, default=None)
    parser.add_argument("--num_proc", type=int, default=32)
    parser.add_argument("--cache", type=str, default="cache")
    args = parser.parse_args()
    if args.num_shards is None and args.partition == "train":
        args.num_shards = len(os.listdir(args.dataset))
    elif args.num_shards is None and args.partition == "test":
        args.num_shards = 1
    args.output_dir = os.path.join(args.output_dir, args.partition)
    return args

if __name__ == "__main__":
    args = parse_args()
    os.makedirs(args.output_dir, exist_ok=True)
    dataset = load_dataset(args.dataset, split="train", cache_dir=args.cache)
    if args.partition == "train":
        filtered_dataset = dataset.filter(filter_train_dataset,num_proc =args.num_proc)
    elif args.partition == "test":
        filtered_dataset = dataset.filter(filter_test_dataset,num_proc =args.num_proc)
    output_path = os.path.join(args.output_dir,"data-{index:05d}-of-{num_shards:05d}.parquet")
    for index in range(args.num_shards):
        shard = filtered_dataset.shard(index=index, num_shards=args.num_shards, contiguous=True)
        shard.to_parquet(output_path.format(index=index,num_shards=args.num_shards))