parquet-converter commited on
Commit
0f79915
·
1 Parent(s): 00d84f7

Update parquet files

Browse files
.gitattributes DELETED
@@ -1,53 +0,0 @@
1
- *.7z filter=lfs diff=lfs merge=lfs -text
2
- *.arrow filter=lfs diff=lfs merge=lfs -text
3
- *.bin filter=lfs diff=lfs merge=lfs -text
4
- *.bz2 filter=lfs diff=lfs merge=lfs -text
5
- *.ftz filter=lfs diff=lfs merge=lfs -text
6
- *.gz filter=lfs diff=lfs merge=lfs -text
7
- *.h5 filter=lfs diff=lfs merge=lfs -text
8
- *.joblib filter=lfs diff=lfs merge=lfs -text
9
- *.lfs.* filter=lfs diff=lfs merge=lfs -text
10
- *.lz4 filter=lfs diff=lfs merge=lfs -text
11
- *.model filter=lfs diff=lfs merge=lfs -text
12
- *.msgpack filter=lfs diff=lfs merge=lfs -text
13
- *.npy filter=lfs diff=lfs merge=lfs -text
14
- *.npz filter=lfs diff=lfs merge=lfs -text
15
- *.onnx filter=lfs diff=lfs merge=lfs -text
16
- *.ot filter=lfs diff=lfs merge=lfs -text
17
- *.parquet filter=lfs diff=lfs merge=lfs -text
18
- *.pb filter=lfs diff=lfs merge=lfs -text
19
- *.pickle filter=lfs diff=lfs merge=lfs -text
20
- *.pkl filter=lfs diff=lfs merge=lfs -text
21
- *.pt filter=lfs diff=lfs merge=lfs -text
22
- *.pth filter=lfs diff=lfs merge=lfs -text
23
- *.rar filter=lfs diff=lfs merge=lfs -text
24
- saved_model/**/* filter=lfs diff=lfs merge=lfs -text
25
- *.tar.* filter=lfs diff=lfs merge=lfs -text
26
- *.tflite filter=lfs diff=lfs merge=lfs -text
27
- *.tgz filter=lfs diff=lfs merge=lfs -text
28
- *.wasm filter=lfs diff=lfs merge=lfs -text
29
- *.xz filter=lfs diff=lfs merge=lfs -text
30
- *.zip filter=lfs diff=lfs merge=lfs -text
31
- *.zst filter=lfs diff=lfs merge=lfs -text
32
- *tfevents* filter=lfs diff=lfs merge=lfs -text
33
- # Audio files - uncompressed
34
- *.pcm filter=lfs diff=lfs merge=lfs -text
35
- *.sam filter=lfs diff=lfs merge=lfs -text
36
- *.raw filter=lfs diff=lfs merge=lfs -text
37
- # Audio files - compressed
38
- *.aac filter=lfs diff=lfs merge=lfs -text
39
- *.flac filter=lfs diff=lfs merge=lfs -text
40
- *.mp3 filter=lfs diff=lfs merge=lfs -text
41
- *.ogg filter=lfs diff=lfs merge=lfs -text
42
- *.wav filter=lfs diff=lfs merge=lfs -text
43
- # Image files - uncompressed
44
- *.bmp filter=lfs diff=lfs merge=lfs -text
45
- *.gif filter=lfs diff=lfs merge=lfs -text
46
- *.png filter=lfs diff=lfs merge=lfs -text
47
- *.tiff filter=lfs diff=lfs merge=lfs -text
48
- # Image files - compressed
49
- *.jpg filter=lfs diff=lfs merge=lfs -text
50
- *.jpeg filter=lfs diff=lfs merge=lfs -text
51
- *.webp filter=lfs diff=lfs merge=lfs -text
52
- data/test.jsonl filter=lfs diff=lfs merge=lfs -text
53
- data/train.jsonl filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
20Q.py DELETED
@@ -1,69 +0,0 @@
1
- import os
2
- import json
3
- import datasets
4
-
5
-
6
- # TODO: Add description of the dataset here
7
- # You can copy an official description
8
- _DESCRIPTION = """20Q"""
9
-
10
- # TODO: Add a link to an official homepage for the dataset here
11
- _HOMEPAGE = ""
12
-
13
- # TODO: Add the licence for the dataset here if you can find it
14
- _LICENSE = ""
15
-
16
-
17
-
18
- class NewDataset(datasets.GeneratorBasedBuilder):
19
- """TODO: Short description of my dataset."""
20
-
21
- VERSION = datasets.Version("1.0.0")
22
-
23
- def _info(self):
24
- features = datasets.Features(
25
- {
26
- "subject": datasets.Value("string"),
27
- "question": datasets.Value("string"),
28
- "label": datasets.Value("bool"),
29
- "label_fine_grained": datasets.Value("string")
30
- }
31
- )
32
-
33
- return datasets.DatasetInfo(
34
- description=_DESCRIPTION,
35
- features=features,
36
- homepage=_HOMEPAGE,
37
- license=_LICENSE,
38
- citation="_CITATION",
39
- )
40
-
41
- def _split_generators(self, dl_manager):
42
- data_dir = dl_manager.download_and_extract(["data/train.jsonl", "data/test.jsonl"])
43
- return [
44
- datasets.SplitGenerator(
45
- name=datasets.Split.TRAIN,
46
- gen_kwargs={
47
- "filepath": data_dir[0],
48
- "split": "train",
49
- },
50
- ),
51
- datasets.SplitGenerator(
52
- name=datasets.Split.TEST,
53
- gen_kwargs={
54
- "filepath": data_dir[1],
55
- "split": "test"
56
- },
57
- ),
58
- ]
59
-
60
- def _generate_examples(self, filepath, split):
61
- with open(filepath, encoding="utf-8") as f:
62
- for key, row in enumerate(f):
63
- data = json.loads(row)
64
- yield key, {
65
- "subject": data["subject"],
66
- "question": data["question"],
67
- "label": data["label"],
68
- "label_fine_grained": data["label_fine_grained"]
69
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
README.md DELETED
@@ -1,22 +0,0 @@
1
- ---
2
- annotations_creators: []
3
- language:
4
- - en
5
- language_creators: []
6
- license: []
7
- multilinguality:
8
- - monolingual
9
- pretty_name: 20Q - World Knowledge Benchmark
10
- size_categories:
11
- - 1K<n<10K
12
- source_datasets: []
13
- tags:
14
- - 20Q
15
- - Twenty Questions
16
- - 20 Questions
17
- task_categories:
18
- - question-answering
19
- task_ids: []
20
- ---
21
-
22
- # Dataset Card for 20Q
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
dataset_infos.json DELETED
@@ -1 +0,0 @@
1
- {"default": {"description": "20Q", "citation": "_CITATION", "homepage": "", "license": "", "features": {"subject": {"dtype": "string", "id": null, "_type": "Value"}, "question": {"dtype": "string", "id": null, "_type": "Value"}, "label": {"dtype": "bool", "id": null, "_type": "Value"}, "label_fine_grained": {"dtype": "string", "id": null, "_type": "Value"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "new_dataset", "config_name": "default", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 40740, "num_examples": 815, "dataset_name": "new_dataset"}, "test": {"name": "test", "num_bytes": 120391, "num_examples": 2500, "dataset_name": "new_dataset"}}, "download_checksums": {"data/train.jsonl": {"num_bytes": 90784, "checksum": "e84b36b32ca487767a1a7852769e7d23446124ffb5874b5dd1152ecfc94a53f4"}, "data/test.jsonl": {"num_bytes": 274029, "checksum": "7988520fd212c07487b2a6ec55a3e498722c0c926120414492d45f2395a854bb"}}, "download_size": 364813, "post_processing_size": null, "dataset_size": 161131, "size_in_bytes": 525944}}
 
 
data/train.jsonl → default/20_q-test.parquet RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e84b36b32ca487767a1a7852769e7d23446124ffb5874b5dd1152ecfc94a53f4
3
- size 90784
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:85c024890282fe9d20696c5994de723666502882b5678cec94b7e4ff252ccdc1
3
+ size 51334
data/test.jsonl → default/20_q-train.parquet RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7988520fd212c07487b2a6ec55a3e498722c0c926120414492d45f2395a854bb
3
- size 274029
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:de4d6e3a128a3d64984ef6ef0164de0f602b65c0af6a55191431c54ab905fa01
3
+ size 18990