Commit
·
05c9169
1
Parent(s):
184ac90
Update parquet files
Browse files
.gitattributes
DELETED
|
@@ -1,38 +0,0 @@
|
|
| 1 |
-
*.7z filter=lfs diff=lfs merge=lfs -text
|
| 2 |
-
*.arrow filter=lfs diff=lfs merge=lfs -text
|
| 3 |
-
*.bin filter=lfs diff=lfs merge=lfs -text
|
| 4 |
-
*.bin.* filter=lfs diff=lfs merge=lfs -text
|
| 5 |
-
*.bz2 filter=lfs diff=lfs merge=lfs -text
|
| 6 |
-
*.ftz filter=lfs diff=lfs merge=lfs -text
|
| 7 |
-
*.gz filter=lfs diff=lfs merge=lfs -text
|
| 8 |
-
*.h5 filter=lfs diff=lfs merge=lfs -text
|
| 9 |
-
*.joblib filter=lfs diff=lfs merge=lfs -text
|
| 10 |
-
*.lfs.* filter=lfs diff=lfs merge=lfs -text
|
| 11 |
-
*.model filter=lfs diff=lfs merge=lfs -text
|
| 12 |
-
*.msgpack filter=lfs diff=lfs merge=lfs -text
|
| 13 |
-
*.onnx filter=lfs diff=lfs merge=lfs -text
|
| 14 |
-
*.ot filter=lfs diff=lfs merge=lfs -text
|
| 15 |
-
*.parquet filter=lfs diff=lfs merge=lfs -text
|
| 16 |
-
*.pb filter=lfs diff=lfs merge=lfs -text
|
| 17 |
-
*.pt filter=lfs diff=lfs merge=lfs -text
|
| 18 |
-
*.pth filter=lfs diff=lfs merge=lfs -text
|
| 19 |
-
*.rar filter=lfs diff=lfs merge=lfs -text
|
| 20 |
-
saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
| 21 |
-
*.tar.* filter=lfs diff=lfs merge=lfs -text
|
| 22 |
-
*.tflite filter=lfs diff=lfs merge=lfs -text
|
| 23 |
-
*.tgz filter=lfs diff=lfs merge=lfs -text
|
| 24 |
-
*.xz filter=lfs diff=lfs merge=lfs -text
|
| 25 |
-
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 26 |
-
*.zstandard filter=lfs diff=lfs merge=lfs -text
|
| 27 |
-
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
| 28 |
-
# Audio files - uncompressed
|
| 29 |
-
*.pcm filter=lfs diff=lfs merge=lfs -text
|
| 30 |
-
*.sam filter=lfs diff=lfs merge=lfs -text
|
| 31 |
-
*.raw filter=lfs diff=lfs merge=lfs -text
|
| 32 |
-
# Audio files - compressed
|
| 33 |
-
*.aac filter=lfs diff=lfs merge=lfs -text
|
| 34 |
-
*.flac filter=lfs diff=lfs merge=lfs -text
|
| 35 |
-
*.mp3 filter=lfs diff=lfs merge=lfs -text
|
| 36 |
-
*.ogg filter=lfs diff=lfs merge=lfs -text
|
| 37 |
-
*.wav filter=lfs diff=lfs merge=lfs -text
|
| 38 |
-
train.jsonl filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
README.md
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
#amazon reviews multi chinese
|
| 2 |
-
|
| 3 |
-
This dataset is a port of the official ['amazon_reviews_multi' dataset] (https://huggingface.co/datasets/amazon_reviews_multi) on the Hub. It has just the Chinese language version. It has been reduced to just 3 columns (and 4th "label_text") that are relevant to the SetFit task.
|
|
|
|
|
|
|
|
|
|
|
|
train.jsonl → SetFit--amazon_reviews_multi_zh/json-test.parquet
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:31a527d794016a68ddfae08ad1fdc23497a010c7267482a921b8458f2d442af9
|
| 3 |
+
size 604012
|
SetFit--amazon_reviews_multi_zh/json-train.parquet
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d66eb42a6f5e245f9c7b279de8a33813d78c4df50bc718bdfbbf6776b8a52688
|
| 3 |
+
size 23971143
|
SetFit--amazon_reviews_multi_zh/json-validation.parquet
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3ac85a7150c39fb82cfa5a546a9f4c548c1890f04fb59150af219a9aca8edc44
|
| 3 |
+
size 589732
|
create_dataset.py
DELETED
|
@@ -1,20 +0,0 @@
|
|
| 1 |
-
from datasets import load_dataset
|
| 2 |
-
|
| 3 |
-
|
| 4 |
-
|
| 5 |
-
def main():
|
| 6 |
-
raw_data = load_dataset("amazon_reviews_multi", "zh")
|
| 7 |
-
raw_data = raw_data.rename_column("review_id", "id")
|
| 8 |
-
raw_data = raw_data.rename_column("review_body", "text")
|
| 9 |
-
raw_data = raw_data.remove_columns(["product_id", "reviewer_id", "review_title", "language", "product_category"])
|
| 10 |
-
|
| 11 |
-
for split, dataset in raw_data.items():
|
| 12 |
-
dataset = dataset.map(lambda x: {"label": x["stars"]-1}, num_proc=4)
|
| 13 |
-
dataset = dataset.map(lambda x: {"label_text": str(x["label"])}, num_proc=4)
|
| 14 |
-
dataset = dataset.remove_columns(["stars"])
|
| 15 |
-
dataset.to_json(f"{split}.jsonl")
|
| 16 |
-
|
| 17 |
-
if __name__ == "__main__":
|
| 18 |
-
main()
|
| 19 |
-
|
| 20 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
test.jsonl
DELETED
|
The diff for this file is too large to render.
See raw diff
|
|
|
validation.jsonl
DELETED
|
The diff for this file is too large to render.
See raw diff
|
|
|