Datasets:
1T Conte
commited on
Commit
·
e120dd0
1
Parent(s):
d0676a1
first commit
Browse files- .gitignore +1 -0
- Makefile +29 -0
- amazon_reviews_2013/train-0000-of-0017.parquet +3 -0
- amazon_reviews_2013/train-0001-of-0017.parquet +3 -0
- amazon_reviews_2013/train-0002-of-0017.parquet +3 -0
- amazon_reviews_2013/train-0003-of-0017.parquet +3 -0
- amazon_reviews_2013/train-0004-of-0017.parquet +3 -0
- amazon_reviews_2013/train-0005-of-0017.parquet +3 -0
- amazon_reviews_2013/train-0006-of-0017.parquet +3 -0
- amazon_reviews_2013/train-0007-of-0017.parquet +3 -0
- amazon_reviews_2013/train-0008-of-0017.parquet +3 -0
- amazon_reviews_2013/train-0009-of-0017.parquet +3 -0
- amazon_reviews_2013/train-0010-of-0017.parquet +3 -0
- amazon_reviews_2013/train-0011-of-0017.parquet +3 -0
- amazon_reviews_2013/train-0012-of-0017.parquet +3 -0
- amazon_reviews_2013/train-0013-of-0017.parquet +3 -0
- amazon_reviews_2013/train-0014-of-0017.parquet +3 -0
- amazon_reviews_2013/train-0015-of-0017.parquet +3 -0
- amazon_reviews_2013/train-0016-of-0017.parquet +3 -0
- amazon_reviews_2013/train-0017-of-0017.parquet +3 -0
- convert.py +127 -0
.gitignore
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
*.txt.gz
|
Makefile
ADDED
|
@@ -0,0 +1,29 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
download:
|
| 2 |
+
wget -c "https://snap.stanford.edu/data/amazon/Amazon_Instant_Video.txt.gz"
|
| 3 |
+
wget -c "https://snap.stanford.edu/data/amazon/Arts.txt.gz"
|
| 4 |
+
wget -c "https://snap.stanford.edu/data/amazon/Automotive.txt.gz"
|
| 5 |
+
wget -c "https://snap.stanford.edu/data/amazon/Baby.txt.gz"
|
| 6 |
+
wget -c "https://snap.stanford.edu/data/amazon/Beauty.txt.gz"
|
| 7 |
+
wget -c "https://snap.stanford.edu/data/amazon/Books.txt.gz"
|
| 8 |
+
wget -c "https://snap.stanford.edu/data/amazon/Cell_Phones_&_Accessories.txt.gz"
|
| 9 |
+
wget -c "https://snap.stanford.edu/data/amazon/Clothing_&_Accessories.txt.gz"
|
| 10 |
+
wget -c "https://snap.stanford.edu/data/amazon/Electronics.txt.gz"
|
| 11 |
+
wget -c "https://snap.stanford.edu/data/amazon/Gourmet_Foods.txt.gz"
|
| 12 |
+
wget -c "https://snap.stanford.edu/data/amazon/Health.txt.gz"
|
| 13 |
+
wget -c "https://snap.stanford.edu/data/amazon/Home_&_Kitchen.txt.gz"
|
| 14 |
+
wget -c "https://snap.stanford.edu/data/amazon/Industrial_&_Scientific.txt.gz"
|
| 15 |
+
wget -c "https://snap.stanford.edu/data/amazon/Jewelry.txt.gz"
|
| 16 |
+
wget -c "https://snap.stanford.edu/data/amazon/Kindle_Store.txt.gz"
|
| 17 |
+
wget -c "https://snap.stanford.edu/data/amazon/Movies_&_TV.txt.gz"
|
| 18 |
+
wget -c "https://snap.stanford.edu/data/amazon/Musical_Instruments.txt.gz"
|
| 19 |
+
wget -c "https://snap.stanford.edu/data/amazon/Music.txt.gz"
|
| 20 |
+
wget -c "https://snap.stanford.edu/data/amazon/Office_Products.txt.gz"
|
| 21 |
+
wget -c "https://snap.stanford.edu/data/amazon/Patio.txt.gz"
|
| 22 |
+
wget -c "https://snap.stanford.edu/data/amazon/Pet_Supplies.txt.gz"
|
| 23 |
+
wget -c "https://snap.stanford.edu/data/amazon/Shoes.txt.gz"
|
| 24 |
+
wget -c "https://snap.stanford.edu/data/amazon/Software.txt.gz"
|
| 25 |
+
wget -c "https://snap.stanford.edu/data/amazon/Sports_&_Outdoors.txt.gz"
|
| 26 |
+
wget -c "https://snap.stanford.edu/data/amazon/Tools_&_Home_Improvement.txt.gz"
|
| 27 |
+
wget -c "https://snap.stanford.edu/data/amazon/Toys_&_Games.txt.gz"
|
| 28 |
+
wget -c "https://snap.stanford.edu/data/amazon/Video_Games.txt.gz"
|
| 29 |
+
wget -c "https://snap.stanford.edu/data/amazon/Watches.txt.gz"
|
amazon_reviews_2013/train-0000-of-0017.parquet
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6f0d1d457074fcd30e332b12278c520977fadd96f418add7860484a578d16c1d
|
| 3 |
+
size 939927613
|
amazon_reviews_2013/train-0001-of-0017.parquet
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d47af7ac2b3417c6791e913654b995d0fbdd595372bfcaf7d6edd524dc0f955d
|
| 3 |
+
size 1105515091
|
amazon_reviews_2013/train-0002-of-0017.parquet
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9a366801b5d7253a28b142c0791083abceb48258dbf3b3e772c012d936085dd2
|
| 3 |
+
size 1104040669
|
amazon_reviews_2013/train-0003-of-0017.parquet
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7e45572b024a0d999c9695f280534872ddd498a4fb3d7054a6deb02157d78421
|
| 3 |
+
size 1103357354
|
amazon_reviews_2013/train-0004-of-0017.parquet
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:854a98360b8eb4b7f16351122c9f412b296d93c227b79780df0d66353dc8d519
|
| 3 |
+
size 1088789014
|
amazon_reviews_2013/train-0005-of-0017.parquet
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d811c9290cf7436dfae47ef2d608b38eeefa34d04391b2a6b271349fc0bfd1ef
|
| 3 |
+
size 1108575489
|
amazon_reviews_2013/train-0006-of-0017.parquet
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:296e3d316f5458c7a78bdf8d73918fbe6909e221c096337c0c47880b42d4ce6e
|
| 3 |
+
size 1104722517
|
amazon_reviews_2013/train-0007-of-0017.parquet
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:29a867a3b53da6398a4fbaa4578407ebebd4892ecb898376ae519564e545b149
|
| 3 |
+
size 712914822
|
amazon_reviews_2013/train-0008-of-0017.parquet
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a21797891c3eab3e9503cb03ac4167825e5227eb6d150c3330650faf8f20f2c3
|
| 3 |
+
size 636022186
|
amazon_reviews_2013/train-0009-of-0017.parquet
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e972b5ffd01f410a0ebdfaf0f1f7314a7a2c64ce39ed9d676536c8b39258c5ce
|
| 3 |
+
size 1116785087
|
amazon_reviews_2013/train-0010-of-0017.parquet
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1eefee5e93c5858bed9713ce7d1ea22ff971ac3e86459db8d39991ac0cc9282c
|
| 3 |
+
size 1119456878
|
amazon_reviews_2013/train-0011-of-0017.parquet
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:fc10130cab8cc0f93986b1a9ef01232cfa797b259108f8c9799782bd588baaa3
|
| 3 |
+
size 1112711889
|
amazon_reviews_2013/train-0012-of-0017.parquet
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d4601e11d9724c29adf50a84af9f212c08241ac88f5166a32554b3c2000dd3f3
|
| 3 |
+
size 1099536245
|
amazon_reviews_2013/train-0013-of-0017.parquet
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a6eac3e6ca90a39c4ee8aed9888996fe74074b6d5777c76351eb350f620de6e6
|
| 3 |
+
size 999966538
|
amazon_reviews_2013/train-0014-of-0017.parquet
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9d264d263cfe1a1a70b8b14a17721611c40d00db9bf658465686cdc9f0dcaedf
|
| 3 |
+
size 1005727136
|
amazon_reviews_2013/train-0015-of-0017.parquet
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:780ec688809aafb40c018c66c2a2629701464ea4c9c3fdeaee9ea570bc0c19cc
|
| 3 |
+
size 1000589188
|
amazon_reviews_2013/train-0016-of-0017.parquet
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4566de9a4be1308bf546b30a6ee5a92e54f04b82a20f79d2ce508b985b789b64
|
| 3 |
+
size 666796901
|
amazon_reviews_2013/train-0017-of-0017.parquet
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:11e225680cf9fee37a0969a2d1c70caaa6d3e845dc3ca64f16e50b7a57e2b909
|
| 3 |
+
size 528189664
|
convert.py
ADDED
|
@@ -0,0 +1,127 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
"""
|
| 2 |
+
Convert the Amazon reviews dataset to parquet format.
|
| 3 |
+
|
| 4 |
+
Usage:
|
| 5 |
+
$ make download
|
| 6 |
+
$ python convert.py
|
| 7 |
+
"""
|
| 8 |
+
|
| 9 |
+
import os
|
| 10 |
+
import gzip
|
| 11 |
+
|
| 12 |
+
from glob import glob
|
| 13 |
+
|
| 14 |
+
import pandas as pd
|
| 15 |
+
|
| 16 |
+
OUTPUT_DIR = "amazon_reviews_2013"
|
| 17 |
+
CHUNK_SIZE = 2000000
|
| 18 |
+
|
| 19 |
+
CATEGORIES = {
|
| 20 |
+
"Amazon_Instant_Video.txt.gz": "Amazon Instant Video", # 717,651 reviews
|
| 21 |
+
"Arts.txt.gz": "Arts", # 27,980 reviews
|
| 22 |
+
"Automotive.txt.gz": "Automotive", # 188,728 reviews
|
| 23 |
+
"Baby.txt.gz": "Baby", # 184,887 reviews
|
| 24 |
+
"Beauty.txt.gz": "Beauty", # 252,056 reviews
|
| 25 |
+
"Books.txt.gz": "Book", # 12,886,488 reviews
|
| 26 |
+
"Cell_Phones_&_Accessories.txt.gz": "Cell Phone", # 78,930 reviews
|
| 27 |
+
"Clothing_&_Accessories.txt.gz": "Clothing", # 581,933 reviews
|
| 28 |
+
"Electronics.txt.gz": "Electronics", # 1,241,778 reviews
|
| 29 |
+
"Gourmet_Foods.txt.gz": "Gourmet Food", # 154,635 reviews
|
| 30 |
+
"Health.txt.gz": "Health", # 428,781 reviews
|
| 31 |
+
"Home_&_Kitchen.txt.gz": "Home & Kitchen", # 991,794 reviews
|
| 32 |
+
"Industrial_&_Scientific.txt.gz": "Industrial & Scientific", # 137,042 reviews
|
| 33 |
+
"Jewelry.txt.gz": "Jewelry", # 58,621 reviews
|
| 34 |
+
"Kindle_Store.txt.gz": "Kindle Store", # 160,793 reviews
|
| 35 |
+
"Movies_&_TV.txt.gz": "Movie & TV", # 7,850,072 reviews
|
| 36 |
+
"Musical_Instruments.txt.gz": "Musical Instrument", # 85,405 reviews
|
| 37 |
+
"Music.txt.gz": "Music", # 6,396,350 reviews
|
| 38 |
+
"Office_Products.txt.gz": "Office", # 138,084 reviews
|
| 39 |
+
"Patio.txt.gz": "Patio", # 206,250 reviews
|
| 40 |
+
"Pet_Supplies.txt.gz": "Pet Supply", # 217,170 reviews
|
| 41 |
+
"Shoes.txt.gz": "Shoe", # 389,877 reviews
|
| 42 |
+
"Software.txt.gz": "Software", # 95,084 reviews
|
| 43 |
+
"Sports_&_Outdoors.txt.gz": "Sports & Outdoor", # 510,991 reviews
|
| 44 |
+
"Tools_&_Home_Improvement.txt.gz": "Tools & Home Improvement", # 409,499 reviews
|
| 45 |
+
"Toys_&_Games.txt.gz": "Toy & Game", # 435,996 reviews
|
| 46 |
+
"Video_Games.txt.gz": "Video Game", # 463,669 reviews
|
| 47 |
+
"Watches.txt.gz": "Watch", # 68,356 reviews
|
| 48 |
+
}
|
| 49 |
+
|
| 50 |
+
|
| 51 |
+
def to_parquet():
|
| 52 |
+
"""
|
| 53 |
+
Convert a single file to parquet
|
| 54 |
+
"""
|
| 55 |
+
n_chunks = 0
|
| 56 |
+
train_data = []
|
| 57 |
+
|
| 58 |
+
for filename in CATEGORIES:
|
| 59 |
+
|
| 60 |
+
for entry in parse_file(filename):
|
| 61 |
+
train_data.append(entry)
|
| 62 |
+
|
| 63 |
+
if len(train_data) == CHUNK_SIZE:
|
| 64 |
+
save_parquet(train_data, "train", n_chunks)
|
| 65 |
+
train_data = []
|
| 66 |
+
n_chunks += 1
|
| 67 |
+
|
| 68 |
+
if train_data:
|
| 69 |
+
save_parquet(train_data, "train", n_chunks)
|
| 70 |
+
|
| 71 |
+
return n_chunks
|
| 72 |
+
|
| 73 |
+
|
| 74 |
+
def save_parquet(data, split, chunk):
|
| 75 |
+
"""
|
| 76 |
+
Save data to parquet
|
| 77 |
+
"""
|
| 78 |
+
fname = os.path.join(OUTPUT_DIR, f"{split}-{chunk:04d}-of-nchunks.parquet")
|
| 79 |
+
df = pd.DataFrame(data)
|
| 80 |
+
df.to_parquet(fname)
|
| 81 |
+
|
| 82 |
+
|
| 83 |
+
def parse_file(filename):
|
| 84 |
+
"""
|
| 85 |
+
Parse a single file
|
| 86 |
+
"""
|
| 87 |
+
f = gzip.open(filename, "r")
|
| 88 |
+
entry = {}
|
| 89 |
+
for line in f:
|
| 90 |
+
line = line.decode().strip()
|
| 91 |
+
colon_pos = line.find(":")
|
| 92 |
+
if colon_pos == -1:
|
| 93 |
+
entry["product/category"] = CATEGORIES[filename]
|
| 94 |
+
yield entry
|
| 95 |
+
entry = {}
|
| 96 |
+
continue
|
| 97 |
+
e_name = line[:colon_pos]
|
| 98 |
+
rest = line[colon_pos + 2 :]
|
| 99 |
+
entry[e_name] = rest
|
| 100 |
+
|
| 101 |
+
yield entry
|
| 102 |
+
|
| 103 |
+
|
| 104 |
+
def rename_chunks(n_chunks):
|
| 105 |
+
"""
|
| 106 |
+
Replace nchunks in filename by the actual number of chunks
|
| 107 |
+
"""
|
| 108 |
+
for fname in glob(os.path.join(OUTPUT_DIR, "train-*-of-nchunks.parquet")):
|
| 109 |
+
new_fname = fname.replace("-nchunks", f"-{n_chunks:04d}")
|
| 110 |
+
os.rename(fname, new_fname)
|
| 111 |
+
|
| 112 |
+
|
| 113 |
+
def run():
|
| 114 |
+
"""
|
| 115 |
+
Convert all files to parquet
|
| 116 |
+
"""
|
| 117 |
+
if not os.path.exists(OUTPUT_DIR):
|
| 118 |
+
os.makedirs(OUTPUT_DIR)
|
| 119 |
+
|
| 120 |
+
n_chunks = to_parquet()
|
| 121 |
+
print(f"{n_chunks} chunks saved")
|
| 122 |
+
|
| 123 |
+
rename_chunks(n_chunks)
|
| 124 |
+
|
| 125 |
+
|
| 126 |
+
if __name__ == "__main__":
|
| 127 |
+
run()
|