| --- |
| dataset_info: |
| features: |
| - name: x |
| dtype: float64 |
| - name: 'y' |
| dtype: float64 |
| - name: language |
| dtype: string |
| - name: corpus |
| dtype: string |
| splits: |
| - name: train |
| num_bytes: 247037602 |
| num_examples: 5785741 |
| download_size: 112131877 |
| dataset_size: 247037602 |
| license: apache-2.0 |
| --- |
| |
| What follows is research code. It is by no means optimized for speed, efficiency, or readability. |
|
|
| ## Data loading, tokenizing and sharding |
|
|
| ```python |
| import os |
| import numpy as np |
| import pandas as pd |
| from sklearn.feature_extraction.text import TfidfTransformer |
| from sklearn.decomposition import TruncatedSVD |
| from tqdm.notebook import tqdm |
| from openTSNE import TSNE |
| import datashader as ds |
| import colorcet as cc |
| |
| |
| from dask.distributed import Client |
| import dask.dataframe as dd |
| import dask_ml |
| import dask.bag as db |
| |
| from transformers import AutoTokenizer |
| from datasets import load_dataset |
| from datasets.utils.py_utils import convert_file_size_to_int |
| |
| |
| def batch_tokenize(batch): |
| return {'tokenized': [' '.join(e.tokens) for e in tokenizer(batch['text']).encodings]} # "text" column hard encoded |
| |
| # The original viz used a subset of the ROOTS Corpus. |
| # More info on the entire dataset here: https://huggingface.co/bigscience-data |
| # And here: https://arxiv.org/abs/2303.03915 |
| |
| dset = load_dataset(..., split="train") |
| |
| dset = dset.map(batch_tokenize, batched=True, batch_size=64, num_proc=28) |
| |
| dset_name = "roots_subset" |
| max_shard_size = convert_file_size_to_int('300MB') |
| dataset_nbytes = dset.data.nbytes |
| num_shards = int(dataset_nbytes / max_shard_size) + 1 |
| num_shards = max(num_shards, 1) |
| print(f"Sharding into {num_shards} files.") |
| os.makedirs(f"{dset_name}/tokenized", exist_ok=True) |
| for shard_index in tqdm(range(num_shards)): |
| shard = dset.shard(num_shards=num_shards, index=shard_index, contiguous=True) |
| shard.to_parquet(f"{dset_name}/tokenized/tokenized-{shard_index:03d}.parquet") |
| |
| ``` |
|
|
| ## Embedding |
|
|
| ```python |
| client = Client() # To keep track of dask computation |
| client |
| |
| df = dd.read_parquet(f'{dset_name}/tokenized/') |
| vect = dask_ml.feature_extraction.text.CountVectorizer(tokenizer=str.split, |
| token_pattern=None, |
| vocabulary=vocab) |
| tokenized_bag = df['tokenized'].to_bag() |
| X = vect.transform(tokenized_bag) |
| |
| counts = X.compute() |
| client.shutdown() |
| |
| tfidf_transformer = TfidfTransformer(sublinear_tf=True, norm="l2") |
| tfidf = tfidf_transformer.fit_transform(counts) |
| |
| svd = TruncatedSVD(n_components=160) |
| X_svd = svd.fit_transform(tfidf) |
| |
| tsne = TSNE( |
| perplexity=30, # not sure what param setting resulted in the plot |
| n_jobs=28, |
| random_state=42, |
| verbose=True, |
| ) |
| |
| tsne_embedding = tsne.fit(X) |
| ``` |
|
|
| ## Plotting |
|
|
| ```python |
| df = pd.DataFrame(data=tsne_embedding, columns=['x','y']) |
| agg = ds.Canvas(plot_height=600, plot_width=600).points(df, 'x', 'y') |
| img = ds.tf.shade(agg, cmap=cc.fire, how='eq_hist') |
| ds.tf.set_background(img, "black") |
| ``` |
|
|
|  |