import os import re import pandas as pd import numpy as np import datasets logger = datasets.logging.get_logger(__name__) import datasets URL = "https://huggingface.co/datasets/thewall/tokenizer/resolve/main" class TokenizerConfig(datasets.BuilderConfig): def __init__(self, **kwargs): super(TokenizerConfig, self).__init__(**kwargs) class Tokenizer(datasets.GeneratorBasedBuilder): BUILDER_CONFIGS = [ TokenizerConfig(name=key) for key in ["esm", "progen2", "raptgen", "aptamer"] ] DEFAULT_CONFIG_NAME = "esm" def _info(self): return datasets.DatasetInfo( features=datasets.Features( { "id": datasets.Value("int32"), "name": datasets.Value("string"), "path": datasets.Value("string"), "tokenizer": datasets.Value("string"), "special_tokens_map": datasets.Value("string"), } ), ) def _split_generators(self, dl_manager): file = dl_manager.download_and_extract(f"{URL}/{self.config.name}.tar.gz") return [ datasets.SplitGenerator(name=datasets.Split.TRAIN, gen_kwargs={"filepath": file}), ] def _generate_examples(self, filepath): """This function returns the examples in the raw (text) form.""" logger.info("generating examples from = %s", filepath) name = self.config.name tokenizer_file = os.path.join(filepath, name, "tokenizer.json") tokens_map_file = os.path.join(filepath, name, "special_tokens_map.json") with open(tokenizer_file) as f: tokenizer = "".join(f.readlines()) with open(tokens_map_file) as f: special_tokens = "".join(f.readlines()) yield 0, {"id": 0, "path": os.path.join(filepath, name), "name": name, "tokenizer": tokenizer, "special_tokens_map": special_tokens,} if __name__=="__main__": from datasets import load_dataset from tokenizers import Tokenizer from transformers import PreTrainedTokenizerFast import json dataset = load_dataset("tokenizer.py", split="all") tokenizer = Tokenizer.from_str(dataset[0]['tokenizer']) token_map = json.loads(dataset[0]['special_tokens_map']) fast_tokenizer = PreTrainedTokenizerFast(tokenizer_object=tokenizer, **token_map)