date-data / date-data.py
nguyenminh4099's picture
Upload date-data.py
9b6b71a verified
import os
import re
import sys
import datasets
import pandas as pd
from huggingface_hub import HfFileSystem
from typing import List
logger = datasets.logging.get_logger(name=__name__)
fs = HfFileSystem()
_CITATION = """
"""
_DESCRIPTION = """\
This dataset contain file about datetime date.
It's created with purpose is practice loading dataset from hugging face hub."""
_HOMEPAGE = """\
https://github.com/minhnv4099
"""
_REPO = "datasets/nguyenminh4099/date-data"
_BRANCH = "main"
_REPO_BRANCH = f"{_REPO}@{_BRANCH}"
_REPO_URL = f"https://huggingface.co/{_REPO}/resolve/{_BRANCH}"
_URLS = {
'zipfile': os.path.join(_REPO_URL, "data", "{}.zip"),
'metadata': _REPO_URL + "/metadata.parquet",
}
_CONFIGS = ['all']
_CONFIGS.extend(
os.path.basename(file)[:-4]
for file in fs.listdir(_REPO_BRANCH + "/data/", detail=False)
if file.endswith('.zip')
)
# TODO: Define Dataset Builder config
class DateDataConfig(datasets.BuilderConfig):
def __init__(
self,
name: str,
**kwargs,
):
super(DateDataConfig, self).__init__(
name=name,
version=datasets.Version("1.0.0"),
)
# self.metadata = metadata
# self.url = kwargs.get('url', "https://huggingface.co/datasets/nguyenminh4099/date-data")
# self.data_url = kwargs.get('data_url', None)
# self.description = kwargs.get('description', _DESCRIPTION)
# logger.info('call BuilderConfig')
# TODO: Define Dataset Builder
class DateData(datasets.GeneratorBasedBuilder):
logger.info('call dataset builder')
BUILDER_CONFIGS = [
DateDataConfig(
name=name,
# metadata=_URLS['metadata'],
# data_url=_URLS['zipfile'].format(name),
)
for name in _CONFIGS
]
DEFAULT_CONFIG_NAME = 'all'
def _info(self) -> datasets.DatasetInfo:
features = datasets.Features({
"id": datasets.Value('string'),
"dow": datasets.Value('string'),
"month": datasets.Value('string'),
"dom": datasets.Value('string'),
"hour": datasets.Value('string'),
"min": datasets.Value('string'),
"second": datasets.Value('string'),
"timezone": datasets.Value('string'),
"year": datasets.Value('string'),
"file_path": datasets.Value('string'),
})
print(self.config)
return datasets.DatasetInfo(
features=features,
description=_DESCRIPTION,
citation=_CITATION,
homepage=_HOMEPAGE,
)
def _split_generators(
self,
dl_manager: datasets.DownloadManager,
) -> List[datasets.SplitGenerator]:
logger.info("Call _split_generators")
configs = _CONFIGS[1:5] if self.config.name == 'all' else [self.config.name]
data_files = {
config : _URLS['zipfile'].format(config)
for config in configs
}
data_dict = dl_manager.download_and_extract(data_files)
print(data_dict)
return [
datasets.SplitGenerator(
name=datasets.Split.TRAIN,
gen_kwargs={
"metadata": _URLS['metadata'],
"data_dict": data_dict,
}
)
]
def _generate_examples(
self,
metadata: str,
data_dict: dict,
) -> dict:
logger.info("Call _generate_examples")
infos = datasets.load_dataset(
"parquet",
data_files=[metadata],
split='train',
)
metadata_df = infos.to_pandas()
data_df = pd.DataFrame(
{
"shard" : list(data_dict.keys()),
"data_dir" : list(data_dict.values()),
},
columns=['shard', 'data_dir'],
index=range(len(data_dict))
)
metadata_df = metadata_df.merge(
right=data_df,
how='right',
left_on='shard',
right_on='shard',
sort=True,
)
for i, sample in enumerate(metadata_df.itertuples()):
file_name = os.path.join(
sample.data_dir, sample.id + ".txt"
)
example = self._read_txt(file_name=file_name)
example['id'] = sample.id
example['file_path'] = file_name
yield i, example
def _read_txt(
self,
file_name: str,
) -> dict:
with open(file=file_name, mode='r') as f:
return self._extract_datetime(f.read())
def _extract_datetime(
self,
datetime_string: str,
) -> dict:
datetime_string = datetime_string.strip("./ ")
components = re.split(pattern=r'[\s\:]+', string=datetime_string)
return {
"dow": components[0],
"month": components[1],
"dom": components[2],
"hour": components[3],
"min": components[4],
"second": components[5],
"timezone": components[6],
"year": components[7],
}
DateData()