| import datasets |
| import csv |
|
|
| _LICENSE = """ |
| TO DO: Licencia |
| """ |
|
|
| with open("README.md", "r") as f: |
| lines = iter(f.readlines()) |
| for line in lines: |
| if "### Dataset Summary" in line: |
| break |
| next(lines) |
| _DESCRIPTION = next(lines) |
|
|
| _CITATION = """ |
| TO DO: Cita |
| """ |
|
|
| _LANGUAGES = { |
| "es": "Spanish", |
| "pt": "Portuguese" |
| } |
|
|
| _ALL_LANGUAGES = "all_languages" |
|
|
| _VERSION = "1.0.0" |
|
|
| _HOMEPAGE_URL = "https://github.com/lpsc-fiuba/MeLiSA" |
|
|
| _DOWNLOAD_URL = "./{lang}/{split}.csv" |
|
|
|
|
| class MeLiSAConfig(datasets.BuilderConfig): |
| """BuilderConfig for MeLiSA.""" |
|
|
| def __init__(self, languages=None, **kwargs): |
| """Constructs a MeLiSAConfig. |
| Args: |
| **kwargs: keyword arguments forwarded to super. |
| """ |
| super(MeLiSAConfig, self).__init__(version=datasets.Version(_VERSION, ""), **kwargs) |
| self.languages = languages |
|
|
|
|
|
|
| class MeLiSA(datasets.GeneratorBasedBuilder): |
| """MeLiSA dataset.""" |
|
|
| BUILDER_CONFIGS = [ |
| MeLiSAConfig( |
| name=_ALL_LANGUAGES, |
| languages=_LANGUAGES, |
| description="A collection of Mercado Libre reviews specifically designed to aid research in spanish and portuguese sentiment classification.", |
| ) |
| ] + [ |
| MeLiSAConfig( |
| name=lang, |
| languages=[lang], |
| description=f"{_LANGUAGES[lang]} examples from a collection of Mercado Libre reviews specifically designed to aid research in sentiment classification", |
| ) |
| for lang in _LANGUAGES |
| ] |
| BUILDER_CONFIG_CLASS = MeLiSAConfig |
| DEFAULT_CONFIG_NAME = _ALL_LANGUAGES |
|
|
| def _info(self): |
| return datasets.DatasetInfo( |
| description=_DESCRIPTION, |
| features=datasets.Features( |
| { |
| "country": datasets.Value("string"), |
| "category": datasets.Value("string"), |
| "review_content": datasets.Value("string"), |
| "review_title": datasets.Value("string"), |
| "review_rate": datasets.Value("int32") |
| } |
| ), |
| supervised_keys=None, |
| license=_LICENSE, |
| homepage=_HOMEPAGE_URL, |
| citation=_CITATION, |
| ) |
| |
| def _split_generators(self, dl_manager): |
| train_urls = [_DOWNLOAD_URL.format(split="train", lang=lang) for lang in self.config.languages] |
| dev_urls = [_DOWNLOAD_URL.format(split="validation", lang=lang) for lang in self.config.languages] |
| test_urls = [_DOWNLOAD_URL.format(split="test", lang=lang) for lang in self.config.languages] |
|
|
| train_paths = dl_manager.download_and_extract(train_urls) |
| dev_paths = dl_manager.download_and_extract(dev_urls) |
| test_paths = dl_manager.download_and_extract(test_urls) |
|
|
| return [ |
| datasets.SplitGenerator(name=datasets.Split.TRAIN, gen_kwargs={"file_paths": train_paths}), |
| datasets.SplitGenerator(name=datasets.Split.VALIDATION, gen_kwargs={"file_paths": dev_paths}), |
| datasets.SplitGenerator(name=datasets.Split.TEST, gen_kwargs={"file_paths": test_paths}), |
| ] |
|
|
| def _generate_examples(self, file_paths): |
| """Generate features given the directory path. |
| Args: |
| file_path: path where the tsv file is stored |
| Yields: |
| The features. |
| """ |
|
|
| for file_path in file_paths: |
| with open(file_path, "r", encoding="utf-8") as csvfile: |
| reader = csv.DictReader(csvfile) |
| for i, row in enumerate(reader): |
| yield i, row |