| """HashSet dataset."""
|
|
|
| import datasets
|
| import pandas as pd
|
| import json
|
|
|
| _CITATION = """
|
| @article{kodali2022hashset,
|
| title={HashSet--A Dataset For Hashtag Segmentation},
|
| author={Kodali, Prashant and Bhatnagar, Akshala and Ahuja, Naman and Shrivastava, Manish and Kumaraguru, Ponnurangam},
|
| journal={arXiv preprint arXiv:2201.06741},
|
| year={2022}
|
| }
|
| """
|
|
|
| _DESCRIPTION = """
|
| Hashset is a new dataset consisiting on 1.9k manually annotated and 3.3M loosely supervised tweets for testing the
|
| efficiency of hashtag segmentation models. We compare State of The Art Hashtag Segmentation models on Hashset and other
|
| baseline datasets (STAN and BOUN). We compare and analyse the results across the datasets to argue that HashSet can act
|
| as a good benchmark for hashtag segmentation tasks.
|
|
|
| HashSet Manual: contains 1.9k manually annotated hashtags. Each row consists of the hashtag, segmented
|
| hashtag ,named entity annotations, a list storing whether the hashtag contains mix of hindi and english
|
| tokens and/or contains non-english tokens.
|
| """
|
| _URL = "https://raw.githubusercontent.com/prashantkodali/HashSet/master/datasets/hashset/HashSet-Manual.csv"
|
|
|
| class HashSetManual(datasets.GeneratorBasedBuilder):
|
|
|
| VERSION = datasets.Version("1.0.0")
|
|
|
| def _info(self):
|
| return datasets.DatasetInfo(
|
| description=_DESCRIPTION,
|
| features=datasets.Features(
|
| {
|
| "index": datasets.Value("int32"),
|
| "hashtag": datasets.Value("string"),
|
| "segmentation": datasets.Value("string"),
|
| "spans": datasets.Sequence(
|
| {
|
| "start": datasets.Value("int32"),
|
| "end": datasets.Value("int32"),
|
| "text": datasets.Value("string")
|
| }
|
| ),
|
| "source": datasets.Value("string"),
|
| "gold_position": datasets.Value("int32"),
|
| "mix": datasets.Value("bool"),
|
| "other": datasets.Value("bool"),
|
| "ner": datasets.Value("bool"),
|
| "annotator_id": datasets.Value("int32"),
|
| "annotation_id": datasets.Value("int32"),
|
| "created_at": datasets.Value("timestamp[us]"),
|
| "updated_at": datasets.Value("timestamp[us]"),
|
| "lead_time": datasets.Value("float64"),
|
| "rank": datasets.Sequence(
|
| {
|
| "position": datasets.Value("int32"),
|
| "candidate": datasets.Value("string")
|
| }
|
| )
|
| }
|
| ),
|
| supervised_keys=None,
|
| homepage="https://github.com/prashantkodali/HashSet/",
|
| citation=_CITATION,
|
| )
|
|
|
| def _split_generators(self, dl_manager):
|
| downloaded_files = dl_manager.download(_URL)
|
| return [
|
| datasets.SplitGenerator(name=datasets.Split.TEST, gen_kwargs={"filepath": downloaded_files }),
|
| ]
|
|
|
| def _generate_examples(self, filepath):
|
|
|
| def read_language_labels(field):
|
| mix_label = "Hashtag has a mix of english and hindi tokens"
|
| other_label = "Hashtag has non english token "
|
| ner_label = "Hashtag has named entities"
|
| try:
|
| record = json.loads(field)
|
| except json.decoder.JSONDecodeError:
|
| record = {"choices": [field]}
|
|
|
| mix = False
|
| other = False
|
| ner = False
|
| if mix_label in record["choices"]:
|
| mix = True
|
| if other_label in record["choices"]:
|
| other = True
|
| if ner_label in record["choices"]:
|
| ner = True
|
| return mix, other, ner
|
|
|
| def read_entities(field):
|
| try:
|
| record = json.loads(field)
|
| except json.decoder.JSONDecodeError:
|
| return []
|
| output = []
|
| for row in record:
|
| output.append({
|
| "start": row.get("start", None),
|
| "end": row.get("end", None),
|
| "text": row.get("text", None)
|
| })
|
| return output
|
|
|
| def read_rank(row):
|
| output = []
|
| for i in range(10):
|
| output.append({
|
| "position": str(i+1),
|
| "candidate": row[str(i+1)]
|
| })
|
| return output
|
|
|
| def get_gold_position(field):
|
| output = field.strip("$")
|
| try:
|
| return int(output)
|
| except ValueError:
|
| return None
|
|
|
| records = pd.read_csv(filepath).to_dict("records")
|
| for idx, row in enumerate(records):
|
| mix, other, ner = read_language_labels(row["mutlitoken"])
|
| yield idx, {
|
| "index": row["Unnamed: 0"],
|
| "hashtag": row["Hashtag"],
|
| "segmentation": row["Final Segmentation"],
|
| "spans": read_entities(row["charner"]),
|
| "source": row["Source"],
|
| "gold_position": get_gold_position(row["topk"]),
|
| "mix": mix,
|
| "other": other,
|
| "ner": ner,
|
| "annotator_id": int(row["annotator"]),
|
| "annotation_id": int(row["annotation_id"]),
|
| "created_at": row["created_at"],
|
| "updated_at": row["updated_at"],
|
| "lead_time": row["lead_time"],
|
| "rank": read_rank(row)
|
| }
|
|
|