| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| |
|
| | |
| | """The Sequence labellIng evaLuatIon benChmark fOr spoken laNguagE (SILICONE) benchmark.""" |
| |
|
| |
|
| | import textwrap |
| |
|
| | import pandas as pd |
| |
|
| | import datasets |
| |
|
| |
|
| | _SILICONE_CITATION = """\ |
| | @inproceedings{chapuis-etal-2020-hierarchical, |
| | title = "Hierarchical Pre-training for Sequence Labelling in Spoken Dialog", |
| | author = "Chapuis, Emile and |
| | Colombo, Pierre and |
| | Manica, Matteo and |
| | Labeau, Matthieu and |
| | Clavel, Chlo{\'e}", |
| | booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2020", |
| | month = nov, |
| | year = "2020", |
| | address = "Online", |
| | publisher = "Association for Computational Linguistics", |
| | url = "https://www.aclweb.org/anthology/2020.findings-emnlp.239", |
| | doi = "10.18653/v1/2020.findings-emnlp.239", |
| | pages = "2636--2648", |
| | abstract = "Sequence labelling tasks like Dialog Act and Emotion/Sentiment identification are a |
| | key component of spoken dialog systems. In this work, we propose a new approach to learn |
| | generic representations adapted to spoken dialog, which we evaluate on a new benchmark we |
| | call Sequence labellIng evaLuatIon benChmark fOr spoken laNguagE benchmark (SILICONE). |
| | SILICONE is model-agnostic and contains 10 different datasets of various sizes. |
| | We obtain our representations with a hierarchical encoder based on transformer architectures, |
| | for which we extend two well-known pre-training objectives. Pre-training is performed on |
| | OpenSubtitles: a large corpus of spoken dialog containing over 2.3 billion of tokens. We |
| | demonstrate how hierarchical encoders achieve competitive results with consistently fewer |
| | parameters compared to state-of-the-art models and we show their importance for both |
| | pre-training and fine-tuning.", |
| | } |
| | """ |
| |
|
| | _SILICONE_DESCRIPTION = """\ |
| | The Sequence labellIng evaLuatIon benChmark fOr spoken laNguagE (SILICONE) benchmark is a collection |
| | of resources for training, evaluating, and analyzing natural language understanding systems |
| | specifically designed for spoken language. All datasets are in the English language and cover a |
| | variety of domains including daily life, scripted scenarios, joint task completion, phone call |
| | conversations, and televsion dialogue. Some datasets additionally include emotion and/or sentimant |
| | labels. |
| | """ |
| |
|
| | _URL = "https://raw.githubusercontent.com/eusip/SILICONE-benchmark/main" |
| |
|
| | SWDA_DA_DESCRIPTION = { |
| | "sd": "Statement-non-opinion", |
| | "b": "Acknowledge (Backchannel)", |
| | "sv": "Statement-opinion", |
| | "%": "Uninterpretable", |
| | "aa": "Agree/Accept", |
| | "ba": "Appreciation", |
| | "fc": "Conventional-closing", |
| | "qw": "Wh-Question", |
| | "nn": "No Answers", |
| | "bk": "Response Acknowledgement", |
| | "h": "Hedge", |
| | "qy^d": "Declarative Yes-No-Question", |
| | "bh": "Backchannel in Question Form", |
| | "^q": "Quotation", |
| | "bf": "Summarize/Reformulate", |
| | 'fo_o_fw_"_by_bc': "Other", |
| | 'fo_o_fw_by_bc_"': "Other", |
| | "na": "Affirmative Non-yes Answers", |
| | "ad": "Action-directive", |
| | "^2": "Collaborative Completion", |
| | "b^m": "Repeat-phrase", |
| | "qo": "Open-Question", |
| | "qh": "Rhetorical-Question", |
| | "^h": "Hold Before Answer/Agreement", |
| | "ar": "Reject", |
| | "ng": "Negative Non-no Answers", |
| | "br": "Signal-non-understanding", |
| | "no": "Other Answers", |
| | "fp": "Conventional-opening", |
| | "qrr": "Or-Clause", |
| | "arp_nd": "Dispreferred Answers", |
| | "t3": "3rd-party-talk", |
| | "oo_co_cc": "Offers, Options Commits", |
| | "aap_am": "Maybe/Accept-part", |
| | "t1": "Downplayer", |
| | "bd": "Self-talk", |
| | "^g": "Tag-Question", |
| | "qw^d": "Declarative Wh-Question", |
| | "fa": "Apology", |
| | "ft": "Thanking", |
| | "+": "Unknown", |
| | "x": "Unknown", |
| | "ny": "Unknown", |
| | "sv_fx": "Unknown", |
| | "qy_qr": "Unknown", |
| | "ba_fe": "Unknown", |
| | } |
| |
|
| | MRDA_DA_DESCRIPTION = { |
| | "s": "Statement/Subjective Statement", |
| | "d": "Declarative Question", |
| | "b": "Backchannel", |
| | "f": '"Follow-me"', |
| | "q": "Question", |
| | } |
| |
|
| | IEMOCAP_E_DESCRIPTION = { |
| | "ang": "Anger", |
| | "dis": "Disgust", |
| | "exc": "Excitement", |
| | "fea": "Fear", |
| | "fru": "Frustration", |
| | "hap": "Happiness", |
| | "neu": "Neutral", |
| | "oth": "Other", |
| | "sad": "Sadness", |
| | "sur": "Surprise", |
| | "xxx": "Unknown", |
| | } |
| |
|
| |
|
| | class SiliconeConfig(datasets.BuilderConfig): |
| | """BuilderConfig for SILICONE.""" |
| |
|
| | def __init__( |
| | self, |
| | text_features, |
| | label_column, |
| | data_url, |
| | citation, |
| | url, |
| | label_classes=None, |
| | **kwargs, |
| | ): |
| | """BuilderConfig for SILICONE. |
| | Args: |
| | text_features: `dict[string, string]`, map from the name of the feature |
| | dict for each text field to the name of the column in the tsv file |
| | label_column: `string`, name of the column in the csv/txt file corresponding |
| | to the label |
| | data_url: `string`, url to download the csv/text file from |
| | citation: `string`, citation for the data set |
| | url: `string`, url for information about the data set |
| | label_classes: `list[string]`, the list of classes if the label is |
| | categorical. If not provided, then the label will be of type |
| | `datasets.Value('float32')`. |
| | **kwargs: keyword arguments forwarded to super. |
| | """ |
| | super(SiliconeConfig, self).__init__(version=datasets.Version("1.0.0", ""), **kwargs) |
| | self.text_features = text_features |
| | self.label_column = label_column |
| | self.label_classes = label_classes |
| | self.data_url = data_url |
| | self.citation = citation |
| | self.url = url |
| |
|
| |
|
| | class Silicone(datasets.GeneratorBasedBuilder): |
| | """The Sequence labellIng evaLuatIon benChmark fOr spoken laNguagE (SILICONE) benchmark.""" |
| |
|
| | BUILDER_CONFIGS = [ |
| | SiliconeConfig( |
| | name="dyda_da", |
| | description=textwrap.dedent( |
| | """\ |
| | The DailyDialog Act Corpus contains multi-turn dialogues and is supposed to reflect daily |
| | communication by covering topics about daily life. The dataset is manually labelled with |
| | dialog act and emotions. It is the third biggest corpus of SILICONE with 102k utterances.""" |
| | ), |
| | text_features={ |
| | "Utterance": "Utterance", |
| | "Dialogue_Act": "Dialogue_Act", |
| | "Dialogue_ID": "Dialogue_ID", |
| | }, |
| | label_classes=["commissive", "directive", "inform", "question"], |
| | label_column="Dialogue_Act", |
| | data_url={ |
| | "train": _URL + "/dyda/train.csv", |
| | "dev": _URL + "/dyda/dev.csv", |
| | "test": _URL + "/dyda/test.csv", |
| | }, |
| | citation=textwrap.dedent( |
| | """\ |
| | @InProceedings{li2017dailydialog, |
| | author = {Li, Yanran and Su, Hui and Shen, Xiaoyu and Li, Wenjie and Cao, Ziqiang and Niu, Shuzi}, |
| | title = {DailyDialog: A Manually Labelled Multi-turn Dialogue Dataset}, |
| | booktitle = {Proceedings of The 8th International Joint Conference on Natural Language Processing (IJCNLP 2017)}, |
| | year = {2017} |
| | }""" |
| | ), |
| | url="http://yanran.li/dailydialog.html", |
| | ), |
| | SiliconeConfig( |
| | name="dyda_e", |
| | description=textwrap.dedent( |
| | """\ |
| | The DailyDialog Act Corpus contains multi-turn dialogues and is supposed to reflect daily |
| | communication by covering topics about daily life. The dataset is manually labelled with |
| | dialog act and emotions. It is the third biggest corpus of SILICONE with 102k utterances.""" |
| | ), |
| | text_features={ |
| | "Utterance": "Utterance", |
| | "Emotion": "Emotion", |
| | "Dialogue_ID": "Dialogue_ID", |
| | }, |
| | label_classes=["anger", "disgust", "fear", "happiness", "no emotion", "sadness", "surprise"], |
| | label_column="Emotion", |
| | data_url={ |
| | "train": _URL + "/dyda/train.csv", |
| | "dev": _URL + "/dyda/dev.csv", |
| | "test": _URL + "/dyda/test.csv", |
| | }, |
| | citation=textwrap.dedent( |
| | """\ |
| | @InProceedings{li2017dailydialog, |
| | author = {Li, Yanran and Su, Hui and Shen, Xiaoyu and Li, Wenjie and Cao, Ziqiang and Niu, Shuzi}, |
| | title = {DailyDialog: A Manually Labelled Multi-turn Dialogue Dataset}, |
| | booktitle = {Proceedings of The 8th International Joint Conference on Natural Language Processing (IJCNLP 2017)}, |
| | year = {2017} |
| | }""" |
| | ), |
| | url="http://yanran.li/dailydialog.html", |
| | ), |
| | SiliconeConfig( |
| | name="iemocap", |
| | description=textwrap.dedent( |
| | """\ |
| | The IEMOCAP database is a multi-modal database of ten speakers. It consists of dyadic |
| | sessions where actors perform improvisations or scripted scenarios. Emotion categories |
| | are: anger, happiness, sadness, neutral, excitement, frustration, fear, surprise, and other. |
| | There is no official split of this dataset.""" |
| | ), |
| | text_features={ |
| | "Dialogue_ID": "Dialogue_ID", |
| | "Utterance_ID": "Utterance_ID", |
| | "Utterance": "Utterance", |
| | "Emotion": "Emotion", |
| | }, |
| | label_classes=list(IEMOCAP_E_DESCRIPTION.keys()), |
| | label_column="Emotion", |
| | data_url={ |
| | "train": _URL + "/iemocap/train.csv", |
| | "dev": _URL + "/iemocap/dev.csv", |
| | "test": _URL + "/iemocap/test.csv", |
| | }, |
| | citation=textwrap.dedent( |
| | """\ |
| | @article{busso2008iemocap, |
| | title={IEMOCAP: Interactive emotional dyadic motion capture database}, |
| | author={Busso, Carlos and Bulut, Murtaza and Lee, Chi-Chun and Kazemzadeh, Abe and Mower, |
| | Emily and Kim, Samuel and Chang, Jeannette N and Lee, Sungbok and Narayanan, Shrikanth S}, |
| | journal={Language resources and evaluation}, |
| | volume={42}, |
| | number={4}, |
| | pages={335}, |
| | year={2008}, |
| | publisher={Springer} |
| | }""" |
| | ), |
| | url="https://sail.usc.edu/iemocap/", |
| | ), |
| | SiliconeConfig( |
| | name="maptask", |
| | description=textwrap.dedent( |
| | """\ |
| | The HCRC MapTask Corpus was constructed through the verbal collaboration of participants |
| | in order to construct a map route. This corpus is small (27k utterances). As there is |
| | no standard train/dev/test split performance depends on the split.""" |
| | ), |
| | text_features={ |
| | "Speaker": "Speaker", |
| | "Utterance": "Utterance", |
| | "Dialogue_Act": "Dialogue_Act", |
| | }, |
| | label_classes=[ |
| | "acknowledge", |
| | "align", |
| | "check", |
| | "clarify", |
| | "explain", |
| | "instruct", |
| | "query_w", |
| | "query_yn", |
| | "ready", |
| | "reply_n", |
| | "reply_w", |
| | "reply_y", |
| | ], |
| | label_column="Dialogue_Act", |
| | data_url={ |
| | "train": _URL + "/maptask/train.txt", |
| | "dev": _URL + "/maptask/dev.txt", |
| | "test": _URL + "/maptask/test.txt", |
| | }, |
| | citation=textwrap.dedent( |
| | """\ |
| | @inproceedings{thompson1993hcrc, |
| | title={The HCRC map task corpus: natural dialogue for speech recognition}, |
| | author={Thompson, Henry S and Anderson, Anne H and Bard, Ellen Gurman and Doherty-Sneddon, |
| | Gwyneth and Newlands, Alison and Sotillo, Cathy}, |
| | booktitle={HUMAN LANGUAGE TECHNOLOGY: Proceedings of a Workshop Held at Plainsboro, New Jersey, March 21-24, 1993}, |
| | year={1993} |
| | }""" |
| | ), |
| | url="http://groups.inf.ed.ac.uk/maptask/", |
| | ), |
| | SiliconeConfig( |
| | name="meld_e", |
| | description=textwrap.dedent( |
| | """\ |
| | The Multimodal EmotionLines Dataset enhances and extends the EmotionLines dataset where |
| | multiple speakers participate in the dialogue.""" |
| | ), |
| | text_features={ |
| | "Utterance": "Utterance", |
| | "Speaker": "Speaker", |
| | "Emotion": "Emotion", |
| | "Dialogue_ID": "Dialogue_ID", |
| | "Utterance_ID": "Utterance_ID", |
| | }, |
| | label_classes=["anger", "disgust", "fear", "joy", "neutral", "sadness", "surprise"], |
| | label_column="Emotion", |
| | data_url={ |
| | "train": _URL + "/meld/train.csv", |
| | "dev": _URL + "/meld/dev.csv", |
| | "test": _URL + "/meld/test.csv", |
| | }, |
| | citation=textwrap.dedent( |
| | """\ |
| | @article{chen2018emotionlines, |
| | title={Emotionlines: An emotion corpus of multi-party conversations}, |
| | author={Chen, Sheng-Yeh and Hsu, Chao-Chun and Kuo, Chuan-Chun and Ku, Lun-Wei and others}, |
| | journal={arXiv preprint arXiv:1802.08379}, |
| | year={2018} |
| | }""" |
| | ), |
| | url="https://affective-meld.github.io/", |
| | ), |
| | SiliconeConfig( |
| | name="meld_s", |
| | description=textwrap.dedent( |
| | """\ |
| | The Multimodal EmotionLines Dataset enhances and extends the EmotionLines dataset where |
| | multiple speakers participate in the dialogue.""" |
| | ), |
| | text_features={ |
| | "Utterance": "Utterance", |
| | "Speaker": "Speaker", |
| | "Sentiment": "Sentiment", |
| | "Dialogue_ID": "Dialogue_ID", |
| | "Utterance_ID": "Utterance_ID", |
| | }, |
| | label_classes=["negative", "neutral", "positive"], |
| | label_column="Sentiment", |
| | data_url={ |
| | "train": _URL + "/meld/train.csv", |
| | "dev": _URL + "/meld/dev.csv", |
| | "test": _URL + "/meld/test.csv", |
| | }, |
| | citation=textwrap.dedent( |
| | """\ |
| | @article{chen2018emotionlines, |
| | title={Emotionlines: An emotion corpus of multi-party conversations}, |
| | author={Chen, Sheng-Yeh and Hsu, Chao-Chun and Kuo, Chuan-Chun and Ku, Lun-Wei and others}, |
| | journal={arXiv preprint arXiv:1802.08379}, |
| | year={2018} |
| | }""" |
| | ), |
| | url="https://affective-meld.github.io/", |
| | ), |
| | SiliconeConfig( |
| | name="mrda", |
| | description=textwrap.dedent( |
| | """\ |
| | ICSI MRDA Corpus consist of transcripts of multi-party meetings hand-annotated with dialog |
| | acts. It is the second biggest dataset with around 110k utterances.""" |
| | ), |
| | text_features={ |
| | "Utterance_ID": "Utterance_ID", |
| | "Dialogue_Act": "Dialogue_Act", |
| | "Channel_ID": "Channel_ID", |
| | "Speaker": "Speaker", |
| | "Dialogue_ID": "Dialogue_ID", |
| | "Utterance": "Utterance", |
| | }, |
| | label_classes=list(MRDA_DA_DESCRIPTION.keys()), |
| | label_column="Dialogue_Act", |
| | data_url={ |
| | "train": _URL + "/mrda/train.csv", |
| | "dev": _URL + "/mrda/dev.csv", |
| | "test": _URL + "/mrda/test.csv", |
| | }, |
| | citation=textwrap.dedent( |
| | """\ |
| | @techreport{shriberg2004icsi, |
| | title={The ICSI meeting recorder dialog act (MRDA) corpus}, |
| | author={Shriberg, Elizabeth and Dhillon, Raj and Bhagat, Sonali and Ang, Jeremy and Carvey, Hannah}, |
| | year={2004}, |
| | institution={INTERNATIONAL COMPUTER SCIENCE INST BERKELEY CA} |
| | }""" |
| | ), |
| | url="https://www.aclweb.org/anthology/W04-2319", |
| | ), |
| | SiliconeConfig( |
| | name="oasis", |
| | description=textwrap.dedent( |
| | """\ |
| | The Bt Oasis Corpus (Oasis) contains the transcripts of live calls made to the BT and |
| | operator services. This corpus is rather small (15k utterances). There is no standard |
| | train/dev/test split.""" |
| | ), |
| | text_features={ |
| | "Speaker": "Speaker", |
| | "Utterance": "Utterance", |
| | "Dialogue_Act": "Dialogue_Act", |
| | }, |
| | label_classes=[ |
| | "accept", |
| | "ackn", |
| | "answ", |
| | "answElab", |
| | "appreciate", |
| | "backch", |
| | "bye", |
| | "complete", |
| | "confirm", |
| | "correct", |
| | "direct", |
| | "directElab", |
| | "echo", |
| | "exclaim", |
| | "expressOpinion", |
| | "expressPossibility", |
| | "expressRegret", |
| | "expressWish", |
| | "greet", |
| | "hold", |
| | "identifySelf", |
| | "inform", |
| | "informCont", |
| | "informDisc", |
| | "informIntent", |
| | "init", |
| | "negate", |
| | "offer", |
| | "pardon", |
| | "raiseIssue", |
| | "refer", |
| | "refuse", |
| | "reqDirect", |
| | "reqInfo", |
| | "reqModal", |
| | "selfTalk", |
| | "suggest", |
| | "thank", |
| | "informIntent-hold", |
| | "correctSelf", |
| | "expressRegret-inform", |
| | "thank-identifySelf", |
| | ], |
| | label_column="Dialogue_Act", |
| | data_url={ |
| | "train": _URL + "/oasis/train.txt", |
| | "dev": _URL + "/oasis/dev.txt", |
| | "test": _URL + "/oasis/test.txt", |
| | }, |
| | citation=textwrap.dedent( |
| | """\ |
| | @inproceedings{leech2003generic, |
| | title={Generic speech act annotation for task-oriented dialogues}, |
| | author={Leech, Geoffrey and Weisser, Martin}, |
| | booktitle={Proceedings of the corpus linguistics 2003 conference}, |
| | volume={16}, |
| | pages={441--446}, |
| | year={2003}, |
| | organization={Lancaster: Lancaster University} |
| | }""" |
| | ), |
| | url="http://groups.inf.ed.ac.uk/oasis/", |
| | ), |
| | SiliconeConfig( |
| | name="sem", |
| | description=textwrap.dedent( |
| | """\ |
| | The SEMAINE database comes from the Sustained Emotionally coloured Human-Machine Interaction |
| | using Nonverbal Expression project. This dataset has been annotated on three sentiments |
| | labels: positive, negative and neutral. It is built on Multimodal Wizard of Oz experiment |
| | where participants held conversations with an operator who adopted various roles designed |
| | to evoke emotional reactions. There is no official split on this dataset.""" |
| | ), |
| | text_features={ |
| | "Utterance": "Utterance", |
| | "NbPairInSession": "NbPairInSession", |
| | "Dialogue_ID": "Dialogue_ID", |
| | "SpeechTurn": "SpeechTurn", |
| | "Speaker": "Speaker", |
| | "Sentiment": "Sentiment", |
| | }, |
| | label_classes=["Negative", "Neutral", "Positive"], |
| | label_column="Sentiment", |
| | data_url={ |
| | "train": _URL + "/sem/train.csv", |
| | "dev": _URL + "/sem/dev.csv", |
| | "test": _URL + "/sem/test.csv", |
| | }, |
| | citation=textwrap.dedent( |
| | """\ |
| | @article{mckeown2011semaine, |
| | title={The semaine database: Annotated multimodal records of emotionally colored conversations |
| | between a person and a limited agent}, |
| | author={McKeown, Gary and Valstar, Michel and Cowie, Roddy and Pantic, Maja and Schroder, Marc}, |
| | journal={IEEE transactions on affective computing}, |
| | volume={3}, |
| | number={1}, |
| | pages={5--17}, |
| | year={2011}, |
| | publisher={IEEE} |
| | }""" |
| | ), |
| | url="https://ieeexplore.ieee.org/document/5959155", |
| | ), |
| | SiliconeConfig( |
| | name="swda", |
| | description=textwrap.dedent( |
| | """\ |
| | Switchboard Dialog Act Corpus (SwDA) is a telephone speech corpus consisting of two-sided |
| | telephone conversations with provided topics. This dataset includes additional features |
| | such as speaker id and topic information.""" |
| | ), |
| | text_features={ |
| | "Utterance": "Utterance", |
| | "Dialogue_Act": "Dialogue_Act", |
| | "From_Caller": "From_Caller", |
| | "To_Caller": "To_Caller", |
| | "Topic": "Topic", |
| | "Dialogue_ID": "Dialogue_ID", |
| | "Conv_ID": "Conv_ID", |
| | }, |
| | label_classes=list(SWDA_DA_DESCRIPTION.keys()), |
| | label_column="Dialogue_Act", |
| | data_url={ |
| | "train": _URL + "/swda/train.csv", |
| | "dev": _URL + "/swda/dev.csv", |
| | "test": _URL + "/swda/test.csv", |
| | }, |
| | citation=textwrap.dedent( |
| | """\ |
| | @article{stolcke2000dialogue, |
| | title={Dialogue act modeling for automatic tagging and recognition of conversational speech}, |
| | author={Stolcke, Andreas and Ries, Klaus and Coccaro, Noah and Shriberg, Elizabeth and |
| | Bates, Rebecca and Jurafsky, Daniel and Taylor, Paul and Martin, Rachel and Ess-Dykema, |
| | Carol Van and Meteer, Marie}, |
| | journal={Computational linguistics}, |
| | volume={26}, |
| | number={3}, |
| | pages={339--373}, |
| | year={2000}, |
| | publisher={MIT Press} |
| | }""" |
| | ), |
| | url="https://web.stanford.edu/~jurafsky/ws97/", |
| | ), |
| | ] |
| |
|
| | def _info(self): |
| | features = {text_feature: datasets.Value("string") for text_feature in self.config.text_features.keys()} |
| | if self.config.label_classes: |
| | features["Label"] = datasets.features.ClassLabel(names=self.config.label_classes) |
| | features["Idx"] = datasets.Value("int32") |
| | return datasets.DatasetInfo( |
| | description=_SILICONE_DESCRIPTION, |
| | features=datasets.Features(features), |
| | homepage=self.config.url, |
| | citation=self.config.citation + "\n" + _SILICONE_CITATION, |
| | ) |
| |
|
| | def _split_generators(self, dl_manager): |
| | data_files = dl_manager.download(self.config.data_url) |
| | splits = [] |
| | splits.append( |
| | datasets.SplitGenerator( |
| | name=datasets.Split.TRAIN, |
| | gen_kwargs={ |
| | "data_file": data_files["train"], |
| | "split": "train", |
| | }, |
| | ) |
| | ) |
| | splits.append( |
| | datasets.SplitGenerator( |
| | name=datasets.Split.VALIDATION, |
| | gen_kwargs={ |
| | "data_file": data_files["dev"], |
| | "split": "dev", |
| | }, |
| | ) |
| | ) |
| | splits.append( |
| | datasets.SplitGenerator( |
| | name=datasets.Split.TEST, |
| | gen_kwargs={ |
| | "data_file": data_files["test"], |
| | "split": "test", |
| | }, |
| | ) |
| | ) |
| | return splits |
| |
|
| | def _generate_examples(self, data_file, split): |
| | if self.config.name not in ("maptask", "iemocap", "oasis"): |
| | df = pd.read_csv(data_file, delimiter=",", header=0, quotechar='"', dtype=str)[ |
| | self.config.text_features.keys() |
| | ] |
| |
|
| | if self.config.name == "iemocap": |
| | df = pd.read_csv( |
| | data_file, |
| | delimiter=",", |
| | header=0, |
| | quotechar='"', |
| | names=["Dialogue_ID", "Utterance_ID", "Utterance", "Emotion", "Valence", "Activation", "Dominance"], |
| | dtype=str, |
| | )[self.config.text_features.keys()] |
| |
|
| | if self.config.name in ("maptask", "oasis"): |
| | df = pd.read_csv(data_file, delimiter="|", names=["Speaker", "Utterance", "Dialogue_Act"], dtype=str)[ |
| | self.config.text_features.keys() |
| | ] |
| |
|
| | rows = df.to_dict(orient="records") |
| |
|
| | for n, row in enumerate(rows): |
| | example = row |
| | example["Idx"] = n |
| |
|
| | if self.config.label_column in example: |
| | label = example[self.config.label_column] |
| | example["Label"] = label |
| |
|
| | yield example["Idx"], example |
| |
|