| | --- |
| | language: |
| | - en |
| |
|
| | datasets: |
| | - winogrande |
| |
|
| | widget: |
| | - text: "The roof of Rachel's home is old and falling apart, while Betty's is new. The home value of </s> Rachel is lower." |
| | - text: "The wooden doors at my friends work are worse than the wooden desks at my work, because the </s> desks material is cheaper." |
| | - text: "Postal Service were to reduce delivery frequency. </s> The postal service could deliver less frequently." |
| | - text: "I put the cake away in the refrigerator. It has a lot of butter in it. </s> The cake has a lot of butter in it." |
| |
|
| | --- |
| | # RoBERTa Large model fine-tuned on Winogrande |
| |
|
| | This model was fine-tuned on Winogrande dataset (XL size) in sequence classification task format, meaning that original pairs of sentences |
| | with corresponding options filled in were separated, shuffled and classified independently of each other. |
| |
|
| | ## Model description |
| |
|
| |
|
| | ## Intended use & limitations |
| |
|
| |
|
| | ### How to use |
| |
|
| |
|
| | ## Training data |
| |
|
| | [WinoGrande-XL](https://huggingface.co/datasets/winogrande) reformatted the following way: |
| | 1. Each sentence was split on "`_`" placeholder symbol. |
| | 2. Each option was concatenated with the second part of the split, thus transforming each example into two text segment pairs. |
| | 3. Text segment pairs corresponding to correct and incorrect options were marked with `True` and `False` labels accordingly. |
| | 4. Text segment pairs were shuffled thereafter. |
| |
|
| | For example, |
| |
|
| | ```json |
| | { |
| | "answer": "2", |
| | "option1": "plant", |
| | "option2": "urn", |
| | "sentence": "The plant took up too much room in the urn, because the _ was small." |
| | } |
| | ``` |
| |
|
| | becomes |
| |
|
| | ```json |
| | { |
| | "sentence1": "The plant took up too much room in the urn, because the ", |
| | "sentence2": "plant was small.", |
| | "label": false |
| | } |
| | ``` |
| |
|
| | and |
| |
|
| | ```json |
| | { |
| | "sentence1": "The plant took up too much room in the urn, because the ", |
| | "sentence2": "urn was small.", |
| | "label": true |
| | } |
| | ``` |
| | These sentence pairs are then treated as independent examples. |
| |
|
| | ### BibTeX entry and citation info |
| |
|
| | ```bibtex |
| | @article{sakaguchi2019winogrande, |
| | title={WinoGrande: An Adversarial Winograd Schema Challenge at Scale}, |
| | author={Sakaguchi, Keisuke and Bras, Ronan Le and Bhagavatula, Chandra and Choi, Yejin}, |
| | journal={arXiv preprint arXiv:1907.10641}, |
| | year={2019} |
| | } |
| | |
| | @article{DBLP:journals/corr/abs-1907-11692, |
| | author = {Yinhan Liu and |
| | Myle Ott and |
| | Naman Goyal and |
| | Jingfei Du and |
| | Mandar Joshi and |
| | Danqi Chen and |
| | Omer Levy and |
| | Mike Lewis and |
| | Luke Zettlemoyer and |
| | Veselin Stoyanov}, |
| | title = {RoBERTa: {A} Robustly Optimized {BERT} Pretraining Approach}, |
| | journal = {CoRR}, |
| | volume = {abs/1907.11692}, |
| | year = {2019}, |
| | url = {http://arxiv.org/abs/1907.11692}, |
| | archivePrefix = {arXiv}, |
| | eprint = {1907.11692}, |
| | timestamp = {Thu, 01 Aug 2019 08:59:33 +0200}, |
| | biburl = {https://dblp.org/rec/journals/corr/abs-1907-11692.bib}, |
| | bibsource = {dblp computer science bibliography, https://dblp.org} |
| | } |
| | ``` |