| { |
| "paper_id": "2021", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T01:02:12.985722Z" |
| }, |
| "title": "Gender Bias Hidden Behind Chinese Word Embeddings: The Case of Chinese Adjectives", |
| "authors": [ |
| { |
| "first": "Meichun", |
| "middle": [], |
| "last": "Jiao", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Uppsala University", |
| "location": { |
| "country": "Sweden" |
| } |
| }, |
| "email": "" |
| }, |
| { |
| "first": "Ziyang", |
| "middle": [], |
| "last": "Luo", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Uppsala University", |
| "location": { |
| "country": "Sweden" |
| } |
| }, |
| "email": "" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "Gender bias in word embeddings gradually becomes a vivid research field in recent years. Most studies in this field aim at measurement and debiasing methods with English as the target language. This paper investigates gender bias in static word embeddings from a unique perspective, Chinese adjectives. By training word representations with different models, the gender bias behind the vectors of adjectives is assessed. Through a comparison between the produced results and a human scored data set, we demonstrate how gender bias encoded in word embeddings differentiates from people's attitudes.", |
| "pdf_parse": { |
| "paper_id": "2021", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "Gender bias in word embeddings gradually becomes a vivid research field in recent years. Most studies in this field aim at measurement and debiasing methods with English as the target language. This paper investigates gender bias in static word embeddings from a unique perspective, Chinese adjectives. By training word representations with different models, the gender bias behind the vectors of adjectives is assessed. Through a comparison between the produced results and a human scored data set, we demonstrate how gender bias encoded in word embeddings differentiates from people's attitudes.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "BIAS STATEMENT This paper studies gender bias in Chinese adjectives, captured by word embeddings. For each Chinese adjective, a gender bias score is calculated by w \u2022 ( he \u2212 she) (Bolukbasi et al., 2016) . A positive score represents the Chinese adjective word embeddings is more associated with males, and a negative value refers to the opposite result. In our daily life, we find that gender stereotypes can be conveyed by adjectives. The close association between an adjective and a certain gender could be the accomplice in forming gender stereotypes (Menegatti and Rubini, 2017) . If these stereotypes are learned by the adjective word embeddings, they would be propagated to downstream NLP applications; accordingly, the gender stereotypes would be reinforced in users' mind. For example, the system will tend to use \"smart\" to describe males because of the existed social stereotype in training data that males are good at mathematics; then, the influence of the stereotype would be spread and increased again. Thus, we want to further investigate the bias encoded by the embeddings and how they are different with what in people's mind.", |
| "cite_spans": [ |
| { |
| "start": 179, |
| "end": 203, |
| "text": "(Bolukbasi et al., 2016)", |
| "ref_id": "BIBREF2" |
| }, |
| { |
| "start": 555, |
| "end": 583, |
| "text": "(Menegatti and Rubini, 2017)", |
| "ref_id": "BIBREF11" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "In the deep learning era, a major area of NLP has concerned the representation of words in lowdimensional and continuous vector spaces. People propose different algorithms to achieve such goal, including Word2Vec (Mikolov et al., 2013a) , GloVe (Pennington et al., 2014a) and FastText (Bojanowski et al., 2017) . Word embeddings play an important role in many NLP tasks, such as machine translation (Qi et al., 2018) and sentiment analysis (Yu et al., 2017) . However, several studies point out that word embeddings could capture the gender stereotypes in training data and transmit them to downstream applications (Bolukbasi et al., 2016; Zhao et al., 2017) . The consequence is often unbearable. Take machine translation as an example, if we translate a sentence concerning \"nurse\" from a language with gender-neutral pronouns to English, a female pronoun might be automatically produced to denote \"nurse\" (Prates et al., 2019) . Undoubtedly, this falls into the trap of the typical gender stereotypes. Therefore, the investigation of gender bias in word embeddings is necessary and accordingly attracts scholars' attention in recent years (Bolukbasi et al., 2016; Zhao et al., 2017) .", |
| "cite_spans": [ |
| { |
| "start": 213, |
| "end": 236, |
| "text": "(Mikolov et al., 2013a)", |
| "ref_id": "BIBREF12" |
| }, |
| { |
| "start": 245, |
| "end": 271, |
| "text": "(Pennington et al., 2014a)", |
| "ref_id": "BIBREF14" |
| }, |
| { |
| "start": 285, |
| "end": 310, |
| "text": "(Bojanowski et al., 2017)", |
| "ref_id": "BIBREF1" |
| }, |
| { |
| "start": 399, |
| "end": 416, |
| "text": "(Qi et al., 2018)", |
| "ref_id": "BIBREF17" |
| }, |
| { |
| "start": 440, |
| "end": 457, |
| "text": "(Yu et al., 2017)", |
| "ref_id": "BIBREF20" |
| }, |
| { |
| "start": 615, |
| "end": 639, |
| "text": "(Bolukbasi et al., 2016;", |
| "ref_id": "BIBREF2" |
| }, |
| { |
| "start": 640, |
| "end": 658, |
| "text": "Zhao et al., 2017)", |
| "ref_id": "BIBREF23" |
| }, |
| { |
| "start": 908, |
| "end": 929, |
| "text": "(Prates et al., 2019)", |
| "ref_id": "BIBREF16" |
| }, |
| { |
| "start": 1142, |
| "end": 1166, |
| "text": "(Bolukbasi et al., 2016;", |
| "ref_id": "BIBREF2" |
| }, |
| { |
| "start": 1167, |
| "end": 1185, |
| "text": "Zhao et al., 2017)", |
| "ref_id": "BIBREF23" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Most previous studies concerning gender bias in word embeddings only take English as the target language. Other languages are only included in several multi-lingual projects. For example, Prates et al. (2019) evaluate the gender bias in machine translation by translating 12 gender neutral languages with the Google Translate API; Lewis and Lupyan (2020) examine whether gender stereotypes could be reflected in the large-scale distributional structure of 25 natural languages. Apart from English, other languages have rarely been the target language in the research under this topic. This paper will take Chinese as the target language, investigating gender bias in word embeddings trained with the model designed for special features of Chinese.", |
| "cite_spans": [ |
| { |
| "start": 331, |
| "end": 354, |
| "text": "Lewis and Lupyan (2020)", |
| "ref_id": "BIBREF8" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "The fact that social stereotypes are conveyed in our language is often neglected by the public. From the commonly used adjectives, we could get a glimpse of the social stereotypes of a certain group of people. These stereotypes would confine us to what we should be in the minds of the public. It has been confirmed that when describing different genders, people will choose divergent groups of adjectives even though such a choice might change with the development of society (Garg et al., 2018) . Therefore, this study focuses on the problem of gender stereotypes from the perspective of adjectives. By scoring the gender bias from our trained vectors, we yield a subjective result of the gender preference of a set of adjectives. Through comparing our results with a handcrafted data set of human attitudes towards adjectives (Zhu and Liu, 2020) , we find that what is encoded in word embeddings is, to some extent, inconsistent with people's feelings on the gender preference of these adjectives.", |
| "cite_spans": [ |
| { |
| "start": 477, |
| "end": 496, |
| "text": "(Garg et al., 2018)", |
| "ref_id": "BIBREF5" |
| }, |
| { |
| "start": 829, |
| "end": 848, |
| "text": "(Zhu and Liu, 2020)", |
| "ref_id": "BIBREF25" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Gender could affect the usage of adjectives (Lakoff, 1973) . On the other hand, the attitude of the public towards the social roles of men and women could also be indicated by how adjectives correlates with genders (Zhu and Liu, 2020) . In the past decade, an increasing number of studies investigating adjectives and gender stereotypes from various perspectives are proposed and developed. Baker (2013) reveals the stereotype in the description of boys and girls by analyzing adjectives only used for a certain gender with the aid of corpora covering a range of written genres. Research of Bollywood movies (Madaan et al., 2018) finds that different adjectives are chosen when they try to create impressive male and female roles. The significant divergence between the usage of adjectives for describing men and women has also been confirmed by Hoyle et al. (2019) , and they also notice the variance is consistent with common stereotypes. Zhu and Liu (2020) trace the change of gender bias in Chinese adjectives based on a handcrafted data set that consists of the gender preference score of adjectives. However, the number of studies focusing on Chinese adjectives and gender bias is still limited.", |
| "cite_spans": [ |
| { |
| "start": 44, |
| "end": 58, |
| "text": "(Lakoff, 1973)", |
| "ref_id": "BIBREF7" |
| }, |
| { |
| "start": 215, |
| "end": 234, |
| "text": "(Zhu and Liu, 2020)", |
| "ref_id": "BIBREF25" |
| }, |
| { |
| "start": 391, |
| "end": 403, |
| "text": "Baker (2013)", |
| "ref_id": "BIBREF0" |
| }, |
| { |
| "start": 608, |
| "end": 629, |
| "text": "(Madaan et al., 2018)", |
| "ref_id": "BIBREF10" |
| }, |
| { |
| "start": 846, |
| "end": 865, |
| "text": "Hoyle et al. (2019)", |
| "ref_id": "BIBREF6" |
| }, |
| { |
| "start": 941, |
| "end": 959, |
| "text": "Zhu and Liu (2020)", |
| "ref_id": "BIBREF25" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related work", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Gender bias in word embeddings and the corresponding debiasing methods have been a vivid re-search field in recent years. Bolukbasi et al. (2016) and Caliskan et al. (2017) confirm that word embedding models could precisely capture the social stereotypes concerning people's careers, such as the relationship in an analogy that Man is to Computer Programmer as Woman is to Homemaker. This bias could even be amplified by embedding models (Zhao et al., 2017) . Besides English, other target languages like Swedish (Sahlgren and Olsson, 2019) and Dutch (Wevers, 2019) gradually attract the attention of researchers. Various methods for assessing bias and debiasing are proposed and developed in previous studies. Bolukbasi et al. (2016) firstly measure the gender bias by computing the projection of a word on he \u2212 she direction, which has been confirmed strongly correlated with the public judgment of gender stereotypes. Based on this method, they also develop a debiasing method by post-processing the generated word vectors. and Zhang et al. 2018further propose to debias word embeddings in training procedure by changing the loss of GloVe model (Pennington et al., 2014b) and employing an adversarial network, respectively. Despite a large amount of research having been done in this field, to the best of our knowledge, no one has assessed the underlying gender bias behind adjectives, especially those in non-English languages.", |
| "cite_spans": [ |
| { |
| "start": 122, |
| "end": 145, |
| "text": "Bolukbasi et al. (2016)", |
| "ref_id": "BIBREF2" |
| }, |
| { |
| "start": 150, |
| "end": 172, |
| "text": "Caliskan et al. (2017)", |
| "ref_id": "BIBREF3" |
| }, |
| { |
| "start": 438, |
| "end": 457, |
| "text": "(Zhao et al., 2017)", |
| "ref_id": "BIBREF23" |
| }, |
| { |
| "start": 513, |
| "end": 540, |
| "text": "(Sahlgren and Olsson, 2019)", |
| "ref_id": "BIBREF18" |
| }, |
| { |
| "start": 551, |
| "end": 565, |
| "text": "(Wevers, 2019)", |
| "ref_id": "BIBREF19" |
| }, |
| { |
| "start": 711, |
| "end": 734, |
| "text": "Bolukbasi et al. (2016)", |
| "ref_id": "BIBREF2" |
| }, |
| { |
| "start": 1148, |
| "end": 1174, |
| "text": "(Pennington et al., 2014b)", |
| "ref_id": "BIBREF15" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related work", |
| "sec_num": "2" |
| }, |
| { |
| "text": "To complement the full picture of gender bias encoded in word representation, this paper examines the problem from the perspective of adjectives rather than nouns of occupations that repeatedly appeared in previous studies. Based on the human scoring data set of Zhu and Liu (2020), we investigate the similarities and differences between the automatically captured gender bias in Chinese and people's judgement.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related work", |
| "sec_num": "2" |
| }, |
| { |
| "text": "To uncover the gender stereotypes conveyed by adjectives, we first preprocess a corpus of online Chinese news and train word embeddings on it with two different models. Then, we calculate the gender bias scores based on the generated two vectors and compare them with the human scoring data set, Adjectives list with Gendered Skewness and Sentiment (AGSS) (Zhu and Liu, 2020) .", |
| "cite_spans": [ |
| { |
| "start": 356, |
| "end": 375, |
| "text": "(Zhu and Liu, 2020)", |
| "ref_id": "BIBREF25" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Methodology", |
| "sec_num": "3" |
| }, |
| { |
| "text": "News reports are not only the reflection of social consciousness but also the easily collected corpus Original size 1.54GB Size after preprocessing 2.1GB The number of tokens 375.3M The number of unique words 100.7k for many NLP tasks. Therefore, we choose a corpus of Chinese news reports as our training data set. It was collected and released by Sogou Labs, covering 18 themes of news from various Chinese news websites. 1 The details of the corpus are illustrated in Table 1 . All texts in the data set are cleaned and preprocessed through the following steps.", |
| "cite_spans": [ |
| { |
| "start": 424, |
| "end": 425, |
| "text": "1", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 471, |
| "end": 478, |
| "text": "Table 1", |
| "ref_id": "TABREF0" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Data", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "1. Extract the news content and change the encoding from gbk to utf-8. All the other information and metadata are removed.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Data", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "2. Remove the html tag by the regular expression and conduct Chinese word segmentation with jieba, 2 a widely used Python module.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Data", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "The meaning of Chinese words is usually related to the semantic information carried by the characters (Hanzi) that they are comprised of. Figure 1 shows an example: the word \"xianjing\" means \"demure\", which consists of two characters. The first one, \"xian\", means refined but usually used for describing a woman; the second character \"jing\" means silent and quiet. The word inherits and combines the meaning of each character, even the information concerning gender. This feature of Chinese leads to the development of word embedding models in which word vectors are trained with the characterlevel information. However, no study before has provided any ideas about how the encoding of gender bias information will be affected by training embedding with character-level information. Therefore, we decide to train our vectors with one of such models, namely the character-enhanced word embedding model (CWE) (Chen et al., 2015) . In addition to the word vector, this model also trains a vector for Chinese characters. CWE is developed based on the framework CBOW (Mikolov et al., 2013b) . CBOW aims at predicting the target word by understanding the surrounding context words. Practically, its objective is to maximize the average log probability given a word sequence D = {x 1 , . . . , x M }. CWE modifies the way of representing the context words in the algorithm of CBOW, predicting target words by combining character embedding and word embedding. A context word x j in CWE would be represented as follows,", |
| "cite_spans": [ |
| { |
| "start": 907, |
| "end": 926, |
| "text": "(Chen et al., 2015)", |
| "ref_id": "BIBREF4" |
| }, |
| { |
| "start": 1062, |
| "end": 1085, |
| "text": "(Mikolov et al., 2013b)", |
| "ref_id": "BIBREF13" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 138, |
| "end": 146, |
| "text": "Figure 1", |
| "ref_id": "FIGREF0" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Training and evaluation of word embeddings", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "x j = 1 2 \uf8eb \uf8ed w j + 1 N j N j k=1 c k \uf8f6 \uf8f8 .", |
| "eq_num": "(1)" |
| } |
| ], |
| "section": "Training and evaluation of word embeddings", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "w j refers to the word embedding of x j ; N j represents the number of characters in x j ; c k is the representation of the k-th character in x j . For comparison, we also train vectors on CBOW to show in the influence of character-level information. The Python library Gensim 3 is used for training the representation with CBOW, and the other with CWE is completed by the released code of Chen et al. (2015) . 4 To make the results comparable, we keep the same hyper-parameters for the two models. Detailed information is recorded in Table 2 .", |
| "cite_spans": [ |
| { |
| "start": 390, |
| "end": 408, |
| "text": "Chen et al. (2015)", |
| "ref_id": "BIBREF4" |
| }, |
| { |
| "start": 411, |
| "end": 412, |
| "text": "4", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 535, |
| "end": 542, |
| "text": "Table 2", |
| "ref_id": "TABREF1" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Training and evaluation of word embeddings", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "To ensure the effective of the produced embeddings, we evaluate them by word analogy tasks and the corresponding tools developed by . The test data set of the task includes 17813 questions about morphological or semantic rela-tions. 5 The results are illustrated in table 3. Although the semantic task results are lower than the values given in the paper of , we still assume that they are reliable as the size of our training data is only the half of theirs.", |
| "cite_spans": [ |
| { |
| "start": 233, |
| "end": 234, |
| "text": "5", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Training and evaluation of word embeddings", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "Morphological Semantic 11 ", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Model", |
| "sec_num": null |
| }, |
| { |
| "text": "We employ the method of Bolukbasi et al. (2016) to assess gender bias encoded in the trained embeddings. For each adjective, a gender bias score is calculated by w \u2022 ( he \u2212 she) based on its vector. 6 A positive result presents that the word has a closer association with males, while a negative score implies that the word is more associated with females. The higher the absolute value, the more biased the adjective is. 0 means totally neutral. Adjectives List with Gendered Skewness and Sentiment (AGSS) is a handcrafted data set built by questionnaire in the project of Zhu and Liu (2020) . 6 linguists firstly select 466 Chinese adjectives that could describe people, then 116 gender-balanced respondents score these adjectives by questionnaires. The the scale of score 1 to 5 is used to reflect people's attitude, with 1 being more related to female and 5 more related to male. Table 4 shows some example data from AGSS. Finally, 304 adjectives are scored larger than 3, 153 adjectives get score less than 3, and 9 are believed totally neutral. According to the statistics of AGSS, the adjectives chosen for this data set are more associated with males, so Zhu and Liu (2020) state that AGSS is with gender skewness. To analyze the results, we compare our gender bias scores from word embeddings with the AGSS scores. As they are on different scales, Pearson correlation coefficient is employed here. It could measure the the strength of the linear association between two variables, which returns a value between -1 and 1. 1 indicates strong positive linear", |
| "cite_spans": [ |
| { |
| "start": 24, |
| "end": 47, |
| "text": "Bolukbasi et al. (2016)", |
| "ref_id": "BIBREF2" |
| }, |
| { |
| "start": 199, |
| "end": 200, |
| "text": "6", |
| "ref_id": null |
| }, |
| { |
| "start": 574, |
| "end": 592, |
| "text": "Zhu and Liu (2020)", |
| "ref_id": "BIBREF25" |
| }, |
| { |
| "start": 1163, |
| "end": 1181, |
| "text": "Zhu and Liu (2020)", |
| "ref_id": "BIBREF25" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 884, |
| "end": 891, |
| "text": "Table 4", |
| "ref_id": "TABREF4" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Gender bias measurement and data set", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "Gender skewness in AGSS powerful 4.44 vuglar 3.62 selfless 3.00 cute 2.26 decorous 1.59 4 Results and discussion", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Words", |
| "sec_num": null |
| }, |
| { |
| "text": "We calculate the gender bias score for the same adjectives with AGSS and conclude the basic statistics in Table 5 . More adjectives are categorized into the group close to male. This is identified with what Zhu and Liu (2020) state about AGSS (mentioned in Section 3.3). However, it should be noticed that the average scores of both models result in a negative value. This might suggest that most absolute values of negative gender bias scores are much higher than the positive group. CBOW CWE # pos. score 283 316 # neg. score 183 150 Avg. score -0.02029 -0.02945 ", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 106, |
| "end": 113, |
| "text": "Table 5", |
| "ref_id": "TABREF5" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Gender bias scores from word embeddings", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "The Pearson correlation coefficients presented in Table 6 suggest the two categories of data are positively associated. However, the correlation is not that strong with only around 0.5, since the range of Pearson coefficient is from -1 to 1. Besides, the gender bias scores from the word embeddings trained with CWE are more associated with the human scores. This might suggest that the character-level information could help the model capture gender bias more precisely, or we should say such information could contribute to encoding what is in people's minds.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 50, |
| "end": 57, |
| "text": "Table 6", |
| "ref_id": "TABREF6" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Correlation between word vectors and AGSS", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "In Figure 2 , we can find more details of the correlation between the two categories of data. By com- Figure 2 : Scatter plots of AGSS scores and gender bias scores from word vectors trained with CBOW (left) and CWE (right). AGSS refers to the AGSS scores and bias word and bias char refers to the generated gender bias scores. The distribution of gender bias scores and AGSS scores are on the top and right of the plots respectively. The lines show the linear relation between the two categories. paring the distribution of the two types of scores, we notice that the scores given by people are very concentrated between 2.5 to 3.5, while automatically calculated scores have a wider distribution. This might be caused by different scales, but may also come from people hypocrisy: they spontaneously narrow the extent of gender preference of words when they are asked to score their attitudes. Besides, it is a clear tendency that some words only for males in people's impression are automatically given a negative score, which means they are more close to women in word vectors. Therefore, we conduct further analysis by separating the data into two groups based on the neutral line in AGSS.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 3, |
| "end": 11, |
| "text": "Figure 2", |
| "ref_id": null |
| }, |
| { |
| "start": 102, |
| "end": 110, |
| "text": "Figure 2", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Correlation between word vectors and AGSS", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "We recalculate the Pearson correlation coefficients for the two group of data, presenting results in Table 7 and Table 8 . To give a full picture, sepa- rated scatter plots as shown in Figure 3 and Figure 4 are also included. The increment of coefficients for the group with AGSS scores lower than 3 suggests that most adjectives believed for describing women are closer to females in word vectors as well. What is encoded by word embedding is consistent with people's impressions of these words. In addition, the correlation for scores from vectors trained with CBOW exceeds the results with the CWE model. This finding might indicate the underlying negative influence of covering character-level information in the word embedding. However, a substantial divergence appears in the other group. Based on the scatter plot and the Pearson coefficient, some of the adjectives that almost exclusively connect with male in people's minds could be very neutral according to our word embedding. The coefficients also suggest that the two categories of data do not show linear rela-tions. Additionally, only one-third of the adjectives in this group are closer to males in word embedding, while the others are actually more associated with females. Obviously, what we estimate from embedding disagrees with people's attitudes. This could be explained by the development of language. The study of Zhu and Liu (2020) proves that some Chinese adjectives for describing men in past time gradually become neutral in written language. Since the language used online develops fast and our training data are online news reports, the word embedding we trained is likely affected by the change. However, the public has not realized such development although they might start to use it in the new way. Therefore, when they are queried about the attitude towards attitudes, they might give an answer based on their outdated knowledge.", |
| "cite_spans": [ |
| { |
| "start": 1388, |
| "end": 1406, |
| "text": "Zhu and Liu (2020)", |
| "ref_id": "BIBREF25" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 101, |
| "end": 120, |
| "text": "Table 7 and Table 8", |
| "ref_id": "TABREF7" |
| }, |
| { |
| "start": 185, |
| "end": 193, |
| "text": "Figure 3", |
| "ref_id": "FIGREF1" |
| }, |
| { |
| "start": 198, |
| "end": 206, |
| "text": "Figure 4", |
| "ref_id": "FIGREF2" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Correlation between word vectors and AGSS", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "In this paper, we investigate gender bias in Chinese word embeddings from the perspective of adjectives, and compare automatically calculated gender bias score with human attitudes. We elaborately present the differences between gender bias encoded in word vectors and the people's feeling of the same adjective. For the words that people believe for describing women, the extracted score of gender bias gives an identified results; while for adjectives that should be used for men in people's mind, our results suggest that these group of words are actually more neutral than the crowd judgement. Additionally, how the word embedding models covering character-level information perform in terms of capturing gender bias in Chinese is also examined.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion", |
| "sec_num": "5" |
| }, |
| { |
| "text": "http://www.sogou.com/labs/resource/ca.php 2 https://github.com/fxsjy/jieba", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "https://github.com/RaRe-Technologies/gensim 4 https://github.com/Leonard-Xu/CWE/tree/master/src", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "https://github.com/Embedding/Chinese-Word-Vectors/tree/master/testsets6 We use the Chinese translation of he and she when conducting experiments.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| } |
| ], |
| "back_matter": [ |
| { |
| "text": "Acknowledgments This project grew out of a master course project for the Fall 2020 Uppsala University 5LN714, Language Technology: Research and Development. We would like to thank Sara Stymne and Ali Basirat for some great suggestions and the anonymous reviewers for their excellent feedback.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "acknowledgement", |
| "sec_num": null |
| } |
| ], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "Will ms ever be as frequent as mr? a corpus-based comparison of gendered terms across four diachronic corpora of british english", |
| "authors": [ |
| { |
| "first": "Paul", |
| "middle": [], |
| "last": "Baker", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "Gender and Language", |
| "volume": "", |
| "issue": "1", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Paul Baker. 2013. Will ms ever be as frequent as mr? a corpus-based comparison of gendered terms across four diachronic corpora of british english. Gender and Language, 1(1).", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "Enriching word vectors with subword information", |
| "authors": [ |
| { |
| "first": "Piotr", |
| "middle": [], |
| "last": "Bojanowski", |
| "suffix": "" |
| }, |
| { |
| "first": "Edouard", |
| "middle": [], |
| "last": "Grave", |
| "suffix": "" |
| }, |
| { |
| "first": "Armand", |
| "middle": [], |
| "last": "Joulin", |
| "suffix": "" |
| }, |
| { |
| "first": "Tomas", |
| "middle": [], |
| "last": "Mikolov", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Transactions of the Association for Computational Linguistics", |
| "volume": "5", |
| "issue": "", |
| "pages": "135--146", |
| "other_ids": { |
| "DOI": [ |
| "10.1162/tacl_a_00051" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Piotr Bojanowski, Edouard Grave, Armand Joulin, and Tomas Mikolov. 2017. Enriching word vectors with subword information. Transactions of the Associa- tion for Computational Linguistics, 5:135-146.", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "Man is to computer programmer as woman is to homemaker? debiasing word embeddings", |
| "authors": [ |
| { |
| "first": "Tolga", |
| "middle": [], |
| "last": "Bolukbasi", |
| "suffix": "" |
| }, |
| { |
| "first": "Kai-Wei", |
| "middle": [], |
| "last": "Chang", |
| "suffix": "" |
| }, |
| { |
| "first": "Y", |
| "middle": [], |
| "last": "James", |
| "suffix": "" |
| }, |
| { |
| "first": "Venkatesh", |
| "middle": [], |
| "last": "Zou", |
| "suffix": "" |
| }, |
| { |
| "first": "Adam", |
| "middle": [ |
| "T" |
| ], |
| "last": "Saligrama", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Kalai", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "Advances in neural information processing systems", |
| "volume": "", |
| "issue": "", |
| "pages": "4349--4357", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Tolga Bolukbasi, Kai-Wei Chang, James Y Zou, Venkatesh Saligrama, and Adam T Kalai. 2016. Man is to computer programmer as woman is to homemaker? debiasing word embeddings. In Ad- vances in neural information processing systems, pages 4349-4357.", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "Semantics derived automatically from language corpora contain human-like biases", |
| "authors": [ |
| { |
| "first": "Aylin", |
| "middle": [], |
| "last": "Caliskan", |
| "suffix": "" |
| }, |
| { |
| "first": "Joanna", |
| "middle": [ |
| "J" |
| ], |
| "last": "Bryson", |
| "suffix": "" |
| }, |
| { |
| "first": "Arvind", |
| "middle": [], |
| "last": "Narayanan", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Science", |
| "volume": "356", |
| "issue": "6334", |
| "pages": "183--186", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Aylin Caliskan, Joanna J Bryson, and Arvind Narayanan. 2017. Semantics derived automatically from language corpora contain human-like biases. Science, 356(6334):183-186.", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "Joint learning of character and word embeddings", |
| "authors": [ |
| { |
| "first": "Xinxiong", |
| "middle": [], |
| "last": "Chen", |
| "suffix": "" |
| }, |
| { |
| "first": "Lei", |
| "middle": [], |
| "last": "Xu", |
| "suffix": "" |
| }, |
| { |
| "first": "Zhiyuan", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "" |
| }, |
| { |
| "first": "Maosong", |
| "middle": [], |
| "last": "Sun", |
| "suffix": "" |
| }, |
| { |
| "first": "Huanbo", |
| "middle": [], |
| "last": "Luan", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "Twenty-Fourth International Joint Conference on Artificial Intelligence", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Xinxiong Chen, Lei Xu, Zhiyuan Liu, Maosong Sun, and Huanbo Luan. 2015. Joint learning of charac- ter and word embeddings. In Twenty-Fourth Inter- national Joint Conference on Artificial Intelligence.", |
| "links": null |
| }, |
| "BIBREF5": { |
| "ref_id": "b5", |
| "title": "Proceedings of the National Academy of ences of the United States of America", |
| "authors": [ |
| { |
| "first": "Nikhil", |
| "middle": [], |
| "last": "Garg", |
| "suffix": "" |
| }, |
| { |
| "first": "Londa", |
| "middle": [], |
| "last": "Schiebinger", |
| "suffix": "" |
| }, |
| { |
| "first": "Dan", |
| "middle": [], |
| "last": "Jurafsky", |
| "suffix": "" |
| }, |
| { |
| "first": "James", |
| "middle": [], |
| "last": "Zou", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "", |
| "volume": "115", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Nikhil Garg, Londa Schiebinger, Dan Jurafsky, and James Zou. 2018. Word embeddings quantify 100 years of gender and ethnic stereotypes. Proceed- ings of the National Academy of ences of the United States of America, 115(16):E3635.", |
| "links": null |
| }, |
| "BIBREF6": { |
| "ref_id": "b6", |
| "title": "Unsupervised discovery of gendered language through latent-variable modeling", |
| "authors": [ |
| { |
| "first": "Alexander Miserlis", |
| "middle": [], |
| "last": "Hoyle", |
| "suffix": "" |
| }, |
| { |
| "first": "Lawrence", |
| "middle": [], |
| "last": "Wolf-Sonkin", |
| "suffix": "" |
| }, |
| { |
| "first": "Hanna", |
| "middle": [], |
| "last": "Wallach", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "1706--1716", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Alexander Miserlis Hoyle, Lawrence Wolf-Sonkin, Hanna Wallach, Isabelle Augenstein, and Ryan Cot- terell. 2019. Unsupervised discovery of gendered language through latent-variable modeling. In Pro- ceedings of the 57th Annual Meeting of the Asso- ciation for Computational Linguistics, pages 1706- 1716.", |
| "links": null |
| }, |
| "BIBREF7": { |
| "ref_id": "b7", |
| "title": "Language and woman's place. Language in society", |
| "authors": [ |
| { |
| "first": "Robin", |
| "middle": [], |
| "last": "Lakoff", |
| "suffix": "" |
| } |
| ], |
| "year": 1973, |
| "venue": "", |
| "volume": "2", |
| "issue": "", |
| "pages": "45--79", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Robin Lakoff. 1973. Language and woman's place. Language in society, 2(1):45-79.", |
| "links": null |
| }, |
| "BIBREF8": { |
| "ref_id": "b8", |
| "title": "Gender stereotypes are reflected in the distributional structure of 25 languages", |
| "authors": [ |
| { |
| "first": "Molly", |
| "middle": [], |
| "last": "Lewis", |
| "suffix": "" |
| }, |
| { |
| "first": "Gary", |
| "middle": [], |
| "last": "Lupyan", |
| "suffix": "" |
| } |
| ], |
| "year": 2020, |
| "venue": "Nature human behaviour", |
| "volume": "", |
| "issue": "", |
| "pages": "1--8", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Molly Lewis and Gary Lupyan. 2020. Gender stereo- types are reflected in the distributional structure of 25 languages. Nature human behaviour, pages 1-8.", |
| "links": null |
| }, |
| "BIBREF9": { |
| "ref_id": "b9", |
| "title": "Analogical reasoning on Chinese morphological and semantic relations", |
| "authors": [ |
| { |
| "first": "Shen", |
| "middle": [], |
| "last": "Li", |
| "suffix": "" |
| }, |
| { |
| "first": "Zhe", |
| "middle": [], |
| "last": "Zhao", |
| "suffix": "" |
| }, |
| { |
| "first": "Renfen", |
| "middle": [], |
| "last": "Hu", |
| "suffix": "" |
| }, |
| { |
| "first": "Wensi", |
| "middle": [], |
| "last": "Li", |
| "suffix": "" |
| }, |
| { |
| "first": "Tao", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "" |
| }, |
| { |
| "first": "Xiaoyong", |
| "middle": [], |
| "last": "Du", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics", |
| "volume": "2", |
| "issue": "", |
| "pages": "138--143", |
| "other_ids": { |
| "DOI": [ |
| "10.18653/v1/P18-2023" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Shen Li, Zhe Zhao, Renfen Hu, Wensi Li, Tao Liu, and Xiaoyong Du. 2018. Analogical reasoning on Chi- nese morphological and semantic relations. In Pro- ceedings of the 56th Annual Meeting of the Associa- tion for Computational Linguistics (Volume 2: Short Papers), pages 138-143, Melbourne, Australia. As- sociation for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF10": { |
| "ref_id": "b10", |
| "title": "Analyze, detect and remove gender stereotyping from bollywood movies", |
| "authors": [ |
| { |
| "first": "Nishtha", |
| "middle": [], |
| "last": "Madaan", |
| "suffix": "" |
| }, |
| { |
| "first": "Sameep", |
| "middle": [], |
| "last": "Mehta", |
| "suffix": "" |
| }, |
| { |
| "first": "Taneea", |
| "middle": [], |
| "last": "Agrawaal", |
| "suffix": "" |
| }, |
| { |
| "first": "Vrinda", |
| "middle": [], |
| "last": "Malhotra", |
| "suffix": "" |
| }, |
| { |
| "first": "Aditi", |
| "middle": [], |
| "last": "Aggarwal", |
| "suffix": "" |
| }, |
| { |
| "first": "Yatin", |
| "middle": [], |
| "last": "Gupta", |
| "suffix": "" |
| }, |
| { |
| "first": "Mayank", |
| "middle": [], |
| "last": "Saxena", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "Conference on Fairness, Accountability and Transparency", |
| "volume": "", |
| "issue": "", |
| "pages": "92--105", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Nishtha Madaan, Sameep Mehta, Taneea Agrawaal, Vrinda Malhotra, Aditi Aggarwal, Yatin Gupta, and Mayank Saxena. 2018. Analyze, detect and remove gender stereotyping from bollywood movies. In Conference on Fairness, Accountability and Trans- parency, pages 92-105.", |
| "links": null |
| }, |
| "BIBREF11": { |
| "ref_id": "b11", |
| "title": "Gender bias and sexism in language", |
| "authors": [ |
| { |
| "first": "Michela", |
| "middle": [], |
| "last": "Menegatti", |
| "suffix": "" |
| }, |
| { |
| "first": "Monica", |
| "middle": [], |
| "last": "Rubini", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Oxford Research Encyclopedia of Communication", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Michela Menegatti and Monica Rubini. 2017. Gender bias and sexism in language. In Oxford Research Encyclopedia of Communication.", |
| "links": null |
| }, |
| "BIBREF12": { |
| "ref_id": "b12", |
| "title": "Efficient estimation of word representations in vector space", |
| "authors": [ |
| { |
| "first": "Tomas", |
| "middle": [], |
| "last": "Mikolov", |
| "suffix": "" |
| }, |
| { |
| "first": "Kai", |
| "middle": [], |
| "last": "Chen", |
| "suffix": "" |
| }, |
| { |
| "first": "Greg", |
| "middle": [], |
| "last": "Corrado", |
| "suffix": "" |
| }, |
| { |
| "first": "Jeffrey", |
| "middle": [], |
| "last": "Dean", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Tomas Mikolov, Kai Chen, Greg Corrado, and Jeffrey Dean. 2013a. Efficient estimation of word represen- tations in vector space.", |
| "links": null |
| }, |
| "BIBREF13": { |
| "ref_id": "b13", |
| "title": "Distributed representations of words and phrases and their compositionality", |
| "authors": [ |
| { |
| "first": "Tomas", |
| "middle": [], |
| "last": "Mikolov", |
| "suffix": "" |
| }, |
| { |
| "first": "Ilya", |
| "middle": [], |
| "last": "Sutskever", |
| "suffix": "" |
| }, |
| { |
| "first": "Kai", |
| "middle": [], |
| "last": "Chen", |
| "suffix": "" |
| }, |
| { |
| "first": "Greg", |
| "middle": [ |
| "S" |
| ], |
| "last": "Corrado", |
| "suffix": "" |
| }, |
| { |
| "first": "Jeff", |
| "middle": [], |
| "last": "Dean", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "Advances in neural information processing systems", |
| "volume": "", |
| "issue": "", |
| "pages": "3111--3119", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Tomas Mikolov, Ilya Sutskever, Kai Chen, Greg S Cor- rado, and Jeff Dean. 2013b. Distributed representa- tions of words and phrases and their compositional- ity. In Advances in neural information processing systems, pages 3111-3119.", |
| "links": null |
| }, |
| "BIBREF14": { |
| "ref_id": "b14", |
| "title": "GloVe: Global vectors for word representation", |
| "authors": [ |
| { |
| "first": "Jeffrey", |
| "middle": [], |
| "last": "Pennington", |
| "suffix": "" |
| }, |
| { |
| "first": "Richard", |
| "middle": [], |
| "last": "Socher", |
| "suffix": "" |
| }, |
| { |
| "first": "Christopher", |
| "middle": [], |
| "last": "Manning", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP)", |
| "volume": "", |
| "issue": "", |
| "pages": "1532--1543", |
| "other_ids": { |
| "DOI": [ |
| "10.3115/v1/D14-1162" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jeffrey Pennington, Richard Socher, and Christo- pher Manning. 2014a. GloVe: Global vectors for word representation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 1532-1543, Doha, Qatar. Association for Computational Lin- guistics.", |
| "links": null |
| }, |
| "BIBREF15": { |
| "ref_id": "b15", |
| "title": "Glove: Global vectors for word representation", |
| "authors": [ |
| { |
| "first": "Jeffrey", |
| "middle": [], |
| "last": "Pennington", |
| "suffix": "" |
| }, |
| { |
| "first": "Richard", |
| "middle": [], |
| "last": "Socher", |
| "suffix": "" |
| }, |
| { |
| "first": "Christopher D", |
| "middle": [], |
| "last": "Manning", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP)", |
| "volume": "", |
| "issue": "", |
| "pages": "1532--1543", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jeffrey Pennington, Richard Socher, and Christopher D Manning. 2014b. Glove: Global vectors for word representation. In Proceedings of the 2014 confer- ence on empirical methods in natural language pro- cessing (EMNLP), pages 1532-1543.", |
| "links": null |
| }, |
| "BIBREF16": { |
| "ref_id": "b16", |
| "title": "Assessing gender bias in machine translation: a case study with google translate", |
| "authors": [ |
| { |
| "first": "Pedro", |
| "middle": [ |
| "H" |
| ], |
| "last": "Marcelo Or Prates", |
| "suffix": "" |
| }, |
| { |
| "first": "Lu\u00eds C", |
| "middle": [], |
| "last": "Avelar", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Lamb", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "Neural Computing and Applications", |
| "volume": "", |
| "issue": "", |
| "pages": "1--19", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Marcelo OR Prates, Pedro H Avelar, and Lu\u00eds C Lamb. 2019. Assessing gender bias in machine translation: a case study with google translate. Neural Comput- ing and Applications, pages 1-19.", |
| "links": null |
| }, |
| "BIBREF17": { |
| "ref_id": "b17", |
| "title": "When and why are pre-trained word embeddings useful for neural machine translation?", |
| "authors": [ |
| { |
| "first": "Ye", |
| "middle": [], |
| "last": "Qi", |
| "suffix": "" |
| }, |
| { |
| "first": "Devendra", |
| "middle": [], |
| "last": "Sachan", |
| "suffix": "" |
| }, |
| { |
| "first": "Matthieu", |
| "middle": [], |
| "last": "Felix", |
| "suffix": "" |
| }, |
| { |
| "first": "Sarguna", |
| "middle": [], |
| "last": "Padmanabhan", |
| "suffix": "" |
| }, |
| { |
| "first": "Graham", |
| "middle": [], |
| "last": "Neubig", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies", |
| "volume": "2", |
| "issue": "", |
| "pages": "529--535", |
| "other_ids": { |
| "DOI": [ |
| "10.18653/v1/N18-2084" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ye Qi, Devendra Sachan, Matthieu Felix, Sarguna Pad- manabhan, and Graham Neubig. 2018. When and why are pre-trained word embeddings useful for neu- ral machine translation? In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 2 (Short Papers), pages 529-535, New Orleans, Louisiana. Associa- tion for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF18": { |
| "ref_id": "b18", |
| "title": "Gender bias in pretrained Swedish embeddings", |
| "authors": [ |
| { |
| "first": "Magnus", |
| "middle": [], |
| "last": "Sahlgren", |
| "suffix": "" |
| }, |
| { |
| "first": "Fredrik", |
| "middle": [], |
| "last": "Olsson", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "Proceedings of the 22nd Nordic Conference on Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "35--43", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Magnus Sahlgren and Fredrik Olsson. 2019. Gender bias in pretrained Swedish embeddings. In Pro- ceedings of the 22nd Nordic Conference on Com- putational Linguistics, pages 35-43, Turku, Finland. Link\u00f6ping University Electronic Press.", |
| "links": null |
| }, |
| "BIBREF19": { |
| "ref_id": "b19", |
| "title": "Using word embeddings to examine gender bias in dutch newspapers", |
| "authors": [ |
| { |
| "first": "Melvin", |
| "middle": [], |
| "last": "Wevers", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "Proceedings of the 1st International Workshop on Computational Approaches to Historical Language Change", |
| "volume": "", |
| "issue": "", |
| "pages": "92--97", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Melvin Wevers. 2019. Using word embeddings to ex- amine gender bias in dutch newspapers, 1950-1990. In Proceedings of the 1st International Workshop on Computational Approaches to Historical Language Change, pages 92-97.", |
| "links": null |
| }, |
| "BIBREF20": { |
| "ref_id": "b20", |
| "title": "Refining word embeddings for sentiment analysis", |
| "authors": [ |
| { |
| "first": "Liang-Chih", |
| "middle": [], |
| "last": "Yu", |
| "suffix": "" |
| }, |
| { |
| "first": "Jin", |
| "middle": [], |
| "last": "Wang", |
| "suffix": "" |
| }, |
| { |
| "first": "K", |
| "middle": [ |
| "Robert" |
| ], |
| "last": "Lai", |
| "suffix": "" |
| }, |
| { |
| "first": "Xuejie", |
| "middle": [], |
| "last": "Zhang", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "534--539", |
| "other_ids": { |
| "DOI": [ |
| "10.18653/v1/D17-1056" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Liang-Chih Yu, Jin Wang, K. Robert Lai, and Xuejie Zhang. 2017. Refining word embeddings for sen- timent analysis. In Proceedings of the 2017 Con- ference on Empirical Methods in Natural Language Processing, pages 534-539, Copenhagen, Denmark. Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF21": { |
| "ref_id": "b21", |
| "title": "Mitigating unwanted biases with adversarial learning", |
| "authors": [ |
| { |
| "first": "Brian", |
| "middle": [], |
| "last": "Hu Zhang", |
| "suffix": "" |
| }, |
| { |
| "first": "Blake", |
| "middle": [], |
| "last": "Lemoine", |
| "suffix": "" |
| }, |
| { |
| "first": "Margaret", |
| "middle": [], |
| "last": "Mitchell", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "Proceedings of the", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Brian Hu Zhang, Blake Lemoine, and Margaret Mitchell. 2018. Mitigating unwanted biases with adversarial learning. In Proceedings of the 2018", |
| "links": null |
| }, |
| "BIBREF22": { |
| "ref_id": "b22", |
| "title": "AAAI/ACM Conference on AI, Ethics, and Society", |
| "authors": [], |
| "year": null, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "335--340", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "AAAI/ACM Conference on AI, Ethics, and Society, pages 335-340.", |
| "links": null |
| }, |
| "BIBREF23": { |
| "ref_id": "b23", |
| "title": "Men also like shopping: Reducing gender bias amplification using corpus-level constraints", |
| "authors": [ |
| { |
| "first": "Jieyu", |
| "middle": [], |
| "last": "Zhao", |
| "suffix": "" |
| }, |
| { |
| "first": "Tianlu", |
| "middle": [], |
| "last": "Wang", |
| "suffix": "" |
| }, |
| { |
| "first": "Mark", |
| "middle": [], |
| "last": "Yatskar", |
| "suffix": "" |
| }, |
| { |
| "first": "Vicente", |
| "middle": [], |
| "last": "Ordonez", |
| "suffix": "" |
| }, |
| { |
| "first": "Kai-Wei", |
| "middle": [], |
| "last": "Chang", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "2979--2989", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jieyu Zhao, Tianlu Wang, Mark Yatskar, Vicente Or- donez, and Kai-Wei Chang. 2017. Men also like shopping: Reducing gender bias amplification using corpus-level constraints. In Proceedings of the 2017 Conference on Empirical Methods in Natural Lan- guage Processing, pages 2979-2989.", |
| "links": null |
| }, |
| "BIBREF24": { |
| "ref_id": "b24", |
| "title": "Learning gender-neutral word embeddings", |
| "authors": [ |
| { |
| "first": "Jieyu", |
| "middle": [], |
| "last": "Zhao", |
| "suffix": "" |
| }, |
| { |
| "first": "Yichao", |
| "middle": [], |
| "last": "Zhou", |
| "suffix": "" |
| }, |
| { |
| "first": "Zeyu", |
| "middle": [], |
| "last": "Li", |
| "suffix": "" |
| }, |
| { |
| "first": "Wei", |
| "middle": [], |
| "last": "Wang", |
| "suffix": "" |
| }, |
| { |
| "first": "Kai-Wei", |
| "middle": [], |
| "last": "Chang", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "4847--4853", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jieyu Zhao, Yichao Zhou, Zeyu Li, Wei Wang, and Kai- Wei Chang. 2018. Learning gender-neutral word embeddings. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Process- ing, pages 4847-4853.", |
| "links": null |
| }, |
| "BIBREF25": { |
| "ref_id": "b25", |
| "title": "Great males and stubborn females: A diachronic study of corpusbased gendered skewness in chinese adjectives", |
| "authors": [ |
| { |
| "first": "Shucheng", |
| "middle": [], |
| "last": "Zhu", |
| "suffix": "" |
| }, |
| { |
| "first": "Pengyuan", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "" |
| } |
| ], |
| "year": 2020, |
| "venue": "Proceedings of the 19th Chinese National Conference on Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "31--42", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Shucheng Zhu and Pengyuan Liu. 2020. Great males and stubborn females: A diachronic study of corpus- based gendered skewness in chinese adjectives. In Proceedings of the 19th Chinese National Confer- ence on Computational Linguistics, pages 31-42.", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "FIGREF0": { |
| "type_str": "figure", |
| "text": "An example of semantic relation between Chinese words and characters. Pinying (pronunciation of the word or character) is in the lower right parentheses; English translation is noted directly below the word", |
| "num": null, |
| "uris": null |
| }, |
| "FIGREF1": { |
| "type_str": "figure", |
| "text": "Scatter plots of the data group with AGSS scores <3. AGSS refers to the AGSS scores and bias word and bias char refers to the generated gender bias scores.", |
| "num": null, |
| "uris": null |
| }, |
| "FIGREF2": { |
| "type_str": "figure", |
| "text": "Scatter plots of the data group with AGSS scores >3. AGSS refers to the AGSS scores and bias word and bias char refers to the generated gender bias scores.", |
| "num": null, |
| "uris": null |
| }, |
| "TABREF0": { |
| "content": "<table/>", |
| "type_str": "table", |
| "num": null, |
| "text": "The details of the Chinese news corpus.", |
| "html": null |
| }, |
| "TABREF1": { |
| "content": "<table/>", |
| "type_str": "table", |
| "num": null, |
| "text": "Word embeddings training hyper-parameter details.", |
| "html": null |
| }, |
| "TABREF3": { |
| "content": "<table/>", |
| "type_str": "table", |
| "num": null, |
| "text": "Accuracy scores of different word embeddings in the evaluation tasks. The results are reported as acc\u00d7 100.", |
| "html": null |
| }, |
| "TABREF4": { |
| "content": "<table><tr><td>: Example data from AGSS. Each word is trans-</td></tr><tr><td>lated into English.</td></tr><tr><td>correlation, 0 indicates no linear correlation and</td></tr><tr><td>\u22121 indicates a strong negative linear correlation.</td></tr></table>", |
| "type_str": "table", |
| "num": null, |
| "text": "", |
| "html": null |
| }, |
| "TABREF5": { |
| "content": "<table/>", |
| "type_str": "table", |
| "num": null, |
| "text": "Statistics of the gender bias scores from two embeddings.", |
| "html": null |
| }, |
| "TABREF6": { |
| "content": "<table/>", |
| "type_str": "table", |
| "num": null, |
| "text": "Pearson correlation coefficient between AGSS score and gender bias scores from trained vectors. CBOW score and CWE score refer to the gender bias score from word vectors trained with CBOW and CWE model.", |
| "html": null |
| }, |
| "TABREF7": { |
| "content": "<table><tr><td/><td>CBOW CWE</td></tr><tr><td>Pearson coefficient</td><td>0.036 0.020</td></tr><tr><td>p-value</td><td>0.543 0.724</td></tr></table>", |
| "type_str": "table", |
| "num": null, |
| "text": "Pearson correlation coefficient of the data group with AGSS scores <3.", |
| "html": null |
| }, |
| "TABREF8": { |
| "content": "<table/>", |
| "type_str": "table", |
| "num": null, |
| "text": "Pearson correlation coefficient of the data group with AGSS scores >3.", |
| "html": null |
| } |
| } |
| } |
| } |