| { |
| "paper_id": "2020", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T15:21:02.456128Z" |
| }, |
| "title": "", |
| "authors": [], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "", |
| "pdf_parse": { |
| "paper_id": "2020", |
| "_pdf_hash": "", |
| "abstract": [], |
| "body_text": [ |
| { |
| "text": "SIGTYP 2020 is the second edition of the workshop for typology-related research and its integration into multilingual Natural Language Processing (NLP), which was inaugurated last year at ACL 2019. The workshop is co-located with the 2020 Conference on Empirical Methods in Natural Language Processing, which takes place virtually this year. Our workshop includes a shared task on typological feature prediction, which saw the participation of 5 teams for a total of 8 systems submitted.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": null |
| }, |
| { |
| "text": "Information from topic contexts: the prediction of aspectual coding of verbs in Russian Michael Richter and Tariq Yousef Based on Shannon's coding theorem, we predict that aspectual coding asymmetries of verbs in Russian can be predicted by the verbal feature Average Information Content. We employ the novel Topic Context Model that calculates the verbal information content from the number of topics in the target words' larger discourses and their local discourses. In contrast to a previous study, TCM yielded disappointing results in this study which is, as we conclude, mainly due to the small number of local contexts we utilized.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Non-archival Abstracts", |
| "sec_num": null |
| }, |
| { |
| "text": "Chiara Alzetta, Felice Dell'Orletta, Simonetta Montemagni and Giulia Venturi This contribution presents the results of a method for typological feature identification in multilingual treebanks. The results are exemplified on Italian and English subject relations. Applications of the method for multilingual dependency parsing evaluation are discussed.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Uncovering Typological Context-Sensitive Features", |
| "sec_num": null |
| }, |
| { |
| "text": "Alexander Gutkin, Martin Jansche and Lucy Skidmore This extended abstract provides a summary of our past and ongoing work on assessing the quality of multilingual phoneme inventories derived from typological resources, inducing phonological inventories using distinctive feature representations from the speech data and the important role phonological typology plays in these approaches.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Towards Induction of Structured Phoneme Inventories", |
| "sec_num": null |
| }, |
| { |
| "text": "Is Typology-Based Adaptation Effective for Multilingual Sequence Labelling? Ahmet \u00dcst\u00fcn, Arianna Bisazza, Gosse Bouma and Gertjan van Noord Recent work has shown that a single multilingual model with typologically informed parameter sharing can improve the performance in dependency parsing on both high-resource and zero-shot conditions. In this work, we investigate whether such improvements are also observed in the POS, NER and morphological tagging tasks.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Towards Induction of Structured Phoneme Inventories", |
| "sec_num": null |
| }, |
| { |
| "text": "Aryaman Arora and Nathan Schneider The use of specific case markers and adpositions for particular semantic roles is idiosyncratic to every language. This poses problems in many natural language processing tasks such as machine translation and semantic role labelling. Models for these tasks rely on human-annotated corpora as training data.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "SNACS Annotation of Case Markers and Adpositions in Hindi", |
| "sec_num": null |
| }, |
| { |
| "text": "There is a lack of corpora in South Asian languages for such tasks. Even Hindi, despite being a resourcerich language, is limited in available labelled data. This extended abstract presents the in-progress annotation of case markers and adpositions in a Hindi corpus, employing the cross-lingual scheme proposed by Schneider et al. (2017), Semantic Network of Adposition and Case Supersenses (SNACS). The SNACS guidelines we developed also apply to Urdu. We hope to finalize this corpus and develop NLP tools making use of the dataset, as well as promote NLP for typologically similar South Asian languages.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "SNACS Annotation of Case Markers and Adpositions in Hindi", |
| "sec_num": null |
| }, |
| { |
| "text": "Yushi Hu, Shane Settle and Karen Livescu Acoustic word embeddings (AWEs) are vector representations of spoken word segments. AWEs can be learned jointly with embeddings of character sequences, to generate phonetically meaningful embeddings of written words, or acoustically grounded word embeddings (AGWEs). Such embeddings have been used to improve speech retrieval, recognition, and spoken term discovery. In this work, we extend this idea to multiple low-resource languages. We jointly train an AWE model and an AGWE model, using phonetically transcribed data from multiple languages. The pre-trained models can then be used for unseen zero-resource languages, or fine-tuned on data from low-resource languages. We also investigate distinctive features, as an alternative to phone labels, to better share cross-lingual information. We test our models on word discrimination tasks for twelve languages while varying the amount of target language training data, and find significant benefits to the proposed multilingual approach.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Multilingual Jointly Trained Acoustic and Written Word Embeddings", |
| "sec_num": null |
| }, |
| { |
| "text": "Multilingual BERT learns abstract case representations Isabel Papadimitriou, Ethan A. Chi, Richard Futrell and Kyle Mahowald We investigate how Multilingual BERT (mBERT) encodes grammar by examining how the high-order grammatical feature of morphosyntactic alignment (how different languages define what counts as a \"subject\") is manifested across the embedding spaces of different languages. To understand if and how morphosyntactic alignment affects contextual embedding spaces, we train classifiers to recover the subjecthood of mBERT embeddings in transitive sentences (which do not contain overt information about morphosyntactic alignment) and then evaluate them zero-shot on intransitive sentences (where subjecthood classification depends on alignment), within and across languages. We find that the resulting classifier distributions reflect the morphosyntactic alignment of their training languages. Our results demonstrate that mBERT representations are influenced by high-level grammatical features that are not manifested in any one input sentence, and that this is robust across languages. Further examining the characteristics that our classifiers rely on, we find that features such as passive voice, animacy and case strongly correlate with classification decisions, suggesting that mBERT does not encode a purely syntactic subjecthood, but a continuous subjecthood as is proposed in much of the functional linguistics literature. Together, these results provide insight into how grammatical features manifest in contextual embedding spaces, at a level of abstraction not covered by previous work.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Multilingual Jointly Trained Acoustic and Written Word Embeddings", |
| "sec_num": null |
| }, |
| { |
| "text": "DEmA: the Pavia Diachronic Emergence of Alignment database Sonia Cristofaro and Guglielmo Inglese This paper describes a workflow to impute missing values in a typological database, a subset of the World Atlas of Language Structures (WALS). Using a world-wide phylogeny derived from lexical data, the model assumes a phylogenetic continuous time Markov chain governing the evolution of typological values. Data imputation is performed via a Maximum Likelihood estimation on the basis of this model. As back-off model for languages whose phylogenetic position is unknown, a k-nearest neighbor classification based on geographic distance is performed.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Keyword Spotting: A quick-and-dirty method for extracting typological features of language from grammatical descriptions Harald Hammarstr\u00f6m", |
| "sec_num": null |
| }, |
| { |
| "text": "xiv A dataset and metric to evaluate lexical extraction from parallel corpora Barend Beekhuizen This work presents a novel dataset and a metric for evaluating methods for automated extraction of translation equivalent expressions in massively parallel corpora for the purposes of lexical semantic typology. Patterns in the annotation and the evaluation of the extraction methods were discussed, and directions for future research were indicated.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Keyword Spotting: A quick-and-dirty method for extracting typological features of language from grammatical descriptions Harald Hammarstr\u00f6m", |
| "sec_num": null |
| }, |
| { |
| "text": "The role of community size and network structure in shaping linguistic diversity: experimental evidence Limor Raviv, Antje Mayer and Shiri Lev-Ari Why are there so many different languages in the world? How much do languages differ from each other in terms of their linguistic structure? And how do such differences come about?", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Keyword Spotting: A quick-and-dirty method for extracting typological features of language from grammatical descriptions Harald Hammarstr\u00f6m", |
| "sec_num": null |
| }, |
| { |
| "text": "One possibility is that linguistic diversity stems from differences in the social environments in which languages evolve. Specifically, it has been suggested that small, tightly knit communities can maintain high levels of linguistic complexity, while bigger and sparser communities tend to have languages that are structurally simpler, i.e., languages with more regular and more systematic grammars. However, to date this hypothesis has not been tested experimentally. Moreover, community size and network structure are typically confounded in the real-world, making it hard to evaluate the unique contribution of each social factor to this pattern of variation.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Keyword Spotting: A quick-and-dirty method for extracting typological features of language from grammatical descriptions Harald Hammarstr\u00f6m", |
| "sec_num": null |
| }, |
| { |
| "text": "To address this issue, we used a novel group communication paradigm. This experimental paradigm allowed us to look at the live formation of new languages that were created in the lab by different microsocieties under different social conditions. By analyzing the emerging languages, we could tease apart the causal role of community size and network structure, and see how the process of language evolution and change is shaped by the fact that languages develop in communities of different sizes and different social structures.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Keyword Spotting: A quick-and-dirty method for extracting typological features of language from grammatical descriptions Harald Hammarstr\u00f6m", |
| "sec_num": null |
| }, |
| { |
| "text": "During the group communication game, participants' goal was to communicate successfully about different novel scenes, using only invented nonsense words. A 'speaker' would see one of four shapes moving in some direction on a screen, and would type in a nonsense word to describe the scene (its shape and direction). The 'listener' would then guess which scene their partner was referring to by selecting one of eight scenes on their own screen. Participants received points for every successful interaction (correct guesses), and also feedback to allow them to learn for future interactions. Participants paired up with a different person from their group at every new round, taking turns producing and guessing words.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Keyword Spotting: A quick-and-dirty method for extracting typological features of language from grammatical descriptions Harald Hammarstr\u00f6m", |
| "sec_num": null |
| }, |
| { |
| "text": "At the start of the game, people would randomly guessed meanings and make up new names. Over the course of several hours, participants started to combine words or part-words systematically, creating an actual mini-language. For instance, in one group, 'wowo-ik' meant that a specific shape was going up and right, whereas 'wowo-ii' meant that the same shape was going straight up. With such a 'regular' system, it becomes easier to predict the meaning of new labels ('mop-ik' meant a different shape going up and right).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Keyword Spotting: A quick-and-dirty method for extracting typological features of language from grammatical descriptions Harald Hammarstr\u00f6m", |
| "sec_num": null |
| }, |
| { |
| "text": "In the first experiment, we examined the role of community size by having participants play in either 'small' groups of four participants or 'large' groups of eight participants. Would the large groups invent more structured languages than the small groups? Results showed that larger groups created languages with more systematic grammars, and did so faster and more consistently than small groups. This finding suggested that the number of people in the community can affect the grammar of languages. We suggest that larger groups are under a stronger pressure to create systematic languages because members of larger groups are typically faced with more input variability, and have less shared history with each member of their group.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Keyword Spotting: A quick-and-dirty method for extracting typological features of language from grammatical descriptions Harald Hammarstr\u00f6m", |
| "sec_num": null |
| }, |
| { |
| "text": "xv In contrast, in the second experiment we found no evidence for a similar role of network structure. When we compared the performance of three network conditions (i.e., fully connected networks, small-world networks, scale-free networks) that varied in their degree of connectivity while group size constant was kept constant, we found that all groups developed languages that were highly systematic, communicatively efficient, stable, and shared across members, with dense and sparse groups reaching similar levels of linguistic structure over time. Although there were no significant differences between networks with respect to their degree of systematic grammar, we found that small-world networks showed the most variance in their behaviors. This result suggests that small-world networks may be more sensitive to random events (i.e., drift).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Keyword Spotting: A quick-and-dirty method for extracting typological features of language from grammatical descriptions Harald Hammarstr\u00f6m", |
| "sec_num": null |
| }, |
| { |
| "text": "Together, the findings from the two experiments reported above show that factors in the social environment, and specifically community size, can affect patterns of language diversity and shape the nature and structure of languages.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Keyword Spotting: A quick-and-dirty method for extracting typological features of language from grammatical descriptions Harald Hammarstr\u00f6m", |
| "sec_num": null |
| }, |
| { |
| "text": "xvi", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Keyword Spotting: A quick-and-dirty method for extracting typological features of language from grammatical descriptions Harald Hammarstr\u00f6m", |
| "sec_num": null |
| } |
| ], |
| "back_matter": [], |
| "bib_entries": {}, |
| "ref_entries": {} |
| } |
| } |