| { |
| "paper_id": "2020", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T07:33:12.211960Z" |
| }, |
| "title": "From Linguistic Research Projects to Language Technology Platforms : A Case Study in Learner Data", |
| "authors": [ |
| { |
| "first": "Annanda", |
| "middle": [], |
| "last": "Sousa", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "" |
| }, |
| { |
| "first": "Nicolas", |
| "middle": [], |
| "last": "Ballier", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "nicolas.ballier@u-paris.fr" |
| }, |
| { |
| "first": "Thomas", |
| "middle": [], |
| "last": "Gaillat", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "thomas.gaillat@univ-rennes2.fr" |
| }, |
| { |
| "first": "Bernardo", |
| "middle": [], |
| "last": "Stearns", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "bernardo.stearns@nuigalway.ie" |
| }, |
| { |
| "first": "Manel", |
| "middle": [], |
| "last": "Zarrouk", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "zarrouk@lipn.univ-paris13.fr" |
| }, |
| { |
| "first": "Andrew", |
| "middle": [], |
| "last": "Simpkin", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "andrew.simpkin@nuigalway.ie" |
| }, |
| { |
| "first": "Manon", |
| "middle": [], |
| "last": "Bouy\u00e9", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "mbouye@eila.univ-paris-diderot.fr" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "This paper describes the workflow and architecture adopted by a linguistic research project on learner data. We report our experience and present the research outputs turned into resources that we wish to share with the community. We discuss the current limitations and the next steps that could be taken for the scaling and development of our research project. Allying NLP and language-centric AI, we discuss similar projects and possible ways to start collaborating towards potential platform interoperability.", |
| "pdf_parse": { |
| "paper_id": "2020", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "This paper describes the workflow and architecture adopted by a linguistic research project on learner data. We report our experience and present the research outputs turned into resources that we wish to share with the community. We discuss the current limitations and the next steps that could be taken for the scaling and development of our research project. Allying NLP and language-centric AI, we discuss similar projects and possible ways to start collaborating towards potential platform interoperability.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "This paper illustrates the current shift from language sciences to linguistic data science. We intend to describe the prototype of an Automatic Essay Scoring system (AES) user interface predicting proficiency levels in English and discuss scalability, interoperability and some development issues when sharing our models and other research outputs of our project. Automating language assessment is a task conducted with Automatic Essay Scoring systems (AES). Initially based on rule-based approaches (Page, 1968) , more modern systems now rely on probabilistic models. Some of these models depend on the identification of features that are used as predictors of writing quality. Some of these features operationalise complexity and act as criterial features in L2 language (Hawkins and Filipovi\u0107, 2012) . They help build computer models for error detection and automated assessment and, by using model explanation procedures, their significance and effect can be measured. Recent work on identifying criterial features has been fruitful, as many studies have addressed many types of features. However, most of the studies (one notable exception is found in (Volodina et al., 2016) , with a system designed for Swedish) are experimental and do not include any automated pipeline that can handle user data from input to output. In other words, pre-processing and data analysis are not necessarily connected to any machine learning module and a user interface. Most experiments include several experimental stages of data modeling, which impedes any real-life exploitation of the models such as a student typing a text to have it graded. The work on criterial features has also raised the need to build systems dedicated to linguistic feature extraction. The purpose of such a task is to build datasets reflecting the multi-dimensionality of language. Several tools have been developed to suit the needs of specific projects in the extraction of linguistic complexity features (Lu, 2014; Crossley et al., 2014; Crossley et al., 2019; Kyle et al., 2018) . These tools provide features of different dimensions of language. However, it is not possible to apply them to a single data set in one operation. The researcher who wants to weigh the significance of all these features would benefit from a single tool applied uniformly to any data set. Our proposal stems from a project dedicated to predicting proficiency levels in learner English. This system is made up of a user interface in which learners of English can type in a text and immediately be prompted with an assessment of their proficiency level after submission. The system was designed following a modular approach which provides room for other researchers' models. We show that it is possible to use what we have called 'the DemoIT infrastructure' to implement other models dedicated to processing texts with a view to classify them according to predetermined classes. In addition, we have derived a feature extraction pipeline from the demo and it enables researchers to build datasets by applying several state-of-theart tools for further analysis. In an effort to contributing to the FAIR paradigm, we have made available the code of the interface, the initial dataset and our statistical model (the .sav file). This how-to paper guides the computationally literate linguist from the data modelling to the actual web-interface for the deliverables of her linguistic project. Our case study is at the crossroads of", |
| "cite_spans": [ |
| { |
| "start": 500, |
| "end": 512, |
| "text": "(Page, 1968)", |
| "ref_id": "BIBREF25" |
| }, |
| { |
| "start": 773, |
| "end": 802, |
| "text": "(Hawkins and Filipovi\u0107, 2012)", |
| "ref_id": "BIBREF11" |
| }, |
| { |
| "start": 1157, |
| "end": 1180, |
| "text": "(Volodina et al., 2016)", |
| "ref_id": "BIBREF31" |
| }, |
| { |
| "start": 1974, |
| "end": 1984, |
| "text": "(Lu, 2014;", |
| "ref_id": "BIBREF19" |
| }, |
| { |
| "start": 1985, |
| "end": 2007, |
| "text": "Crossley et al., 2014;", |
| "ref_id": "BIBREF4" |
| }, |
| { |
| "start": 2008, |
| "end": 2030, |
| "text": "Crossley et al., 2019;", |
| "ref_id": "BIBREF5" |
| }, |
| { |
| "start": 2031, |
| "end": 2049, |
| "text": "Kyle et al., 2018)", |
| "ref_id": "BIBREF15" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1." |
| }, |
| { |
| "text": "\u2022 research projects in applied linguistics,", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1." |
| }, |
| { |
| "text": "Language Technology Platforms,", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "\u2022 containerisation and virtualisation technologies for", |
| "sec_num": null |
| }, |
| { |
| "text": "\u2022 development of Language Technology platform interoperability: we present our web application and our workflow, as well as the exchange models, data and metadata...", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "\u2022 containerisation and virtualisation technologies for", |
| "sec_num": null |
| }, |
| { |
| "text": "The rest of the paper is organised as follows: section 2 presents some comparable tools for the analysis of complexity. Section 3. describes the context that triggered the need of the implementation of the infrastructure, which is a project aiming to automatically predict the CEFR level of a language learner. Section 4 details the tool we are proposing in this paper as well as some suggestions of improvements. The solution we adopted, \"DemoIt\", is a web-based infrastructure that allows users to demo their text processing systems easily in a scheduled asynchronous way. Section 5 presents the infrastructure we have adopted and the sub-components that can be re-used. Section 6 details the resulting resources we make available as deliverables of the project. Section 7 discusses our next steps in relation to other similar infrastructures, taking into account interoperability, multilingualism, scalability and legal restrictions (GDPR and copyright).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "\u2022 containerisation and virtualisation technologies for", |
| "sec_num": null |
| }, |
| { |
| "text": "A number of projects already exist in the domain of complexity feature extraction. Specific tools are dedicated to a specific dimension of complexity. A number of tools focus on lexical complexity, e.g. LCA (Lu, 2012) and TAALES (Kyle et al., 2018) . Other tools focus on syntactic complexity, e.g. L2SCA (Lu, 2010) and TAASC (Kyle et al., 2018) . Other tools focus on pragmatic dimensions, e.g. cohesion with TAACO (Crossley et al., 2019) and Coh-Metrix (Mc-Namara et al., 2014) . All these tools provide many metrics of one dimension to build datasets for further analysis. More recently, work has been invested in developing common frameworks to support data interoperability with shared tools. CTAP (Chen and Meurers, 2016) is such a tool and allows a researcher to select various types of linguistic features to extract prior to building a customised data set. This approach provides the benefit of letting researchers choose and apply complexity analyses from a broad set of available features.", |
| "cite_spans": [ |
| { |
| "start": 207, |
| "end": 217, |
| "text": "(Lu, 2012)", |
| "ref_id": "BIBREF18" |
| }, |
| { |
| "start": 229, |
| "end": 248, |
| "text": "(Kyle et al., 2018)", |
| "ref_id": "BIBREF15" |
| }, |
| { |
| "start": 305, |
| "end": 315, |
| "text": "(Lu, 2010)", |
| "ref_id": "BIBREF17" |
| }, |
| { |
| "start": 326, |
| "end": 345, |
| "text": "(Kyle et al., 2018)", |
| "ref_id": "BIBREF15" |
| }, |
| { |
| "start": 416, |
| "end": 439, |
| "text": "(Crossley et al., 2019)", |
| "ref_id": "BIBREF5" |
| }, |
| { |
| "start": 444, |
| "end": 479, |
| "text": "Coh-Metrix (Mc-Namara et al., 2014)", |
| "ref_id": null |
| }, |
| { |
| "start": 703, |
| "end": 727, |
| "text": "(Chen and Meurers, 2016)", |
| "ref_id": "BIBREF3" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Existing tools for linguistic complexity feature extraction", |
| "sec_num": "2." |
| }, |
| { |
| "text": "This section presents the experimental setup, the components of our project.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "The Project: a Machine Learning Driven Language Proficiency Level Prediction", |
| "sec_num": "3." |
| }, |
| { |
| "text": "Our project aims to investigate criterial features in learner English and to build a proof-of-concept system for language proficiency level assessment in English. Our research focus is to identify linguistic features and to integrate them within a system with a machine learning component. The purpose is to create a system to analyse learner English essay writings and map them to specific language levels of the Common European Framework of Reference (CEFR) for Languages (European Council, 2001 ). The proposal is a supervised learning approach in which we build several models designed to assign levels of the CEFR, which, to the best of our knowledge is novel. The system is trained on a database of more than 40,000 texts (approx. 3,298,343 tokens) that have already been labeled and grammatically annotated (Geertzen et al., 2013; Huang et al., 2018) . The model relies on error-independent features of English to build a multi-dimensional feature representation of written essays. Figure 1 recaps the pipeline of the project. ", |
| "cite_spans": [ |
| { |
| "start": 474, |
| "end": 497, |
| "text": "(European Council, 2001", |
| "ref_id": "BIBREF6" |
| }, |
| { |
| "start": 814, |
| "end": 837, |
| "text": "(Geertzen et al., 2013;", |
| "ref_id": "BIBREF10" |
| }, |
| { |
| "start": 838, |
| "end": 857, |
| "text": "Huang et al., 2018)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 989, |
| "end": 997, |
| "text": "Figure 1", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Aims of the Project", |
| "sec_num": "3.1." |
| }, |
| { |
| "text": "The model was trained and tested on the Spanish and French L1 subsets of the Education First-Cambridge Open Language Database (EFCAMDAT) (Geertzen et al., 2013) , an 83-million-word corpus collected and made available by Cambridge University and its partner, the organization Education First. Data 1 from 49,813 texts written by 8,851 learners were extracted. The model was also tested 2 on the CEFR ASAG corpus (Tack et al., 2017) , another collection of learner texts made up of short answers to open-ended questions and written by French L1 learners of English. The texts were graded with CEFR levels by three experts. By using the aforementioned corpus subsets, we implemented a program pipeline which is designed to convert the texts into series of values, subsequently used as features. Several state-of-the-art tools are exploited to extract features of several linguistic dimensions and create three datasets. Two internal sets are created from the 49,813 observations, i.e. the training set (75% randomly extracted from the EFCAMDAT corpus) and a test set (25% randomly extracted). One external data set was created from the ASAG-CEFR corpus including 299 observations. The internal dataset will be made available online for the research community on the EFCAMDAT website. Programming scripts will also be made available via an online software development platform. In order to ensure compliance with the General Data Protection Regulation (GDPR), data will be anonymised and no personal identification of learners will be used and published.", |
| "cite_spans": [ |
| { |
| "start": 137, |
| "end": 160, |
| "text": "(Geertzen et al., 2013)", |
| "ref_id": "BIBREF10" |
| }, |
| { |
| "start": 412, |
| "end": 431, |
| "text": "(Tack et al., 2017)", |
| "ref_id": "BIBREF28" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": ". Corpora and Dataset", |
| "sec_num": null |
| }, |
| { |
| "text": "The model relies on a dataset of linguistic complexity metrics of different dimensions: syntactic, lexical, semantic, accuracy and pragmatic. These metrics form numeric feature vectors of values and characterise the learner texts. The vectors are matched with the CEFR levels assigned to the texts. We use several tools to compute the met-rics. Syntactic complexity measures are computed with the L2 Syntactic Complexity Analyzer (L2SCA). These tools rely on the Tregex module of Stanford CoreNLP (Manning et al., 2014) for phrase constituent retrieval. The Tool for the Automatic analysis of Syntactic Sophistication and Complexity (TAASC) is also used to compute ratios and scores of syntactic complexity such as prepositions per nominal, adjectival modifiers per object of the preposition, and also probability that two items occur together. For lexico-semantic features, the pipeline implements the Lexical Complexity Analyzer (LCA) relying on Treetagger (Schmid, 1994) to compute lexical diversity metrics. The Automatic Assessment of Lexical Proficiency (TAALES) tool computes includes 130 lexical indices with classic lexical complexity metrics and psycholinguistic properties of words. These properties are based on judgments of concreteness, familiarity, imageability, or supposed age of first exposure. The TAALES indices include frequencies, ratios of lexical words and n-grams as well as comparative metrics sourced from reference corpora. The textstat Python library 3 was used to compute readability metrics that indicate the level of difficulty of texts. Accuracy features are computed with the pyenchant Python library (Kelly, 2016) 4 for misspelt words. Regarding pragmatic features, the pipeline includes the Tool for the Automatic Analysis of Cohesion (TAACO) which computes metrics based on referential and discourse characteristics such as pronouns, lexical overlaps and connectives. In total, 768 different features were extracted and merged into one dataset to input into the classification models.", |
| "cite_spans": [ |
| { |
| "start": 959, |
| "end": 973, |
| "text": "(Schmid, 1994)", |
| "ref_id": "BIBREF27" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Feature Extraction", |
| "sec_num": "3.2.2." |
| }, |
| { |
| "text": "The aim was to construct a classification model of learner CEFR levels (A1, A2, B1, B2, C1, C2, ie from beginners to advanced speakers). Among several model types tested, the optimal classification performance in the testing data set was found using multinomial logistic regression . The classifier using all features reached 82% accuracy (0.80 mean F1-score) on a six-point scale classification. Given that the levels are ordinals, one of the reviewers suggested misclassifications B1 for B2 should incur different costs than, say, misclassifying B1 for C2. We did not resort to cost matrix, a system 5 used in the Cap2018 data challenge (Ballier et al., 2020 ) with the same classification task. The cost matrix used in this data challenge penalised a default assignment to A1 (because of a skewed dataset) and rewarded B1 versus B2 distinction (a sensitive boundary for some educational institutions like engineer schools). With hindsight, the robustness of such a cost matrix should be tested among several candidates to assess the consequences of penalisation weights.", |
| "cite_spans": [ |
| { |
| "start": 639, |
| "end": 660, |
| "text": "(Ballier et al., 2020", |
| "ref_id": "BIBREF2" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "CEFR Level Classifier", |
| "sec_num": "3.2.3." |
| }, |
| { |
| "text": "To be able to demonstrate and test our model in order to have a proof-of-concept, we decided to create a web-based infrastructure that i) handles the Input/Output from and to the user and ii) schedules the tasks to and from the classifier. The infrastructure's primary requirement was to process new texts for on-the-fly metric computation followed by classification. As a result, a web interface outputs the CEFR level predicted for the new texts. This infrastructure is composed of Docker modules (Merkel, 2014) , which are interconnected to handle data ingestion, processing and model classification. The infrastructure is built to allow model switching. The system can be modified in three points. The feature extraction pipeline can be modified so as to compute different metrics. The classification model can be changed to match the extracted features. The User Interface (UI, cf. Figure 2 ) can be modified according to the task at hand. Front-end. The UI demo has two main components (Figure 3) , i) the web app component that interacts with the user on the web browser by receiving and responding text, and ii) the background component that manages the text processing system (in our case, extracting features and using them to classify the text's English level). These two components communicate with each other using a message broker -Redis, the third component of the UI demo. We created these different components in order to decouple the web app from the processing part, so the web app interface would not be blocked by a given request. This architecture provides the benefit of availability for new input, while processing feature extraction and classification to get a response for the current user. Back-end. The environment of the web app and the processor are separated into two different docker containers, both of them having an instance of a Celery app. Celery works in an asynchronous way, from one side the Celery scheduler is responsible to create tasks, and from the other side the Celery worker is responsible to process the Celery scheduler created before. The web app is an instance of a Celery scheduler while the processor is an instance of a celery worker. When the web app receives a request with a text to process, it uses Celery to create a task and put it into the Celery queue by sending the text to be processed together with the task ID through Redis. The processor docker, a Celery worker, communicates with Redis polling for tasks to process. When it receives a task from Redis, the worker processes it and uses Redis again to communicate the processed prediction level of the text. In the end, the web app can consult Redis, via the ID task, to get the result of the processing.", |
| "cite_spans": [ |
| { |
| "start": 499, |
| "end": 513, |
| "text": "(Merkel, 2014)", |
| "ref_id": "BIBREF23" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 887, |
| "end": 895, |
| "text": "Figure 2", |
| "ref_id": "FIGREF1" |
| }, |
| { |
| "start": 992, |
| "end": 1002, |
| "text": "(Figure 3)", |
| "ref_id": "FIGREF2" |
| } |
| ], |
| "eq_spans": [], |
| "section": "The Infrastructure", |
| "sec_num": "3.3." |
| }, |
| { |
| "text": "Prior to plugging in a new model, it is necessary to conduct a supervised learning method on some data in order to fit a model. Once the model is tuned, a .sav file can be retrieved and placed into the architecture as shown in Figure 3 . This .sav format stores models as a binary file. It is also essential to match the input features of the new model with the output features of the processing pipeline. The features created by the tools need to be filtered in order to pass only the data with the required features into the model.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 227, |
| "end": 235, |
| "text": "Figure 3", |
| "ref_id": "FIGREF2" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Plugging into the Model", |
| "sec_num": "4.2." |
| }, |
| { |
| "text": "We deployed our docker on an a web platform hosted by the French Linguistic Platform infrastructure HUMA-NUM 6 . The deployment of our UI was tested on HUMA-NUMm for a deployment on a Virtual machine using our docker virtualization. The system is fast and the HUMA-NUM team is very reactive. We needed to change the port 80 as the web interface required a different access port. 7 The feature extraction tool was also tested on this virtual machine. You need at least 32 CPUs 32 Gigabytes. We processed a 2019 version of the entire EFCAMDAT and it processes 100 files every five minutes. This confirms our desire to optimise the feature extraction. The feature extractor is optimised in the sense that each repertory is turned into an individual .csv files for the output data, so that the calculated features are regularly saved in case of crash (in the case of non-utf8 files, or unexpected sequences of commas, as experience seem to show).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Tests and Redeployment", |
| "sec_num": "4.3." |
| }, |
| { |
| "text": "The system relies on the open-source Docker technology 8 . It is made up of three Docker containers, each designed for a specific purpose as detailed in 4.1. In order to run the system, docker must be installed. Docker-compose 9 is also required to run the three container Docker applications. Note that the infrastructure relies on a number of data handling technologies and a framework that do not require installation as they are located within Docker containers. This includes Redis 10 , a database management system, as well as the Flask web application framework 11 .", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Infrastructure Requirements", |
| "sec_num": "4.4." |
| }, |
| { |
| "text": "This section describes the current state of our platformrelated project and research outputs.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Available Resources", |
| "sec_num": "5." |
| }, |
| { |
| "text": "6 https://www.huma-num.fr/about-us 7 http://linguisticdataprocessing.huma-num.fr/ 8 see https://www.docker.com/ 9 see https://docs.docker.com/compose/install/ 10 see https://redislabs.com/ 11 see https://www.palletsprojects.com/p/flask/", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Available Resources", |
| "sec_num": "5." |
| }, |
| { |
| "text": "The project has yielded a number of resources and tools presented in Table 1 and all referenced in the project's web page 12 . The first tool is dedicated to end users, i.e. learners of English. A User Interface (UI) demo program provides a prototype for real-time proficiency assessment of new texts. Texts in English of more than 70 tokens are assigned probabilities of belonging to A1, A2, B1, B2, C1, C2 of CEFR levels. A documentation is available and it can be installed from NUIG Insight's Gitlab repository on request. Once installed, the demo acts as a web server, handles the text processing pipeline and interfaces with the results sent from the classification model as explained in Figure 3 . The second tool aims at the research community and focused on criterial features and AES systems. Researchers in this domain need to test different types of linguistic features to assess whether they are potentially criterial in determining CEFR levels (Hawkins and Filipovi\u0107, 2012) . We provide a tool that allows researchers to process batches of learner texts and output data sets ready for analysis. This tool includes the same processing pipeline as the UI demo and is only operated with the shell or command line. Users only have to place texts in a directory used as program input. It is available from NUIG Insight's Gitlab repository on request. In the course of developing these tools, a number of data resources were created. First, the data set used for modeling in the UI demo system is available and can be exploited for other types of analyses grounded in linguistic complexity. The data set is composed of a list of metric values matched with CEFR levels. Secondly, the classification model implemented in the UI demo program is also available for download. Following the latest version of Nakala, the dataset warehouse hosted and managed by HUMA-NUM, we will upload our datasets with the corresponding permanent handles and DOIs to be attributed by HUMA-NUM 13 .", |
| "cite_spans": [ |
| { |
| "start": 958, |
| "end": 987, |
| "text": "(Hawkins and Filipovi\u0107, 2012)", |
| "ref_id": "BIBREF11" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 69, |
| "end": 76, |
| "text": "Table 1", |
| "ref_id": null |
| }, |
| { |
| "start": 694, |
| "end": 702, |
| "text": "Figure 3", |
| "ref_id": "FIGREF2" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Research Output", |
| "sec_num": "5.1." |
| }, |
| { |
| "text": "For the web architecture, we have adopted Celery to avoid bottlenecks and nevertheless allow several users to query the system at the same time. As one of the reviewers pointed out, this decision has some consequences as opposed to an event-based approach. We intended to create the UI demo as a proof-of-concept for the project with time limitation. Celery seemed to be the best choice for us considering i) we were using a Python web application, ii) our team did not include specialists of event-based technologies, so the learning curve to implement using Celery was shorter than using an event-based approach, iii) our provisional goal for the project's scope was to make the UI demo available for more than one user, although we did not have the budget to support a massive number of users connecting to the system at the same time. We feel we can recommend this solution for a small-scale implementation but definitely, because of scalability limitation , future developments of our project could include an Table 1 : The granularity of availability of the resources event-based distributed approach for services communication/scheduler. This would make our system more robust and able to deal with more users at the same time.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 1015, |
| "end": 1022, |
| "text": "Table 1", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Technical Aspects", |
| "sec_num": "5.2." |
| }, |
| { |
| "text": "For the feature extractor and individual distribution of the UI for researchers, we tested the Docker setup on MS Windows 10, iOS (Mojave) and Linux (Ubuntu 18.06). Because TAALES loads greedy frequency inventories and parsing the data is also memory-demanding, 16 Go of RAM are necessary for the execution (3 CPUs and 7 Gigas recommended for preferences in the docker). We wrote a user's guide in English for the installation of each tool. The whole annotation pipeline is in python, building from the former blocks from the tools. We have opted for a 3.6 version of python.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Technical Aspects", |
| "sec_num": "5.2." |
| }, |
| { |
| "text": "The project was funded by the two partner countries and partners abode by the standard legal framework in use for this binational scheme 14 . Research papers were signed by the members of the projects of the two teams. Though we acknowledged respective percentages of ownership in relation to the input of the programming team, we agreed on Creative Commons Non Commercial Share Alike Licence for all our research outputs. In terms of social aspects and community, we targeted two types of audiences. Our on-line prototype aims at learners of English and teachers in classroom environments. It is maintained until end of 2020 on the Irish partner infrastructure. To ensure sustainability afterwards, we have adopted the HUMA-NUM infrastructure 15 to host our project. As a prototype, it may experience scalability issues. The feature extraction tool was designed to be of interest to a potential consortium of researchers sharing a similar aim. Researchers working on linguistic complexity and conducting feature extraction tasks (for instance, for text classification purposes) may find the tool a useful assistant as it avoids coding. Specialists in Second Language Acquisition or Learner Corpus Research may benefit from the customisable micro-systems implemented in LS2CA_MS. For this tool parsing English data based on LS2CA (Lu, 2014) , simple Tregex syntax (Levy and Andrew, 2006) can be used to create new features for the analysis of micro-systems (Gaillat et al., submitted) .", |
| "cite_spans": [ |
| { |
| "start": 1330, |
| "end": 1340, |
| "text": "(Lu, 2014)", |
| "ref_id": "BIBREF19" |
| }, |
| { |
| "start": 1364, |
| "end": 1387, |
| "text": "(Levy and Andrew, 2006)", |
| "ref_id": "BIBREF16" |
| }, |
| { |
| "start": 1457, |
| "end": 1484, |
| "text": "(Gaillat et al., submitted)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Governance and Uptake", |
| "sec_num": "5.3." |
| }, |
| { |
| "text": "This section discusses developments in the making.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Perspectives and Improvements", |
| "sec_num": "6." |
| }, |
| { |
| "text": "To improve our model, we would like to store users' input to exploit it in further analyses. For the time being, the essays /texts submitted for assessment are not stored. We would like to make sure further crowdsourcing developments are compliant with GDPR and consider publishing general conditions of use warning the user that the texts used as queries will be stored to fine-tune the models. Anonymising the data for crowdsourcing will be carried out in accordance with (Klavan et al., 2012) . More generally, and more theoretically, the question is the compatibility of the models with the features when the system needs to evolve. If we collect more data (and possibly add other features), we will probably need to fine-tune the statistical models. We do not know whether this should lead to a standardization of (linguistic) features and interfaces as one of the reviewers suggested, but we suggest that an adaptation of the models can be reimplemented in the architecture by modifying the .sav file. We are not aware of any interface or standards to cater for this need.", |
| "cite_spans": [ |
| { |
| "start": 474, |
| "end": 495, |
| "text": "(Klavan et al., 2012)", |
| "ref_id": "BIBREF14" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Crowdsourcing Derived Applications and GDPR", |
| "sec_num": "6.1." |
| }, |
| { |
| "text": "This section sums up current (and future) developments in the making of our project, with a view to offer more interoperability with existing platforms or similar projects.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Engineering and Interoperability", |
| "sec_num": "6.2." |
| }, |
| { |
| "text": "We spent much time integrating the various tools, and necessarily more time on feature collection and extraction than on dimensionality reduction. We are currently processing a complete dump of the EFCAMDAT dataset to address these issues and (Gaillat et al., submitted) reports preliminary findings on the French and Spanish datasets. The project was to build a proof-of-concept for the automation of proficiency level assessment. Further developments are required to improve the system. In its current version the model relies on a broad set of features (over 750) which makes it prone to overfitting. It is thus necessary to find a simpler model based on less features. This, in turn, will impact the data processing pipeline as only those tools related to selected features will be kept. We have obtained preliminary results with a new model based on the elastic net regression method (Zou and Hastie, 2005 ) trained on the EFCAMDAT training set and tested on another data set extracted from a totally different corpus, i.e. the CEFR-ASAG corpus (Tack et al., 2017) . This method comes with the benefit of including feature dimensionality reduction. Using just 44 features classification showed 75.0% accuracy (CI [74.3, 75.8] , p<0.001) and 59.2% (CI [53.4, 64.8] , p <0.001) on the EFCAMDAT and CEFR-ASAG test sets respectively For more generic linguistic feature processing and analysis, it would be relevant to design a tool and feature selection assistant for the batch feature extraction tool. This would enable researchers to select features and tools as needed very much like CTAP (Chen and Meurers, 2016 ). The latter tool supports a modular approach to feature extraction allowing for reusability. Users can compose their dataset variables prior to running an NLP pipeline that processes texts to produced the desired variable values. As the authors mention, additional functionality including machine learning modules is required to combine the collected evidence with specific outcome variables such as CEFR levels. An interface between CTAP and our Web demo UI could be developed in order to allow data exchange between the CTAP output and our CEFR classifier. Another advantage of a feature selection assistant would be to support multilingual processing. The current pipeline makes use of a number of tools that are language agnostic for the computation of some of the metrics. By allowing researchers to select language-agnostic metrics, it would be possible to build data sets used for modeling CEFR classification in other languages than English. Conversely, some metrics are language-dependent, as in the case of many lexical sophistication metrics which are based on lexical frequency inventories extracted from reference corpora of English. One line of research is to adapt some metrics to French as a Foreign Language, especially readability metrics (Fran\u00e7ois, 2015) or to Dutch (Tack et al., 2018) . Developing the interoperability of our feature extracting tool in the sense of multilingualism is also made possible by adapting our microsystem features to other languages, probably French as a Foreign Language for the next phase.", |
| "cite_spans": [ |
| { |
| "start": 243, |
| "end": 270, |
| "text": "(Gaillat et al., submitted)", |
| "ref_id": null |
| }, |
| { |
| "start": 889, |
| "end": 910, |
| "text": "(Zou and Hastie, 2005", |
| "ref_id": "BIBREF33" |
| }, |
| { |
| "start": 1050, |
| "end": 1069, |
| "text": "(Tack et al., 2017)", |
| "ref_id": "BIBREF28" |
| }, |
| { |
| "start": 1214, |
| "end": 1224, |
| "text": "(CI [74.3,", |
| "ref_id": null |
| }, |
| { |
| "start": 1225, |
| "end": 1230, |
| "text": "75.8]", |
| "ref_id": null |
| }, |
| { |
| "start": 1252, |
| "end": 1262, |
| "text": "(CI [53.4,", |
| "ref_id": null |
| }, |
| { |
| "start": 1263, |
| "end": 1268, |
| "text": "64.8]", |
| "ref_id": null |
| }, |
| { |
| "start": 1593, |
| "end": 1616, |
| "text": "(Chen and Meurers, 2016", |
| "ref_id": "BIBREF3" |
| }, |
| { |
| "start": 2876, |
| "end": 2892, |
| "text": "(Fran\u00e7ois, 2015)", |
| "ref_id": "BIBREF8" |
| }, |
| { |
| "start": 2905, |
| "end": 2924, |
| "text": "(Tack et al., 2018)", |
| "ref_id": "BIBREF29" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Feature Engineering and Dimensionality Reduction", |
| "sec_num": "6.2.1." |
| }, |
| { |
| "text": "One of the reviewers enquired about the implications of spoken data for our system. The short answer is that some written-based metrics may not be adequate for spoken data but speech data could be pre-processed to be fully tested by our system. The team discussed implementing a speech-totext system, with the proviso that a single acoustic model should be chosen (eg preferably French speakers with available data). (Mariko and Kondo, 2019) reported successful use of IBM Watson Speech-to-text technology to transcribe learner speech for Japanese learners of English and give examples of the output. They reported Word Error Rates on 50 randomised speech samples and concluded that the automatic procedure was worth it. An important caveat for the calculation of the metrics is the absence of punctuation marks and the potentially useful insertion of \"%HESITA-TION\" for filled pauses (no threshold reported for the duration of filled pauses, though). The Watson system runs on Python 3.4. but is in the cloud and is not free. They do not seem to indicate whether the quality of the voice recognition improves over time for their longitudinal data. Following a uniformly python pipeline, we would try to use SpeechRecognition (Zhang, 2017) as a pre-processing stage of spoken data. We would have the added benefit of analysing spoken production, but this would probably imply a semi-automatic solution as the speech-to-text outputs would probably need to be manually edited. We also have initial reservations as to the applicability of writtendesigned metrics to spoken data (Ballier and Gaillat, 2016) , in particular the transferability of the T-unit (a crucial concept for some complexity metrics) for spoken utterances, but we could experiment a speech-to-text module to preprocess learner recordings in order to test our model on spoken data, at least for fluency.", |
| "cite_spans": [ |
| { |
| "start": 417, |
| "end": 441, |
| "text": "(Mariko and Kondo, 2019)", |
| "ref_id": "BIBREF21" |
| }, |
| { |
| "start": 1226, |
| "end": 1239, |
| "text": "(Zhang, 2017)", |
| "ref_id": "BIBREF32" |
| }, |
| { |
| "start": 1575, |
| "end": 1602, |
| "text": "(Ballier and Gaillat, 2016)", |
| "ref_id": "BIBREF0" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Pre-processing", |
| "sec_num": "6.2.2." |
| }, |
| { |
| "text": "Collecting metrics to assess learner performance could be used for didactic purposes. It sounds plausible to select some features to guide learners in ICALL systems for selfassessment of their performance. A member of our team has begun applying some of the features to produce immediate feedback for learners, elaborating on a prototype (Ballier et al., 2019) . This data visualisation application of our feature extractor takes the form of a dashboard where learner scores are compared to means of students of the same cohort and to native scores on similar essays.", |
| "cite_spans": [ |
| { |
| "start": 338, |
| "end": 360, |
| "text": "(Ballier et al., 2019)", |
| "ref_id": "BIBREF1" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Post-processing", |
| "sec_num": "6.2.3." |
| }, |
| { |
| "text": "We have reached the final stage of our one-year project, and have tried to work in line with the FAIR paradigm (Mons, 2018) for our resources to be:", |
| "cite_spans": [ |
| { |
| "start": 111, |
| "end": 123, |
| "text": "(Mons, 2018)", |
| "ref_id": "BIBREF24" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Sustainability and the FAIR Paradigm", |
| "sec_num": "6.3." |
| }, |
| { |
| "text": "\u2022 Findable : the tools and resources are available from the project's web page, possibly from the LREC resource map and our datasets are to be linked with permanent handles and DOIs thanks to the Nakala HUMA-NUM services.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Sustainability and the FAIR Paradigm", |
| "sec_num": "6.3." |
| }, |
| { |
| "text": "\u2022 Accessible: Some copyright restrictions apply to the corpus we used and to some of the tools. Access is either universal or restricted. There are copyright restrictions to the TAALES tool and to the initial EF-CAMDAT corpus data.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Sustainability and the FAIR Paradigm", |
| "sec_num": "6.3." |
| }, |
| { |
| "text": "\u2022 Interoperable: our project is multi-platform and our UI infrastructure could be compared in terms of interoperability with similar existing language platforms. A project is in the making with the curators of the REALEC corpus (Vinogradova, 2016) and of the REALEC-inspector web interface 16 to analyse the relevance of the automatically extracted features for Russian learners of English.", |
| "cite_spans": [ |
| { |
| "start": 228, |
| "end": 247, |
| "text": "(Vinogradova, 2016)", |
| "ref_id": "BIBREF30" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Sustainability and the FAIR Paradigm", |
| "sec_num": "6.3." |
| }, |
| { |
| "text": "\u2022 Re-usable : The UI interface can be customised for on-the-fly processing of texts. For example, to improve comparability with other language platforms, a three-point scale of learner levels could be reimplemented on our system with a different statistical model (.sav file). Maybe our datasets will be reused as well. Data Management Plans were not required for this level of funding, but we tried to produce comparable information for our datasets. ", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Sustainability and the FAIR Paradigm", |
| "sec_num": "6.3." |
| }, |
| { |
| "text": "This paper has presented two tools and a set of of resources implemented in a Language Technology project. These tools rely on a modular implementation of a Docker architecture. As a result, this architecture is reusable in other 16 https://linghub.ru/inspector/ LT contexts such as L1 identification or Text Classification. We provide a web-based user demo tool and a linguistic complexity metric extraction tool. These tools can be modified to accommodate other projects relying on text features and classification. Our idea was to showcase the full workflow from the linguistic modelling to the web-based user interface to help linguists to disseminate their research projects. In this sense, this paper was intended as a 'how-to' for corpus linguistics to possibly publish web-based interfaces exploiting their data modelling.", |
| "cite_spans": [ |
| { |
| "start": 230, |
| "end": 232, |
| "text": "16", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion", |
| "sec_num": "7." |
| }, |
| { |
| "text": "Our project is a case study for linguistics as a cumulative data science. We showcase the data life cycle and some of its uses. We were able to reuse part of the EFCAMDAT data collection, we were able to concatenate several existing tools in a single workflow, we added our own microsystem features based on our analysis of learners' issues (the LS2CA_MS component of our pipeline) and our modelling (the .sav file), we shared the demoIT infrastructure we designed to exploit it. More data production can be expected with the UI and the feature extraction tool. Customisation is expected for our demoIT UI infrastructure and micro-system features. Our collaboration between corpus linguists, computational linguists, statisticians and computational scientists pertake of the current shift towards linguistic data science.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion", |
| "sec_num": "7." |
| }, |
| { |
| "text": "This work was undertaken with a PHC Hubert Currien Ulysse 2019 grant (project ref. 43121RJ) with the financial support of:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Acknowledgements", |
| "sec_num": "8." |
| }, |
| { |
| "text": "\u2022 Science Foundation Ireland (SFI) under Grant Number SFI/12/RC/2289_P2, co-funded by the European Regional Development Fund.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Acknowledgements", |
| "sec_num": "8." |
| }, |
| { |
| "text": "\u2022 The French Ministry of Higher Education, Research and Innovation (Minist\u00e8re de l'Enseignement sup\u00e9rieur, de la Recherche et de l'Innovation, MESRI)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Acknowledgements", |
| "sec_num": "8." |
| }, |
| { |
| "text": "\u2022 The French Ministry for Europe and Foreign Affairs (Minist\u00e8res de l'Europe et des affaires \u00e9trang\u00e8res, MEAE)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Acknowledgements", |
| "sec_num": "8." |
| }, |
| { |
| "text": "We wish to thank the three anonymous reviewers for their comments. We are grateful to Kristopher Kyle and Scott Crossley for providing us with the source code of the TAALES tool and to Xiaofei Lu for showing us his L2SCA as visiting scholar and allowing us to implement the microsystem features. We thank Theodora Alexopoulou and Carlos Balhana for their willingness to help us in our project by hosting our resulting dataset on the EFCAMDAT website, in accordance with the corpus maintenance regulations. We thank Detmar Meurers for his enthusiasm and for letting us access an early stage version of the CTAP annotation platform. We also thank Felipe Arruda who provided us with substantial software development advice. We also thank the audience of the Paris workshop (30 Oct.2019) for their feedback on the presentation of the deliverables of the project.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Acknowledgements", |
| "sec_num": "8." |
| }, |
| { |
| "text": "The University of Cambridge and English First took no part in the data manipulation. The dataset including the EFCAMDAT texts will be hosted by Cambridge, in accordance with the corpus regulations. Access is free for academic non-commercial uses, provided potential users request permission using an academic email address.2 Evaluation results are discussed in(Gaillat et al., submitted)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "see https://pypi.org/project/textstat/ 4 see https://pypi.org/project/pyenchant/ 5 http://cap2018.litislab.fr/competition_en.pdf", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "see www.clillac-arp.univ-paris-diderot.fr/projets/ulysse2019 13 https://humanum.hypotheses.org/5989", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "A Guideline (in French) of recommended good practices for holders of this grant is available here: https://www.campusfrance.org/sites/default/files/medias/documents/ 2017-11/guide_bonnes_pratiques_ulysses.pdf 15 https://www.huma-num.fr/about-us", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| } |
| ], |
| "back_matter": [], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "Classification d'apprenants francophones de l'anglais sur la base des m\u00e9triques de complexit\u00e9 lexicale et syntaxique", |
| "authors": [ |
| { |
| "first": "N", |
| "middle": [], |
| "last": "Ballier", |
| "suffix": "" |
| }, |
| { |
| "first": "T", |
| "middle": [], |
| "last": "Gaillat", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "JEP-TALN-RECITAL 2016", |
| "volume": "9", |
| "issue": "", |
| "pages": "1--14", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ballier, N. and Gaillat, T. (2016). Classification d'apprenants francophones de l'anglais sur la base des m\u00e9triques de complexit\u00e9 lexicale et syntaxique. In JEP- TALN-RECITAL 2016, volume 9, pages 1-14.", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "Prototype de feedback visuel des productions \u00e9crites d'apprenants francophones de l'anglais sous moodle", |
| "authors": [ |
| { |
| "first": "N", |
| "middle": [], |
| "last": "Ballier", |
| "suffix": "" |
| }, |
| { |
| "first": "T", |
| "middle": [], |
| "last": "Gaillat", |
| "suffix": "" |
| }, |
| { |
| "first": "E", |
| "middle": [], |
| "last": "Pacquetet", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "Actes de la 9\u00e8me Conf\u00e9rence sur les Environnements Informatiques pour l'Apprentissage Humain (EIAH2019)", |
| "volume": "", |
| "issue": "", |
| "pages": "395--398", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ballier, N., Gaillat, T., and Pacquetet, E. (2019). Prototype de feedback visuel des productions \u00e9crites d'apprenants francophones de l'anglais sous moodle. In Julien Broisin, et al., editors, Actes de la 9\u00e8me Conf\u00e9rence sur les Environnements Informatiques pour l'Apprentissage Humain (EIAH2019), pages 395-398.", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "Machine learning for learner English", |
| "authors": [ |
| { |
| "first": "N", |
| "middle": [], |
| "last": "Ballier", |
| "suffix": "" |
| }, |
| { |
| "first": "S", |
| "middle": [], |
| "last": "Canu", |
| "suffix": "" |
| }, |
| { |
| "first": "C", |
| "middle": [], |
| "last": "Petitjean", |
| "suffix": "" |
| }, |
| { |
| "first": "G", |
| "middle": [], |
| "last": "Gasso", |
| "suffix": "" |
| }, |
| { |
| "first": "C", |
| "middle": [], |
| "last": "Balhana", |
| "suffix": "" |
| }, |
| { |
| "first": "T", |
| "middle": [], |
| "last": "Alexopoulou", |
| "suffix": "" |
| }, |
| { |
| "first": "T", |
| "middle": [], |
| "last": "Gaillat", |
| "suffix": "" |
| } |
| ], |
| "year": 2020, |
| "venue": "International Journal of Learner Corpus Research", |
| "volume": "6", |
| "issue": "1", |
| "pages": "72--103", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ballier, N., Canu, S., Petitjean, C., Gasso, G., Balhana, C., Alexopoulou, T., and Gaillat, T. (2020). Machine learn- ing for learner English. International Journal of Learner Corpus Research, 6(1):72-103.", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "CTAP: A web-based tool supporting automatic complexity analysis", |
| "authors": [ |
| { |
| "first": "X", |
| "middle": [], |
| "last": "Chen", |
| "suffix": "" |
| }, |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Meurers", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "Proceedings of the Workshop on Computational Linguistics for Linguistic Complexity (CL4LC)", |
| "volume": "", |
| "issue": "", |
| "pages": "113--119", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Chen, X. and Meurers, D. (2016). CTAP: A web-based tool supporting automatic complexity analysis. In Pro- ceedings of the Workshop on Computational Linguistics for Linguistic Complexity (CL4LC), pages 113-119.", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "Linguistic Microfeatures to Predict L2 Writing Proficiency: A Case Study in Automated Writing Evaluation", |
| "authors": [ |
| { |
| "first": "S", |
| "middle": [ |
| "A" |
| ], |
| "last": "Crossley", |
| "suffix": "" |
| }, |
| { |
| "first": "K", |
| "middle": [], |
| "last": "Kyle", |
| "suffix": "" |
| }, |
| { |
| "first": "L", |
| "middle": [ |
| "K" |
| ], |
| "last": "Allen", |
| "suffix": "" |
| }, |
| { |
| "first": "L", |
| "middle": [], |
| "last": "Guo", |
| "suffix": "" |
| }, |
| { |
| "first": "D", |
| "middle": [ |
| "S" |
| ], |
| "last": "Mc-Namara", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "The Journal of Writing Assessment", |
| "volume": "7", |
| "issue": "1", |
| "pages": "1--34", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Crossley, S. A., Kyle, K., Allen, L. K., Guo, L., and Mc- Namara, D. S. (2014). Linguistic Microfeatures to Pre- dict L2 Writing Proficiency: A Case Study in Automated Writing Evaluation. The Journal of Writing Assessment, 7(1):1-34.", |
| "links": null |
| }, |
| "BIBREF5": { |
| "ref_id": "b5", |
| "title": "The Tool for the Automatic Analysis of Cohesion 2.0: Integrating semantic similarity and text overlap", |
| "authors": [ |
| { |
| "first": "S", |
| "middle": [ |
| "A" |
| ], |
| "last": "Crossley", |
| "suffix": "" |
| }, |
| { |
| "first": "K", |
| "middle": [], |
| "last": "Kyle", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Dascalu", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "Behavior Research Methods", |
| "volume": "51", |
| "issue": "1", |
| "pages": "14--27", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Crossley, S. A., Kyle, K., and Dascalu, M. (2019). The Tool for the Automatic Analysis of Cohesion 2.0: Inte- grating semantic similarity and text overlap. Behavior Research Methods, 51(1):14-27.", |
| "links": null |
| }, |
| "BIBREF6": { |
| "ref_id": "b6", |
| "title": "Common European Framework of Reference for Languages: Learning, teaching, assessment", |
| "authors": [ |
| { |
| "first": "European", |
| "middle": [], |
| "last": "Council", |
| "suffix": "" |
| } |
| ], |
| "year": 2001, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "European Council. (2001). Common European Framework of Reference for Languages: Learning, teaching, assess- ment. Cambridge University Press, Cambridge.", |
| "links": null |
| }, |
| "BIBREF7": { |
| "ref_id": "b7", |
| "title": "Guidelines on the implementation of Open Access to scientific publications and research data in projects supported by the European Research Council under Horizon", |
| "authors": [ |
| { |
| "first": "European Research", |
| "middle": [], |
| "last": "Council", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "European Research Council. (2017). Guidelines on the im- plementation of Open Access to scientific publications and research data in projects supported by the European Research Council under Horizon 2020.", |
| "links": null |
| }, |
| "BIBREF8": { |
| "ref_id": "b8", |
| "title": "When readability meets computational linguistics: a new paradigm in readability. Revue fran\u00e7aise de linguistique appliqu\u00e9e", |
| "authors": [ |
| { |
| "first": "T", |
| "middle": [], |
| "last": "Fran\u00e7ois", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "", |
| "volume": "20", |
| "issue": "", |
| "pages": "79--97", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Fran\u00e7ois, T. (2015). When readability meets computa- tional linguistics: a new paradigm in readability. Revue fran\u00e7aise de linguistique appliqu\u00e9e, 20(2):79-97.", |
| "links": null |
| }, |
| "BIBREF9": { |
| "ref_id": "b9", |
| "title": "Predicting CEFR levels in learners of English: the use of microsystem criterial features in a machine learning approach", |
| "authors": [ |
| { |
| "first": "T", |
| "middle": [], |
| "last": "Gaillat", |
| "suffix": "" |
| }, |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Simpkin", |
| "suffix": "" |
| }, |
| { |
| "first": "N", |
| "middle": [], |
| "last": "Ballier", |
| "suffix": "" |
| }, |
| { |
| "first": "B", |
| "middle": [], |
| "last": "Stearns", |
| "suffix": "" |
| }, |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Sousa", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Bouy\u00e9", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Zarrouk", |
| "suffix": "" |
| } |
| ], |
| "year": null, |
| "venue": "Journal With Anonymous Submission", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Gaillat, T., Simpkin, A., Ballier, N., Stearns, B., Sousa, A., Bouy\u00e9, M., and Zarrouk, M. (submitted). Predicting CEFR levels in learners of English: the use of microsys- tem criterial features in a machine learning approach. Journal With Anonymous Submission.", |
| "links": null |
| }, |
| "BIBREF10": { |
| "ref_id": "b10", |
| "title": "Automatic Linguistic Annotation of Large Scale L2 Databases: The EF-Cambridge Open Language Database (EFCamDat)", |
| "authors": [ |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Geertzen", |
| "suffix": "" |
| }, |
| { |
| "first": "T", |
| "middle": [], |
| "last": "Alexopoulou", |
| "suffix": "" |
| }, |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Korhonen", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "Proceeedings of the 31st Second Language Research Forum", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Geertzen, J., Alexopoulou, T., and Korhonen, A. (2013). Automatic Linguistic Annotation of Large Scale L2 Databases: The EF-Cambridge Open Language Database (EFCamDat). In R. T. Miller, et al., editors, Proceeedings of the 31st Second Language Research Fo- rum, Carnegie Mellon. Cascadilla Press.", |
| "links": null |
| }, |
| "BIBREF11": { |
| "ref_id": "b11", |
| "title": "Criterial Features in L2 English: Specifying the Reference Levels of the Common European Framework", |
| "authors": [ |
| { |
| "first": "J", |
| "middle": [ |
| "A" |
| ], |
| "last": "Hawkins", |
| "suffix": "" |
| }, |
| { |
| "first": "L", |
| "middle": [], |
| "last": "Filipovi\u0107", |
| "suffix": "" |
| } |
| ], |
| "year": 2012, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Hawkins, J. A. and Filipovi\u0107, L. (2012). Criterial Features in L2 English: Specifying the Reference Levels of the Common European Framework. Cambridge University Press, United Kingdom.", |
| "links": null |
| }, |
| "BIBREF13": { |
| "ref_id": "b13", |
| "title": "Pyenchant a spellchecking library for python", |
| "authors": [ |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Kelly", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Kelly, R. (2016). Pyenchant a spellchecking library for python. available: https://pythonhosted. org/pyenchant.", |
| "links": null |
| }, |
| "BIBREF14": { |
| "ref_id": "b14", |
| "title": "The legal aspects of using data from linguistic experiments for creating language resources", |
| "authors": [ |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Klavan", |
| "suffix": "" |
| }, |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Tavast", |
| "suffix": "" |
| }, |
| { |
| "first": "Kelli", |
| "middle": [], |
| "last": "", |
| "suffix": "" |
| }, |
| { |
| "first": "A", |
| "middle": [], |
| "last": "", |
| "suffix": "" |
| } |
| ], |
| "year": 2012, |
| "venue": "Human Language Technologies The Baltic Perspective: Proceedings of the Fifth International Conference Baltic HLT 2012", |
| "volume": "", |
| "issue": "", |
| "pages": "71--78", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Klavan, J., Tavast, A., and Kelli, A. (2012). The legal as- pects of using data from linguistic experiments for cre- ating language resources. In Arvi Tavast, et al., editors, Human Language Technologies The Baltic Perspective: Proceedings of the Fifth International Conference Baltic HLT 2012, pages 71-78. IOS Press.", |
| "links": null |
| }, |
| "BIBREF15": { |
| "ref_id": "b15", |
| "title": "The tool for the automatic analysis of lexical sophistication (TAALES): version 2.0", |
| "authors": [ |
| { |
| "first": "K", |
| "middle": [], |
| "last": "Kyle", |
| "suffix": "" |
| }, |
| { |
| "first": "S", |
| "middle": [], |
| "last": "Crossley", |
| "suffix": "" |
| }, |
| { |
| "first": "C", |
| "middle": [], |
| "last": "Berger", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "Behavior Research Methods", |
| "volume": "50", |
| "issue": "3", |
| "pages": "1030--1046", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Kyle, K., Crossley, S., and Berger, C. (2018). The tool for the automatic analysis of lexical sophistication (TAALES): version 2.0. Behavior Research Methods, 50(3):1030-1046.", |
| "links": null |
| }, |
| "BIBREF16": { |
| "ref_id": "b16", |
| "title": "Tregex and Tsurgeon: tools for querying and manipulating tree data structures", |
| "authors": [ |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Levy", |
| "suffix": "" |
| }, |
| { |
| "first": "G", |
| "middle": [], |
| "last": "Andrew", |
| "suffix": "" |
| } |
| ], |
| "year": 2006, |
| "venue": "Proceedings of LREC2006", |
| "volume": "", |
| "issue": "", |
| "pages": "2231--2234", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Levy, R. and Andrew, G. (2006). Tregex and Tsurgeon: tools for querying and manipulating tree data structures. In Proceedings of LREC2006, pages 2231-2234.", |
| "links": null |
| }, |
| "BIBREF17": { |
| "ref_id": "b17", |
| "title": "Automatic analysis of syntactic complexity in second language writing", |
| "authors": [ |
| { |
| "first": "X", |
| "middle": [], |
| "last": "Lu", |
| "suffix": "" |
| } |
| ], |
| "year": 2010, |
| "venue": "International Journal of Corpus Linguistics", |
| "volume": "15", |
| "issue": "4", |
| "pages": "474--496", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Lu, X. (2010). Automatic analysis of syntactic complex- ity in second language writing. International Journal of Corpus Linguistics, 15(4):474-496.", |
| "links": null |
| }, |
| "BIBREF18": { |
| "ref_id": "b18", |
| "title": "The Relationship of Lexical Richness to the Quality of ESL Learners' Oral Narratives. The Modern Language", |
| "authors": [ |
| { |
| "first": "X", |
| "middle": [], |
| "last": "Lu", |
| "suffix": "" |
| } |
| ], |
| "year": 2012, |
| "venue": "Journal", |
| "volume": "96", |
| "issue": "2", |
| "pages": "190--208", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Lu, X. (2012). The Relationship of Lexical Richness to the Quality of ESL Learners' Oral Narratives. The Modern Language Journal, 96(2):190-208.", |
| "links": null |
| }, |
| "BIBREF19": { |
| "ref_id": "b19", |
| "title": "Computational Methods for Corpus Annotation and Analysis", |
| "authors": [ |
| { |
| "first": "X", |
| "middle": [], |
| "last": "Lu", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Lu, X. (2014). Computational Methods for Corpus Anno- tation and Analysis. Springer, Dordrecht.", |
| "links": null |
| }, |
| "BIBREF20": { |
| "ref_id": "b20", |
| "title": "The Stanford CoreNLP Natural Language Processing Toolkit", |
| "authors": [ |
| { |
| "first": "C", |
| "middle": [ |
| "D" |
| ], |
| "last": "Manning", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Surdeanu", |
| "suffix": "" |
| }, |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Bauer", |
| "suffix": "" |
| }, |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Finkel", |
| "suffix": "" |
| }, |
| { |
| "first": "S", |
| "middle": [ |
| "J" |
| ], |
| "last": "Bethard", |
| "suffix": "" |
| }, |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Mcclosky", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics: System Demonstrations", |
| "volume": "", |
| "issue": "", |
| "pages": "55--60", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Manning, C. D., Surdeanu, M., Bauer, J., Finkel, J., Bethard, S. J., and McClosky, D. (2014). The Stanford CoreNLP Natural Language Processing Toolkit. In Pro- ceedings of the 52nd Annual Meeting of the Association for Computational Linguistics: System Demonstrations, pages 55-60.", |
| "links": null |
| }, |
| "BIBREF21": { |
| "ref_id": "b21", |
| "title": "Constructing a longitudinal learner corpus to track l2 spoken English", |
| "authors": [ |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Mariko", |
| "suffix": "" |
| }, |
| { |
| "first": "Y", |
| "middle": [], |
| "last": "Kondo", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "Journal of Modern Languages", |
| "volume": "29", |
| "issue": "", |
| "pages": "23--44", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Mariko, A. and Kondo, Y. (2019). Constructing a longitu- dinal learner corpus to track l2 spoken English. Journal of Modern Languages, 29:23-44.", |
| "links": null |
| }, |
| "BIBREF22": { |
| "ref_id": "b22", |
| "title": "Automated Evaluation of Text and Discourse with Coh-Metrix", |
| "authors": [ |
| { |
| "first": "D", |
| "middle": [ |
| "S" |
| ], |
| "last": "Mcnamara", |
| "suffix": "" |
| }, |
| { |
| "first": "A", |
| "middle": [ |
| "C" |
| ], |
| "last": "Graesser", |
| "suffix": "" |
| }, |
| { |
| "first": "P", |
| "middle": [ |
| "M" |
| ], |
| "last": "Mccarthy", |
| "suffix": "" |
| }, |
| { |
| "first": "Z", |
| "middle": [], |
| "last": "Cai", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "McNamara, D. S., Graesser, A. C., McCarthy, P. M., and Cai, Z. (2014). Automated Evaluation of Text and Dis- course with Coh-Metrix. Cambridge University Press, USA.", |
| "links": null |
| }, |
| "BIBREF23": { |
| "ref_id": "b23", |
| "title": "Docker: lightweight linux containers for consistent development and deployment. Linux journal", |
| "authors": [ |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Merkel", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Merkel, D. (2014). Docker: lightweight linux containers for consistent development and deployment. Linux jour- nal, 2014(239):2.", |
| "links": null |
| }, |
| "BIBREF24": { |
| "ref_id": "b24", |
| "title": "Data stewardship for open science: Implementing FAIR principles", |
| "authors": [ |
| { |
| "first": "B", |
| "middle": [], |
| "last": "Mons", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Mons, B. (2018). Data stewardship for open science: Im- plementing FAIR principles. Chapman and Hall/CRC.", |
| "links": null |
| }, |
| "BIBREF25": { |
| "ref_id": "b25", |
| "title": "The Use of the Computer in Analyzing Student Essays. International Review of Education / Internationale Zeitschrift f\u00fcr Erziehungswissenschaft / Revue Internationale de l'Education", |
| "authors": [ |
| { |
| "first": "E", |
| "middle": [ |
| "B" |
| ], |
| "last": "Page", |
| "suffix": "" |
| } |
| ], |
| "year": 1968, |
| "venue": "", |
| "volume": "14", |
| "issue": "", |
| "pages": "210--225", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Page, E. B. (1968). The Use of the Computer in Analyz- ing Student Essays. International Review of Education / Internationale Zeitschrift f\u00fcr Erziehungswissenschaft / Revue Internationale de l'Education, 14(2):210-225.", |
| "links": null |
| }, |
| "BIBREF26": { |
| "ref_id": "b26", |
| "title": "R\u00e9aliser un plan de gestion de donn\u00e9es \"fair", |
| "authors": [ |
| { |
| "first": "N", |
| "middle": [], |
| "last": "Reymonet", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Moysan", |
| "suffix": "" |
| }, |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Cartier", |
| "suffix": "" |
| }, |
| { |
| "first": "R", |
| "middle": [], |
| "last": "D\u00e9l\u00e9montez", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Reymonet, N., Moysan, M., Cartier, A., and D\u00e9l\u00e9montez, R. (2018). R\u00e9aliser un plan de gestion de donn\u00e9es \"fair\" : mod\u00e8le 2018.", |
| "links": null |
| }, |
| "BIBREF27": { |
| "ref_id": "b27", |
| "title": "Probabilistic Part-of-Speech Tagging Using Decision Trees", |
| "authors": [ |
| { |
| "first": "H", |
| "middle": [], |
| "last": "Schmid", |
| "suffix": "" |
| } |
| ], |
| "year": 1994, |
| "venue": "Proceedings of the International Conference on New Methods in Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "14--16", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Schmid, H. (1994). Probabilistic Part-of-Speech Tagging Using Decision Trees. In Proceedings of the Interna- tional Conference on New Methods in Language Pro- cessing, pages 14-16, Manchester: UK.", |
| "links": null |
| }, |
| "BIBREF28": { |
| "ref_id": "b28", |
| "title": "Human and Automated CEFR-based Grading of Short Answers", |
| "authors": [ |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Tack", |
| "suffix": "" |
| }, |
| { |
| "first": "T", |
| "middle": [], |
| "last": "Fran\u00e7ois", |
| "suffix": "" |
| }, |
| { |
| "first": "S", |
| "middle": [], |
| "last": "Roekhaut", |
| "suffix": "" |
| }, |
| { |
| "first": "C", |
| "middle": [], |
| "last": "Fairon", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Proceedings of the 12th Workshop on Innovative Use of NLP for Building Educational Applications", |
| "volume": "", |
| "issue": "", |
| "pages": "169--179", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Tack, A., Fran\u00e7ois, T., Roekhaut, S., and Fairon, C. (2017). Human and Automated CEFR-based Grading of Short Answers. In Proceedings of the 12th Workshop on In- novative Use of NLP for Building Educational Applica- tions, pages 169-179, Copenhagen, Denmark, Septem- ber. Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF29": { |
| "ref_id": "b29", |
| "title": "NT2Lex: A CEFR-graded lexical resource for Dutch as a foreign language linked to open Dutch WordNet", |
| "authors": [ |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Tack", |
| "suffix": "" |
| }, |
| { |
| "first": "T", |
| "middle": [], |
| "last": "Fran\u00e7ois", |
| "suffix": "" |
| }, |
| { |
| "first": "P", |
| "middle": [], |
| "last": "Desmet", |
| "suffix": "" |
| }, |
| { |
| "first": "C", |
| "middle": [], |
| "last": "Fairon", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "Proceedings of the Thirteenth Workshop on Innovative Use of NLP for Building Educational Applications", |
| "volume": "", |
| "issue": "", |
| "pages": "137--146", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Tack, A., Fran\u00e7ois, T., Desmet, P., and Fairon, C. (2018). NT2Lex: A CEFR-graded lexical resource for Dutch as a foreign language linked to open Dutch WordNet. In Proceedings of the Thirteenth Workshop on Innova- tive Use of NLP for Building Educational Applications, pages 137-146, New Orleans, Louisiana, June. Associa- tion for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF30": { |
| "ref_id": "b30", |
| "title": "The role and applications of expert error annotation in a corpus of English learner texts. Computational Linguisitics and Intellectual Technologies", |
| "authors": [ |
| { |
| "first": "O", |
| "middle": [], |
| "last": "Vinogradova", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "Proceedings of Dialog", |
| "volume": "15", |
| "issue": "", |
| "pages": "740--751", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Vinogradova, O. (2016). The role and applications of ex- pert error annotation in a corpus of English learner texts. Computational Linguisitics and Intellectual Technolo- gies. Proceedings of Dialog 2016, 15:740-751.", |
| "links": null |
| }, |
| "BIBREF31": { |
| "ref_id": "b31", |
| "title": "Classification of Swedish learner essays by CEFR levels", |
| "authors": [ |
| { |
| "first": "E", |
| "middle": [], |
| "last": "Volodina", |
| "suffix": "" |
| }, |
| { |
| "first": "I", |
| "middle": [], |
| "last": "Pil\u00e1n", |
| "suffix": "" |
| }, |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Alfter", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "Salomi Papadima-Sophocleous", |
| "volume": "", |
| "issue": "", |
| "pages": "456--461", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Volodina, E., Pil\u00e1n, I., and Alfter, D. (2016). Classification of Swedish learner essays by CEFR levels. In Salomi Papadima-Sophocleous, et al., editors, CALL communi- ties and culture ; short papers from EUROCALL 2016, pages 456-461. Research-publishing.net.", |
| "links": null |
| }, |
| "BIBREF32": { |
| "ref_id": "b32", |
| "title": "Speech recognition (version 3.8) python library", |
| "authors": [ |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Zhang", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Zhang, A. (2017). Speech recognition (ver- sion 3.8) python library. Available from https://pypi.org/project/SpeechRecognition/.", |
| "links": null |
| }, |
| "BIBREF33": { |
| "ref_id": "b33", |
| "title": "Regularization and variable selection via the Elastic Net", |
| "authors": [ |
| { |
| "first": "H", |
| "middle": [], |
| "last": "Zou", |
| "suffix": "" |
| }, |
| { |
| "first": "T", |
| "middle": [], |
| "last": "Hastie", |
| "suffix": "" |
| } |
| ], |
| "year": 2005, |
| "venue": "Journal of the Royal Statistical Society, Series B", |
| "volume": "67", |
| "issue": "2", |
| "pages": "301--320", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Zou, H. and Hastie, T. (2005). Regularization and variable selection via the Elastic Net. Journal of the Royal Statis- tical Society, Series B, 67(2):301-320.", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "FIGREF0": { |
| "num": null, |
| "uris": null, |
| "type_str": "figure", |
| "text": "Figure 1: The pipeline for our project" |
| }, |
| "FIGREF1": { |
| "num": null, |
| "uris": null, |
| "type_str": "figure", |
| "text": "The user interface prototype 4. A Web-based Scheduling Infrastructure for Easy Text-Processing System Demonstration 4.1. Description of the Infrastructure Flow" |
| }, |
| "FIGREF2": { |
| "num": null, |
| "uris": null, |
| "type_str": "figure", |
| "text": "The customisable infrastructure" |
| } |
| } |
| } |
| } |