top of page

13th Workshop on Innovative Use of NLP

for Building Educational Applications

Conference:

Organization:

​

​

Contact Email:

Date:

Venue:

NAACL 2018

Joel Tetreault (Grammarly), Jill Burstein (Educational Testing Service),

Ekaterina Kochmar (University of Cambridge), Claudia Leacock (Grammarly), Helen Yannakoudakis (University of Cambridge)

bea.nlp.workshop@gmail.com

June 05, 2018

New Orleans, USA

Sponsors                   Workshop description                   SLAM Shared Task                   CWI Shared Task                   Important Dates 
 
             Submission Information              Organizing Committee              Program Committee              Related Links

SPONSORS

We're excited to announce our sponsors for the BEA13 Workshop! Our Gold Sponsors are DuolingoGrammarlyNational Board of Medical Examinders and Turnitin / Lightside Labs. Our Silver Sponsors are Educational Testing Service and iLexIRCognii is our bronze level sponsor. If you or your company or institution are interested in sponsoring the BEA12, please send us an email at bea.nlp.workshop@gmail.com. Sponsorship goes toward subsidizing dinner for students attending the workshop and free t-shirts with registration.

Gold Level Sponsors

Silver Level Sponsors

Bronze Level Sponsors

WORKSHOP DESCRIPTION

​

The BEA Workshop is a leading venue for NLP innovation for educational applications. It is one of the largest one-day workshops in the ACL community. The workshop’s continuous growth illustrates an alignment between societal need and technology advances. NLP capabilities now support an array of learning domain knowledge, including writing, speaking, reading, science, and mathematics, and the related intra- (e.g., self-confidence) and inter-personal (e.g., peer collaboration) domains that support achievement in the learning domains. Within these domains, the community continues to develop and deploy innovative NLP approaches for use in educational settings. In the writing and speech domains, automated writing evaluation (AWE) and speech scoring applications, respectively, are commercially deployed in high-stakes assessment, and instructional contexts (including Massive Open Online Courses (MOOCs), and K-12 settings). Commercially-deployed plagiarism detection in K-12 and higher education settings is also prevalent. The current educational and assessment landscape in K-12, higher education, and adult learning (in academic and workplace settings) fosters a strong interest in technologies that yield user log data that can be leveraged for analytics that support proficiency measures for complex constructs across learning domains. For writing, there is a focus on innovation that supports writing tasks requiring source use, argumentative discourse, and factual content accuracy. For speech, there is an interest in advancing automated scoring to include the evaluation of discourse and content features in responses to spoken assessments. General advances in speech technology have promoted a renewed interest in spoken dialog and multimodal systems for instruction and assessment, for instance, for workplace interviews and simulated teaching environments. The explosive growth of mobile applications for game-based and simulation applications for instruction and assessment is another place where NLP has begun to play a large role, especially in language learning.

​

NLP for educational applications has gained visibility outside of the NLP community. First, the Hewlett Foundation reached out to public and private sectors and sponsored two competitions: one for automated essay scoring, and the other for scoring of short response items. The motivation driving these competitions was to engage the larger scientific community in this enterprise. Learning @ Scale is a relatively new venue for NLP research in education. MOOCs now incorporate AWE systems to manage several thousand assignments that may be received during a single MOOC course. MOOCs for Refugees have more recently popped up in response to the current social situations. Courses include language learning, and we can imagine that AWE and other NLP capabilities could support coursework. Another breakthrough for educational applications within the CL community is the presence of a number of shared-task competitions over the last four years – including three shared tasks on grammatical error detection and correction alone. NLP/Education shared tasks, typically in the area of grammar-error detection, have seen new areas of research, such as the “Automated Evaluation of Scientific Writing” at BEA11 and Native Language Identification at BEA12. All of these competitions increased the visibility of, and interest in, our field. In conjunction with the International Joint Conference on Natural Language Processing (ACL-IJCNLP) 2015, the Natural Language Processing Techniques for Educational Applications (NLP-TEA) workshop had a shared task in Chinese error diagnosis, and NLP-TEA had additional shared tasks at the 2016, and a fourth workshop in 2017 co-located with IJCNLP.

​

The 13th BEA workshop will have oral presentation sessions and a large poster session in order to maximize the amount of original work presented. We expect that the workshop will continue to expose the NLP community to technologies that identify novel opportunities for the use of NLP in education in English, and languages other than English. The workshop will solicit both full papers and short papers for either oral or poster presentation. We will solicit papers that incorporate NLP methods, including, but not limited to: automated scoring of open-ended textual and spoken responses; game-based instruction and assessment; educational data mining; intelligent tutoring; peer review, grammatical error detection; learner cognition; spoken dialog; multimodal applications; tools for teachers and test developers; and use of corpora. Research that incorporates NLP methods for use with mobile and game-based platforms will be of special interest. Specific topics include:

 

  • Automated scoring/evaluation for written student responses (across multiple genres)

    • Content analysis for scoring/assessment

    • Detection and correction of grammatical and other types of errors (such as, spelling and word usage)

    • Argumentation, discourse, sentiment, stylistic analysis, & non-literal language 

    • Plagiarism detection

    • Detection of features related to interest, motivation, and values in writing tasks

     

  • Intelligent Tutoring (IT), Collaborative Learning Environments

    • Educational Data Mining: Collection of user log data from educational applications

    • Game-based learning

    • Multimodal communication (including dialog systems) between students and computers

    • Knowledge representation in learning systems

    • Concept visualization in learning systems

    ​

  • Learner cognition

    • Assessment of learners' language and cognitive skill levels

    • Systems that detect and adapt to learners' cognitive or emotional states

    • Tools for learners with special needs

     

  • Use of corpora in educational tools

    • Data mining of learner and other corpora for tool building

    • Annotation standards and schemas / annotator agreement

     

  • Tools and applications for classroom teachers and/or test developers

    • NLP tools for second and foreign language learners

    • Semantic-based access to instructional materials to identify appropriate texts

    • Tools that automatically generate test questions

    • Processing of and access to lecture materials across topics and genres

    • Adaptation of instructional text to individual learners’ grade levels

    • Tools for text-based curriculum development

Description
Sponsors
Shared Task 1

SHARED TASK ON SECOND LANGUAGE ACQUISITION MODELING

​

Task Description

​

The workshop will host a Shared Task on Second Language Acquisition Modeling (SLAM), using data provided by Duolingo, a popular free online computer-aided language learning (CALL) platform. Teams will be provided with “traces” of all translation and transcription exercises from 800+ language learners — annotated for errors — spanning their first 100 days of activity on Duolingo. The task is then to predict errors made by a held-out set of 100+ language learners over their first 100 days. There will be four tracks for learners of English, Spanish, French and German. We believe that this task presents several new and interesting dimensions for research in Second Language Acquisition modeling: (1) subjects are mostly beginners in their respective L2s, (2) success will likely require teams to model learning — and forgetting — over time, and (3) teams are encouraged to use features which generalize across a variety of languages (hence 4 tracks). 

​

URLhttp://sharedtask.duolingo.com

​

Task Organizers

​

Burr Settles (Duolingo), Erin Gustafson (Duolingo), Masato Hagiwara (Duolingo), Bozena Pajak (Duolingo), Joseph Rollinson (Duolingo), Chris Brust (Duolingo), Hideki Shima (Duolingo), Nitin Madnani (Educational Testing Service) 

SHARED TASK ON COMPLEX WORD IDENTIFICATION

​

Task Description

​

Over the past decade a number of studies have been published on automatic text simplification (Specia, 2010; Saggion et al. 2015; Štajner, 2015). Text simplification systems aim to facilitate reading comprehension to different target readerships such as foreign language learners, and native speakers with low literacy levels or various kinds of reading impairments. Two main factors that impact reading comprehension addressed by these systems are lexical complexity and syntactic complexity.

Many lexical simplification systems have been proposed up to this date (Glavaš and Štajner, 2015; Paetzold and Specia, 2016). It has been shown that those systems which have a separate complex word identification (CWI) module at the beginning of their pipeline outperform those systems which treat all words as potentially complex (Paetzold and Specia, 2015). Therefore, automatic identification of words that are difficult for a given target population is an important step for building better performing lexical simplification systems. This step is known as complex word identification (CWI) (Shardlow, 2013).

​

The first shared task on CWI was organized at the SemEval 2016 (Paetzold and Specia, 2016b). It featured 21 teams that competed submitting 42 systems trained to predict whether words in a given context were complex or non-complex for a non-native English speaker. Following the success of the first CWI shared task at SemEval 2016 we propose the organization of a second edition of the challenge at the BEA workshop 2018.

​

The first edition of the CWI challenge included only English data aimed at non-native English speakers, whereas the second edition will feature a multilingual dataset (Yimam, 2017a, 2017b) and four individual tracks: (1) English monolingual CWI, (2) Spanish monolingual CWI, (3) German monolingual CWI, (4) Multilingual CWI with a French test set.

​

URLhttps://sites.google.com/view/cwisharedtask2018/

​

Task Organizers

​

Chris Biemann (University of Hamburg), Shervin Malmasi (Harvard Medical School), Gustavo Paetzold (University of Sheffield), Lucia Specia (University of Sheffield), Sanja Štajner (University of Mannheim), Anais Tack (KU Leuven), Seid Muhie Yimam (University of Hamburg), Marcos Zampieri (University of Wolverhampton)

Shared Task 2

IMPORTANT DATES

​

  • Submission Deadline: Tuesday, March 20 - 23:59 EST (New York City Time) [ Current EST ]

  • Notification of Acceptance: Wednesday, April 04

  • Camera-ready Papers Due: Monday, April 16

  • Workshop: June 05

Important Dates

SUBMISSION INFORMATION

​

We will be using the NAACL submission guidelines and style files for the BEA13 Workshop this year. Authors are invited to submit a full paper of up to 8 pages of content with unlimited pages for references. We also invite short papers of up to 4 pages of content, including unlimited pages for references. Final camera ready versions of accepted papers will be given an additional page of content to address reviewer comments.

 

Previously published papers cannot be accepted. The submissions will be reviewed by the program committee. As reviewing will be blind, please ensure that papers are anonymous. Self-references that reveal the author's identity, e.g., "We previously showed (Smith, 1991) ...", should be avoided. Instead, use citations such as "Smith previously showed (Smith, 1991) ...".

​

We have also included conflict of interest in the submission form. You should mark all potential reviewers who have been authors on the paper, are from the same research group or institution, or who have seen versions of this paper or discussed it with you.

 

We will be using the START conference system to manage submissions:
https://www.softconf.com/naacl2018/Education-NLP18.

Submission Info

ORGANIZING COMMITTEE

​

  • Joel Tetreault, Grammarly (primary contact)

  • Jill Burstein, Educational Testing Services

  • Ekaterina Kochmar, University of Cambridge

  • Claudia Leacock, Grammarly

  • Helen Yannakoudakis, University of Cambridge

Organising Committee

PROGRAM COMMITTEE

​

  • Lars Ahrenberg, Linköping University

  • David Alfter, University of Gothenburg

  • Dimitris Alikaniotis, Grammarly

  • Homa B. Hashemi, Intelligent System Program, University of Pittsburgh

  • Rafael E Banchs, Institute for Infocomm Research

  • Sagnik Banerjee, Iowa State University

  • Rajendra Banjade, Audible inc. (an Amazon company)

  • Lee Becker, Pearson

  • Beata Beigman Klebanov, Educational Testing Service

  • Lisa Beinborn UKP Lab, Technische Universität Darmstadt

  • Delphine Bernhard, Université de Strasbourg

  • Sameer Bhatnagar ,Polytechnique Montreal

  • Serge Bibauw, KU Leuven & Université Catholique de Louvain

  • Joachim Bingel, University of Copenhagen

  • Johannes Bjerva, University of Copenhagen

  • Kristy Boyer, University of Florida

  • Ted Briscoe, University of Cambridge

  • Dominique Brunato, Institute of Computational Linguistics (ILC-CNR)

  • Chris Bryant, University of Cambridge

  • Andrew Caines, University of Cambridge

  • Mei-Hua Chen, Tunghai University

  • Martin Chodorow, Hunter College of CUNY

  • Shamil Chollampatt, National University of Singapore

  • Mark Core, University of Southern California

  • Robert Dale, Language Technology Group

  • Vidas Daudaravicius, VTEX Research

  • Kordula De Kuthy, University of Tübingen

  • Barbara Di Eugenio, University of Illinois at Chicago

  • Yo Ehara, National Institute of Advanced Industrial Science and Technology

  • Noureddine Elouazizi, Faculty of Science (Skylight/Dean Office), University of British Columbia

  • Keelan Evanini, Educational Testing Service

  • Cédrick Fairon, Université Catholique de Louvain

  • Youmna Farag, University of Cambridge

  • Mariano Felice, University of Cambridge

  • Oliver Ferschke, M*Modal

  • Michael Flor, Educational Testing Service

  • Thomas François, UCLouvain

  • Michael Gamon, Microsoft Research

  • Kallirroi Georgila, University of Southern California

  • Andrew Gibson, University of Technology Sydney

  • Jonathan Gordon, University of Southern California

  • Floriana Grasso, University of Liverpool, UK

  • Gintare Grigonyte, Stockholm University

  • Iryna Gurevych, UKP Lab, Technische Universität Darmstadt

  • Na-Rae Han, University of Pittsburgh

  • Jiangang Hao, Educational Testing Service

  • Marti Hearst, University of California, Berkeley

  • Trude Heift, Simon Fraser University

  • Derrick Higgins, American Family Insurance

  • Andrea Horbach, University Duisburg-Essen

  • Chung-Chi Huang, Frostburg State University

  • Radu Tudor Ionescu, University of Bucharest

  • Ross Israel, Factual Inc

  • Lifeng Jin, The Ohio State University

  • Pamela Jordan, University of Pittsburgh

  • Marcin Junczys-Dowmunt, Adam Mickiewicz University

  • John Kelleher, Dublin Institute of Technology

  • Levi King, Indiana University

  • Mamoru Komachi, Tokyo Metropolitan University

  • Sandra Kuebler, Indiana University

  • Girish Kumar, Carousell

  • Ji-Ung Lee, UKP Lab Technische Universität Darmstadt

  • John Lee, City University of Hong Kong

  • Lung-Hao Lee, National Taiwan Normal University

  • James Lester, North Carolina State University

  • Wen Li, Indiana University

  • Maria Liakata, University of Warwick

  • Chen Liang, Pennsylvania State University

  • Diane Litman, University of Pittsburgh

  • Peter Ljunglöf, University of Gothenburg and Chalmers University

  • Anastassia Loukina, Educational Testing Service

  • Xiaofei Lu, Pennsylvania State University

  • Luca Lugini, University of Pittsburgh

  • Nitin Madnani, Educational Testing Service

  • Montse Maritxalar, University of the Basque Country

  • Ilia Markov, Center for Computing Research, Instituto Politécnico Nacional

  • James Martin, University of Colorado Boulder

  • Ditty Mathew, IIT Madras

  • Julie Medero, Harvey Mudd College

  • Beata Megyesi, Uppsala University

  • Detmar Meurers, University of Tübingen

  • Maria Moritz, University of Goettingen

  • Smaranda Muresan, Columbia University

  • Courtney Napoles, Grammarly

  • Diane Napolitano, Educational Testing Service

  • Hwee Tou Ng, National University of Singapore

  • Huy Nguyen, Graduate Student at University of Pittsburgh

  • Rodney Nielsen, University of North Texas

  • Nobal Niraula, Boeing Research & Technology

  • Yoo Rhee Oh, Electronics and Telecommunications Research Institute (ETRI)

  • Constantin Orasan, University of Wolverhampton

  • Robert Östling, Stockholm University

  • Ulrike Pado, Hochschule fuer Technik Stuttgart

  • Ted Pedersen, University of Minnesota, Duluth

  • Isaac Persing , The University of Texas at Dallas

  • Ildikó Pilán, University of Gothenburg

  • Patti Price, PPRICE Speech and Language Technology Consulting

  • Taraka Rama, University of Oslo

  • Lakshmi Ramachandran, A9.com Inc

  • Vikram Ramanarayanan, Educational Testing Service R&D and UC San Francisco

  • Sudha Rao, University of Maryland, College Park

  • Hanumant Redkar, Indian Institute of Technology Bombay (IIT Bombay)

  • Livy Real, IBM Research

  • Marek Rei, University of Cambridge

  • Robert Reynolds, Brigham Young University

  • Brian Riordan, Educational Testing Service

  • Andrew Rosenberg, IBM Research AI

  • Mark Rosenstein, Pearson

  • Mihai Rotaru, Textkernel

  • Alla Rozovskaya, City University of New York

  • C. Anton Rytting, University of Maryland

  • Allen Schmaltz, Harvard University

  • Claudia Schulz, Technische Universität Darmstadt

  • Burr Settles, Duolingo

  • Grigori Sidorov, Instituto Politecnico Nacional 

  • Anders Søgaard, University of Copenhagen

  • Helmer Strik Linguistics, Centre for Language Studies (CLS), Centre for Language and Speech Technology (CLST), Radboud University Nijmegen; NovoLanguage Nijmegen

  • Jan Švec, Department of Cybernetics, University of West Bohemia

  • Anaïs Tack , Université Catholique de Louvain & KU Leuven

  • Yuen-Hsien, Tseng National Taiwan Normal University

  • Sowmya Vajjala, Iowa State University

  • Giulia Venturi, Institute of Computational Linguistics "A. Zampolli" (ILC-CNR), Pisa

  • Aline Villavicencio, Federal University of Rio Grande do Sul (Brazil) and University of Essex (UK)

  • Elena Volodina, University of Gothenburg, Sweden

  • Shuting Wang, Facebook

  • Michael White, The Ohio State University

  • David Wible, National Central University

  • Alistair Willis, Open University, UK

  • Michael Wojatzki , University of Duisburg-Essen

  • Huichao Xue, LinkedIn

  • Victoria Yaneva, University of Wolverhampton

  • Zheng Yuan, University of Cambridge

  • Marcos Zampieri, University of Wolverhampton

  • Klaus Zechner, Educational Testing Service​

Program Committee
Related Links
bottom of page