top of page

The 11th Workshop on Innovative Use of NLP

for Building Educational Applications

Conference:

Organization:

 

 

Contact Email:

Date:

Venue:

NAACL 2016

Joel Tetreault (Yahoo Labs), Jill Burstein (Educational Testing Service),

Claudia Leacock (Consultant), Helen Yannakoudakis (University of Cambridge)

bea.nlp.workshop@gmail.com

June 16, 2016

San Diego, CA, USA

SPONSORS

We are pleased to announce that Educational Testing ServiceGrammarlyTurnitin Lightside LabsPacific Metrics,  Cambridge Assessments and American Institutes for Research are all gold level sponsors of the BEA11 workshop and iLexIR and Cognii are our silver level sponsors! If you or your company or institution are interested in sponsoring the BEA11, please send us an email at bea.nlp.workshop@gmail.com. Sponsorship goes toward subsidizing dinner for students attending the workshop and free t-shirts with registration.

Gold Level Sponsors

Silver Level Sponsors

Bronze Level Sponsors

WORKSHOP DESCRIPTION

​

The BEA Workshop is a leading venue for NLP innovation for educational applications. It is one of the largest one-day workshops in the ACL community. The workshop’s continuous growth illustrates an alignment between societal need and technology advances. NLP capabilities now support an array of learning domains, including writing, speaking, reading, science, and mathematics. Within these domains, the community continues to develop and deploy innovative NLP approaches for use in educational settings. In the writing and speech domains, automated writing evaluation (AWE) and speech scoring applications, respectively, are commercially deployed in high-stakes assessment, and instructional contexts (including Massive Open Online Courses (MOOCs), and K-12 settings). Commercially-deployed plagiarism detection in K-12 and higher education settings is also prevalent. The current educational and assessment landscape in K-12 and higher education fosters a strong interest in technologies that yield analytics to support proficiency measures for complex constructs across learning domains. For writing, there is a focus innovation that supports writing tasks requiring source use, argumentative discourse, and factual content accuracy. For speech, there is an interest in advancing automated scoring to include the evaluation of discourse and content features in responses to spoken assessments. General advances in speech technology have promoted a renewed interest in spoken dialog and multimodal systems for instruction and assessment. The explosive growth of mobile applications for game-based and simulation applications for instruction and assessment is another place where NLP can play a large role. 

​

The use of NLP in educational applications has gained visibility outside of the NLP community. First, the Hewlett Foundation reached out to public and private sectors and sponsored two competitions: one for automated essay scoring, and the other for scoring of short response items. The motivation driving these competitions was to engage the larger scientific community in this enterprise. MOOCs are now also beginning to incorporate AWE systems to manage the thousands of assignments that may be received during a single MOOC course. Learning @ Scale is a relatively new venue for NLP research in education. Another breakthrough for educational applications within the CL community is the presence of a number of shared-task competitions over the last four years – including three shared tasks on grammatical error correction alone. In 2014 alone, there were four shared tasks in NLP/Education related areas. Most recently, the 2015 ACL-IJCNLP Workshop on Natural Language Processing Techniques for Educational Applications workshop had a shared task in Chinese error diagnosis. All of these competitions increased the visibility of, and interest in, our field. 

​

The workshop will have oral presentation sessions and a large poster session in order to maximize the amount of original work presented. This year, we are planning an invited Industry Panel comprised of representatives of companies that work in the NLP and Education space. We expect that the workshop will continue to expose the NLP community to technologies that identify novel opportunities for the use of NLP in education in English, and languages other than English. The workshop will solicit both full papers and short papers for either oral or poster presentation. We will solicit papers that incorporate NLP methods, including, but not limited to: automated scoring of open-ended textual and spoken responses; game-based instruction and assessment; intelligent tutoring; peer review, grammatical error detection; learner cognition; spoken dialog; multimodal applications; tools for teachers and test developers; and use of corpora. Research that incorporates NLP methods for use with mobile and game-based platforms will be of special interest. Specific topics include:

​

  • Automated scoring/evaluation for written student responses

    • Content analysis for scoring/assessment 

    • Analysis of the structure of argumentation

    • Grammatical error detection and correction 

    • Discourse and stylistic analysis 

    • Plagiarism detection 

    • Machine translation for assessment, instruction and curriculum development 

    • Detection of non-literal language (e.g., metaphor)

    • Sentiment analysis

    • Non-traditional genres (beyond essay scoring)

​

  • Intelligent Tutoring (IT) and Game-based assessment that incorporates NLP

    • Dialogue systems in education 

    • Hypothesis formation and testing 

    • Multi-modal communication between students and computers 

    • Generation of tutorial responses 

    • Knowledge representation in learning systems 

    • Concept visualization in learning systems

​

  • Learner cognition

    • Assessment of learners' language and cognitive skill levels 

    • Systems that detect and adapt to learners' cognitive or emotional states 

    • Tools for learners with special needs

​

  • Use of corpora in educational tools

    • Data mining of learner and other corpora for tool building 

    • Annotation standards and schemas / annotator agreement

​

  • Tools and applications for classroom teachers and/or test developers

    • NLP tools for second and foreign language learners 

    • Semantic-based access to instructional materials to identify appropriate texts 

    • Tools that automatically generate test questions 

    • Processing of and access to lecture materials across topics and genres 

    • Adaptation of instructional text to individual learners' grade levels 

    • Tools for text-based curriculum development 

    • E-learning tools for personalized course content 

    • Language-based educational games

​

  • Descriptions and proposals for shared tasks 

​

  • Retrospective or survey papers on a particular NLP/Edu topic or field 

 

  • Vision papers about ideas discussing how the field should develop

Description
Sponsors
AESW Shared Task

AESW SHARED TASK

​

We are pleased to announce that the first edition of the "Automated Evaluation of Scientific Writing" Shared Task on grammatical error detection will be co-located with BEA11 this year. The Shared Task will be organized independently from the BEA11. System description papers submitted and accepted to the AESW Shared Task will be presented as posters at the BEA Poster Session. In addition, the AESW organizers will summarize the results of the Shared Task in an oral presentation during the BEA. For more information on the task, as well as important dates and submission information, please go to: http://textmining.lt/aesw/index.html. Registration closes February 1st.

SUBMISSION INFORMATION

​

We will be using the NAACL Submission Guidelines for the BEA11 Workshop this year. Authors are invited to submit a full paper of up to 9 pages of content with up to 2 additional pages for references. We also invite short papers of up to 5 pages of content, including 2 additional pages for references. Please note that unlike previous years, final, camera ready versions of accepted papers will not be given an additional page to address reviewer comments.​

​

Papers which describe systems are also invited to give a demo of their system. If you would like to present a demo in addition to presenting the paper, please make sure to select either "full paper + demo" or "short paper + demo" under "Submission Category" in the START submission page. 

​

Previously published papers cannot be accepted. The submissions will be reviewed by the program committee. As reviewing will be blind, please ensure that papers are anonymous. Self-references that reveal the author's identity, e.g., "We previously showed (Smith, 1991) ...", should be avoided. Instead, use citations such as "Smith previously showed (Smith, 1991) ...".

​

We have also included conflict of interest in the submission form. You should mark all potential reviewers who have been authors on the paper, are from the same research group or institution, or who have seen versions of this paper or discussed it with you.

​

Please use the NAACL style sheets for composing your paper: http://naacl.org/naacl-pubs/ .

​

We will be using the START conference system to manage submissions: https://www.softconf.com/naacl2016/BEA11/.

Submission Info

IMPORTANT DATES

​

  • Submission Deadline: March 08 - 23:59 EST (New York City Time) [ Current EST ]

  • Notification of Acceptance: March 25

  • Camera-ready Papers Due: Apr 07

  • Workshop: June 16

Important Dates

PRESENTATION INFORMATION

​

Oral Presentations: Long papers accepted for oral presentations are allotted 20 minutes for the talk and 5 minutes for questions. Short papers that are accepted for oral presentations are allotted 15 minutes for the talk and 5 minutes for questions.

​

Poster Presentations: All papers accepted for a poster presentation will be presented in the session after lunch between 2:00 and 3:30. The posterboards will be self-standing, on top of tables (giving room for laptops, business cards, handouts, etc). The posterboards measure 36 inches high and 48 inches wide . Double-sided tape, pushpins, etc. for affixing the posters to the boards will be provided.

Presentation Info
Program

WORKSHOP PROGRAM

​

​

8:45 - 9:00

​

9:00 - 9:15

​

9:15 - 9:40

​

​

9:40 - 10:05

​

​

10:05 - 10:30

​

​

​

10:30 - 11:00

​

11:00 - 11:25

​

​

11:25 - 11:50

​

​

11:50 - 12:15

​

​

12:25 - 14:00

​

14:00 - 15:30

​

14:00 - 14:45

​

​

​

​

​

​

​

​

​

​

​

​

​

​

​

​

​

​

​

​

​

​

​

​

​

​

​

​

​

​

14:45 - 15:30

​

​

​

​

​

​

​

​

​

​

​

​

​

​

​

​

​

​

​

​

​

​

​

​

​

​

​

​

​

​

​

15:30

​

16:00 - 16:20

​

​

16:20 - 16:45

​

​

 

16:45 - 17:10

​

​

17:10 - 17:25

​

18:30

​

​

BEA11 Proceedings 

​

Loading in of Oral Presentations

 

Opening Remarks

 

The Effect of Multiple Grammatical Errors on Processing Non-Native Writing 
Courtney Napoles, Aoife Cahill and Nitin Madnani

 

Text Readability Assessment for Second Language Learners 
Menglin Xia, Ekaterina Kochmar and Ted Briscoe

 

Automatic Generation of Context-Based Fill-in-the-Blank Exercises Using Co-occurrence Likelihoods and Google n-grams 
Jennifer Hill and Rahul Simha

 

Break

 

Automated classification of collaborative problem solving interactions in simulated science tasks 
Michael Flor, Su-Youn Yoon, Jiangang Hao, Lei Liu and Alina von Davier

 

Computer-assisted stylistic revision with incomplete and noisy feedback. A pilot study 
Christian M. Meyer and Johann Frerik Koch

 

A Report on the Automatic Evaluation of Scientific Writing Shared Task 
Vidas Daudaravicius, Rafael E. Banchs, Elena Volodina and Courtney Napoles

 

Lunch

 

BEA11 Poster and Demo Session

 

BEA11 Poster and Demo Session A

​

Bundled Gap Filling: A New Paradigm for Unambiguous Cloze Exercises 
Michael Wojatzki, Oren Melamud and Torsten Zesch

Unsupervised Modeling of Topical Relevance in L2 Learner Text 
Ronan Cummins, Helen Yannakoudakis and Ted Briscoe

Pictogrammar: an AAC device based on a semantic grammar 
Fernando Martínez-Santiago, Miguel Ángel García Cumbreras, Arturo Montejo Ráez and Manuel Carlos Díaz Galiano

Detecting Context Dependence in Exercise Item Candidates Selected from Corpora 
Ildikó Pilán

Model Combination for Correcting Preposition Selection Errors 
Nitin Madnani, Michael Heilman and Aoife Cahill

Automated scoring across different modalities 
Anastassia Loukina and Aoife Cahill

Predicting the Spelling Difficulty of Words for Language Learners 
Lisa Beinborn, Torsten Zesch and Iryna Gurevych

Topicality-Based Indices for Essay Scoring 
Beata Beigman Klebanov, Michael Flor and Binod Gyawali

Shallow Semantic Reasoning from an Incomplete Gold Standard for Learner Language 
Levi King and Markus Dickinson

Characterizing Text Difficulty with Word Frequencies 
Xiaobin Chen and Detmar Meurers

The NTNU-YZU System in the AESW Shared Task: Automated Evaluation of Scientific Writing Using a Convolutional Neural Network 
Lung-Hao Lee, Bo-Lin Lin, Liang-Chih Yu and Yuen-Hsien Tseng

Feature-Rich Error Detection in Scientific Writing Using Logistic Regression 
Madeline Remse, Mohsen Mesgar and Michael Strube

UW-Stanford System Description for AESW 2016 Shared Task on Grammatical Error Detection 
Dan Flickinger, Michael Goodman and Woodley Packard

​

BEA11 Poster and Demo Session B

 

Automatically Scoring Tests of Proficiency in Music Instruction 
Nitin Madnani, Aoife Cahill and Brian Riordan

Automatically Extracting Topical Components for a Response-to-Text Writing Assessment 
Zahra Rahimi and Diane Litman

Cost-Effectiveness in Building a Low-Resource Morphological Analyzer for Learner Language 
Scott Ledbetter and Markus Dickinson 

Spoken Text Difficulty Estimation Using Linguistic Features 
Su-Youn Yoon, Yeonsuk Cho and Diane Napolitano

Enhancing STEM Motivation through Personal and Communal Values: NLP for Assessment of Utility Value in Student Writing 
Beata Beigman Klebanov, Jill Burstein, Judith Harackiewicz, Stacy Priniski and Matthew Mulholland 

Evaluation Dataset (DT-Grade) and Word Weighting Approach towards Constructed Short Answers Assessment in Tutorial Dialogue Context 
Rajendra Banjade, Nabin Maharjan, Nobal Bikram Niraula, Dipesh Gautam, Borhan Samei and Vasile Rus

Candidate re-ranking for SMT-based grammatical error correction 
Zheng Yuan, Ted Briscoe and Mariano Felice

Augmenting Course Material with Open Access Textbooks 
Smitha Milli and Marti A. Hearst

Exploring the Intersection of Short Answer Assessment, Authorship Attribution, and Plagiarism Detection 
Björn Rudzewitz

Linguistically Aware Information Retrieval: Providing Input Enrichment for Second Language Learners 
Maria Chinkina and Detmar Meurers

Combined Tree Kernel-based classifiers for Assessing Quality of Scientific Text 
Liliana Mamani Sanchez and Hector-Hugo Franco-Penya

Combining Off-the-shelf Grammar and Spelling Tools for the Automatic Evaluation of Scientific Writing (AESW) Shared Task 2016 
René Witte and Bahar Sateli

Sentence-Level Grammatical Error Identification as Sequence-to-Sequence Correction 
Allen Schmaltz, Yoon Kim, Alexander M. Rush and Stuart Shieber

​

Break

 

Sentence Similarity Measures for Fine-Grained Estimation of Topical Relevance in Learner Essays 
Marek Rei and Ronan Cummins

 

Insights from Russian second language readability classification: complexity-dependent training requirements, and feature evaluation of multiple categories 
Robert Reynolds

​

Investigating Active Learning for Short-Answer Scoring 
Andrea Horbach and Alexis Palmer

​

Closing Remarks 

 

Post-workshop dinner 
Stone Brewing World Bistro & Gardens (Liberty Station) at 2816 Historic Decatur Road, Ste 116
directions from hotel to restaurant ] 

ORGANIZING COMMITTEE

​

BEA11 Workshop

 

  • Joel Tetreault, Yahoo Labs (primary contact)

  • Jill Burstein, Educational Testing Services

  • Claudia Leacock, Consultant

  • Helen Yannakoudakis, University of Cambridge

​

BEA Organization

 

  • Webmasters: 

    • Sowmya V.B., Iowa State University

    • Ekaterina Kochmar, University of Cambridge

  • SIG Research: 

    • Sowmya V.B., Iowa State University

    • Helen Yannakoudakis, University of Cambridge

  • Newsletter: 

    • Ildikó Pilán, University of Gothenburg

    • Ekaterina Kochmar, University of Cambridge

    • Sowmya V.B., Iowa State University

    • Helen Yannakoudakis, University of Cambridge

    • Joel Tetreault, Yahoo Labs

Organising Committee

PROGRAM COMMITTEE

​

  • Laura Allen, Arizona State University

  • Rafael Banchs, I2R

  • Timo Baumann, Universität Hamburg

  • Lee Becker, Hapara

  • Beata Beigman Klebanov, Educational Testing Service

  • Lisa Beinborn, TU Darmstadt

  • Kay Berkling, Cooperative State University Karlsruhe 

  • Suma Bhat, University of Illinois, Urbana-Champaign

  • Serge Bibauw, KU Leuven & Université Catholique de Louvain

  • David Bloom, Pacific Metrics

  • Chris Brew, Thomson Reuters

  • Ted Briscoe, University of Cambridge

  • Chris Brockett, Microsoft Research

  • Julian Brooke, University of Melbourne

  • Aoife Cahill, Educational Testing Service

  • Lei Chen, Educational Testing Service

  • Min Chi, NCSU

  • Martin Chodorow, CUNY & Educational Testing Service

  • Mark Core, University of Southern California

  • Scott Crossley, Georgia State University

  • Luis Fernando D'Haro, Human Language Technology - Institute for Infocomm Research

  • Daniel Dahlmeier, SAP

  • Barbara Di Eugenio, University of Illinois Chicago

  • Markus Dickinson, Indiana University

  • Yo Ehara, Tokyo Metropolitan University

  • Keelan Evanini, Educational Testing Service

  • Mariano Felice, University of Cambridge

  • Michael Flor, Educational Testing Service

  • Thomas François, Université Catholique de Louvain

  • Michael Gamon, Microsoft Research NeXT

  • Binyam Gebrekidan Gebre, Max Planck Computing and Data Facility

  • Kallirroi Georgila, University of Southern California

  • Dan Goldwasser, Purdue University 

  • Cyril Goutte, National Research Council Canada

  • Iryna Gurevych, UKP Lab, Technische Universität Darmstadt

  • Na-Rae Han, University of Pittsburgh

  • Andrea Horbach, Saarland University

  • Chung-Chi Huang, National Institutes of Health

  • Radu Tudor Ionescu, University of Bucharest

  • Ross Israel, Factual

  • Fazel Keshtkar, Southeast Missouri State University

  • Ekaterina Kochmar, University of Cambridge

  • Mamoru Komachi, Tokyo Metropolitan University

  • Bob Krovetz, Lexical Research

  • Lun-Wei Ku, Academia Sinica

  • Kristopher Kyle, Georgia State University

  • John Lee, City University of Hong Kong

  • Ben Leong, Educational Testing Service

  • James Lester, North Carolina State University

  • Diane Litman, University of Pittsburgh

  • Annie Louis, University of Essex

  • Anastassia Loukina, Educational Testing Service

  • Xiaofei Lu, Pennsylvania State University

  • Wencan Luo, University of Pittsburgh

  • Nitin Madnani, Educational Testing Service

  • Shervin Malmasi, Macquarie University

  • Montse Maritxalar, University of the Basque Country

  • Julie Medero, Harvey Mudd College

  • Detmar Meurers, Universität Tübingen

  • Lisa Michaud, Aspect Software

  • Rada Mihalcea, U. Michigan

  • Michael Mohler, Language Computer Corp.

  • Smaranda Muresan, Columbia University

  • Courtney Napoles, JHU

  • Hwee Tou Ng, National University of Singapore

  • Vincent Ng, UT Dallas

  • Huy Nguyen, University of Pittsburgh

  • Rodney Nielsen, University of North Texas

  • Nobal Niraula, The University of Memphis

  • Simon Ostermann, Saarland University

  • Alexis Palmer, Heidelberg University

  • Ted Pedersen, University of Minnesota, Duluth

  • Ildikó Pilán, University of Gothenburg

  • Zahra Rahimi, University of Pittsburgh

  • Lakshmi Ramachandran, Pearson

  • Arti Ramesh, University of Maryland, College Park

  • Marek Rei, University of Cambridge

  • Robert Reynolds, University of Tromsø

  • Brian Riordan, Educational Testing Service

  • Mark Rosenstein, Pearson

  • Mihai Rotaru, Textkernel

  • Alla Rozovskaya, Virginia Tech

  • C. Anton Rytting, University of Maryland College Park

  • Keisuke Sakaguchi, Johns Hopkins University

  • Mathias Schulze, University of Waterloo

  • Swapna Somasundaran, Educational Testing Service

  • Helmer Strik, Centre for Language Studies (CLS), Centre for Language and Speech Technology (CLST), Radboud University, Nijmegen

  • David Suendermann-Oeft, Educational Testing Service

  • Sowmya Vajjala, Iowa State University

  • Giulia Venturi, Institute of Computational Linguistics "Antonio Zampolli" (ILC-CNR)

  • Elena Volodina, University of Gothenburg

  • Carl Vogel, Trinity College

  • Xinhao Wang, Educational Testing Service

  • Michael White, Department of Linguistics, The Ohio State University

  • David Wible, National Central University

  • Alistair Willis, The Open University, UK

  • Magdalena Wolska, Universität Tübingen

  • Peter Wood, University of Saskatchewan

  • Huichao Xue, Google

  • Helen Yannakoudakis, University of Cambridge

  • Marcos Zampieri, Saarland University

  • Klaus Zechner, Educational Testing Service

  • Torsten Zesch, University of Duisburg-Essen

  • Fan Zhang, University of Pittsburgh

  • Xiaodan Zhu, National Research Council Canada

Program Committee
Related Links

OTHER UPCOMING EDUCATIONAL/NLP EVENTS

​

​

​

​

Other Events
bottom of page