Here Are the Best Natural Language Processing Papers From ACL 2022

Best Research Papers in Natural Language Processing

Since its inception in 1962, the Association for Computational Linguistics (ACL) has been “the premier international scientific and professional society for people working on computational problems involving human language, a field often referred to as either computational linguistics (CL) or natural language processing (NLP).”

ACL activities include the annual summer meeting and publication (via MIT Press) of the Computational Linguistics journal. The 60th Annual Meeting of the ACL took place from May 22–27, 2022 as a hybrid event, in Dublin and online.

These were the Top 10 themes in this year’s conference.

  • Machine Learning for NLP
  • Machine Translation and “Multilinguality” (Multilingualism)
  • Interpretability and Analysis of Models for NLP
  • Information Extraction
  • NLP Applications
  • Dialogue and Interactive Systems
  • Resources and Evaluation 
  • Semantics
  • Question Answering
  • Generation

Summarization, low-resource languages, speech technologies, multimodality, and ethics were also covered, albeit to a lesser extent.

And the Award for Best Paper Goes to…

More than 2,000 authors from all over the world contributed to the conference, either via long or short paper (604 long papers, 98 short papers).

Best Paper Award for 2022
Learned Incremental Representations for Parsing

Nikita Kitaev, Ph.D. student; Thomas Lu, undergraduate student, UC Berkeley; Dan Klein, Technical Fellow, Microsoft Semantic Machines and Professor, UC Berkeley

The authors of this paper designed a representation for parsing that is “maximally speculation free.” As they noted, “human capabilities suggest that we should also be able to build accurate parsers that […] operate incrementally.”

Best Special Theme Paper Award
Requirements and Motivations of Low-Resource Speech Synthesis for Language Revitalization

Aidan Pine and Patrick William Littell, National Research Council Canada; Dan Wells, PhD student and Korin Richmond, Associate Professor, University of Edinburgh; Nathan Brinklow, Professor, Queen’s University

The authors aimed to revitalize three indigenous languages by developing speech synthesis systems for them, reevaluating the question of “how much data is required to build low-resource speech synthesis systems featuring state-of-the-art neural models.”

Best Resource Paper Award

DiBiMT: A Novel Benchmark for Measuring Word Sense Disambiguation Biases in Machine Translation
Niccolò Campolungo, Federico Martelli, and Roberto Navigli, Sapienza University of Rome; Francesco Saina, SSML Carlo Bo

This paper highlights the importance of bias in machine translation (MT). DiBiMT, is a “novel benchmark for measuring and understanding semantic biases in NMT, which goes beyond simple accuracy and provides novel metrics that summarize how biased NMT models are.”

Best Linguistic Insight Paper Award

KinyaBERT: a Morphology-aware Kinyarwanda Language Model

Antoine Nzeyimana, University of Massachusetts Amherst; Andre Niyongabo Rubungo, Polytechnic University of Catalonia

This paper demonstrated the effectiveness of explicitly incorporating morphological information in language-model pretraining. The authors proposed a two-tier BERT architecture (the first tier encodes morphological information; the second one, sentence-level information) and evaluated it on the low-resource, morphologically rich Kinyarwanda language. The authors said this work “should motivate more research into morphology-aware language models.”

Outstanding Papers

The richness and quality of the conference papers are reflected in the Outstanding Papers​​ list as judged by the Best Paper Committee.

The 60-60 Diversity and Inclusion Initiative

To celebrate its 60th year, ACL launched the 60-60 Initiative as an effort to “remove the ingrained linguistic bias in the scientific landscape in general and CL science in particular.”

The inauguration of this Diversity & Inclusion Special Initiative has already been embraced by a core group comprising academic teams from across the globe (the National University of Singapore, Yale University, the University of Illinois Urbana Champaign, NYUAD, and King Saud University), big tech (Baidu, Meta), medium-tech (AppTek) companies, non-profit organizations (AI2), startups (aiXplain), as well as annotation companies (YaiGlobal).

The 60-60 Initiative features some strategic milestones to be reached by 2023, including

  • the complete translation of the entire ACL Anthology into 60 languages;
  • a comprehensive standardized scientific and CL terminology list in 60 languages;
  • live cross-lingual closed captions and dubbing into 60 languages;
  • a comprehensive repository for all the talks and videos from the CL community curated and translated into 60 languages.

Volunteers are encouraged to join this initiative, which aims to democratize CL sciences and maximize CL global reach.

Also read: Product Labeling Requirements: What You Need To Know