Abstracts
-
Evaluating AMR-to-English NLG Evaluation
Emma Manning, Shira Wein and Nathan Schneider (full paper link) -
Informative Manual Evaluation of Machine Translation Output
Maja Popović (full paper link) -
Automatic Machine Translation Evaluation in Many Languages via Zero-Shot Paraphrasing
Brian Thompson and Matt Post (full paper link) -
Evaluating Semantic Accuracy of Data-to-Text Generation with Natural Language Inference
Ondřej Dušek and Zdeněk Kasner (full paper link) -
Studying the Effects of Cognitive Biases in Evaluation of Conversational Agents
Sashank Santhanam and Samira Shaikh (full paper link)
Archival
-
A proof of concept on triangular test evaluation for Natural Language Generation
Javier González Corbelle, José María Alonso Moral and Alberto Bugarín Diz -
This is a Problem, Don’t You Agree? Framing and Bias in Human Evaluation for Natural Language Generation
Stephanie Schoch, Diyi Yang and Yangfeng Ji -
NUBIA: NeUral Based Interchangeability Assessor for Text Generation
Hassan Kane, Muhammed Yusuf Kocyigit, Ali Abdalla, Pelkins Ajanoh and Mohamed Coulibali -
On the interaction of automatic evaluation and task framing in headline style transfer
Lorenzo De Mattei, Michele Cafagna, Huiyuan Lai, Felice Dell’Orletta, Malvina Nissim and Albert Gatt -
Evaluation rules! On the use of grammars and rule-based systems for NLG evaluation
Emiel van Miltenburg, Chris van der Lee, Thiago Castro-Ferreira and Emiel Krahmer