Tasnim Mohiuddin, Prathyusha Jwalapuram, Xiang Lin, Shafiq Joty
We conduct experiments on benchmarking well-known traditional and neural coherence models on synthetic sentence ordering tasks, and contrast this with their performance on three downstream applications: coherence evaluation for MT and summarization, and next utterance prediction in retrieval-based dialog.
Prathyusha Jwalapuram, Shafiq Joty, Youlin Shen
Through a combination of targeted fine-tuning objectives and intuitive re-use of the training data the model has failed to adequately learn from, we improve the model performance of both a sentence-level and a contextual model without using any additional data. We target the improvement of pronoun translations through our fine-tuning and evaluate our models on a pronoun benchmark testset.
[EMNLP 2019] Evaluating Pronominal Anaphora in Machine Translation: An Evaluation Measure and a Test Suite
Prathyusha Jwalapuram, Shafiq Joty, Irina Temnikova, Preslav Nakov
We contribute an extensive, targeted dataset that can be used as a test suite for pronoun translation, covering multiple source languages and different pronoun errors drawn from real system translations, for English. We further propose an evaluation measure to differentiate good and bad pronoun translations.
Xiang Lin, Shafiq Joty, Prathyusha Jwalapuram, M Saiful Bari
We propose an efficient neural framework for sentence-level discourse analysis in accordance with Rhetorical Structure Theory (RST).