Article

Language Resources and Evaluation

, Volume 47, Issue 3, pp 661-694

Coreference resolution: an empirical study based on SemEval-2010 shared Task 1

  • Lluís MàrquezAffiliated withDepartament de Llenguatges i Sistemes Informàtics, TALP Research Center, Universitat Politècnica de Catalunya Email author 
  • , Marta RecasensAffiliated withDepartament de Lingüística, CLiC Research Center, Universitat de Barcelona
  • , Emili SapenaAffiliated withDepartament de Llenguatges i Sistemes Informàtics, TALP Research Center, Universitat Politècnica de Catalunya

Rent the article at a discount

Rent now

* Final gross prices may vary according to local VAT.

Get Access

Abstract

This paper presents an empirical evaluation of coreference resolution that covers several interrelated dimensions. The main goal is to complete the comparative analysis from the SemEval-2010 task on Coreference Resolution in Multiple Languages. To do so, the study restricts the number of languages and systems involved, but extends and deepens the analysis of the system outputs, including a more qualitative discussion. The paper compares three automatic coreference resolution systems for three languages (English, Catalan and Spanish) in four evaluation settings, and using four evaluation measures. Given that our main goal is not to provide a comparison between resolution algorithms, these are merely used as tools to shed light on the different conditions under which coreference resolution is evaluated. Although the dimensions are strongly interdependent, making it very difficult to extract general principles, the study reveals a series of interesting issues in relation to coreference resolution: the portability of systems across languages, the influence of the type and quality of input annotations, and the behavior of the scoring measures.

Keywords

Coreference resolution and evaluation NLP system analysis Machine learning based NLP tools SemEval-2010 (Task 1) Discourse entities