Anaphora Resolution. João Marques

Similar documents
Anaphora Resolution. Nuno Nobre

Identifying Anaphoric and Non- Anaphoric Noun Phrases to Improve Coreference Resolution

08 Anaphora resolution

Towards a more consistent and comprehensive evaluation of anaphora resolution algorithms and systems

Reference Resolution. Regina Barzilay. February 23, 2004

Reference Resolution. Announcements. Last Time. 3/3 first part of the projects Example topics

Anaphora Resolution in Biomedical Literature: A

Hybrid Approach to Pronominal Anaphora Resolution in English Newspaper Text

Automatic Evaluation for Anaphora Resolution in SUPAR system 1

ANAPHORIC REFERENCE IN JUSTIN BIEBER S ALBUM BELIEVE ACOUSTIC

Anaphora Resolution in Hindi Language

Dialogue structure as a preference in anaphora resolution systems

Anaphora Resolution in Biomedical Literature: A Hybrid Approach

TEXT MINING TECHNIQUES RORY DUTHIE

Coreference Resolution Lecture 15: October 30, Reference Resolution

Outline of today s lecture

Performance Analysis of two Anaphora Resolution System for Hindi Language

Keywords Coreference resolution, anaphora resolution, cataphora, exaphora, annotation.

807 - TEXT ANALYTICS. Anaphora resolution: the problem

Resolving Direct and Indirect Anaphora for Japanese Definite Noun Phrases

Introduction to the Special Issue on Computational Anaphora Resolution

A Machine Learning Approach to Resolve Event Anaphora

An Introduction to Anaphora

ANAPHORA RESOLUTION IN HINDI LANGUAGE USING GAZETTEER METHOD

AliQAn, Spanish QA System at multilingual

Question Answering. CS486 / 686 University of Waterloo Lecture 23: April 1 st, CS486/686 Slides (c) 2014 P. Poupart 1

A Survey on Anaphora Resolution Toolkits

Anaphora Resolution Exercise: An overview

ZAC: Zero Anaphora Corpus A Corpus for Zero Anaphora Resolution in Portuguese

Palomar & Martnez-Barco the latter being the abbreviating form of the reference to an entity. This paper focuses exclusively on the resolution of anap

Models of Anaphora Processing and the Binding Constraints

1. Read, view, listen to, and evaluate written, visual, and oral communications. (CA 2-3, 5)

ADDIS ABABA UNIVERSITY SCHOOL OF GRADUATE STUDIES. Design of Amharic Anaphora Resolution Model. Temesgen Dawit

INFORMATION EXTRACTION AND AD HOC ANAPHORA ANALYSIS

SEVENTH GRADE RELIGION

Statistical anaphora resolution in biomedical texts

On "deep and surface. anaphora. Eunice Pontes

Anaphora Resolution. Nuno Ricardo Pedruco Nobre. Dissertação para obtenção do Grau de Mestre em Engenharia Informática e de Computadores

ANAPHORA RESOLUTION IN MACHINE TRANSLATION

Information Extraction. CS6200 Information Retrieval (and a sort of advertisement for NLP in the spring)

Anaphoric Deflationism: Truth and Reference

Artificial Intelligence Prof. Deepak Khemani Department of Computer Science and Engineering Indian Institute of Technology, Madras

HS01: The Grammar of Anaphora: The Study of Anaphora and Ellipsis An Introduction. Winkler /Konietzko WS06/07

Anaphora Resolution in Portuguese An hybrid approach

The UPV at 2007

An Analysis of Reference in J.K. Rowling s Novel: Harry Potter and the Half-Blood Prince

Houghton Mifflin English 2001 Houghton Mifflin Company Grade Three Grade Five

INTRODUCTION TO THE Holman Christian Standard Bible

Informalizing Formal Logic

Could have done otherwise, action sentences and anaphora

Houghton Mifflin Harcourt Collections 2015 Grade 8. Indiana Academic Standards English/Language Arts Grade 8

Natural Language Processing (NLP) 10/30/02 CS470/670 NLP (10/30/02) 1

This report is organized in four sections. The first section discusses the sample design. The next

Brazilian Portuguese Bare Singulars and Discourse Referents

Houghton Mifflin English 2004 Houghton Mifflin Company Level Four correlated to Tennessee Learning Expectations and Draft Performance Indicators

Russell: On Denoting

PAGE(S) WHERE TAUGHT (If submission is not text, cite appropriate resource(s))

PHILOSOPHY AND RELIGIOUS STUDIES

Visual Analytics Based Authorship Discrimination Using Gaussian Mixture Models and Self Organising Maps: Application on Quran and Hadith

An Efficient Indexing Approach to Find Quranic Symbols in Large Texts

DO YOU WANT TO WRITE:

Semantics and Pragmatics of NLP DRT: Constructing LFs and Presuppositions

Impact of Anaphora Resolution on Opinion Target Identification

Gesture recognition with Kinect. Joakim Larsson

In the name of Allah, the Beneficent and Merciful S/5/100 report 1/12/1982 [December 1, 1982] Towards a worldwide strategy for Islamic policy (Points

Distinctively Christian values are clearly expressed.

OUTSTANDING GOOD SATISFACTORY INADEQUATE

World Religions. These subject guidelines should be read in conjunction with the Introduction, Outline and Details all essays sections of this guide.

Macmillan/McGraw-Hill SCIENCE: A CLOSER LOOK 2011, Grade 1 Correlated with Common Core State Standards, Grade 1

Entailment as Plural Modal Anaphora

A Computational Model for Resolving Pronominal Anaphora in Turkish Using Hobbs Naïve Algorithm

Appendix 1. Towers Watson Report. UMC Call to Action Vital Congregations Research Project Findings Report for Steering Team

Empty Names and Two-Valued Positive Free Logic

Prioritizing Issues in Islamic Economics and Finance

Circumscribing Inconsistency

The Reliability of Anaphoric Annotation, Reconsidered: Taking Ambiguity into Account

THE EFFECT OF PULPITS IN THE RASTI VALUES WITHIN CHURCHES

Introduction to Statistical Hypothesis Testing Prof. Arun K Tangirala Department of Chemical Engineering Indian Institute of Technology, Madras

Artificial Intelligence: Valid Arguments and Proof Systems. Prof. Deepak Khemani. Department of Computer Science and Engineering

Pronominal, temporal and descriptive anaphora

Prentice Hall Literature: Timeless Voices, Timeless Themes, Bronze Level '2002 Correlated to: Oregon Language Arts Content Standards (Grade 7)

Overview of College Board Noncognitive Work Carol Barry

Artificial Intelligence. Clause Form and The Resolution Rule. Prof. Deepak Khemani. Department of Computer Science and Engineering

Bayesian Probability

Macmillan/McGraw-Hill SCIENCE: A CLOSER LOOK 2011, Grade 3 Correlated with Common Core State Standards, Grade 3

***** [KST : Knowledge Sharing Technology]

Studying Adaptive Learning Efficacy using Propensity Score Matching

Prentice Hall Literature: Timeless Voices, Timeless Themes, Silver Level '2002 Correlated to: Oregon Language Arts Content Standards (Grade 8)

TECHNICAL WORKING PARTY ON AUTOMATION AND COMPUTER PROGRAMS. Twenty-Fifth Session Sibiu, Romania, September 3 to 6, 2007

Factivity and Presuppositions David Schueler University of Minnesota, Twin Cities LSA Annual Meeting 2013

2007 HSC Notes from the Marking Centre Classical Hebrew

ON THE ROLE OF METHODOLOGY: ADVICE TO THE ADVISORS

Syllabus for GBIB 507 Biblical Hermeneutics 3 Credit Hours Spring 2015

Anaphora Resolution in Hindi: Issues and Directions

Bounded Rationality. Gerhard Riener. Department of Economics University of Mannheim. WiSe2014

Presupposition and Rules for Anaphora

Sentiment Flow! A General Model of Web Review Argumentation

KEEP THIS COPY FOR REPRODUCTION Pý:RPCS.15i )OCUMENTATION PAGE 0 ''.1-AC7..<Z C. in;2re PORT DATE JPOTTYPE AND DATES COVERID

Scott Foresman Reading Street Common Core 2013

Argument Harvesting Using Chatbots

Transcription:

Anaphora Resolution João Marques IST Instituto Superior Técnico L 2 F Spoken Language Systems Laboratory INES ID Lisboa Rua Alves Redol 9, 1000-029 Lisboa, Portugal jsmarques@l2f.inesc-id.pt Abstract This paper describes the implementation of an hybrid approach to Anaphora Resolution (AR) in Portuguese. ARM 2.0. has been incorporated in a fully-pledged Natural Language Processing system (STRING) and evaluated on a large, manually annotated corpus. 1 Introduction In a time when Natural Language Processing (NLP) draws more and more attention, the task of anaphora resolution presents itself as critical for many applications such as machine translation, information extraction and question answering. For a machine, it is difficult to select the correct entity (antecedent) to which the anaphor (mention) refers, mostly due to the ambiguous nature of natural languages. To overcome this drawback, a great amount of linguistic knowledge (morphological, lexical, syntactic, semantic, and even world knowledge) may be required. Anaphora is a major discursive device used to avoid repetition and increase the cohesion of the text, making the interpretation of sentences depend upon the interpretation of the previous ones (1.1). (1.1) Luís Figo é um ex-futebolista português. Em 2001, ele foi distinguido como melhor jogador do Mundo. Luís Figo is a former Portuguese football player. In 2001, he was distinguished as the world s best player. A human reader immediately understands that, in the second sentence, it was Luís Figo that was distinguished as the world s best player in 2001. However, this deduction actually requires that a link be established between Luís Figo in the first sentence and ele (he) in the second. Only then, can the prize mentioned in the second sentence be attributed to Luís Figo in the first. Therefore, the interpretation of the second sentence is dependent of the former ensuring in this way, the cohesion between the two sentences of this discourse. Besides contributing to the cohesion of the discourse, the two expressions are co-referential since they both refer to the same person in the real world, Luís Figo. Anaphora can also be classified according to the antecedents location:

2 intrasentential, if the antecedent is on the same sentence of the anaphor, or intersentential, if the anaphoric relation is made across sentence boundaries. In addition to the immense knowledge needed to perform anaphora resolution, the various forms that anaphora can assume make it a very challenging task, especially when one intends to teach computers how to solve anaphora. For this work, we consider pronominal anaphora. This includes personal (1.1), possessive (1.2), relative (1.3), demonstrative and reflexive and numerals, which can be used as pronominal-like anaphors. (1.2) O Pedro levou a sua mochila e eu levei a minha. Pedro took his backpack and I took mine. (1.3) O jogador, que tinha cabelo comprido, foi o melhor em campo. The player, that had long hair, was the man of the match. Note that Portuguese possessive pronouns, v.g. {meu/minha, teu/tua, seu/sua, nosso/nossa, vosso/vossa} do not agree in gender and number with their antecedent, but with the non they determine. Hence, minha, in (1.2) refers to a 1 st person singular, but its gender-number agrees with a zeroed noun mochila (backpack), which occured in the previous clause. Personal pronouns resolution usually deals only with e rd person pronouns, both singular and plural, since the first and second person refer to the dialog interlocutors in direct speech. In this paper, we will not consider dialogues, and will only cover pronominal anaphora in indirect speech. In this paper, we intend to choose an annotation framework, build an annotated Portuguese corpus and developing an hybrid approach that extended the scope and improved the performance of the previous AR module (33.5% f-measure). This paper is structured as follows: chapter 2 describes different approaches and systems that attempted to resolve the problem of anaphora; chapter 3 presents the description of the standard golden corpus developed and the process of annotating it; in chapter 4 we propose our methods to resolve the problem of pronominal anaphora resolution; chapter 5 discusses the role of evaluation as well as the different forms of assessing the efficiency of our system. The chapter then presents the results obtained using this methodology; finally, chapter 6 concludes this document, pointing to new directions of study and the further development of the AR module. 2 State of the Art AR algorithms can be broadly classed into rule-based and machine learning approaches. Initially, it was the rule-based approaches such as Hobbs s algorithm [Hobbs, 1978] and Lappin and Leass s [Lappin and Leass, 1994] resolution of anaphora procedure (RAP), which gained popularity. In the 1990s and 2000s, as people grew aware of the complexity of the job at hand, research started to

3 be limited to specific types of anaphora in view of ultimately achieving better results. Dagan and Itai s collocation pattern-based approach [Dagan and Itai, 1991]; Kennedy and Boguraev s parse-free approach [Kennedy and Boguraev, 1996]; Paraboni and Lima s research on Portuguese possessive pronominal anaphora [Paraboni and Strube-de-Lima, 1998]; Mitkov s algorithm [Mitkov, 2002] and haves and Rino s adaptation of Mitkov s algorithm for anaphora resolution in (Brazilian) Portuguese [haves and Rino, 2008]; all these approaches brought new insights about AR and new ways to approach the task. Machine learning approaches to pronoun (and, in general, to anaphora and coreference) resolution ( [Mcarthy and Lehnert, 1995]; [ardie and Wagstaff, 1999]; [Soon et al., 2001]; [Rahman and Ng, 2009]) have been an important direction of research. One of the first methods relying mostly on syntactical knowledge was Hobbs s approach [Hobbs, 1978]. His algorithm is based on various syntactic constraints on pronominalisation, which are used to traverse the syntactic tree. The search is performed in an optimal order so that the NP upon which it terminates is regarded as the probable antecedent of the pronoun at which the algorithm starts. Hobbs evaluated his algorithm in 300 pronouns from three texts with different structures and reached the success rate of 88.3% and, with the inclusion of selectional restrictions, it arrived at 91.7%. Even without any pre-processing errors, the success rate that it achieved is impressive, which consolidates Hobbs s early research as an important benchmark for the scientific community. In 1991, Dagan and Itai presented an innovative statistical approach for resolving 3 rd person pronouns based on collocation (co-occurrence) patterns [Dagan and Itai, 1991]. The candidates substituted the anaphor and the one with most frequent co-occurence patterns was preferred over the others. The experiment was conducted on 59 sentences retrieved from the Hansard corpus containing it and in the 38 sentences that the system responded, it selected the correct antecedent 33 times (87%). In 2001, Soon et al. presented a machine learning approach to noun phrases co-reference resolution in unrestricted text [Soon et al., 2001]. The algorithm used was the 5 classifier, an updated version of 4.5 [Quinlan, 1993]. Soon et al. devised a twelve-feature vector for training and evaluation. The features used were generic enough to be used across different domains and included gender, number and semantic class agreement, distance, appositive syntactic context, and string matching, among others. The system operated automatically and was evaluated on MU-6 and MU-7 corpora, reaching a 62.7% f-measure rate. In 2002, Ruslan Mitkov presented MARS [Mitkov, 2002, p. 145]: a knowledgepoor, heuristic-based, inexpensive and fast approach for pronominal anaphora resolution designed to meet NLP systems practical demands. After identifying anaphors and compiling a set of candidates from the preceding noun phrases taken from the current and the two previous sentences. The algorithm then applied gender and number filters to reduce the number of candidates. Next, it applied a set of antecedent indicators that gave each candidate positive or negative scores, according to the likelhood of their being the antecedent of a

4 given anaphor. For instance, the closer candidates; the first NP in a sentence; the candidate that has an identical collocation pattern as the anaphor; and the ones that were repeated; all these candidates were given a bonus. On the other hand, the PPs or the indefinite candidates or the farther ones received a penalty. MARS was evaluated in technical manuals achieving an overall success rate of 59.35%. In 2008, Rino and haves reported RAPM, an adaptation of Mitkov s algorithm for Brazilian Portuguese [haves and Rino, 2008]. They chose and add antecedent factors that better fit the language. The new factores consists on bonuses given to proper noun candidates, candidates that exhibit the same syntactic role as the anaphor and to the closer candidate. RAPM was evaluated in a law, literary and newswire corpora containing over 1,000 anaphoras. The system operated fully automatically and attained a success rate of 67.01%, which represents a 7.66% boost over normal-mode MARS. In 2009, Rahman and Ng reported a cluster ranking model for co-reference resolution [Rahman and Ng, 2009], which ranks preceding clusters (set of coreferent NPs), rather than candidate antecedents, for an NP to be resolved. For evaluation, 599 documents were selected from the AE 2005 data set. The cluster ranker scored 76% f-measure in true mentions (manually corrected) and 69.3% when the mentions were extracted automatically and, therefore, had an error associated. In 2011, Nobre implemented ARM1.0, an adaptation of the Mitkov s algorithm for resolving Portuguese pronominal anaphora. He achieved a 33.5% f-measure, a value our system aims to improve. 3 orpus To develop a machine learning approach to anaphora resolution, we needed to build a corpus annotated with anaphoric relations, both to supply the training instances to the system, and to serve as a golden standard for the system s evaluation. The dataset used to train and evaluate our system is a fragment of the European Portuguese LE-PAROLE corpus [do Nascimento et al., 1998]. The corpus is quite heterogeneous, being composed of texts from different genres: novels, pieces of news, magazine news and newspaper columns, among others. In total, it contains 290,000 words. The corpus was automatically POS-annotated by STRING [Mamede et al., 2012] and manually corrected. The annotation campaign identified 9,268 anaphoric relations (94.3%) and 560 cataphoras (5.7%). The breakdown of the anaphoras by anaphor type is shown in Table 1: The type of anaphor was identified based on the NLP chain STRING output [Mamede et al., 2012]. This comprises an error margin that is associated with the annotation errors, as some anaphors were identified with a unexpected POS type (such as preposition a, for instance). There were 7,001 anaphoras (75.5%) with the antecedent in the same sentence as the anaphor, while for 2,267 the antecedent is further distant from the anaphor (24.5%). From these, for 1,028

5 Type of anaphor Number of anaphoras Percentage Relative 3,663 39.52% Personal 3,470 37.44% Pronouns Possessive 689 7.43% Indefinite 607 6.55% Demonstrative 188 2.03% Total 8,617 92.97% Articles 338 3.65% Numerals 74 0.80% TOTAL 9,268 100% Table 1. orpus anaphoras composition. anaphors the antecedent was in the previous sentence, for 364 there was a sentence in between, and 223 with two sentences between them. This hints that the majority of anaphoras do not surpass the three-sentence distance window between anaphor and antecedent. The annotated anaphora in which the anaphor is farther from the antecedent reports a distance of 146 sentences between them, which is extremely rare and only happened once. This is extremely rare as only 7 times the number of sentences between anaphor and antecedent surpasses the 50-sentence mark. A rapid analysis of Table 1 confirms that pronouns are the most representative category of anaphors, particularly the personal (37.4%) and relative pronouns (39.5%). 3.1 Annotation Process To perform the annotation, we needed an adequate annotation framework and found it in Glozz [Widlöcher and Mathet, 2012]. Glozz is a free Java-based annotation platform, developed by Antoine Widlöcher and Yann Mathet. It provides a friendly interface, with the possibility of annotating different types of annotation units (e.g. subject relations and anaphoric relations), and coloring annotation units for an easy visualization of the different annotation targets; it also allows to save annotations in XML files and provides hiding options. Attending to the time-consuming nature of the process of annotating corpora, and the need for annotation consistency and reproducible results, we considered that this should not be a one-person task. Thus, it was necessary to define a set of of annotation directives to guarantee the consistency of the whole process. In other words, it was necessary to make sure that each and every annotator performs this task in the same way. These guidelines are provided in [Marques et al., 2013]. To improve the consistency of the process, specific guidelines were devised in order to clearly state the general principles governing the annotation campaign (and to be renewed/reviewed if necessary). Though one cannot state here all the guidelines, one can already state some basic annotation schemata. Thus, we

6 define that zero anaphora should not be annotated at this time. In the case of coordinated antecedents, an anaphoric relation should link the anaphor to each of the antecedents that compose the coordinated antecedent. Furthermore, when two (or more) antecedents refer to the same entity, the one closest to should be preferred over the others. The annotation process was carried out by 5 annotators with expertise in Portuguese Linguistics and NLP. In order to calculate the inter-annotator agreement, we partitioned the corpus into 5+1 parts. Each annotator took the task of annotating one of the parts, but before that, all annotators worked on the same part to calculate inter-annotator agreement. For this, we used an adaptation of the Fleiss kappa coefficient (k) [Fleiss, 1971]. Since Fleiss kappa coefficient required the hypothetical probability of chance agreement (using the observed data to calculate the probabilities of each observer randomly attributing each category rate), and taking into account the specificity of anaphora annotation, particularly the fact that there is no fixed number of categories since the number of candidates vary in each case, it is not possible to calculate k in the same way. Therefore, the general formula of k was adapted as follows: let N be the total number of anaphors, let n be the number of annotators, and let c be the number of candidates for each anaphor. The anaphors are indexed by i = 1,...N and the candidates are indexed by j = 1,...c + 1, where c + 1 represents the case where an anaphor has not been annotated. Let n ij represent the number of raters who assigned the i th anaphor to the j th candidate. The k calculation will thus take the form of equation 1, k = P r(a) = 1 N N 1 N P i = Nn(n 1) ( c+1 i=1 i=1 j=1 n 2 ij Nn) (1) where P i is the extent to which raters agree for the i th anaphor (i.e., compute how many rater-rater pairs are in agreement, relative to the number of all possible rater-rater pairs). After two rounds of annotation in a part of the corpus and the manual group correction, four annotators attained an accuracy greater than 81% which translated in a k of 78.7%, which can be considered reliable. The remaining annotator managed a sub-par accuracy, and was excluded from the rest of the campaign. The annotation directives were also reviewed and updated along the process, in order to clarify some more complex/difficult cases. 4 Architecture We divide the problem of anaphora resolution in three stages: Identification of anaphors; ompilation of the list of candidates; hoice of the most probable candidate;

7 The two first stages were implemented through manual rules, while the latter was based on a learning model that orders the candidates by the probability of their being the anaphor s antecedent. The most probable candidate is then selected as the antecedent of the anaphor. The annotated corpus (see section 3) served as a golden standard for the system s evaluation. 4.1 Anaphor Identification Articles that constitute a single node, that is, articles that are not incorporated in NPs or PPs (1.4); (1.4) Duas universidades: a de Lisboa e a do Porto. Two universities: the one from Lisbon and the one from Porto. Nodes named REL in STRING are also retrieved, as they represent relative pronouns; Pronouns incorporated on a NP or PP that do not violate any of the following rules: Pronouns cannot be 1 st or 2 nd person. 1 st or 2 nd person pronouns refer to the participants in a dialog, and are not addressed in this dissertation; Pronouns cannot be in an attributive (or predicative) position, that is, the pronouns cannot be preceded by the Portuguese verb ser (in English, the to be verb); In a coordination, only if the pronoun is not a demonstrative nor a possessive. This rule excludes coordinated determiners in a NP or PP such as (1.5), so that the pronoun is not considered an anaphor; (1.5) Estas e outras coisas são perigosas. These and other things are dangerous. litic pronoun se attached to a verb with PASS-PRON feature, corresponding to the pronominal passive-like construction, are discarded as they are being used in a expletive way (1.6). (1.6) Dizia-se que era uma decisão irrevogável. It was said to be an irrevocable decision. Also, we compiled a list of pronouns that are traditionally not used in a non-anaphoric manner and, therefore, are automatically excluded as anaphors. This list contains the tokens {toda a gente (everybody), mesmo (the same), o tal (such), um certo (certain), próprio (self), o porquê (because), isto (this), isso (that), aquilo (that), tudo (everything), nenhum (none), nada (nothing), alguém (someone), ninguém (no one/nobody) and algo (something)}. It also includes the locative adverbs with anaphoric value cá (here), l a and ali (there); as well as indefinite pronouns algures (somewhere) and nenhures (nowhere), with locative meaning.

8 4.2 andidate Identification Like the anaphor identification stage, the candidate identification is also made throughout the parsing of the text. Nouns that are heads of NPs and PPs are identified as potential candidates. When STRING identifies that two or more nouns are present in a coordination, they also constitute a coordinate candidate (1.7). (1.7) O João e o Pedro foram a casa da Rita. João and Pedro went to Rita s home. Besides, if a pronoun is (left-side) closer to a relative pronoun anaphor than any other candidate, it is also identified as a candidate for that anaphor to account for cases such as (1.8), where the indefinite aquilo (that) is to be considered the antecedent of the relative que (what): (1.8) Foi aquilo que nos levou a agir assim. It was that what made us act like that. At last, the span of text from which the candidates are to be retrieved is limited to a two sentence window only the candidates that are on the same sentence at the left of the anaphor, or in the previous two sentences, are selected. Exception is made to the relative pronoun anaphors, whose candidates must be selected from the same sentence and at the anaphor s left side 1. 4.3 Selection of the best candidate The ordering of the candidate list (and the choice of the most probable one) is based on the model generated through the application of a machine learning method applied to the corpus we annotated. To do this, we used the WEKA software 2 [Witten et al., 2005]. Our system identifies the anaphors and candidates for each anaphor, and creates an instance for each pair anaphor-candidate with several features displayed in Table 2 (page 14). As we implemented a supervised learning (based on the annotation), each instance contains the target feature (T) is antecedent that could be either true if the candidate is the antecedent for the anaphor, or false otherwise. The remaining features values are retrieved from the STRING output. The features are grouped in three types: anaphor-related features (A), candidate-related features () and features related to the relationship between anaphor and the candidates (R). 1 The process of annotation presented strong evidences that the relative pronoun anaphor s antecedent is almost every time in the same sentence as the anaphor, and often immediately at its left side. 2 WEKA(Waikato Environment for Knowledge Analysis) is a popular suite of machine learning software written in Java, developed at the University of Waikato, New Zealand.

9 A machine learning method adequate to our task had to be defined. We chose the Expectation-Maximization algorithm (EM) [Dempster et al., 1977]. EM is a soft-clustering convergence method, which means, in our case, that it provides the probabilities of an instance (pair anaphor-candidate) to belong to a cluster. Running EM in two clusters (one representing that the candidate is the antecedent for that anaphor, and the other representing that it is not), we are able to get the probabilities of each candidate to be the antecedent and therefore we are able to choose the best one. 5 Evaluation It is important to analyze AR along each stage, since each step s efficiency constitutes a ceiling to the performance of the next phase. In other words, if the anaphor is not successfully found or if the antecedent is not present in the candidate list, the anaphora will not be resolved however good the model may be. Figure 1 allows us to compare the ceilings that are being carried from the anaphor identification stage to the candidates identification and, in turn, to anaphora resolution. Note that this figure only assesses recall as it does not consider the misidentified anaphors. oncerning anaphor identification, Relative and reflexive se pronouns stand out as the best results, due to the usually greater proximity between anaphor and its antecedent in this cases (3.38 and 20.11 words between anaphor and antecedent, respectively). Possessive pronouns also achieve a very interesting 60.8% recall. Nonetheless, it is clear that the ceiling from the previous stage of candidates identification influence AR results for all anaphors. On the other hand, personal pronouns other than the reflexive se show significant lower recall (44.3%). This is explained by the lower ceiling inherited from the previous steps of processing, but also by the larger number of candidates, resulting from a wider search space in the candidate selection (two sentences window). ARM 2.0. applies gender and number filters to the candidates (those that do not agree in gender and number with the anaphor are discarded). Evaluation showed that the application of filters improved the results, which is explained by the lower number of candidates the model has to choose from. Table 3 (page 15) displays the results at each processing stage. The application of gender and number filters propel ARM 2.0. to overall results above 50% on precision and recall, bordering the 60% mark. It is clear that the precision is typically lower than the recall, a fact that can be explained by the decision of not annotating co-referent anaphoras (the manual rules that were developed cannot discern co-referent anaphoras from identity-ofsense anaphoras; in this way, ARM 2.0. considers what is in reality an identityof-sense anaphor, that was not resolved in annotation, as an anaphor, evaluated as incorrectly identified, thus decreasing the preicision). ataphora events also help to lower the recall, as potential anaphors are, in reality, cataphors: from the 560 cataphoras present in the corpus, 264 cataphors were incorrectly considered as potential anaphors, which represents 2.3% of all the anaphors identified by

10 Fig. 1. Recall performance of the different stages of ARM2.0 for each type of anaphor. ARM 2.0. Special cases of annotation or XIP errors are also among the reasons that prevent a higher precision and recall. The Recall column under andidate identification presents the ceiling for the machine learning model. Our program identified the antecedent 84.7% of the times, while in 9.2% of the remaining cases, the program could not found it due to the antecedent being out-of-range. As expected, the large majority of the relative pronouns antecedents are on the same sentence as the anaphor. The candidate list compilation causes a fall in the ceiling of personal pronouns as they are the type of anaphors whose antecedents can be farther from its anaphor, generating out-of-range absence of the antecedent in the candidate list. Finally, we take a look at the model efficiency, that is, how does the model perform when it has all the conditions necessary to resolve the anaphora. Figure 2 compares the model efficiency with critical success rate, i.e., the efficiency of the model when it discards all the anaphoras that can be resolved in a trivial way, namely, when there is only a single candidate antecedent for the anaphor or all other candidates but one are excluded on the basis of gender and number agreement 3. Relative pronouns are the type of anaphor that take most advantage of these cases (682 cases, 26.58%), since the one sentence window applied in these type of anaphors promote single candidate anaphoras, hence registering the major drop-off when discarding the gender-number agreement and single 3 This measure has been proposed by Mitkov [Mitkov, 2002].

11 candidate solvable anaphoras. A little portion of personal pronouns, excluding se, are also resolved under these terms (76 cases, 9.57%), which is natural if we consider that only the accusative and nominative 3 rd person are marked for gender and number. On the other hand, se pronouns rarely are resolved on the basis of a single candidate or gender and number agreement. This can be explained by the fact that this type of anaphor compiles a list of candidates, whose range reports a two sentence window, minimizing the single candidate scenario. onsidering that se pronouns are also not marked for gender and number, it is natural the little impact of critical success rate in this type of anaphor (10 cases, 0.38%). The remaining types of anaphors are only very rarely resolved under these conditions, the possessive pronouns are not even submitted to gender and number filters (each of the remaining types of anaphora registered under 10 gender-number or single-candidate solvable anaphoras). Fig. 2. Performance of the ARM2.0 AR model for each type of anaphor. The model efficiency is relatively good ranging between 64.6% in personal pronouns (excluding se) and 86.7% in demonstrative pronouns resolution. We consider an overall efficiency of 77.8% a very solid value. Even when considering only tougher anaphoras, ARM 2.0. AR model attains a 72.7%, which continues to be a reliable rating. In face of the results reported by most of the aforementioned systems (section 2), it could be posited that ARM 2.0. still has a significant room for improvement. However, it is important to notice that these systems use manually corrected input data, limited textual diversity of relatively small number of anaphora instances in their evaluation. This contrasts with the approach adopted in this paper, which aims at getting raw texts and resolving its anaphors in an entirely automatically way, something that is much closer to a real scenario of a NLP system in use. On the other hand, the diverse textual genres included

12 in the evaluation corpus and the sheer number of anaphora instances manually annotated, along with the process of annotation itself led us to believe that these results may reflect better the difficulty of the task in a realistic scenario. Nonetheless, ARM 2.0 represents a step forward as it improved ARM 1.0 not only in performance but in resolving a more extensive scope of anaphoras and evaluating them in an extensive and unprecedentedly large Portuguese annotated corpus. 6 onclusions and Future Work The results are deemed as satisfactory, as they met the goals of choosing an annotation framework, building an annotated Portuguese corpus and developing an hybrid approach that extended the scope and improved the performance of the previous AR module. The annotation of an over 9,000 anaphoras on a 290,000 tokens corpus adds value to this work and significance to the results achieved. The gap between ARM 2.0 results and the ones reported by some of the systems presented in section 2, even taking into account their different scope, the different corpora they used, and the fact that their input was previously corrected, shows that there is still room for improvement. In future work, it would be interesting to provide the corpus with a wider range of anaphoric relations, such as co-reference, metonymy, subset/superset relations, zero and identity-of-sense anaphora. This could help to better assess the specific problems posed by each type of anaphora and, ultimately, to devise better strategies to resolve it. The introduction of new knowledge sources, namely at semantic and pragmatic level, and the exploration of collocation patterns [Dagan and Itai, 1991] could also enrich the model and, extensively, the AR task. References [ardie and Wagstaff, 1999] ardie,. and Wagstaff, K. (1999). Noun Phrase oreference as lustering. In Proceedings of the 1999 Joint SIGDAT onference on Empirical Methods in Natural Language Processing and Very Large orpora, EMNLP/VL 99, pages 82 89, ollege Park, Maryland, USA. [haves and Rino, 2008] haves, A. R. and Rino, L. H. (2008). The Mitkov Algorithm for Anaphora Resolution in Portuguese. In Proceedings of the 8 th International onference on omputational Processing of the Portuguese Language, PROPOR 08, pages 51 60, Aveiro, Portugal. Springer-Verlag. [Dagan and Itai, 1991] Dagan, I. and Itai, A. (1991). A Statistical Filter for Resolving Pronoun References. In Feldman, Y. A. and Bruckstein, A., editors, Artificial Intelligence and omputer Vision, pages 125 135. Elsevier Science Publishers B.V. [Dempster et al., 1977] Dempster, A., Laird, N., and Rubin, D. (1977). Maximum Likelihood from Incomplete Data via the em Algorithm. Journal of the Royal Statistical Society, 39(1):1 38. [do Nascimento et al., 1998] do Nascimento, M., Veloso, R., Marrafa, P., Pereira, L., Ribeiro, R., and Wittmann, L. (1998). LE-PAROLE: do orpus à Modelização da

13 Informação Lexical num Sistema Multifunção. Actas do XIII Encontro Nacional da Associação Portuguesa de Linguística, 2:115 134. [Fleiss, 1971] Fleiss, J. (1971). Measuring Nominal Scale Agreement among many Raters. Psychological Bulletin, 76(5):378 382. [Hobbs, 1978] Hobbs, J. R. (1978). Resolving Pronoun References. Lingua, 44:311 338. [Kennedy and Boguraev, 1996] Kennedy,. and Boguraev, B. (1996). Anaphora for Everyone: Pronominal Anaphora Resolution without a Parser. In Proceedings of the 16 th International onference on omputational Linguistics, OLING 96, pages 113 118, openhagen, Denmark. John Wiley and Sons, Ltd. [Lappin and Leass, 1994] Lappin, S. and Leass, H. J. (1994). An Algorithm for Pronominal Anaphora Resolution. omputational Linguistics, 20(4):535 561. [Mamede et al., 2012] Mamede, N., Baptista, J., Diniz,., and abarrão, V. (2012). STRING: an Hybrid Statistical and Rule-based Natural Language Processing hain for Portuguese. PROPOR 12 (Demo Session), oimbra, Portugal. https://string. l2f.inesc-id.pt//. [Marques et al., 2013] Marques, J., Baptista, J., and Mamede, N. (2013). Anaphora Annotation Guidelines. Technical report, INES-ID, Lisboa. [Mcarthy and Lehnert, 1995] Mcarthy, J. F. and Lehnert, W. G. (1995). Using Decision Trees for oreference Resolution. In Proceedings of the 8 th International Joint onference on Artificial Intelligence, IJAI 95, pages 1050 1055, Montreal, Québec, anada. Morgan Kaufmann Publishers Inc. [Mitkov, 2002] Mitkov, R. (2002). Anaphora Resolution. Pearson. [Paraboni and Strube-de-Lima, 1998] Paraboni, I. and Strube-de-Lima, V. L. (1998). Possessive Pronominal Anaphor Resolution in Portuguese Written Texts. In Proceedings of the 17 th International onference on omputational Linguistics, OLING 98, pages 1010 1014, Montreal, Québec, anada. Association for omputational Linguistics. [Quinlan, 1993] Quinlan, J. R. (1993). 4.5: Programs for Machine Learning. Morgan Kaufmann Publishers Inc. [Rahman and Ng, 2009] Rahman, A. and Ng, V. (2009). Supervised Models for oreference Resolution. In Proceedings of Empirical Methods in Natural Language Processing, EMNLP 09, pages 968 977, Singapore. Association for omputational Linguistics. [Soon et al., 2001] Soon, W. M., Ng, H. T., and Lim, D.. Y. (2001). A Machine Learning Approach to oreference Resolution of Noun Phrases. omputational Linguistics, 27(4):521 544. [Widlöcher and Mathet, 2012] Widlöcher, A. and Mathet, Y. (2012). The Glozz Platform: a orpus Annotation and Mining Tool. In Proceedings of the 2012 Association for omputational Liguistics Symposium on Document Engineering, DocEng 12, pages 171 180, Paris, France. Telecom ParisTech, Association for omputational Liguistics. [Witten et al., 2005] Witten, I., Frank, E., and Hall, M. (2005). Data Mining: Practical Machine Learning Tools and Techniques. Morgan Kaufmann Publishers Inc., San Francisco, USA, Second edition.

14 Type Feature Description Possible values A A complement is anaphor indirect complement is anaphor subject R distance number of sentences between anaphor and candidate numeric R same sentence verifies if the anaphor and candidate are in the same sentence R gender agreement verifies if anaphor and candidate agree in gender R number agreement verifies if anaphor and candidate agree in number verifies if the anaphor and R share relation candidate have a relation with verb with the same verb (e.g. subject, direct complement) A anaphor gender gender of the anaphor {MAS(uline), FEM(inine), IND(efinite)} A anaphor number number of the anaphor {SG (singular), PL(ural), IND(efinite)} A anaphor type type of the anaphor {PRON(oun), ART(icle), NUM(eral)} A {PERS(onal), POSS(essive), anaphor pronoun type of the pronoun DEM(onstrative), IND(efinite), NULL type (anaphor is not a pronoun)} A is anaphor clitic verifies if the anaphor is a clitic A is anaphor direct verifies if the anaphor is a di- rect complement verifies if the anaphor is an indirect complement verifies if the anaphor is a subject candidate gender gender of the candidate {MAS(uline), FEM(inine), IND(efinite)} candidate number number of the candidate {SG (singular), PL(ural), IND(efinite)} is candidate a verifies if the candidate has location is candidate an organization is candidate composed is candidate demonstrative a location feature verifies if the candidate has an organization feature verifies if the candidate comprehends more then one entity verifies if the candidate is preceded by a demonstrative pronoun verifies if the candidate has is candidate human a human feature is candidate a verifies if the candidate is a proper noun proper noun is candidate verifies if the candidate is indefinite indefinite is candidate a verifies if the candidate has location a location feature is candidate NE verifies if the candidate is a named entity is candidate direct verifies if the candidate is a complement direct complement is candidate indirect verifies if the candidate is an complement indirect complement verifies if the candidate is a is candidate subject subject is candidate NP verifies if the candidate is or PP NP or PP order of the candidate; 1 if order of candidate is the closest candidate (regarding the anaphor), 2 if is the second closest, and so on T number of candidates is antecedent number of candidates for the anaphor verifies if the candidate is the antecedent for the anaphor Table 2. Features used in ARM2.0. {true, false, null} {numeric} {numeric}

15 Type Anaphor ident. andidates ident. Anaphora resolution of anaphor R P F R P F R P F se 98.8% 61.7% 76.0% 82.5% 51.6% 63.5% 67.0% 41.9% 51.5% Personal All exc. se 92.7% 90.5% 91.6% 63.9% 62.4% 63.1% 44.3% 43.3% 43.8% Pronouns All 95.6% 74.1% 83.5% 72.5% 56.2% 63.3% 54.7% 42.5% 47.8% Relative Pronouns que 84.6% 81.6% 83.1% 80.2% 77.5% 78.8% 64.2% 62.0% 63.1% onde 90.7% 91.4% 91.0% 87.4% 88.1% 87.8% 63.4% 63.9% 63.7% All 84.0% 82.6% 83.3% 78.8% 77.5% 78.1% 62.6% 61.6% 62.1% Possessive pronouns 89.0% 95.9% 92.3% 79.6% 85.7% 82.5% 60.8% 65.5% 63.1% Dem. pronouns 84.8% 47.7% 61.0% 62.8% 35.3% 45.2% 54.5% 30.6% 39.2% Articles 95.2% 23.1% 37.2% 61.1% 15.0% 24.2% 53.2% 12.9% 20.8% TOTAL 89.3% 76.2% 82.2% 75.8% 64.7% 69.8% 59.0% 50.3% 54.3% Table 3. Precision, recall and f-measure of all AR stages of the final ARM 2.0. model in the entire corpus (including gender-number filters).