Outline of today s lecture

Similar documents
Coreference Resolution Lecture 15: October 30, Reference Resolution

Identifying Anaphoric and Non- Anaphoric Noun Phrases to Improve Coreference Resolution

Reference Resolution. Regina Barzilay. February 23, 2004

Reference Resolution. Announcements. Last Time. 3/3 first part of the projects Example topics

Anaphora Resolution in Hindi Language

Anaphora Resolution in Biomedical Literature: A

08 Anaphora resolution

ANAPHORIC REFERENCE IN JUSTIN BIEBER S ALBUM BELIEVE ACOUSTIC

TEXT MINING TECHNIQUES RORY DUTHIE

A Machine Learning Approach to Resolve Event Anaphora

Resolving Direct and Indirect Anaphora for Japanese Definite Noun Phrases

Dialogue structure as a preference in anaphora resolution systems

807 - TEXT ANALYTICS. Anaphora resolution: the problem

Towards a more consistent and comprehensive evaluation of anaphora resolution algorithms and systems

HS01: The Grammar of Anaphora: The Study of Anaphora and Ellipsis An Introduction. Winkler /Konietzko WS06/07

Anaphora Resolution. Nuno Nobre

Keywords Coreference resolution, anaphora resolution, cataphora, exaphora, annotation.

Houghton Mifflin English 2001 Houghton Mifflin Company Grade Three Grade Five

Semantics and Pragmatics of NLP DRT: Constructing LFs and Presuppositions

SEVENTH GRADE RELIGION

Anaphora Resolution in Biomedical Literature: A Hybrid Approach

Statistical anaphora resolution in biomedical texts

10. Presuppositions Introduction The Phenomenon Tests for presuppositions

Automatic Evaluation for Anaphora Resolution in SUPAR system 1

Presupposition and Rules for Anaphora

Version 1.0. General Certificate of Education June Religious Studies Religion and Contemporary Society AS Unit H. Final.

Discourse Constraints on Anaphora Ling 614 / Phil 615 Sponsored by the Marshall M. Weinberg Fund for Graduate Seminars in Cognitive Science

1. Introduction Formal deductive logic Overview

Anaphoric Deflationism: Truth and Reference

A Survey on Anaphora Resolution Toolkits

Lecture 2.1 INTRO TO LOGIC/ ARGUMENTS. Recognize an argument when you see one (in media, articles, people s claims).

Macmillan/McGraw-Hill SCIENCE: A CLOSER LOOK 2011, Grade 4 Correlated with Common Core State Standards, Grade 4

Hybrid Approach to Pronominal Anaphora Resolution in English Newspaper Text

An Introduction to Anaphora

Palomar & Martnez-Barco the latter being the abbreviating form of the reference to an entity. This paper focuses exclusively on the resolution of anap

Question Answering. CS486 / 686 University of Waterloo Lecture 23: April 1 st, CS486/686 Slides (c) 2014 P. Poupart 1

Pragmatic Presupposition

CS224W Project Proposal: Characterizing and Predicting Dogmatic Networks

ANAPHORA RESOLUTION IN HINDI LANGUAGE USING GAZETTEER METHOD

Coordination Problems

Prentice Hall Literature: Timeless Voices, Timeless Themes, Silver Level '2002 Correlated to: Oregon Language Arts Content Standards (Grade 8)

Anaphora Resolution Exercise: An overview

Prentice Hall Literature: Timeless Voices, Timeless Themes, Bronze Level '2002 Correlated to: Oregon Language Arts Content Standards (Grade 7)

Natural Language Processing

NICHOLAS J.J. SMITH. Let s begin with the storage hypothesis, which is introduced as follows: 1

An Analysis of Reference in J.K. Rowling s Novel: Harry Potter and the Half-Blood Prince

Diving In: Getting the Most from God s Word Investigate the Word (Observation and Study) Teaching: Paul Lamey

Performance Analysis of two Anaphora Resolution System for Hindi Language

ROBERT STALNAKER PRESUPPOSITIONS

Houghton Mifflin English 2004 Houghton Mifflin Company Grade Six. correlated to. TerraNova, Second Edition Level 16

Pronominal, temporal and descriptive anaphora

Houghton Mifflin English 2004 Houghton Mifflin Company Grade Five. correlated to. TerraNova, Second Edition Level 15

Models of Anaphora Processing and the Binding Constraints

Could have done otherwise, action sentences and anaphora

Understanding Truth Scott Soames Précis Philosophy and Phenomenological Research Volume LXV, No. 2, 2002

QCAA Study of Religion 2019 v1.1 General Senior Syllabus

ADDIS ABABA UNIVERSITY SCHOOL OF GRADUATE STUDIES. Design of Amharic Anaphora Resolution Model. Temesgen Dawit

Russell on Descriptions

That's Your Evidence?: Using Mechanical Turk To Develop A Computational Account Of Debate And Argumentation In Online Forums

Unit. Science and Hypothesis. Downloaded from Downloaded from Why Hypothesis? What is a Hypothesis?

ECE 5424: Introduction to Machine Learning

Houghton Mifflin Harcourt Collections 2015 Grade 8. Indiana Academic Standards English/Language Arts Grade 8

Information Extraction. CS6200 Information Retrieval (and a sort of advertisement for NLP in the spring)

StoryTown Reading/Language Arts Grade 2

Georgia Quality Core Curriculum 9 12 English/Language Arts Course: American Literature/Composition

CORRELATION FLORIDA DEPARTMENT OF EDUCATION INSTRUCTIONAL MATERIALS CORRELATION COURSE STANDARDS/BENCHMARKS

ADAIR COUNTY SCHOOL DISTRICT GRADE 03 REPORT CARD Page 1 of 5

4) When are complex discourse entities constructed in the process of text comprehension?

Implied (Unstated) Main Ideas

(Refer Slide Time 03:00)

Philosophy 240: Symbolic Logic

The summer solstice is generally understood to mark the first day of summer.

Houghton Mifflin English 2004 Houghton Mifflin Company Level Four correlated to Tennessee Learning Expectations and Draft Performance Indicators

INFORMATION EXTRACTION AND AD HOC ANAPHORA ANALYSIS

A-LEVEL RELIGIOUS STUDIES

CRUCIAL TOPICS IN THE DEBATE ABOUT THE EXISTENCE OF EXTERNAL REASONS

Factivity and Presuppositions David Schueler University of Minnesota, Twin Cities LSA Annual Meeting 2013

2004 by Dr. William D. Ramey InTheBeginning.org

Critical Appreciation of Jonathan Schaffer s The Contrast-Sensitivity of Knowledge Ascriptions Samuel Rickless, University of California, San Diego

Richard L. W. Clarke, Notes REASONING

Parents Seminar English Language Sharing 11 February 2017

GCE Religious Studies

Year 4 Medium Term Planning

Russell on Plurality

In general, the simplest of argument maps will take the form of something like this:

What would count as Ibn Sīnā (11th century Persia) having first order logic?

Based on the translation by E. M. Edghill, with minor emendations by Daniel Kolak.

1 Clarion Logic Notes Chapter 4

Lesson 7: Pain. In today's chapters Jonas receives painful memories from The Giver. How do you think he will respond to these memories?

Correlation. Mirrors and Windows, Connecting with Literature, Level II

The Reliability of Anaphoric Annotation, Reconsidered: Taking Ambiguity into Account

On Interpretation. Section 1. Aristotle Translated by E. M. Edghill. Part 1

abc Mark Scheme Religious Studies 1061 General Certificate of Education Philosophy of Religion 2009 examination - January series

Year 4 Medium Term Planning

Norva Y S Lo Produced by Norva Y S Lo Edited by Andrew Brennan

A Correlation of. To the. Language Arts Florida Standards (LAFS) Grade 3

ELA CCSS Grade Three. Third Grade Reading Standards for Literature (RL)

THE MEANING OF OUGHT. Ralph Wedgwood. What does the word ought mean? Strictly speaking, this is an empirical question, about the

Grade 6 correlated to Illinois Learning Standards for Mathematics

INTRODUCTION TO LOGIC 1 Sets, Relations, and Arguments

Transcription:

Outline of today s lecture Putting sentences together (in text). Coherence Anaphora (pronouns etc) Algorithms for anaphora resolution

Document structure and discourse structure Most types of document are highly structured, implicitly or explicitly: Scientific papers: conventional structure (differences between disciplines). News stories: first sentence is a summary. Blogs, etc etc Topics within documents. Relationships between sentences.

Rhetorical relations Max fell. John pushed him. can be interpreted as: or 1. Max fell because John pushed him. EXPLANATION 2 Max fell and then John pushed him. NARRATION Implicit relationship: discourse relation or rhetorical relation because, and then are examples of cue phrases

Coherence Lecture 9: Discourse Coherence Anaphora (pronouns etc) Algorithms for anaphora resolution

Coherence Coherence Discourses have to have connectivity to be coherent: Kim got into her car. Sandy likes apples. Can be OK in context: Kim got into her car. Sandy likes apples, so Kim thought she d go to the farm shop and see if she could get some.

Coherence Coherence Discourses have to have connectivity to be coherent: Kim got into her car. Sandy likes apples. Can be OK in context: Kim got into her car. Sandy likes apples, so Kim thought she d go to the farm shop and see if she could get some.

Coherence Coherence in generation Language generation needs to maintain coherence. In trading yesterday: Dell was up 4.2%, Safeway was down 3.2%, HP was up 3.1%. Better: Computer manufacturers gained in trading yesterday: Dell was up 4.2% and HP was up 3.1%. But retail stocks suffered: Safeway was down 3.2%. More about generation in the next lecture.

Coherence Coherence in interpretation Discourse coherence assumptions can affect interpretation: Kim s bike got a puncture. She phoned the AA. Assumption of coherence (and knowledge about the AA) leads to bike interpreted as motorbike rather than pedal cycle. John likes Bill. He gave him an expensive Christmas present. If EXPLANATION - he is probably Bill. If JUSTIFICATION (supplying evidence for first sentence), he is John.

Coherence Factors influencing discourse interpretation 1. Cue phrases. 2. Punctuation (also prosody) and text structure. Max fell (John pushed him) and Kim laughed. Max fell, John pushed him and Kim laughed. 3. Real world content: Max fell. John pushed him as he lay on the ground. 4. Tense and aspect. Max fell. John had pushed him. Max was falling. John pushed him. Hard problem, but surfacy techniques (punctuation and cue phrases) work to some extent.

Coherence Rhetorical relations and summarization Analysis of text with rhetorical relations generally gives a binary branching structure: nucleus and satellite: e.g., EXPLANATION, JUSTIFICATION equal weight: e.g., NARRATION Max fell because John pushed him.

Coherence Rhetorical relations and summarization Analysis of text with rhetorical relations generally gives a binary branching structure: nucleus and satellite: e.g., EXPLANATION, JUSTIFICATION equal weight: e.g., NARRATION Max fell because John pushed him.

Coherence Summarisation by satellite removal If we consider a discourse relation as a relationship between two phrases, we get a binary branching tree structure for the discourse. In many relationships, such as Explanation, one phrase depends on the other: e.g., the phrase being explained is the main one and the other is subsidiary. In fact we can get rid of the subsidiary phrases and still have a reasonably coherent discourse.

Coherence Summarisation by satellite removal If we consider a discourse relation as a relationship between two phrases, we get a binary branching tree structure for the discourse. In many relationships, such as Explanation, one phrase depends on the other: e.g., the phrase being explained is the main one and the other is subsidiary. In fact we can get rid of the subsidiary phrases and still have a reasonably coherent discourse.

Coherence Summarisation by satellite removal If we consider a discourse relation as a relationship between two phrases, we get a binary branching tree structure for the discourse. In many relationships, such as Explanation, one phrase depends on the other: e.g., the phrase being explained is the main one and the other is subsidiary. In fact we can get rid of the subsidiary phrases and still have a reasonably coherent discourse. We get a binary branching tree structure for the discourse. In many relationships one phrase depends on the other. In fact we can get rid of the subsidiary phrases and still have a reasonably coherent discourse.

Anaphora (pronouns etc) Lecture 9: Discourse Coherence Anaphora (pronouns etc) Algorithms for anaphora resolution

Anaphora (pronouns etc) Referring expressions Niall Ferguson is prolific, well-paid and a snappy dresser. Stephen Moss hated him at least until he spent an hour being charmed in the historian s Oxford study. referent a real world entity that some piece of text (or speech) refers to. the actual Prof. Ferguson referring expressions bits of language used to perform reference by a speaker. Niall Ferguson, he, him antecedent the text initially evoking a referent. Niall Ferguson anaphora the phenomenon of referring to an antecedent.

Anaphora (pronouns etc) Pronoun resolution Pronouns: a type of anaphor. Pronoun resolution: generally only consider cases which refer to antecedent noun phrases. Niall Ferguson is prolific, well-paid and a snappy dresser. Stephen Moss hated him at least until he spent an hour being charmed in the historian s Oxford study.

Anaphora (pronouns etc) Pronoun resolution Pronouns: a type of anaphor. Pronoun resolution: generally only consider cases which refer to antecedent noun phrases. Niall Ferguson is prolific, well-paid and a snappy dresser. Stephen Moss hated him at least until he spent an hour being charmed in the historian s Oxford study.

Anaphora (pronouns etc) Pronoun resolution Pronouns: a type of anaphor. Pronoun resolution: generally only consider cases which refer to antecedent noun phrases. Niall Ferguson is prolific, well-paid and a snappy dresser. Stephen Moss hated him at least until he spent an hour being charmed in the historian s Oxford study.

Anaphora (pronouns etc) Hard constraints: Pronoun agreement A little girl is at the door see what she wants, please? My dog has hurt his foot he is in a lot of pain. * My dog has hurt his foot it is in a lot of pain. Complications: The team played really well, but now they are all very tired. Kim and Sandy are asleep: they are very tired. Kim is snoring and Sandy can t keep her eyes open: they are both exhausted.

Anaphora (pronouns etc) Hard constraints: Reflexives John i cut himself i shaving. (himself = John, subscript notation used to indicate this) # John i cut him j shaving. (i j a very odd sentence) Reflexive pronouns must be coreferential with a preceeding argument of the same verb, non-reflexive pronouns cannot be.

Anaphora (pronouns etc) Hard constraints: Pleonastic pronouns Pleonastic pronouns are semantically empty, and don t refer: It is snowing It is not easy to think of good examples. It is obvious that Kim snores. It bothers Sandy that Kim snores.

Anaphora (pronouns etc) Soft preferences: Salience Recency Kim has a big car. Sandy has a smaller one. Lee likes to drive it. Grammatical role Subjects > objects > everything else: Fred went to the Grafton Centre with Bill. He bought a CD. Repeated mention Entities that have been mentioned more frequently are preferred. Parallelism Entities which share the same role as the pronoun in the same sort of sentence are preferred: Bill went with Fred to the Grafton Centre. Kim went with him to Lion Yard. Him=Fred Coherence effects (mentioned above)

Anaphora (pronouns etc) World knowledge Sometimes inference will override soft preferences: Andrew Strauss again blamed the batting after England lost to Australia last night. They now lead the series three-nil. they is Australia. But a discourse can be odd if strong salience effects are violated: The England football team won last night. Scotland lost.? They have qualified for the World Cup with a 100% record.

Anaphora (pronouns etc) World knowledge Sometimes inference will override soft preferences: Andrew Strauss again blamed the batting after England lost to Australia last night. They now lead the series three-nil. they is Australia. But a discourse can be odd if strong salience effects are violated: The England football team won last night. Scotland lost.? They have qualified for the World Cup with a 100% record.

Algorithms for anaphora resolution Lecture 9: Discourse Coherence Anaphora (pronouns etc) Algorithms for anaphora resolution

Algorithms for anaphora resolution Anaphora resolution as supervised classification Classification: training data labelled with class and features, derive class for test data based on features. For potential pronoun/antecedent pairings, class is TRUE/FALSE. Assume candidate antecedents are all NPs in current sentence and preceeding 5 sentences (excluding pleonastic pronouns)

Algorithms for anaphora resolution Example Niall Ferguson is prolific, well-paid and a snappy dresser. Stephen Moss hated him at least until he spent an hour being charmed in the historian s Oxford study. Issues: detecting pleonastic pronouns and predicative NPs, deciding on treatment of possessives (the historian and the historian s Oxford study), named entities (e.g., Stephen Moss, not Stephen and Moss), allowing for cataphora,...

Algorithms for anaphora resolution Example Niall Ferguson is prolific, well-paid and a snappy dresser. Stephen Moss hated him at least until he spent an hour being charmed in the historian s Oxford study. Issues: detecting pleonastic pronouns and predicative NPs, deciding on treatment of possessives (the historian and the historian s Oxford study), named entities (e.g., Stephen Moss, not Stephen and Moss), allowing for cataphora,...

Algorithms for anaphora resolution Features Cataphoric Binary: t if pronoun before antecedent. Number agreement Binary: t if pronoun compatible with antecedent. Gender agreement Binary: t if gender agreement. Same verb Binary: t if the pronoun and the candidate antecedent are arguments of the same verb. Sentence distance Discrete: { 0, 1, 2... } Grammatical role Discrete: { subject, object, other } The role of the potential antecedent. Parallel Binary: t if the potential antecedent and the pronoun share the same grammatical role. Linguistic form Discrete: { proper, definite, indefinite, pronoun }

Algorithms for anaphora resolution Feature vectors pron ante cat num gen same dist role par form him Niall F. f t t f 1 subj f prop him Ste. M. f t t t 0 subj f prop him he t t t f 0 subj f pron he Niall F. f t t f 1 subj t prop he Ste. M. f t t f 0 subj t prop he him f t t f 0 obj f pron

Algorithms for anaphora resolution Training data, from human annotation class cata num gen same dist role par form TRUE f t t f 1 subj f prop FALSE f t t t 0 subj f prop FALSE t t t f 0 subj f pron FALSE f t t f 1 subj t prop TRUE f t t f 0 subj t prop FALSE f t t f 0 obj f pron

Algorithms for anaphora resolution Naive Bayes Classifier Choose most probable class given a feature vector f : Apply Bayes Theorem: Constant denominator: ĉ = argmax P(c f ) c C P(c f ) = P( f c)p(c) P( f ) ĉ = argmax P( f c)p(c) c C Independent feature assumption ( naive ): n ĉ = argmax P(c) P(f i c) c C i=1

Algorithms for anaphora resolution Problems with simple classification model Cannot implement repeated mention effect. Cannot use information from previous links: Sturt think they can perform better in Twenty20 cricket. It requires additional skills compared with older forms of the limited over game. it should refer to Twenty20 cricket, but looked at in isolation could get resolved to Sturt. If linkage between they and Sturt, then number agreement is pl. Not really pairwise: really need discourse model with real world entities corresponding to clusters of referring expressions.

Algorithms for anaphora resolution Evaluation Simple approach is link accuracy. Assume the data is previously marked-up with pronouns and possible antecedents, each pronoun is linked to an antecedent, measure percentage correct. But: Identification of non-pleonastic pronouns and antecendent NPs should be part of the evaluation. Binary linkages don t allow for chains: Sally met Andrew in town and took him to the new restaurant. He was impressed. Multiple evaluation metrics exist because of such problems.

Algorithms for anaphora resolution Classification in NLP Also sentiment classification, word sense disambiguation and many others. POS tagging (sequences). Feature sets vary in complexity and processing needed to obtain features. Statistical classifier allows some robustness to imperfect feature determination. Acquiring training data is expensive. Few hard rules for selecting a classifier: e.g., Naive Bayes often works even when independence assumption is clearly wrong (as with pronouns). Experimentation, e.g., with WEKA toolkit.

Algorithms for anaphora resolution Next time Natural language generation Overview of a generation system (and more about cricket). Generation of referring expressions.