Index of papers in Proc. ACL that mention
  • coreference
Bergsma, Shane and Lin, Dekang and Goebel, Randy
Conclusion
A consequence of this research was the creation of It-Bank, a collection of thousands of labelled examples of the pronoun it, which will benefit other coreference resolution researchers.
Conclusion
Another avenue of study will look at the interaction between coreference resolution and machine translation.
Conclusion
In general, jointly optimizing translation and coreference is an exciting and largely unexplored research area, now partly enabled by our portable non-referential detection methodology.
Evaluation
Standard coreference resolution data sets annotate all noun phrases that have an antecedent noun phrase in the text.
Introduction
The goal of coreference resolution is to determine which noun phrases in a document refer to the same real-world entity.
Introduction
As part of this task, coreference resolution systems must decide which pronouns refer to preceding noun phrases (called antecedents) and which do not.
Introduction
In sentence (1), it is an anaphoric pronoun referring to some previous noun phrase, like “the sauce” or “an appointment.” In sentence (2), it is part of the idiomatic expression “make it” meaning “succeed.” A coreference resolution system should find an antecedent for the first it but not the second.
Methodology
Although coreference evaluations, such as the MUC (1997) tasks, also make this distinction, it is not necessarily used by all researchers.
Related Work
First of all, research in coreference resolution has shown the benefits of modules for general noun anaphoricity determination (Ng and Cardie, 2002; Denis and Baldridge, 2007).
Related Work
Bergsma and Lin (2006) determine the likelihood of coreference along the syntactic path connecting a pronoun to a possible antecedent, by looking at the distribution of the path in text.
Results
Notably, the first noun-phrase before the context is the word “software.” There is strong compatibility between the pronoun-parent “install” and the candidate antecedent “software.” In a full coreference resolution system, when the anaphora resolution module has a strong preference to link it to an antecedent (which it should when the pronoun is indeed referential), we can override a weak non-referential probability.
coreference is mentioned in 11 sentences in this paper.
Topics mentioned in this paper:
Martschat, Sebastian
Abstract
We present an unsupervised model for coreference resolution that casts the problem as a clustering task in a directed labeled weighted multigraph.
Introduction
Coreference resolution is the task of determining which mentions in a text refer to the same entity.
Introduction
Quite recently, however, rule-based approaches regained popularity due to Stanford’s multi-pass sieve approach which exhibits state-of-the-art performance on many standard coreference data sets (Raghunathan et al., 2010) and also won the CoNLL-2011 shared task on coreference resolution (Lee et al., 2011; Pradhan et al., 2011).
Introduction
In this paper we present a graph-based approach for coreference resolution that models a document to be processed as a graph.
Related Work
Graph-based coreference resolution.
Related Work
Nicolae and Nicolae (2006) phrase coreference resolution as a graph clustering problem: they first perform pairwise classification and then construct a graph using the derived confidence values as edge weights.
Related Work
(2010) and Cai and Strube (2010) perform coreference resolution in one step using graph partitioning approaches.
coreference is mentioned in 30 sentences in this paper.
Topics mentioned in this paper:
Li, Peifeng and Zhu, Qiaoming and Zhou, Guodong
Abstract
To resolve such problem, this paper proposes a novel global argument inference model to explore specific relationships, such as Coreference , Sequence and Parallel, among relevant event mentions to recover those inter-sentence arguments in the sentence, discourse and document layers which represent the cohesion of an event or a topic.
Inferring Inter-Sentence Arguments on Relevant Event Mentions
In this paper, we divide the relations among relevant event mentions into three categories: Coreference , Sequence and Parallel.
Inferring Inter-Sentence Arguments on Relevant Event Mentions
An event may have more than one mention in a document and coreference event mentions refer to the same event, as same as the definition in the ACE evaluations.
Inferring Inter-Sentence Arguments on Relevant Event Mentions
Those coreference event mentions always have the same arguments and roles.
Introduction
extractor, it is really challenging to recognize these entities as the arguments of its corefered mention E3 since to reduce redundancy in a Chinese discourse, the later Chinese sentences omit many of these entities already mentioned in previous sentences.
coreference is mentioned in 12 sentences in this paper.
Topics mentioned in this paper:
Lassalle, Emmanuel and Denis, Pascal
Abstract
This paper proposes a new method for significantly improving the performance of pairwise coreference models.
Abstract
In effect, our approach finds an optimal feature space (derived from a base feature set and indicator set) for discriminating coreferential mention pairs.
Introduction
Coreference resolution is the problem of partitioning a sequence of noun phrases (or mentions), as they occur in a natural language text, into a set of referential entities.
Introduction
A common approach to this problem is to separate it into two modules: on the one hand, one defines a model for evaluating coreference links, in general a discriminative classifier that detects coreferential mention pairs.
Introduction
In this kind of architecture, the performance of the entire coreference system strongly depends on the quality of the local pairwise classifier.1 Consequently, a lot of research effort on coreference resolution has focused on trying to boost the performance of the pairwise classifier.
Modeling pairs
Pairwise models basically employ one local classifier to decide whether two mentions are coreferential or not.
Modeling pairs
For instance, some coreference resolution systems process different kinds of anaphors separately, which suggests for example that pairs containing an anaphoric pronoun behave differently from pairs with non-
Modeling pairs
where Q classically represents randomness, X is the space of objects (“mention pairs”) that is not directly observable and yij(w) E 3/ = {+1, —1} are the labels indicating whether mi and 7m are coreferential or not.
System description
We tested 3 classical greedy link selection strategies that form clusters from the classifier decision: Closest-First (merge mentions with their closest coreferent mention on the left) (Soon et al., 2001),
coreference is mentioned in 13 sentences in this paper.
Topics mentioned in this paper:
Guinaudeau, Camille and Strube, Michael
Experiments
We also propose to use a coreference resolution system and consider coreferent entities to be the same discourse entity.
Experiments
As the coreference resolution system is trained on well-formed textual documents and expects a correct sentence ordering, we use in all our experiments only features that do not rely on sentence order (e.g.
Experiments
Second, we want to evaluate the influence of automatically performed coreference resolution in a controlled fashion.
The Entity Grid Model
Finally, they include a heuristic coreference resolution component by linking mentions which share a
coreference is mentioned in 19 sentences in this paper.
Topics mentioned in this paper:
Durrett, Greg and Hall, David and Klein, Dan
Abstract
Efficiently incorporating entity-level information is a challenge for coreference resolution systems due to the difficulty of exact inference over partitions.
Abstract
We describe an end-to-end discriminative probabilistic model for coreference that, along with standard pairwise features, enforces structural agreement constraints between specified properties of coreferent mentions.
Example
One way is to exploit the correct coreference decision we have already made, they A referring to people, since people are not as likely to have a price as art items are.
Example
Because even these six mentions have hundreds of potential partitions into coreference chains, we cannot search over partitions exhaustively, and therefore we must design our model to be able to use this information while still admitting an efficient inference scheme.
Introduction
The inclusion of entity-level features has been a driving force behind the development of many coreference resolution systems (Luo et al., 2004; Rahman and Ng, 2009; Haghighi and Klein, 2010; Lee et al., 2011).
Introduction
However, such systems may be locked into bad coreference decisions and are difficult to directly optimize for standard evaluation metrics.
Introduction
structural agreement factors softly drive properties of coreferent mentions to agree with one another.
Models
,i—1,<neW>}; this variable specifies mention i’s selected antecedent or indicates that it begins a new coreference chain.
Models
Note that a set of coreference chains 0 (the final desired output) can be uniquely determined from a, but a is not uniquely determined by C.
Models
Figure 1: Our BASIC coreference model.
coreference is mentioned in 35 sentences in this paper.
Topics mentioned in this paper:
Andrews, Nicholas and Eisner, Jason and Dredze, Mark
Abstract
In this paper, we propose a model for cross-document coreference resolution that achieves robustness by learning similarity from unlabeled data.
Introduction
even identical—do not necessarily corefer .
Introduction
In this paper, we propose a method for jointly (1) learning similarity between names and (2) clustering name mentions into entities, the two major components of cross-document coreference resolution systems (Baron and Freedman, 2008; Finin et al., 2009; Rao et al., 2010; Singh et al., 2011; Lee et al., 2012; Green et al., 2012).
Introduction
Such creative spellings are especially common on Twitter and other social media; we give more examples of coreferents learned by our model in Section 8.4.
Overview and Related Work
Cross-document coreference resolution (CDCR) was first introduced by Bagga and Baldwin (1998b).
Overview and Related Work
Most approaches since then are based on the intuitions that coreferent names tend to have “similar” spellings and tend to appear in “similar” contexts.
Overview and Related Work
We adopt a “phylogenetic” generative model of coreference .
coreference is mentioned in 21 sentences in this paper.
Topics mentioned in this paper:
Wick, Michael and Singh, Sameer and McCallum, Andrew
Abstract
Methods that measure compatibility between mention pairs are currently the dominant approach to coreference .
Abstract
These trees succinctly summarize the mentions providing a highly compact, information-rich structure for reasoning about entities and coreference uncertainty at massive scales.
Abstract
We demonstrate that the hierarchical model is several orders of magnitude faster than pairwise, allowing us to perform coreference on six million author mentions in under four hours on a single CPU.
Introduction
Coreference resolution, the task of clustering mentions into partitions representing their underlying real-world entities, is fundamental for high-level information extraction and data integration, including semantic search, question answering, and knowledge base construction.
Introduction
For example, coreference is Vital for determining author publication lists in bibliographic knowledge bases such as CiteSeer and Google Scholar, where the repository must know if the “R.
Introduction
31 for Fast Coreference at Large Scale
coreference is mentioned in 58 sentences in this paper.
Topics mentioned in this paper:
Björkelund, Anders and Kuhn, Jonas
Abstract
We investigate different ways of learning structured perceptron models for coreference resolution when using nonlocal features and beam search.
Background
Coreference resolution is the task of grouping referring expressions (or mentions) in a text into disjoint clusters such that all mentions in a cluster refer to the same entity.
Background
In recent years much work on coreference resolution has been devoted to increasing the ex-pressivity of the classical mention-pair model, in which each coreference classification decision is limited to information about two mentions that make up a pair.
Background
This shortcoming has been addressed by entity-mention models, which relate a candidate mention to the full cluster of mentions predicted to be coreferent so far (for more discussion on the model types, see, e.g., (Ng, 2010)).
Introduction
We show that for the task of coreference resolution the straightforward combination of beam search and early update (Collins and Roark, 2004) falls short of more limited feature sets that allow for exact search.
Introduction
Coreferent mentions in a document are usually annotated as sets of mentions, where all mentions in a set are coreferent .
Introduction
This approach provides a powerful boost to the performance of coreference resolvers, but we find that it does not combine well with the LaSO learning strategy.
coreference is mentioned in 25 sentences in this paper.
Topics mentioned in this paper:
Bansal, Mohit and Klein, Dan
Abstract
To address semantic ambiguities in coreference resolution, we use Web n-gram features that capture a range of world knowledge in a diffuse but robust way.
Abstract
When added to a state-of-the-art coreference baseline, our Web features give significant gains on multiple datasets (ACE 2004 and ACE 2005) and metrics (MUC and B3), resulting in the best results reported to date for the end-to-end task of coreference resolution.
Baseline System
Reconcile is one of the best implementations of the mention-pair model (Soon et al., 2001) of coreference resolution.
Baseline System
The mention-pair model relies on a pairwise function to determine whether or not two mentions are coreferent .
Baseline System
Pairwise predictions are then consolidated by transitive closure (or some other clustering method) to form the final set of coreference clusters (chains).
Introduction
Many of the most difficult ambiguities in coreference resolution are semantic in nature.
Introduction
For resolving coreference in this example, a system would benefit from the world knowledge that Obama is the president.
Introduction
There have been multiple previous systems that incorporate some form of world knowledge in coreference resolution tasks.
coreference is mentioned in 31 sentences in this paper.
Topics mentioned in this paper:
Singh, Sameer and Subramanya, Amarnag and Pereira, Fernando and McCallum, Andrew
Abstract
Cross-document coreference , the task of grouping all the mentions of each entity in a document collection, arises in information extraction and automated knowledge base construction.
Abstract
To solve the problem we propose two ideas: (a) a distributed inference technique that uses parallelism to enable large scale processing, and (b) a hierarchical model of coreference that represents uncertainty over multiple granular—ities of entities to facilitate more effective approximate inference.
Introduction
Given a collection of mentions of entities extracted from a body of text, coreference or entity resolution consists of clustering the mentions such that two mentions belong to the same cluster if and only if they refer to the same entity.
Introduction
While significant progress has been made in within-document coreference (Ng, 2005; Culotta et al., 2007; Haghighi and Klein, 2007; Bengston and Roth, 2008; Haghighi and Klein,
Introduction
2009; Haghighi and Klein, 2010), the larger problem of cross-document coreference has not received as much attention.
coreference is mentioned in 34 sentences in this paper.
Topics mentioned in this paper:
Huang, Hongzhao and Cao, Yunbo and Huang, Xiaojiang and Ji, Heng and Lin, Chin-Yew
Introduction
From a system-to-system perspective, wikification has demonstrated its usefulness in a variety of applications, including coreference resolution (Ratinov and Roth, 2012) and classification (Vitale et al., 2012).
Principles and Approach Overview
_ _. Coreference . '
Principles and Approach Overview
Principle 2 (Coreference): Two coreferential mentions should be linked to the same concept.
Principles and Approach Overview
For example, if we know “nc” and “North Carolina” are coreferential , then they should both be linked to North Carolina.
Relational Graph Construction
In this subsection, we introduce the concept meta path which will be used to detect coreference (section 4.3) and semantic relatedness relations (section 4.4).
Relational Graph Construction
4.3 Coreference
Relational Graph Construction
A coreference relation (Principle 2) usually occurs across multiple tweets due to the highly redundant information in Twitter.
coreference is mentioned in 15 sentences in this paper.
Topics mentioned in this paper:
Kobdani, Hamidreza and Schuetze, Hinrich and Schiehlen, Michael and Kamp, Hans
Abstract
In this paper, we present an unsupervised framework that bootstraps a complete coreference resolution (CoRe) system from word associations mined from a large unlabeled corpus.
Abstract
We show that word associations are useful for CoRe — e. g., the strong association between Obama and President is an indicator of likely coreference .
Introduction
Coreference resolution (CoRe) is the process of finding markables (noun phrases) referring to the same real world entity or concept.
Introduction
Until recently, most approaches tried to solve the problem by binary classification, where the probability of a pair of markables being coreferent is estimated from labeled data.
Introduction
Alternatively, a model that determines whether a markable is coreferent with a preceding cluster can be used.
Related Work
We use the term semi-supervised for approaches that use some amount of human-labeled coreference pairs.
Related Work
(2002) used co-training for coreference resolution, a semi-supervised method.
coreference is mentioned in 31 sentences in this paper.
Topics mentioned in this paper:
Iida, Ryu and Poesio, Massimo
Abstract
We present an ILP-based model of zero anaphora detection and resolution that builds on the joint determination of anaphoricity and coreference model proposed by Denis and Baldridge (2007), but revises it and extends it into a three-way ILP problem also incorporating subject detection.
Introduction
The felicitousness of zero anaphoric reference depends on the referred entity being sufficiently salient, hence this type of data—particularly in Japanese and Italian—played a key role in early work in coreference resolution, e.g., in the development of Centering (Kameyama, 1985; Walker et a1., 1994; Di Eugenio, 1998).
Introduction
(2010)), and their use in competitions such as SEMEVAL 2010 Task 1 on Multilingual Coreference (Recasens et a1., 2010), is leading to a renewed interest in zero anaphora resolution, particularly at the light of the mediocre results obtained on zero anaphors by most systems participating in SEMEVAL.
Introduction
We integrate the zero anaphora resolver with a coreference resolver and demonstrate that the approach leads to improved results for both Italian and Japanese.
coreference is mentioned in 23 sentences in this paper.
Topics mentioned in this paper:
Liu, Changsong and She, Lanbo and Fang, Rui and Chai, Joyce Y.
Evaluation and Discussion
We first applied the semantic parser and coreference classifier as described in Section 4.1 to process each dialogue, and then built a graph representation based on the automatic processing results at the end of the dialogue.
Probabilistic Labeling for Reference Grounding
Our system first processes the data using automatic semantic parsing and coreference resolution.
Probabilistic Labeling for Reference Grounding
We then perform pairwise coreference resolution on the discourse entities to find out the discourse relations between entities from different utterances.
Probabilistic Labeling for Reference Grounding
Based on the semantic parsing and pairwise coreference resolution results, our system further builds a graph representation to capture the collaborative discourse and formulate referential grounding as a probabilistic labeling problem, as described next.
coreference is mentioned in 15 sentences in this paper.
Topics mentioned in this paper:
Luo, Xiaoqiang and Pradhan, Sameer and Recasens, Marta and Hovy, Eduard
Abstract
BLANC is a link-based coreference evaluation metric for measuring the quality of coreference systems on gold mentions.
Introduction
Coreference resolution aims at identifying natural language expressions (or mentions) that refer to the same entity.
Introduction
A critically important problem is how to measure the quality of a coreference resolution system.
Introduction
In particular, MUC measures the degree of agreement between key coreference links (i.e., links among mentions within entities) and response coreference links, while non-coreference links (i.e., links formed by mentions from different entities) are not explicitly taken into account.
Notations
Let and Or be the set of coreference links formed by mentions in 19, and 73-:
Notations
Note that when an entity consists of a single mention, its coreference link set is empty.
Original BLANC
When Tk, = Tr, Rand Index can be applied directly since coreference resolution reduces to a clustering problem where mentions are partitioned into clusters (entities):
coreference is mentioned in 30 sentences in this paper.
Topics mentioned in this paper:
Pitler, Emily and Louis, Annie and Nenkova, Ani
Abstract
Focus, coherence and referential clarity are best evaluated by a class of features measuring local coherence on the basis of cosine similarity between sentences, coreference information, and summarization specific features.
Indicators of linguistic quality
This class of linguistic quality indicators is a combination of factors related to coreference , adjacent sentence similarity, and summary-specific context of surface cohesive devices.
Indicators of linguistic quality
Coreference Steinberger et al.
Indicators of linguistic quality
(2007) compare the coreference chains in input documents and in summaries in order to locate potential problems.
Results and discussion
For all four other questions, the best feature set is Continuity, which is a combination of summarization specific features, coreference features and cosine similarity of adjacent sentences.
Results and discussion
We now investigate to what extent each of its components—summary-specific features, coreference , and cosine similarity between adjacent sentences—contribute to performance.
Results and discussion
However, the coreference features do not seem to contribute much towards predicting summary linguistic quality.
coreference is mentioned in 12 sentences in this paper.
Topics mentioned in this paper:
Mirkin, Shachar and Dagan, Ido and Pado, Sebastian
Abstract
Discourse references, notably coreference and bridging, play an important role in many text understanding applications, but their impact on textual entailment is yet to be systematically understood.
Background
The simplest form of information that discourse provides is coreference , i.e., information that two linguistic expressions refer to the same entity or event.
Background
Coreference is particularly important for processing pronouns and other anaphoric expressions, such as he in Example 1.
Background
While coreference indicates equivalence, bridging points to the existence of a salient semantic relation between two distinct entities or events.
Introduction
The detection and resolution of discourse references such as coreference and bridging anaphora play an important role in text understanding applications, like question answering and information extraction.
Introduction
The understanding that the second sentence of the text entails the hypothesis draws on two coreference relationships, namely that he is Oswald, and
Introduction
However, the utilization of discourse information for such inferences has been so far limited mainly to the substitution of nominal coreferents , while many aspects of the interface between discourse and semantic inference needs remain unexplored.
coreference is mentioned in 40 sentences in this paper.
Topics mentioned in this paper:
Yang, Xiaofeng and Su, Jian and Lang, Jun and Tan, Chew Lim and Liu, Ting and Li, Sheng
Abstract
The traditional mention-pair model for coreference resolution cannot capture information beyond mention pairs for both learning and testing.
Abstract
To deal with this problem, we present an expressive entity-mention model that performs coreference resolution at an entity level.
Abstract
The solution can explicitly express relations between an entity and the contained mentions, and automatically learn first-order rules important for coreference decision.
Introduction
Coreference resolution is the process of linking multiple mentions that refer to the same entity.
Introduction
Most of previous work adopts the mention-pair model, which recasts coreference resolution to a binary classification problem of determining whether or not two mentions in a document are co-referring (e.g.
Introduction
An alternative learning model that can overcome this problem performs coreference resolution based on entity-mention pairs (Luo et al., 2004; Yang et al., 2004b).
coreference is mentioned in 37 sentences in this paper.
Topics mentioned in this paper:
Raghavan, Preethi and Fosler-Lussier, Eric and Elhadad, Noémie and Lai, Albert M.
Abstract
The cross-narrative coreference and temporal relation weights used in both these approaches are learned from a corpus of clinical narratives.
Introduction
These cross-narrative coreferences act as important anchors for reasoning with information across narratives.
Introduction
We leverage cross-narrative coreference information along with confident cross-narrative temporal relation predictions and learn to align and temporally order medical event sequences across longitudinal clinical narratives.
Introduction
The cross-narrative coreference and temporal relation scores used in both these approaches are learned from a corpus of patient narratives from The Ohio State University Wexner Medical Center.
Problem Description
elstart 2 628mm; and elstop = 6287501,, when 61 and 62 corefer .
Problem Description
Thus, in order to align event sequences, we need to compute scores corresponding to cross-narrative medical event coreference resolution and cross-narrative temporal relations.
Problem Description
4 Cross-Narrative Coreference Resolution and Temporal Relation Learning
Related Work
We use dynamic programming to compute the best alignment, given the temporal and coreference information between medical events across these sequences.
coreference is mentioned in 44 sentences in this paper.
Topics mentioned in this paper:
Gerber, Matthew and Chai, Joyce
Conclusions and future work
Although we consistently observed development gains from using automatic coreference resolution, this process creates errors that need to be studied more closely.
Discussion
First, the system identified coreferent mentions of Olivetti that participated in exporting and supplying events (not shown).
Implicit argument identification
A candidate constituent c will often form a coreference chain with other constituents in the discourse.
Implicit argument identification
When determining whether 0 is the iargg of investment, one can draw evidence from other mentions in 0’s coreference chain.
Implicit argument identification
Thus, the unit of classification for a candidate constituent c is the three-tuple (p, iargn, c’), where c’ is a coreference chain comprising 0 and its coreferent constituents.3 We defined a binary classification function Pr(+| (p,iargn,c’ that predicts the probability that the entity referred to by c fills the missing argument position iargn of predicate instance p. In the remainder of this paper, we will refer to c as the primary filler, differentiating it from other mentions in the coreference chain c’ .
Related work
(2005) suggested approaches to implicit argument identification based on observed coreference patterns; however, the authors did not implement and evaluate such methods.
Related work
analysis of naturally occurring coreference patterns to aid implicit argument identification.
coreference is mentioned in 13 sentences in this paper.
Topics mentioned in this paper:
Yang, Bishan and Cardie, Claire
Approach
Opinion Coreference Sentences in a discourse can be linked by many types of coherence relations (Jurafsky et al., 2000).
Approach
Coreference is one of the commonly used relations in written text.
Approach
In this work, we explore coreference in the context of sentence-level sentiment analysis.
Introduction
(2008) defines coreference relations on opinion targets and applies them to constrain the polarity of sentences.
coreference is mentioned in 17 sentences in this paper.
Topics mentioned in this paper:
Dubey, Amit
Abstract
This paper introduces a novel sentence processing model that consists of a parser augmented with a probabilistic logic-based model of coreference resolution, which allows us to simulate how context interacts with syntax in a reading task.
Introduction
This is the first model we know of which introduces a broad-coverage sentence processing model which takes the effect of coreference and discourse into account.
Introduction
There are three main parts of the model: a syntactic processor, a coreference resolution system, and a simple pragmatics processor which computes certain limited forms of discourse coherence.
Introduction
The coreference resolution system is implemented
Model
The model comprises three parts: a parser, a coreference resolution system, and a pragmatics subsystem.
Model
However, as the coreference processor takes trees as input, we must therefore unpack parses before resolving referential ambiguity.
Model
the agent), get the -LGS label; (iv) non-recursive NPs are renamed NPbase (the coreference system treats each NPbase as a markable).
coreference is mentioned in 25 sentences in this paper.
Topics mentioned in this paper:
Cheung, Jackie Chi Kit and Penn, Gerald
Introduction
The authors remark that extracted sentences with VFs that are referentially related to previous context (e. g., they contain a coreferential noun phrase or a discourse relation like “therefore”) are reinserted at higher accuracies.
Introduction
The main focus of that work, however, was to adapt the model for use in a low-resource situation when perfect coreference information is not available.
Introduction
Table 3: Accuracy of automatic annotations of noun phrases with coreferents .
coreference is mentioned in 14 sentences in this paper.
Topics mentioned in this paper:
Huang, Jian and Taylor, Sarah M. and Smith, Jonathan L. and Fotiadis, Konstantinos A. and Giles, C. Lee
Abstract
Coreferencing entities across documents in a large corpus enables advanced document understanding tasks such as question answering.
Abstract
This paper presents a novel cross document coreference approach that leverages the profiles of entities which are constructed by using information extraction tools and reconciled by using a within-document coreference module.
Abstract
We compare the kernelized clustering method with a popular fuzzy relation clustering algorithm (FRC) and show 5% improvement in coreference performance.
Introduction
Cross document coreference (CDC) is the task of consolidating named entities that appear in multiple documents according to their real referents.
Introduction
document coreference (WDC), which limits the scope of disambiguation to within the boundary of a document.
Introduction
Cross document coreference , on the other hand, is a more challenging task because these linguistics cues and sentence structures no longer apply, given the wide variety of context and styles in different documents.
Methods 2.1 Document Level and Profile Based CDC
We make distinctions between document level and profile based cross document coreference .
Methods 2.1 Document Level and Profile Based CDC
persons in this work), a within-document coreference (WDC) module then links the entities deemed as referring to the same underlying identity into a WDC chain.
Methods 2.1 Document Level and Profile Based CDC
Therefore, the chained entities placed in a name cluster are deemed as coreferent .
coreference is mentioned in 30 sentences in this paper.
Topics mentioned in this paper:
Stoyanov, Veselin and Gilbert, Nathan and Cardie, Claire and Riloff, Ellen
Introduction
As is common for many natural language processing problems, the state-of-the-art in noun phrase (NP) coreference resolution is typically quantified based on system performance on manually annotated text corpora.
Introduction
MUC-6 (1995), ACE NIST (2004)) and their use in many formal evaluations, as a field we can make surprisingly few conclusive statements about the state-of-the-art in NP coreference resolution.
Introduction
In particular, it remains difi‘icult to assess the effectiveness of diflerent coreference resolution approaches, even in relative terms.
coreference is mentioned in 80 sentences in this paper.
Topics mentioned in this paper:
Wolfe, Travis and Van Durme, Benjamin and Dredze, Mark and Andrews, Nicholas and Beller, Charley and Callison-Burch, Chris and DeYoung, Jay and Snyder, Justin and Weese, Jonathan and Xu, Tan and Yao, Xuchen
Evaluation
For richer annotations that include lemmatiza-tions, part of speech, NER, and in-doc coreference , we preprocessed each of the datasets using tools7 similar to those used to create the Annotated Gigaword corpus (Napoles et al., 2012).
Evaluation
Extended Event Coreference Bank Based on the dataset of Bejan and Harabagiu (2010), Lee et al.
Evaluation
(2012) introduced the Extended Event Coreference Bank (EECB) to evaluate cross-document event coreference .
Introduction
Similar to entity coreference resolution, almost all of this work assumes unanchored mentions: predicate argument tuples are grouped together based on coreferent events.
Introduction
The first work on event coreference dates back to Bagga and Baldwin (1999).
PARMA
Predicates are represented as mention spans and arguments are represented as coreference chains (sets of mention spans) provided by in-document coreference resolution systems such as included in the Stanford NLP toolkit.
PARMA
For argument coref chains we heuristically choose a canonical mention to represent each chain, and some features only look at this canonical mention.
PARMA
The canonical mention is chosen based on length,4 information about the head word,5 and position in the document.6 In most cases, coref chains that are longer than one are proper nouns and the canonical mention is the first and longest mention (outranking pronominal references and other name shortenings).
coreference is mentioned in 10 sentences in this paper.
Topics mentioned in this paper:
Laparra, Egoitz and Rigau, German
Conclusions and Future Work
For instance, our system can also profit from additional annotations like coreference , that has proved its utility in previous works.
Evaluation
For each missing argument, the gold-standard includes the whole coreference chain of the filler.
Evaluation
Therefore, the scorer selects from all coreferent mentions the highest Dice value.
ImpAr algorithm
Filling the implicit arguments of a predicate has been identified as a particular case of coreference , very close to pronoun resolution (Silberer and Frank, 2012).
Related Work
This work applied selectional restrictions together with coreference chains, in a very specific domain.
Related Work
These early works agree that the problem is, in fact, a special case of anaphora or coreference resolution.
Related Work
Silberer and Frank (2012) adapted an entity-based coreference resolution model to extend automatically the training corpus.
coreference is mentioned in 9 sentences in this paper.
Topics mentioned in this paper:
Chambers, Nathanael and Jurafsky, Dan
Background
The Chambers and Jurafsky (2008) model learns chains completely unsupervised, (albeit after parsing and resolving coreference in the text) by counting pairs of verbs that share corefer-ring arguments within documents and computing the pointwise mutual information (PMI) between these verb-argument pairs.
Background
Even more telling is that these arguments are jointly shared (the same or coreferent ) across all three events.
Evaluation: Cloze
We use the OpenNLP1 coreference engine to resolve entity mentions.
Narrative Schemas
As mentioned above, narrative chains are learned by parsing the text, resolving coreference , and extracting chains of events that share participants.
Narrative Schemas
In our new model, argument types are learned simultaneously with narrative chains by finding salient words that represent coreferential arguments.
Narrative Schemas
We record counts of arguments that are observed with each pair of event slots, build the referential set for each word from its coreference chain, and then represent each observed argument by the most frequent head word in its referential set (ignoring pronouns and mapping entity mentions with person pronouns to a constant PERSON identifier).
Sample Narrative Schemas
We parse the text into dependency graphs and resolve coreferences .
coreference is mentioned in 8 sentences in this paper.
Topics mentioned in this paper:
Mírovský, Jiří
Introduction
Coreference relations between nodes of certain category types are captured.
Introduction
Attributes coref_text.rf and coref_gram.rf contain ids of coreferential nodes of the respective types.
Phenomena and Requirements
2.1.7 Coreferences
Phenomena and Requirements
Two types of coreferences are annotated on the tectogrammatical layer:
Phenomena and Requirements
0 grammatical coreference
Summary of the Features
0 secondary edges, secondary dependencies, coreferences , long-range relations
coreference is mentioned in 8 sentences in this paper.
Topics mentioned in this paper:
Sammons, Mark and Vydiswaran, V.G.Vinod and Roth, Dan
Annotation Proposal and Pilot Study
From the tables it is apparent that good performance on a range of phenomena in our inference model are likely to have a significant effect on RTE results, with coreference being deemed essential to the inference process for 35% of examples, and a number of other phenomena are sufficiently well represented to merit near-future attention (assuming that RTE systems do not already handle these phenomena, a question we address in section 4).
Annotation Proposal and Pilot Study
Phenomenon Occurrence Agreement coreference 35.00% 0.698 simple rewrite rule 32.62% 0.580 lexical relation 25.00% 0.738 implicit relation 23.33% 0.633 factoid 15.00% 0.412 parent-sibling 1 1.67% 0.500 genetive relation 9.29% 0.608 nominalization 8.33% 0.514 event chain 6.67% 0.589 coerced relation 6.43% 0.540 passive-active 5.24% 0.583 numeric reasoning 4.05% 0.847 spatial reasoning 3.57% 0.720
Annotation Proposal and Pilot Study
The results confirmed our initial intuition about some phenomena: for example, that coreference resolution is central to RTE, and that detecting the connecting structure is crucial in discerning negative from positive examples.
Introduction
Tasks such as Named Entity and coreference resolution, syntactic and shallow semantic parsing, and information and relation extraction have been identified as worthwhile tasks and pursued by numerous researchers.
Introduction
relevant NLP tasks such as NER, Coreference , parsing, data acquisition and application, and others.
NLP Insights from Textual Entailment
ported by their designers were the use of structured representations of shallow semantic content (such as augmented dependency parse trees and semantic role labels); the application of NLP resources such as Named Entity recognizers, syntactic and dependency parsers, and coreference resolvers; and the use of special-purpose ad-hoc modules designed to address specific entailment phenomena the researchers had identified, such as the need for numeric reasoning.
NLP Insights from Textual Entailment
As the example in figure 1 illustrates, most RTE examples require a number of phenomena to be correctly resolved in order to reliably determine the correct label (the Interaction problem); a perfect coreference resolver might as a result yield little improvement on the standard RTE evaluation, even though coreference resolution is clearly required by human readers in a significant percentage of RTE examples.
coreference is mentioned in 8 sentences in this paper.
Topics mentioned in this paper:
Bamman, David and Underwood, Ted and Smith, Noah A.
Data
While previous work uses the Stanford CoreNLP toolkit to identify characters and extract typed dependencies for them, we found this approach to be too slow for the scale of our data (a total of 1.8 billion tokens); in particular, syntactic parsing, with cubic complexity in sentence length, and out-of-the-box coreference resolution (with thousands of potential antecedents) prove to be
Data
It includes the following components for clustering character name mentions, resolving pronominal coreference , and reducing vocabulary dimensionality.
Data
3.2 Pronominal Coreference Resolution
Introduction
(2013) explicitly learn character types (or “personas”) in a dataset of Wikipedia movie plot summaries; and entity-centric models form one dominant approach in coreference resolution (Durrett et al., 2013; Haghighi and Klein, 2010).
coreference is mentioned in 6 sentences in this paper.
Topics mentioned in this paper:
Mazidi, Karen and Nielsen, Rodney D.
Approach
Coreference resolution, which could help avoid vague question generation, is discussed in Section 5.
Linguistic Challenges
Here we briefly describe three challenges: negation detection, coreference resolution, and verb forms.
Linguistic Challenges
5.2 Coreference Resolution
Linguistic Challenges
Currently, our system does not use any type of coreference resolution.
coreference is mentioned in 6 sentences in this paper.
Topics mentioned in this paper:
Ji, Heng and Grishman, Ralph
Conclusion and Future Work
The aggregation approach described here can be easily extended to improve relation detection and coreference resolution (two argument mentions referring to the same role of related events are likely to corefer ).
Global Inference
c for each event argument string and the names coreferential with or related to the argument, the frequency of the event type;
Global Inference
c for each event argument string and the names coreferential with or related to the argument, the frequency of the event type and role.
Related Work
Almost all the current event extraction systems focus on processing single documents and, except for coreference resolution, operate a sentence at a time (Grishman et al., 2005; Ahn, 2006; Hardy et al., 2006).
System Approach Overview
For each argument we also add other names coreferential with or bearing some ACE relation to the argument.
Task and Baseline System
2 In this paper we don’t consider event mention coreference resolution and so don’t distinguish event mentions and events.
coreference is mentioned in 6 sentences in this paper.
Topics mentioned in this paper:
Wang, Lu and Raghavan, Hema and Castelli, Vittorio and Florian, Radu and Cardie, Claire
Experimental Setup
Documents are processed by a full NLP pipeline, including token and sentence segmentation, parsing, semantic role labeling, and an information extraction pipeline consisting of mention detection, NP coreference , cross-document resolution, and relation detection (Florian et al., 2004; Luo et al., 2004; Luo and Zitouni, 2005).
The Framework
Finally, the postprocessing stage applies coreference resolution and sentence reordering to build the summary.
The Framework
Then we conduct simple query expansion based on the title of the topic and cross-document coreference resolution.
The Framework
And for each mention in the query, we add other mentions within the set of documents that corefer with this mention.
coreference is mentioned in 6 sentences in this paper.
Topics mentioned in this paper:
Elson, David and Dames, Nicholas and McKeown, Kathleen
Extracting Conversational Networks from Literature
We then clustered the noun phrases into coreferents for the same entity (person or organization).
Extracting Conversational Networks from Literature
For each named entity, we generate variations on the name that we would expect to see in a coreferent .
Extracting Conversational Networks from Literature
For each named entity, we compile a list of other named entities that may be coreferents , either because they are identical or because one is an expected variation on the other.
coreference is mentioned in 6 sentences in this paper.
Topics mentioned in this paper:
Li, Qi and Ji, Heng
Background
3Throughout this paper we refer to relation mention as relation since we do not consider relation mention coreference .
Experiments
Roth (2011), we excluded the D I SC relation type, and removed relations in the system output which are implicitly correct via coreference links for fair comparison.
Features
Coreference consistency Coreferential entity mentions should be assigned the same entity type.
Features
We determine high-recall coreference links between two segments in the same sentence using some simple heuristic rules:
Features
Then we encode a global feature to check whether two coreferential segments share the same entity type.
coreference is mentioned in 5 sentences in this paper.
Topics mentioned in this paper:
Chambers, Nathanael and Jurafsky, Dan
Learning Templates from Raw Text
This paper extends this intuition by introducing a new vector-based approach to coreference similarity.
Learning Templates from Raw Text
In the sentence, he ran and then he fell, the subjects of run and fall corefer , and so they likely belong to the same scenario-specific semantic role.
Learning Templates from Raw Text
For instance, arguments of the relation go_ofi”:s were seen coreferring with mentions in plant:o, set_ofi”:o and injures We represent go_ofi”:s as a vector of these relation counts, calling this its coref vector representation.
coreference is mentioned in 5 sentences in this paper.
Topics mentioned in this paper:
Liao, Shasha and Grishman, Ralph
Cross-event Approach
For every event, we collect its trigger and event type; for every argument, we use coreference information and record every entity and its role(s) in events of a certain type.
Task Description
( coreferential ) entity mentions.
Task Description
Event extraction depends on previous phases entity mention classification and coreference .
Task Description
Note that entity mentions that share the same EntityID are coreferential and treated as the same object.
coreference is mentioned in 5 sentences in this paper.
Topics mentioned in this paper:
Abend, Omri and Rappoport, Ari
The UCCA Scheme
Unlike common practice in grammatical annotation, linkage relations in UCCA can cross sentence boundaries, as can relations represented in other layers (e.g., coreference ).
The UCCA Scheme
Another immediate extension to UCCA’s foundational layer can be the annotation of coreference relations.
The UCCA Scheme
A coreference layer would annotate a relation between “John” and “his” by introducing a new node whose descendants are these two units.
coreference is mentioned in 4 sentences in this paper.
Topics mentioned in this paper:
Garrido, Guillermo and Peñas, Anselmo and Cabaleiro, Bernardo and Rodrigo, Álvaro
Document Representation
0 Coreference : indicates that two chunks refer to
Document Representation
The processing includes dependency parsing, named entity recognition and coreference resolution, done with the Stanford CoreNLP software (Klein and Manning, 2003); and events and temporal information extraction, via the TARSQI Toolkit (Verhagen et al., 2005).
Document Representation
Each node of GO clusters together coreferent nodes, representing a discourse referent.
coreference is mentioned in 4 sentences in this paper.
Topics mentioned in this paper:
Krishnamurthy, Jayant and Mitchell, Tom
Prior Work
Concept discovery is also related to coreference resolution (Ng, 2008; Poon and Domingos, 2008).
Prior Work
The difference between the two problems is that coreference resolution finds noun phrases that refer to the same concept within a specific document.
Prior Work
We think the concepts produced by a system like ConceptResolver could be used to improve coreference resolution by providing prior knowledge about noun phrases that can refer to the same concept.
coreference is mentioned in 4 sentences in this paper.
Topics mentioned in this paper:
Iida, Ryu and Kobayashi, Syumpei and Tokunaga, Takenobu
Introduction
The task of identifying reference relations including anaphora and coreferences within texts has received a great deal of attention in natural language processing, from both theoretical and empirical perspectives.
Introduction
In these data sets, coreference relations are defined as a limited version of a typical coreference; this generally means that only the relations where expressions refer to the same named entities are addressed, because it makes the coreference resolution task more information extraction-oriented.
Introduction
In other words, the coreference task as defined by MUC and ACE is geared toward only identifying coreference relations anchored to an entity within the text.
Reference Resolution using Extra-linguistic Information
These features have been examined by approaches to anaphora or coreference resolution (Soon et al., 2001; Ng and Cardie, 2002, etc.)
coreference is mentioned in 4 sentences in this paper.
Topics mentioned in this paper:
Tian, Ran and Miyao, Yusuke and Matsuzaki, Takuya
Generating On-the-fly Knowledge
For a TH pair, apply dependency parsing and coreference resolution.
Generating On-the-fly Knowledge
Parsing H Abstract I T/H Coreference DCS trees denotations -
The Idea
DCS trees can be extended to represent linguistic phenomena such as quantification and coreference , with additional markers introducing additional operations on tables.
The Idea
Coreference We use Stanford CoreNLP to resolve coreferences (Raghunathan et al., 2010), whereas coreference is implemented as a special type of selection.
coreference is mentioned in 4 sentences in this paper.
Topics mentioned in this paper:
Yogatama, Dani and Sim, Yanchuan and Smith, Noah A.
Experiments
Therefore, besides canonicalizing named entities, the model also resolves within—document and cross-document coreference , since it assigned a row index for every mention.
Introduction
As a result, the model discovers parts of names—(Mrs., Michelle, Obama)—while simultaneously performing coreference resolution for named entity mentions.
Related Work
Our model is focused on the problem of canonicalizing mention strings into their parts, though its 7“ variables (which map mentions to rows) could be interpreted as (within-document and cross-document) coreference resolution, which has been tackled using a range of probabilistic models (Li et al., 2004; Haghighi and Klein, 2007; Poon and Domingos, 2008; Singh et al., 2011).
coreference is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Christensen, Janara and Soderland, Stephen and Bansal, Gagan and Mausam
Summarizing Within the Hierarchy
An edge from sentence 3,- to sj with positive weight indicates that sj may follow 3,- in a coherent summary, e. g. continued mention of an event or entity, or coreference link between 3,- and sj.
Summarizing Within the Hierarchy
A negative edge indicates an unfulfilled discourse cue or coreference mention.
Summarizing Within the Hierarchy
These are coreference mentions 0r discourse cues where none of the sentences read before (either in an ancestor summary or in the current summary) contain an antecedent:
coreference is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Chan, Yee Seng and Roth, Dan
Experiments
In that work, we also highlight that ACE annotators rarely duplicate a relation link for coreferent mentions.
Experiments
For instance, assume mentions mi, mj, and mk, are in the same sentence, mentions mi and mj are coreferent , and the annotators tag the mention pair mj, mk, with a particular relation r. The annotators will rarely duplicate the same (implicit)
Experiments
Of course, using this scoring method requires coreference information, which is available in the ACE data.
coreference is mentioned in 3 sentences in this paper.
Topics mentioned in this paper: