Index of papers in Proc. ACL 2008 that mention
  • NER
Kazama, Jun'ichi and Torisawa, Kentaro
Abstract
We demonstrated with the IREX dataset for the Japanese NER that using the constructed clusters as a gazetteer (cluster gazetteer) is a effective way of improving the accuracy of NER .
Abstract
Moreover, we demonstrate that the combination of the cluster gazetteer and a gazetteer extracted from Wikipedia, which is also useful for NER , can further improve the accuracy in several cases.
Gazetteer Induction 2.1 Induction by MN Clustering
Kazama and Torisawa (2007) extracted hyponymy relations from the first sentences (i.e., defining sentences) of Wikipedia articles and then used them as a gazetteer for NER .
Gazetteer Induction 2.1 Induction by MN Clustering
Although this Wikipedia gazetteer is much smaller than the English version used by Kazama and Torisawa (2007) that has over 2,000,000 entries, it is the largest gazetteer that can be freely used for Japanese NER .
Gazetteer Induction 2.1 Induction by MN Clustering
Our experimental results show that this Wikipedia gazetteer can be used to improve the accuracy of Japanese NER .
Introduction
Gazetteers, or entity dictionaries, are important for performing named entity recognition ( NER ) accurately.
Introduction
Most studies using gazetteers for NER are based on the assumption that a gazetteer is a mapping
Introduction
For instance, Kazama and Torisawa (2007) used the hyponymy relations extracted from Wikipedia for the English NER , and reported improved accuracies with such a gazetteer.
Using Gazetteers as Features of NER
Since Japanese has no spaces between words, there are several choices for the token unit used in NER .
Using Gazetteers as Features of NER
The NER task is then treated as a tagging task, which assigns IOB tags to each character in a sentence.10 We use Conditional Random Fields (CRFs) (Lafferty et al., 2001) to perform this tagging.
NER is mentioned in 21 sentences in this paper.
Topics mentioned in this paper:
Saha, Sujan Kumar and Mitra, Pabitra and Sarkar, Sudeshna
Introduction
Named Entity Recognition ( NER ) involves locating and classifying the names in a text.
Introduction
NER is an important task, having applications in information extraction, question answering, machine translation and in most other Natural Language Processing (NLP) applications.
Introduction
NER systems have been developed for English and few other languages with high accuracy.
Maximum Entropy Based Model for Hindi NER
MaxEnt computes the probability p(0| h) for any 0 from the space of all possible outcomes 0, and for every h from the space of all possible histories H. In NER , history can be viewed as all information derivable from the training corpus relative to the current token.
Maximum Entropy Based Model for Hindi NER
The training data for the Hindi NER task is composed of about 243K words which is collected from the popular daily Hindi newspaper “Dainik Jagaran”.
NER is mentioned in 21 sentences in this paper.
Topics mentioned in this paper:
Vadas, David and Curran, James R.
Abstract
We also implement novel NER features that generalise the lexical information needed to parse NPs and provide important semantic information.
Experiments
Our experiments are run with the C&C CCG parser (Clark and Curran, 2007b), and will evaluate the changes made to CCGbank, as well as the effectiveness of the NER features.
Experiments
Table 5: Parsing results with NER features
Experiments
5.3 NER features results
Introduction
In particular, we implement new features using NER tags from the BBN Entity Type Corpus (Weischedel and Brunstein, 2005).
Introduction
Applying the NER features results in a total increase of 1.51%.
NER features
Named entity recognition ( NER ) provides information that is particularly relevant for NP parsing, simply because entities are nouns.
NER features
There has also been recent work combining NER and parsing in the biomedical field.
NER features
Lewin (2007) experiments with detecting base-NPs using NER information, while Buyko et al.
NER is mentioned in 19 sentences in this paper.
Topics mentioned in this paper:
Richman, Alexander E. and Schone, Patrick
Abstract
In this paper, we describe a system by which the multilingual characteristics of Wikipedia can be utilized to annotate a large corpus of text with Named Entity Recognition ( NER ) tags requiring minimal human intervention and no linguistic expertise.
Abstract
language daut can be used u) bootstrap the NER process in other languages.
Introduction
Named Entity Recognition ( NER ) has long been a major task of natural language processing.
Training Data Generation
Our approach to multilingual NER is to pull back the decision-making process to English whenever possible, so that we could apply some level of linguistic expertise.
Wikipedia 2.1 Structure
The authors noted that their results would need to pass a manual supervision step before being useful for the NER task, and thus did not evaluate their results in the context of a full NER system.
Wikipedia 2.1 Structure
phrases to the classical NER tags (PERSON, LOCATION, etc.)
Wikipedia 2.1 Structure
For eXample, they used the sentence “Franz Fischler is an Austrian politician” to associate the label “politician” to the surface form “Franz Fischler.” They proceeded to show that the dictionaries generated by their method are useful when integrated into an NER system.
NER is mentioned in 8 sentences in this paper.
Topics mentioned in this paper:
Arnold, Andrew and Nallapati, Ramesh and Cohen, William W.
Conclusions, related & future work
Thus hierarchical priors seem a natural, effective and robust choice for transferring learning across NER datasets and tasks.
Introduction
Consider the task of named entity recognition ( NER ).
Introduction
In many NER problems, features are often constructed as a series of transformations of the input training data, performed in sequence.
NER is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Szpektor, Idan and Dagan, Ido and Bar-Haim, Roy and Goldberger, Jacob
Contextual Preferences Models
We identify entity types using the default Lingpipe2 Named-Entity Recognizer ( NER ), which recognizes the types Location, Person and Organization.
Contextual Preferences Models
To construct cpv;n(r), we currently use a simple approach where each individual term in cpv;e(r) is analyzed by the NER system, and its type (if any) is added to ammo“).
Experimental Settings
The Contextual Preferences for h were constructed manually: the named-entity types for cpvm(h) were set by adapting the entity types given in the guidelines to the types supported by the Ling-pipe NER (described in Section 3.2).
NER is mentioned in 3 sentences in this paper.
Topics mentioned in this paper: