Introduction | The NF (Nachfeld or “post-field”) contains prosodically heavy elements such as postposed prepositional phrases or relative clauses, and occasionally postposed noun phrases . |
Introduction | The model of Elsner and Charniak (2007) uses syntactic cues to model the discourse-newness of noun phrases . |
Introduction | Since noun phrases can be embedded in other noun phrases , overlaps can occur. |
Abstract | Our method correctly determines the matching words between two sentences using corresponding noun phrases . |
Automatic Evaluation Method using Noun-Phrase Chunking | spondences of noun phrases between MT outputs and references using chunking. |
Automatic Evaluation Method using Noun-Phrase Chunking | Secondly, the system calculates word-level scores based on the correct matched words using the determined correspondences of noun phrases . |
Automatic Evaluation Method using Noun-Phrase Chunking | 2.1 Correspondence of Noun Phrases by Chunking |
Introduction | Using noun phrases produced by chunking, our method yields the correct word correspondences and determines the similarity between two sentences in terms of the noun phrase order of appearance. |
Abstract | We present ConceptResolver, a component for the N ever-Ending Language Learner (NELL) (Carlson et al., 2010) that handles both phenomena by identifying the latent concepts that noun phrases refer to. |
Abstract | When ConceptResolver is run on N ELL’s knowledge base, 87% of the word senses it creates correspond to real-world concepts, and 85% of noun phrases that it suggests refer to the same concept are indeed synonyms. |
Introduction | A major limitation of many of these systems is that they fail to distinguish between noun phrases and the underlying concepts they refer to. |
Introduction | Furthermore, two synonymous noun phrases like “apple” and “Apple |
Introduction | Figure 1: An example mapping from noun phrases (left) to a set of underlying concepts (right). |
Abstract | This paper presents a supervised approach for identifying generic noun phrases in context. |
Introduction | Generic expressions come in two basic forms: generic noun phrases and generic sentences. |
Introduction | According to the second view, generic noun phrases denote kinds. |
Introduction | We are not aware of any detailed assessment of the proportion of generic noun phrases in educational text genres or ency-clopaedic resources like Wikipedia. |
Abstract | The idea draws on the observation that the lemmatisation of almost all Polish noun phrases may be decomposed into transformation of singular words (tokens) that make up each phrase. |
Conclusions and further work | We presented a novel approach to lemmatisation of Polish noun phrases . |
Introduction | Similar task may be defined for whole noun phrases (Degorski, 2011). |
Introduction | By lemmatisation of noun phrases (NPs) we will understand assigning each NP a grammatically correct NP corresponding to the same phrase that could stand as a dictionary entry. |
Phrase lemmatisation as a tagging problem | One of the assumptions of KPWr annotation is that actual noun phrases and prepositional phrases are labelled collectively as NP chunks. |
Phrase lemmatisation as a tagging problem | To obtain real noun phrases , phrase-initial prepositions must be stripped off3. |
Related works | Other named entity types may be realised as arbitrary noun phrases . |
Related works | As he notes, organisation names are often built of noun phrases , hence it is important to understand their internal structure. |
Detection of New Entities | To detect noun phrases that potentially refer to entities, we apply a part-of-speech tagger to the input text. |
Introduction | However, state-of-the-art open IE methods extract all noun phrases that are likely to denote entities. |
Introduction | ture are typed noun phrases . |
Introduction | Therefore, our setting resembles the established task of fine-grained typing for noun phrases (Fleis-chmann 2002), with the difference being that we disregard common nouns and phrases for prominent in-KB entities and instead exclusively focus on the difficult case of phrases that likely denote new entities. |
Related Work | Most well-known is the Stanford named entity recognition (NER) tagger (Finkel 2005) which assigns coarse-grained types like person, organization, location, and other to noun phrases that are likely to denote entities. |
Related Work | Noun phrases in the subject role in a large collection of fact triples are heuristically linked to Freebase entities. |
Introduction | In this paper, we focus on improving case prediction for noun phrases (NPs) in German translations. |
Introduction | German sentences exhibit a freer constituent order, and thus case is an important indicator of the grammatical functions of noun phrases . |
Introduction | In all four examples, the verb and the participating noun phrases Mitarbeiter (employee), Kollege (colleague) and Bericht (report) are identical, and the noun phrases are assigned the same case. |
Using subcategorization information | Verb—noun tuples referring to specific syntactic functions within verb subcategorization (verb—noun subcat case prediction) are integrated with an associated probability for accusative (direct object), dative (indirect object) and nominative (subject).6 Further to the subject and object noun phrases , the subcategorization information provides quantitative triples for verb—preposition—noun pairs, thus predicting the case of NPs within prepositional phrases (we do this only when the prepositions are ambiguious, i.e., they could subcategorize either a dative or an accusative NP). |
Using subcategorization information | In addition to modelling subcategorization information, it is also important to differentiate between subcategorized noun phrases (such as object or subject), and noun phrases |
Evaluation | Standard coreference resolution data sets annotate all noun phrases that have an antecedent noun phrase in the text. |
Evaluation | Of course, full coreference-annotated data is a precious resource, with the pronoun it making up only a small portion of the marked-up noun phrases . |
Introduction | The goal of coreference resolution is to determine which noun phrases in a document refer to the same real-world entity. |
Introduction | As part of this task, coreference resolution systems must decide which pronouns refer to preceding noun phrases (called antecedents) and which do not. |
Introduction | that do not refer to preceding noun phrases are called non-anaphoric or non-referential pronouns. |
Results | We thus provide these same nine-token windows to our human subjects, and ask them to decide whether the pronouns refer to previous noun phrases or not, based on these contexts. |
Background | That is, aside from the constraint that verbal clauses require a clitic cluster (marking subject and object agreement and tense, aspect and mood) in second position, the word order is otherwise free, to the point that noun phrases can be noncontiguous, with head nouns and their modifiers separated by unrelated words. |
Background | To relate such discontinuous noun phrases to appropriate semantic representations where ‘having- |
Wambaya grammar | 0 Word order: second position clitic cluster, otherwise free word order, discontinuous noun phrases |
Wambaya grammar | o Derived event modifiers: nominals (nouns, adjectives, noun phrases ) used as event modifiers with meaning dependent on their case marking |
Wambaya grammar | 0 Coordination: of clauses and noun phrases |
Model | 0 Generating Specification Tree: For each text specification, draw a specification tree 75 from all possible trees over the sequence of noun phrases in this specification. |
Model | For example, at the unigram level we aim to capture that noun phrases containing specific words such as “cases” and “lines” may be key phrases (correspond to data chunks appear in the input), and that verbs such as “contain” may indicate that the next noun phrase is a key phrase. |
Model | Total # of words 7330 Total # of noun phrases 1829 Vocabulary size 781 Avg. |
Problem Formulation | As input, we are given a set of text specifications w = {2121, - - - ,wN}, where each w is a text specification represented as a sequence of noun phrases We use UIUC shallow parser to preprocess each text specificaton into a sequence of the noun phrases.4 In addition, we are given a set of input examples for each wi. |
Problem Formulation | Our model predicts specification trees 1: = {751, - - - ,tN } for the text specifications, where each specification tree ti is a dependency tree over noun phrases In general many program input formats are nested tree structures, in which the tree root denotes the entire chunk of program input data and each chunk (tree node) can be further divided into sub-chunks or primitive fields that appear in the program input (see Figure 3). |
Distributional Semantic Hidden Markov Models | Given a document consisting of a sequence of T clauses headed by propositional heads H (verbs or event nouns), and argument noun phrases fl, a DSHMM models the joint probability of observations H, fl, and latent random variables E and g representing domain events and slots respectively; i.e., P(H, fl, E, g |
Distributional Semantic Hidden Markov Models | We assume that event heads are verbs or event nouns, while arguments are the head words of their syntactically dependent noun phrases . |
Guided Summarization Slot Induction | First, the maximal noun phrases are extracted from the contributors and clustered based on the TAC slot of the contributor. |
Guided Summarization Slot Induction | These clusters of noun phrases then become the gold standard clusters against which automatic systems are compared. |
Guided Summarization Slot Induction | Noun phrases are considered to be matched if the lemmata of their head words are the same and they are extracted from the same summary. |
Problem Definition | An open information extractor is a function from a document, d, to a set of triples, {(argl, rel, arg2>}, where the args are noun phrases and rel is a textual fragment indicating an implicit, semantic relation between the two noun phrases . |
Wikipedia-based Open IE | Given the article on “Stanford University,” for example, the matcher should associate (established, 1 8 91) with the sentence “The university was founded in 1891 by Given a Wikipedia page with an infobox, the matcher iterates through all its attributes looking for a unique sentence that contains references to both the subject of the article and the attribute value; these noun phrases will be annotated argl and argg in the training set. |
Wikipedia-based Open IE | Second, it rejects the sentence if the subject and/or attribute value are not heads of the noun phrases containing them. |
Wikipedia-based Open IE | WOEparse uses a pattern learner to classify whether the shortest dependency path between two noun phrases indicates a semantic relation. |
Aspects of linguistic quality | Referential clarity: It should be easy to identify who or what the pronouns and noun phrases in the summary are referring to. |
Indicators of linguistic quality | In this class, we include features that reflect the modification properties of noun phrases (NPs) in the summary that are first mentions to people. |
Indicators of linguistic quality | Noun phrases can include pre-modifiers, apposi-tives, prepositional phrases, etc. |
Indicators of linguistic quality | These include sentence length, number of fragments, average lengths of the a’iflerent types of syntactic phrases, total length of modifiers in noun phrases , and various other syntactic features. |
Experiments | This resulted in a vocabulary of about 32,000 noun phrases , and a set of about 2.4 million tuples in our generalization corpus. |
Experiments | For each of the 500 observed tuples in the test-set we generated a pseudo-negative tuple by randomly sampling two noun phrases from the distribution of NPs in both corpora. |
Topic Models for Selectional Prefs. | 2 Our task is to compute, for each argument ai of each relation r, a set of usual argument values ( noun phrases ) that it takes. |
Topic Models for Selectional Prefs. | Readers familiar with topic modeling terminology can understand our approach as follows: we treat each relation as a document whose contents consist of a bags of words corresponding to all the noun phrases observed as arguments of the relation in our corpus. |
Target-dependent Sentiment Classification | In this paper, we first regard all noun phrases , including the target, as extended targets for simplicity. |
Target-dependent Sentiment Classification | In addition to the noun phrases including the target, we further expand the extended target set with the following three methods: |
Target-dependent Sentiment Classification | It is common that people use definite or demonstrative noun phrases or pronouns referring to the target in a tweet and express sentiments directly on them. |
Target Candidate Ranking | Then we apply a hierarchical Hidden Markov Model (HMM) based Chinese lexical analyzer ICTCLAS (Zhang et al., 2003) to extract named entities, noun phrases and events. |
Target Candidate Ranking | Therefore we limited the types of vertices into: Morph (M), Entity(E), which includes target candidates, Event (EV), and NonEntity Noun Phrases (NP); and used co-occnrrence as the edge type. |
Target Candidate Ranking | We extract entities, events, and nonentity noun phrases that occur in more than one tweet as neighbors. |
Combining CCGbank corrections | Compound noun phrases can nest inside each other, creating bracketing ambiguities: |
Combining CCGbank corrections | The structure of such compound noun phrases is left underspecified in the Penn Treebank (PTB), because the annotation procedure involved stitching together partial parses produced by the Fid-ditch parser (Hindle, 1983), which produced flat brackets for these constructions. |
Combining CCGbank corrections | Vadas and Curran (2007) addressed this by manually annotating all of the ambiguous noun phrases in the PTB, and went on to use this information to correct 20,409 dependencies (1.95%) in CCGbank (Vadas and Curran, 2008). |
Reanalysing partitive constructions | Partitive constructions are not given special treatment in the PTB, and were analysed as noun phrases with a PP modifier in CCGbank: |
CD | As a result, many structures that in other treebanks would be prepositional phrases with embedded noun phrases — and thus nonlocal constituents — are flat prepositional phrases here. |
Introduction | The task for these models is chunking, so we evaluate performance on identification of multiword chunks of all constituent types as well as only noun phrases . |
Tasks and Benchmark | We also evaluate our models based on their performance at identifying base noun phrases , NPs that do not contain nested NPs. |
Introduction | Named entity recognizers perform semantic tagging on proper name noun phrases , and |
Introduction | The mention detection task was introduced in recent ACE evaluations (e.g., (ACE, 2007; ACE, 2008)) and requires systems to identify all noun phrases (proper names, nominals, and pronouns) that correspond to 5-7 semantic classes. |
Related Work | We defined annotation guidelines6 for each semantic category and conducted an inter-annotator agreement study to measure the consistency of the two domain experts on 30 message board posts, which contained 1,473 noun phrases . |
Large-Scale Harvesting of Semantic Predicates | We search the English Wikipedia for all the token sequences which match n, resulting in a list of noun phrases filling the * argument. |
Large-Scale Harvesting of Semantic Predicates | As can be seen, a wide range of noun phrases are extracted, from quantities such as glass and cap to other aspects, such as brand and constituent. |
Preliminaries | While in principle * could match any sequence of words, since we aim at generalizing nouns, in what follows we allow * to match only noun phrases (e.g., glass, hot cup, very big bottle, etc. |
Extracting Rules from Wikipedia | As Wikipedia’s titles are mostly noun phrases, the terms we extract as RHSs are the nouns and noun phrases in the definition. |
Extracting Rules from Wikipedia | It also enables us extracting additional rules by splitting conjoined noun phrases and by taking both the head noun and the complete base noun phrase as the RHS for separate rules (examples 1—3 in Table 1). |
Extracting Rules from Wikipedia | Therefore, we further create rules for all head nouns and base noun phrases within the definition (example 4). |
Language Model 2.1 The General Approach | Three tags are used for different types of noun phrases : pronominal NPs, non-pronominal NPs and prenominal genitives. |
Language Model 2.1 The General Approach | The model for noun phrases is based on the joint probability of the head type (either noun, adjective or proper name), the presence of a determiner and the presence of pre-and postnominal modifiers. |
Linguistic Resources | sentences, subordinate clauses, relative and interrogative clauses, noun phrases , prepositional phrases, adjective phrases and expressions of date and time. |
Introduction | Take the example of translating noun phrases from English to Greek (or German, Czech, etc.). |
Introduction | However, Greek words in noun phrases are inflected based on their role in the sentence. |
Introduction | A purely lexical mapping of English noun phrases to Greek noun phrases suffers from the lack of information about its role in the sentence, making it hard to choose the right inflected forms. |