Background and Motivation | (2011) successfully apply this idea to the transfer of dependency parsers, using part-of-speech tags as the shared representation of words. |
Model Transfer | This setup requires that we use the same feature representation for both languages, for example part-of-speech tags and dependency relation labels should be from the same inventory. |
Model Transfer | In this study we will confine ourselves to those features that are applicable to all languages in question, namely: part-of-speech tags, syntactic dependency structures and representations of the word’s identity. |
Model Transfer | Part-of-speech Tags. |
Setup | We also assume that the predicate identification information is available — in most languages it can be obtained using a relatively simple heuristic based on part-of-speech tags. |
Setup | (2011), we assume that a part-of-speech tagger is available for the target language. |
Abstract | We study substitute vectors to solve the part-of-speech ambiguity problem in an unsupervised setting. |
Abstract | Part-of-speech tagging is a crucial preliminary process in many natural language processing applications. |
Abstract | Because many words in natural languages have more than one part-of-speech tag, resolving part-of-speech ambiguity is an important task. |
Algorithm | Previous work (Yatbaz et al., 2012) demonstrates that clustering substitute vectors of all word types alone has limited success in predicting part-of-speech tag of a word. |
Algorithm | The output of clustering induces part-of-speech categories of words tokens. |
Introduction | part-of-speech or POS tagging) is an important preprocessing step for many natural language processing applications because grammatical rules are not functions of individual words, instead, they are functions of word categories. |
Introduction | In addition, we suggest that the occurrences with different part-of-speech categories of a word should be seen in different contexts. |
Experiments | We use the features of Zhang and Nivre (2011), except that all lexical identities are dropped from the templates during training and testing, hence inducing a ‘delexicalized’ model that employs only ‘universal’ properties from source-side treebanks, such as part-of-speech tags, labels, head-modifier distance, etc. |
Introduction | In the context of part-of-speech tagging, universal representations, such as that of Petrov et al. |
Towards A Universal Treebank | (2012) as the underlying part-of-speech representation. |
Towards A Universal Treebank | For both English and Swedish, we mapped the language-specific part-of-speech tags to universal tags using the mappings of Petrov et al. |
Towards A Universal Treebank | Note that relative to the universal part-of-speech tagset of Petrov et al. |
Related Work | Subsequent improvements use the P(0|b, b’) formula, for example, for incorporating various linguistics feature like part-of-speech (Zens and Ney, 2006), syntactic (Chang et al., 2009), dependency information (Bach et al., 2009) and predicate-argument structure (Xiong et al., 2012). |
Training | In total, we consider 2l part-of-speech tags; some of which are as follow: VC (copula), DEG, DEG, DER, DEV (de-related), PU (punctuation), AD (adjectives) and P (prepositions). |
Training | We train the classifiers on a rich set of binary features ranging from lexical to part-of-speech (POS) and to syntactic features. |
Training | 1. anchor-related: slex (the actual word of 33.12), spos ( part-of-speech (POS) tag of slex), sparent (spos’s parent in the parse tree), tlex (6:: ’s actual target word).. |
Two-Neighbor Orientation Model | In our experiments, we use a simple heuristics based on part-of-speech tags which will be described in Section 7. |
Conclusion | We proposed FWD (Frames, BOW, and part-of-speech specific DAL) features and SemTree data representations. |
Experiments | We remove stop words and use Stanford CoreNLP for part-of-speech tagging and named entity recognition. |
Methods | (2009) introduced part-of-speech specific DAL features for sentiment analysis. |
Related Work | Table 1: FWD features (Frame, bag-of-Words, part-of-speech DAL score) and their value types. |
Abstract | Standard methods for part-of-speech tagging suffer from data sparseness when used on highly inflectional languages (which require large lexical tagset inventories). |
Abstract | Several neural network architectures have been proposed for the task of part-of-speech tagging. |
Abstract | We presented a new approach for large tagset part-of-speech tagging using neural networks. |
Problem Definition and Notation | Structured output prediction encompasses a wide variety of NLP problems like part-of-speech tagging, parsing and machine translation. |
Problem Definition and Notation | Figure 1 illustrates this observation in the context of part-of-speech tagging. |
Problem Definition and Notation | Figure 1: Comparison of number of instances and the number of unique observed part-of-speech structures in the Gi-gaword corpus. |
Experiments | To extract part-of-speech tags, phrase structure trees, and typed dependencies, we use the Stanford parser (Klein and Manning, 2003; de Marneffe et al., 2006) on both train and test sets. |
Experiments | 14MAll’s features are similar to part-of-speech tags and untyped dependency relations. |
Our framework | The part-of-speech (POS) tag of the head of chunk The lexical item of the head noun |
Detection of New Entities | To detect noun phrases that potentially refer to entities, we apply a part-of-speech tagger to the input text. |
Evaluation | HYENA’s relatively poor performance can be attributed to the fact that its features are mainly syntactic such as bi-grams and part-of-speech tags. |
Related Work | All methods use trained classifiers over a variety of linguistic features, most importantly, words and bigrams with part-of-speech tags in a mention and in the textual context preceding and following the mention. |
Abstract | We present a new perceptron learning algorithm using antagonistic adversaries and compare it to previous proposals on 12 multilingual cross-domain part-of-speech tagging datasets. |
Experiments | We consider part-of-speech (POS) tagging, i.e. |
Introduction | Most learning algorithms assume that training and test data are governed by identical distributions; and more specifically, in the case of part-of-speech (POS) tagging, that training and test sentences were sampled at random and that they are identically and independently distributed. |
Instantiation | Part-of-speech: Part-of-speech information can be used to produce features that encourage certain behavior, such as avoiding the deletion of noun phrases. |
Instantiation | We generate part-of-speech information over the original raw text using a Twitter part-of-speech tagger (Ritter et al., 2011). |
Instantiation | Of course, the part-of-speech information obtained this way is likely to be noisy, and we expect our learning algorithm to take that into account. |