Introduction | Furthermore, the learned classes are not directly linked to existing resources such as WordNet (Fellbaum, 1998) or Wikipedia. |
Introduction | 0 We propose SPred, a novel approach which harvests predicates from Wikipedia and generalizes them by leveraging core concepts from WordNet . |
Large-Scale Harvesting of Semantic Predicates | As explained below, we assume the set C to be made up of representative synsets from WordNet . |
Large-Scale Harvesting of Semantic Predicates | We perform this in two substeps: we first link all our disambiguated arguments to WordNet (Section 3.3.1) and then leverage the WordNet taxonomy to populate the semantic classes in 0 (Section 3.3.2). |
Large-Scale Harvesting of Semantic Predicates | 3.3.1 Linking to WordNet |
Experiments | All these systems incorporated lexical semantics features derived from WordNet and named entity features. |
Experiments | Features used in the experiments can be categorized into six types: identical word matching (I), lemma matching (L), WordNet (WN), enhanced Lexical Semantics (LS), Named Entity matching (NE) and Answer type checking (Ans). |
Experiments | Arguably the most common source of word relations, WordNet (WN) provides the primitive features of whether two words could belong to the same synset in WordNet , could be antonyms and whether one is a hypernym of the other. |
Lexical Semantic Models | Although sets of synonyms can be easily found in thesauri or WordNet synsets, such resources typically cover only strict synonyms. |
Lexical Semantic Models | Traditionally, WordNet taxonomy is the linguistic resource for identifying hypernyms and hy-ponyms, applied broadly to many NLP problems. |
Lexical Semantic Models | However, WordNet has a number of well-known limitations including its rather limited or skewed concept distribution and the lack of the coverage of the IsA relation (Song et al., 2011). |
Related Work | Although lexical semantic information derived from WordNet has been used in some of these approaches, the research has mainly focused on modeling the mapping between the syntactic structures of questions and sentences, produced from syntactic analysis. |
Abstract | Structured resources such as WordNet offer a convenient hierarchical means for converging on a common ground for comparison, but offer little support for the divergent thinking that is needed to creatively view one concept as another. |
Abstract | These lateral views complement the vertical views of WordNet , and support a system for idea exploration called Thesaurus Rex. |
Abstract | We show also how Thesaurus Rex supports a novel, generative similarity measure for WordNet . |
Related Work and Ideas | WordNet’s taxonomic organization of noun-senses and verb-senses — in which very general categories are successively divided into increasingly informative subcategories or instance-level ideas — allows us to gauge the overlap in information content, and thus of meaning, of two lexical concepts. |
Related Work and Ideas | Wu & Palmer (1994) use the depth of a lexical concept in the WordNet hierarchy as such a proxy, and thereby estimate the similarity of two lexical concepts as twice the depth of their LCS divided by the sum of their individual depths. |
Related Work and Ideas | Rather, when using Resnick’s metric (or that of Lin, or Jiang and Conrath) for measuring the similarity of lexical concepts in WordNet, one can use the category structure of WordNet itself to estimate information content. |
Seeing is Believing (and Creating) | This reliance on the consensus viewpoint explains why WordNet (Fellbaum, 1998) has proven so useful as a basis for computational measures of lexico-semantic similarity |
Seeing is Believing (and Creating) | Using WordNet , for instance, a similarity measure can vertically converge on a common superordinate category of both inputs, and generate a single numeric result based on their distance to, and the information content of, this common generalization. |
Seeing is Believing (and Creating) | Though WordNet is ideally structured to support vertical, convergent reasoning, its comprehensive nature means it can also be used as a solid foundation for building a more lateral and divergent model of similarity. |
Experiments | The gold standards used in the evaluation are hypernym taxonomies extracted from WordNet and GDP (Open Directory Project), and meronym taxonomies extracted from WordNet . |
Experiments | In WordNet taxonomy extraction, we only use the word senses within a particular taxonomy to ensure no ambiguity. |
Experiments | In total, there are 100 hypernym taxonomies, 50 each extracted from WordNet3 and ODP4, and 50 meronym taxonomies from WordNetS . |
Introduction | It has been receiving increasing afienfion.because senuuunztaxononues,such as WordNet (Fellbaum, 1998), play an important role in solving knowledge-rich problems, including question answering (Harabagiu et a1., 2003) and textual entailment (Geffet and Dagan, 2005). |
Related Work | Pattern quality control is also investigated by using WordNet (Girju et al., 2006), graph structures built among terms (Widdows and Dorow, 2002; Kozareva et al., 2008), and pattern clusters (Davidov and Rappoport, 2008). |
Abstract | Expensive feature engineering based on WordNet senses has been shown to be useful for document level sentiment classification. |
Introduction | WordNet is a byproduct of such an analysis. |
Introduction | In WordNet , paradigms are manually generated based on the principles of lexical and semantic relationship among words (Fellbaum, 1998). |
Introduction | WordNets are primarily used to address the problem of word sense disambiguation. |
A Unified Semantic Representation | As our sense inventory, we use WordNet 3.0 (Fellbaum, 1998). |
A Unified Semantic Representation | The WordNet ontology provides a rich network structure of semantic relatedness, connecting senses directly with their hypemyms, and providing information on semantically similar senses by virtue of their nearby locality in the network. |
A Unified Semantic Representation | To extend beyond a single sense, the random walk may be initialized and restarted from a set of senses (seed nodes), rather than just one; this multi-seed walk produces a multinomial distribution over all the senses in WordNet with higher probability assigned to senses that are frequently visited from the seeds. |
Experiment 1: Textual Similarity | Additionally, because the texts often contain named entities which are not present in WordNet , we incorporated the similarity values produced by four string-based measures, which were used by other teams in the STS task: (1) longest common substring which takes into account the length of the longest overlapping contiguous sequence of characters (substring) across two strings (Gusfield, 1997), (2) longest common subsequence which, instead, finds the longest overlapping subsequence of two strings (Allison and Dix, 1986), (3) Greedy String Tiling which allows reordering in strings (Wise, 1993), and (4) the character/word n-gram similarity proposed by Barron-Cedefio et al. |
Experiment 1: Textual Similarity | o Explicit Semantic Analysis (Gabrilovich and Markovitch, 2007) where the high-dimensional vectors are obtained on WordNet , Wikipedia and Wiktionary. |
Background | WordNet (Fellbaum, 1998), by far the most widely used resource, specifies relations such as hyponymy, derivation, and entailment that can be used for semantic inference (Budanitsky and Hirst, 2006). |
Background | WordNet has also been exploited to automatically generate a training set for a hyponym classifier (Snow et al., 2005), and we make a similar use of WordNet in Section 5.1. |
Background | Recently, Szpektor and Dagan (2009) presented the resource Argument-mapped WordNet, providing entailment relations for predicates in WordNet . |
Learning Entailment Graph Edges | The first step is preprocessing: We use a large corpus and WordNet to train an entailment classifier that estimates the likelihood that one propositional template entails another. |
Learning Entailment Graph Edges | We describe a procedure for learning an entailment classifier, given a corpus and a lexicographic resource ( WordNet ). |
Learning Entailment Graph Edges | Last, we use WordNet to automatically generate a training set and train a classifier. |
FrameNet — Wiktionary Alignment | They align senses in WordNet to Wikipedia entries in a supervised setting using semantic similarity measures. |
FrameNet — Wiktionary Alignment | The PPR measure (Agirre and Soroa, 2009) maps the glosses of the two senses to a semantic vector space spanned up by WordNet synsets and then compares them using the chi-square measure. |
FrameNet — Wiktionary Alignment | The semantic vectors ppr are computed using the personalized PageRank algorithm on the WordNet graph. |
Related Work | (2008) map FrameNet frames to WordNet synsets based on the embedding of FrameNet lemmas in WordNet . |
Related Work | They use Multi-WordNet, an English-Italian wordnet , to induce an Italian FrameNet lexicon with 15,000 entries. |
Related Work | To create MapNet, Tonelli and Pianta (2009) align FrameNet senses with WordNet synsets by exploiting the textual similarity of their glosses. |
Abstract | In recent years, the increasing availability of large-scale, rich semantic knowledge sources (such as Wikipedia and WordNet ) creates new opportunities to enhance the named entity disambiguation by developing algorithms which can exploit these knowledge sources at best. |
Introduction | Fortunately, in recent years, due to the evolution of Web (e.g., the Web 2.0 and the Semantic Web) and many research efforts for the construction of knowledge bases, there is an increasing availability of large-scale knowledge sources, such as Wikipedia and WordNet . |
Introduction | The key point of our method is a reliable semantic relatedness measure between concepts (including WordNet concepts, NEs and Wikipedia concepts), called Structural Semantic Relatedness, which can capture both the explicit semantic relations between concepts and the implicit semantic knowledge embedded in graphs and networks. |
The Structural Semantic Relatedness Measure | We extract three types of semantic relations (semantic relatedness between Wikipedia concepts, lexical relatedness between WordNet concepts and social relatedness between NEs) correspondingly from three knowledge sources: Wikipedia, WordNet and NE Co-occurrence Corpus. |
The Structural Semantic Relatedness Measure | WordNet 3.02 (Fellbaum et al., 1998), a lexical knowledge source includes over 110,000 WordNet concepts (word senses about English words). |
The Structural Semantic Relatedness Measure | Various lexical relations are recorded between WordNet concepts, such as hyponyms, holonym and synonym. |
Related Work | (2004) construct a network based on WordNet synonyms and then use the shortest paths between any given word and the words ’good’ and ’bad’ to determine word polarity. |
Related Work | ’ good’ and ’bad’ themselves are closely related in WordNet with a 5-long sequence “good, sound, heavy, big, bad”. |
Related Work | Hu and Liu (2004) use WordNet synonyms and antonyms to predict the polarity of words. |
Word Polarity | One such important source is WordNet (Miller, 1995). |
Word Polarity | WordNet is a large lexical database of English. |
Word Polarity | The simplest approach is to connect words that occur in the same WordNet synset. |
Abstract | Our rule-base yields comparable performance to WordNet while providing largely complementary information. |
Application Oriented Evaluations | N0 Expansion 0.19 0.54 0.28 WikiBL 0.19 0.53 0.28 Sn0w400K 0.19 0.54 0.28 Lin 0.25 0.39 0.30 WordNet 0.30 0.47 0.37 Extraction Methods from Wikipedia: |
Application Oriented Evaluations | WordNet + WikiAuJules+mce 0.35 0.47 0.40 |
Background | Ponzetto and Strube (2007) identified the subsumption (ISA) relation from Wikipedia’s category tags, while in Yago (Suchanek et al., 2007) these tags, redirect links and WordNet were used to identify instances of 14 predefined specific semantic relations. |
Background | classifiers, whose training examples are derived automatically from WordNet . |
Background | They use these classifiers to suggest extensions to the WordNet hierarchy, the largest one consisting of 400K new links. |
Introduction | A prominent available resource is WordNet (Fellbaum, 1998), from which classical relations such as synonyms, hyponyms and some cases of meronyms may be used as LR rules. |
Introduction | An extension to WordNet was presented by (Snow et al., 2006). |
Introduction | The rule base utility was evaluated within two lexical expansion applications, yielding better results than other automatically constructed baselines and comparable results to WordNet . |
The asterisk denotes an incorrect rule | This is in contrast to (Snow et al., 2005) which focused only on hyponymy and synonymy relations and could therefore extract positive and negative examples from WordNet . |
Conclusion | In addition, by virtue of its WordNet and FrameNet annotations, MASC will be linked to parallel WordNets and FrameNets in languages other than English, thus creating a global resource for multilingual technologies, including machine translation. |
MASC Annotations | The MASC project is itself producing annotations for portions of the corpus for WordNet senses and FrameNet frames and frame elements. |
MASC Annotations | 4This includes WordNet sense annotations, which are not listed in Table 2 because they are not applied to full texts; see Section 3.1 for a description of the WordNet sense annotations in MASC. |
MASC Annotations | 3.1 WordNet Sense Annotations |
Experimental Setup | The data were annotated with coarse-grained senses which were obtained by clustering senses from the WordNet 2.1 sense inventory based on the procedure proposed by Navigli (2006). |
Introduction | Previous approaches using topic models for sense disambiguation either embed topic features in a supervised model (Cai et al., 2007) or rely heavily on the structure of hierarchical lexicons such as WordNet (Boyd-Graber et al., 2007). |
Related Work | (2007) enhance the basic LDA algorithm by incorporating WordNet senses as an additional latent variable. |
Related Work | Instead of generating words directly from a topic, each topic is associated with a random walk through the WordNet hierarchy which generates the observed word. |
Related Work | iosyncracies in the hierarchical structure of WordNet can harm performance. |
The Sense Disambiguation Model | These paraphrases can be taken from an existing resource such as WordNet (Miller, 1995) or supplied by the user (see Section 4). |
The Sense Disambiguation Model | In Model I and Model 11, the sense paraphrases are obtained from WordNet , and both the context and the sense paraphrases are treated as documents, 0 2 dc and s 2 d8. |
The Sense Disambiguation Model | WordNet is a fairly rich resource which provides detailed information about word senses (glosses, example sentences, synsets, semantic relations between senses, etc.). |
Abstract | The resource is automatically constructed by means of a methodology that integrates lexicographic and encyclopedic knowledge from WordNet and Wikipedia. |
BabelNet | Concepts and relations in BabelNet are harvested from the largest available semantic lexicon of English, WordNet , and a wide-coverage collaboratively edited encyclopedia, the English Wikipedia (Section 3.1). |
BabelNet | We collect (a) from WordNet , all available word senses (as concepts) and all the semantic pointers between synsets (as relations); (b) from Wikipedia, all encyclopedic entries (i.e. |
BabelNet | their concepts in common) by establishing a mapping between Wikipedia pages and WordNet senses (Section 3.2). |
Introduction | A pioneering endeavor was WordNet (Fellbaum, 1998), a computational lexicon of English based on psycholinguistic theories. |
Introduction | Wikipedia represents the perfect complement to WordNet , as it provides multilingual lexical knowledge of a mostly encyclopedic nature. |
Introduction | But while a great deal of work has been recently devoted to the automatic extraction of structured information from Wikipedia (Wu and Weld, 2007; Ponzetto and Strube, 2007; Suchanek et al., 2008; Medelyan et al., 2009, inter alia), the knowledge extracted is organized in a looser way than in a computational lexicon such as WordNet . |
Methodology | WordNet . |
Methodology | The most popular lexical knowledge resource in the field of NLP is certainly WordNet , a computational lexicon of the English language. |
Abstract | We create an open multilingual wordnet with large wordnets for over 26 languages and smaller ones for 57 languages. |
Abstract | It is made by combining wordnets with open li-cences, data from Wiktionary and the Unicode Common Locale Data Repository. |
Introduction | One of the many attractions of the semantic network WordNet (Fellbaum, 1998), is that there are numerous wordnets being built for different languages. |
Introduction | There are, in addition, many projects for groups of languages: Euro WordNet (Vossen, 1998), BalkaNet (Tufis et al., 2004), Asian Wordnet (Charoenporn et al., 2008) and more. |
Introduction | Although there are over 60 languages for which wordnets exist in some state of development (Fellbaum and Vossen, 2012, 316), less than half of these have released any data, and for those that have, the data is often not freely accessible (Bond and Paik, 2012). |
Abstract | It is a linked structure of wordnets of 18 different Indian languages, Universal Word dictionary and the Suggested Upper Merged Ontology (SUMO). |
Introduction | Past couple of decades have shown an immense growth in the development of lexical resources such as wordnet , Wikipedia, ontologies etc. |
Introduction | In this paper we present IndoNet, a lexical resource created by merging wordnets of 18 dif- |
Introduction | Suggested Upper Merged Ontology (SUMO) is the largest freely available ontology which is linked to the entire English WordNet (Niles and Pease, 2003). |
Related Work | Over the years wordnet has emerged as the most widely used lexical resource. |
Related Work | Though most of the wordnets are built by following the standards laid by English Wordnet (Fellbaum, 1998), their conceptualizations differ because of the differences in lexicalization of concepts across languages. |
Related Work | Wordnets are available in following Indian languages: Assamese, Bodo, Bengali, English, Gujarati, Hindi, Kashmiri, Konkani, Kannada, Malayalam, Ma-nipuri, Marathi, Nepali, Punjabi, Sanskrit, Tamil, Telugu and Urdu. |
Related Work | HAHAcronym is mainly based on lexical substitution via semantic field opposition, rhyme, rhythm and semantic relations such as antonyms retrieved from WordNet (Stark and Riesenfeld, 1998) for adjectives. |
System Description | To further increase the size of the ingredient list, we utilize another resource called WordNet (Miller, 1995), which is a large lexical database for English. |
System Description | In WordNet , nouns, verbs, adjectives and adverbs are grouped into sets of cognitive synonyms called synsets. |
System Description | Each synset in WordNet expresses a different concept and they are connected to each other with lexical, semantic and conceptual relations. |
Creative Text Retrieval | A generic, lightweight resource like WordNet can provide these relations, or a richer ontology can be used if one is available (e.g. |
Creative Text Retrieval | But ad-hoc categories do not replace natural kinds; rather, they supplement an existing system of more-or-less rigid categories, such as the categories found in WordNet . |
Creative Text Retrieval | member of the category named by C. AC can denote a fixed category in a resource like WordNet or even Wikipedia; thus, Afruit matches any member of {apple, orange, pear, lemon} and Aanimal any member of {dog, cat, mouse, deer, fox}. |
Empirical Evaluation | and @ as category builders to a handcrafted gold standard like WordNet . |
Empirical Evaluation | Other researchers have likewise used WordNet as a gold standard for categorization experiments, and we replicate here the experimental setup of Almuhareb and Poesio (2004, 2005), which is designed to measure the effectiveness of web-acquired conceptual descriptions. |
Empirical Evaluation | Almuhareb and Poesio choose 214 English nouns from 13 of WordNet’s upper-level semantic categories, and proceed to harvest property values for these concepts from the web using the Hearst-like pattern “alanlthe * C islwas”. |
Related Work and Ideas | Techniques vary, from the use of stemmers and morphological analysis to the use of thesauri (such as WordNet ; see Fellbaum, 1998; Voorhees, 1998) to pad a query with synonyms, to the use of statistical analysis to identify more appropriate context-sensitive associations and near-synonyms (e.g. |
Related Work and Ideas | Hearst (1992) shows how a pattern like “Xs and other Ys” can be used to construct more fluid, context-specific taxonomies than those provided by WordNet (e.g. |
Abstract | Our system combines multiple preexisting sources of similarity data (a standard thesaurus, WordNet , contextual similarity), enabling it to capture many types of similarity groups (“synonyms of crash,” “types of car,” etc.). |
Set Expansion | We consider three similarity data sources: the Moby thesaurus1 , WordNet (Fellbaum, 1998), and distributional similarity based on a large corpus of text (Lin, 1998). |
Set Expansion | Unlike some other thesauri (such as WordNet and thesaurus.com), entries are not broken down by word sense. |
Set Expansion | Unlike some other thesauri (including WordNet and thesaurus.com), the entries are not broken down by word sense or part of speech. |
Experimental Setup | degree of Wordnet polysemy for polysemous words |
Experimental Setup | Table 4: Average degree of Wordnet polysemy per category in the 2 domains for Hindi |
Experimental Setup | degree of Wordnet polysemy for polysemous words |
Introduction | This is achieved with the help of a novel synset-aligned multilingual dictionary which facilitates the projection of parameters learned from the Wordnet and annotated corpus of L1 to L2. |
Parameter Projection | Wordnet-dependent parameters depend on the structure of the Wordnet whereas the Corpus-dependent parameters depend on various statistics learned from a sense marked corpora. |
Parameter Projection | Both the tasks of (a) constructing a Wordnet from scratch and (b) collecting sense marked corpora for multiple languages are tedious and expensive. |
Parameter Projection | (2009) observed that by projecting relations from the Wordnet of a language and by projecting corpus statistics from the sense marked corpora of the language to those of the target language, the efi‘ort required in constructing semantic graphs for multiple Wordnets and collecting sense marked corpora for multiple languages can be avoided or reduced. |
Related Work | They showed that it is possible to project the parameters learned from the annotation work of one language to another language provided aligned Wordnets for the two languages are available. |
Related Work | However, they do not address situations where two resource deprived languages have aligned Wordnets but neither has sufficient annotated data. |
Verb Class Model 2.1 Probabilistic Model | The selectional preferences are expressed in terms of semantic concepts from WordNet , rather than a set of individual words. |
Verb Class Model 2.1 Probabilistic Model | 4. selecting a WordNet concept r for each argument slot, e.g. |
Verb Class Model 2.1 Probabilistic Model | and Light (1999) and turn WordNet into a Hidden Markov model (HMM). |
Background | RRR consists of 20,081 training and 3,097 test quadruples of the form (v, n1 , p, n2) , where the attachment decision is either v or n l. The best published results over RRR are those of Stetina and Nagao (1997), who employ WordNet sense predictions from an unsupervised WSD method within a decision tree classifier. |
Background | (2005) experimented with first-sense and hypernym features from HowNet and CiLin (both WordNets for Chinese) in a generative parse model applied to the Chinese Penn Treebank. |
Experimental setting | We experimented with a range of semantic representations, all of which are based on WordNet 2.1. |
Experimental setting | As mentioned above, words in WordNet are organised into sets of synonyms, called synsets. |
Experimental setting | Note that these are the two extremes of semantic granularity in WordNet , and we plan to experiment with intermediate representation levels in future research (c.f. |
Integrating Semantics into Parsing | Our choice for this work was the WordNet 2.1 lexical database, in which synonyms are grouped into synsets, which are then linked via an ISA hierarchy. |
Integrating Semantics into Parsing | WordNet contains other types of relations such as meronymy, but we did not use them in this research. |
Integrating Semantics into Parsing | mallet, square and steel-wool pad are also descendants of TOOL in WordNet , none of which would conventionally be used as the manner adjunct of cat). |
Introduction | We explore several models for semantic representation, based around WordNet (Fellbaum, 1998). |
Abstract | Our NR classification evaluation strictly follows the ACL SemEval-07 Task 4 datasets and protocol, obtaining an f-score of 70.6, as opposed to 64.8 of the best previous work that did not use the manually provided WordNet sense disambiguation tags. |
Experimental Setup | Nouns in this pair were manually labeled with their corresponding WordNet 3 labels and the web queries used to |
Experimental Setup | The 15 submitted systems were assigned into 4 categories according to whether they use the WordNet and Query tags (some systems were assigned to more than a single category, since they reported experiments in several settings). |
Experimental Setup | In our evaluation we do not utilize WordNet or Query tags, hence we compare ourselves with the corresponding group (A), containing 6 systems. |
Introduction | To improve results, some systems utilize additional manually constructed semantic resources such as WordNet (WN) (Beamer et al., 2007). |
Introduction | Furthermore, usage of such resources frequently requires disambiguation and connection of the data to the resource (word sense disambiguation in the case of WordNet ). |
Introduction | We evaluated our algorithm on SemEval-07 Task 4 data, showing superior results over participating algorithms that did not utilize WordNet disambiguation tags. |
Related Work | Many relation classification algorithms utilize WordNet . |
Related Work | Among the 15 systems presented by the 14 SemEval teams, some utilized the manually provided WordNet tags for the dataset pairs (e.g., (Beamer et al., 2007)). |
Results | Method P R F Acc Unsupervised clustering (4.3.3) 64.5 61.3 62.0 64.5 Cluster Labeling (4.3.1) 65.1 69.0 67.2 68.5 HITS Features (4.3.2) 69.1 70.6 70.6 70.1 Best Task 4 (no WordNet) 66.1 66.7 64.8 66.0 Best Task 4 (with WordNet ) 79.7 69.8 72.4 76.3 |
Results | Table 1 shows our results, along with the best Task 4 result not using WordNet labels (Costello, 2007). |
Abstract | Our approach leverages a similarity measure that enables the structural comparison of senses across lexical resources, achieving state-of-the-art performance on the task of aligning WordNet to three different collaborative resources: Wikipedia, Wiktionary and OmegaWiki. |
Experiments | To enable a comparison with the state of the art, we followed Matuschek and Gurevych (2013) and performed an alignment of WordNet synsets (WN) to three different collaboratively-constructed resources: Wikipedia |
Experiments | As mentioned in Section 2.1.1, we build the WN graph by including all the synsets and semantic relations defined in WordNet (e.g., hypernymy and meronymy) and further populate the relation set by connecting a synset to all the other synsets that appear in its disambiguated gloss. |
Introduction | Notable examples are WordNet , Wikipedia and, more recently, collaboratively-curated resources such as OmegaWiki and Wiktionary (Hovy et al., 2013). |
Introduction | When a lexical resource can be viewed as a semantic graph, as with WordNet or Wikipedia, this limit can be overcome by means of alignment algorithms that exploit the network structure to determine the similarity of concept pairs. |
Introduction | We report state-of-the-art performance when aligning WordNet to Wikipedia, OmegaWiki and Wiktionary. |
Related Work | A good example is WordNet , which has been exploited as a semantic network in dozens of NLP tasks (Fellbaum, 1998). |
Resource Alignment | For instance, WordNet can be readily represented as an undirected graph G whose nodes are synsets and edges are modeled after the relations between synsets defined in WordNet (e. g., hypernymy, meronymy, etc. |
Resource Alignment | Any semantic network with a dense relational structure, providing good coverage of the words appearing in the definitions, is a suitable candidate for H. For this purpose we used the WordNet (Fellbaum, 1998) graph which was further enriched by connecting |
Resource Alignment | As an example, assume we are given two semantic signatures computed for two concepts in WordNet and Wiktionary. |
Abstract | We conduct a thorough evaluation of the proposed methodology both manually as well as through comparison with WordNet . |
Abstract | Remarkably, in 44% cases the birth of a novel sense is attested by WordNet, while in 46% cases and 43% cases split and join are respectively confirmed by WordNet . |
Evaluation framework | 6.2 Automated evaluation with WordNet |
Evaluation framework | We chose WordNet for automated evaluation because not only does it have a wide coverage of word senses but also it is being maintained and updated regularly to incorporate new senses. |
Evaluation framework | For our evaluation, we developed an aligner to align the word clusters obtained with WordNet senses. |
Introduction | Remarkably, comparison with the English WordNet indicates that in 44% cases, as identified by our algorithm, there has been a birth of a completely novel sense, in 46% cases a new sense has split off from an older sense and in 43% cases two or more older senses have merged in to form a new sense. |
Related work | A few approaches suggested by (Bond et al., 2009; Paakko' and Linden, 2012) attempt to augment WordNet synsets primarily using methods of annotation. |
Abstract | Although handcrafted lexical resources, such as WordNet, could provide more reliable related terms, previous studies showed that query expansion using only WordNet leads to very limited performance improvement. |
Introduction | Intuitively, compared with co-occurrence-based thesauri, handcrafted thesauri, such as WordNet , could provide more reliable terms for query expansion. |
Introduction | However, previous studies failed to show any significant gain in retrieval performance when queries are expanded with terms selected from WordNet (Voorhees, 1994; Stairmand, 1997). |
Introduction | In this paper, we study several term similarity functions that exploit various information from two lexical resources, i.e., WordNet |
Related Work | Although the use of WordNet in query expansion has been studied by various researchers, the improvement of retrieval performance is often limited. |
Related Work | Voorhees (Voorhees, 1994) expanded queries using a combination of synonyms, hypemyms and hyponyms manually selected from WordNet , and achieved limited improvement (i.e., around —2% to |
Related Work | Stairmand (Stairmand, 1997) used WordNet for query expansion, but they concluded that the improvement was restricted by the coverage of the WordNet and no empirical results were reported. |
Abstract | We examine the differences in content between the 1911 and 1987 versions of Roget’s, and we test both versions with each other and WordNet on problems such as synonym identification and word relatedness. |
Abstract | We also present a novel method for measuring sentence relatedness that can be implemented in either version of Roget’s or in WordNet . |
Abstract | Although the 1987 version of the Thesaurus is better, we show that the 1911 version performs surprisingly well and that often the differences between the versions of R0-get’s and WordNet are not statistically significant. |
Introduction | We compare two versions, the 1987 and 1911 editions of the Thesaurus with each other and with WordNet 3.0. |
Introduction | Roget’s Thesaurus has a unique structure, quite different from WordNet , of which the NLP community has yet to take full advantage. |
Introduction | In this paper we demonstrate that although the 1911 version of the Thesaurus is very old, it can give results comparable to systems that use WordNet or newer versions of Roget’s Thesaurus. |
Evaluation | WordNet ? |
Evaluation | Table 1: Class labels found in WordNet in original form, or found in WordNet after removal of leading words, or not found in WordNet at all |
Evaluation | Accuracy of Class Labels: Built over many years of manual construction efforts, lexical gold standards such as WordNet (Fellbaum, 1998) provide wide-coverage upper ontologies of the English language. |
Introduction | WordNet (Fellbaum, 2010) has been used in numerous natural language processing tasks such as word sense disambiguation and information extraction with considerable success. |
Introduction | Kurdish is a less-resourced language for which, among other resources, no wordnet has been built yet. |
KurdNet: State-of-the-Art | 1. highlighted the main challenges in building a wordnet for the Kurdish language (including its inherent diversity and morphological complexity), |
KurdNet: State-of-the-Art | 2. built the first prototype of KurdNet, the Kurdish WordNet (see a summary below), and |
KurdNet: State-of-the-Art | There are two well-known models for building wordnets for a language (Vossen, 1998): |
Background and Related Work | Typically, word frequency distributions are estimated with respect to a sense-tagged corpus such as SemCor (Miller et al., 1993), a 220,000 word corpus tagged with WordNet (Fellbaum, 1998) senses. |
Background and Related Work | The distributional similarity scores of the nearest neighbours are associated with the respective target word senses using a WordNet similarity measure, such as those proposed by J iang and Conrath (1997) and Banerjee and Pedersen (2002). |
Introduction | (2004b) to remove low-frequency senses from WordNet , we focus on finding senses that are unattested in the corpus on the premise that, given accurate disambiguation, rare senses in a corpus contribute to correct interpretation. |
Macmillan Experiments | For the purposes of this research, the choice of Macmillan is significant in that it is a conventional dictionary with sense definitions and examples, but no linking between senses.11 In terms of the original research which gave rise to the sense-tagged dataset, Macmillan was chosen over WordNet for reasons including: (l) the well-documented difficulties of sense tagging with fine-grained WordNet senses (Palmer et al., 2004; Navigli et al., 2007); (2) the regular update cycle of Macmillan (meaning it contains many recently-emerged senses); and (3) the finding in a preliminary sense-tagging task that it better captured Twitter usages than WordNet (and also OntoNotes: Hovy et al. |
Macmillan Experiments | The average sense ambiguity of the 20 target nouns in Macmillan is 5.6 (but 12.3 in WordNet ). |
Macmillan Experiments | We first notice that, despite the coarser-grained senses of Macmillan as compared to WordNet , the upper bound WSD accuracy using Macmillan is comparable to that of the WordNet-based datasets over the balanced BNC, and quite a bit lower than that of the two domain corpora of Koeling et al. |
Methodology | the WordNet hierarchy). |
WordNet Experiments | For each domain, annotators were asked to sense-annotate a random selection of sentences for each of 40 target nouns, based on WordNet v1.7. |
WordNet Experiments | For each dataset, we use HDP to induce topics for each target lemma, compute the similarity between the topics and the WordNet senses (Equation (1)), and rank the senses based on the prevalence scores (Equation (2)). |
WordNet Experiments | It is important to bear in mind that MKWC in these experiments makes use of full-text parsing in calculating the distributional similarity thesaurus, and the WordNet graph structure in calculating the similarity between associated words and different senses. |
Empirical Evaluation: Simile-derived Representations | Almuhareb and Poesio (2004) used as their experimental basis a sampling of 214 English nouns from 13 of WordNet’s upper-level semantic categories, and proceeded to harvest adjectival features for these noun-concepts from the web using the textual pattern “[a | an | the] * C [is | was]”. |
Harvesting Knowledge from Similes: English and Chinese | Veale and Hao (2007) use the Google API in conjunction with Princeton WordNet (Fellbaum, 1998) as the basis of their harvesting system. |
Harvesting Knowledge from Similes: English and Chinese | They first extracted a list of antonymous adjectives, such as “hot” or “cold”, from WordNet , the intuition being that explicit similes will tend to exploit properties that occupy an exemplary point on a scale. |
Harvesting Knowledge from Similes: English and Chinese | To harvest a comparable body of Chinese similes from the web, we also use the Google API, in conjunction with both WordNet and HowNet (Dong and Dong, 2006). |
Related Work | (1999), in which each of the textual glosses in WordNet (Fellbaum, 1998) is linguistically analyzed to yield a sense-tagged logical form, is an example of the former approach. |
Related Work | Almuhareb and Poesio go on to demonstrate that the values and attributes that are found for word-concepts on the web yield a sufficiently rich representation for these word-concepts to be automatically clustered into a form resembling that assigned by WordNet (see Fellbaum, 1998). |
Tagging and Mapping of Similes | In the case of English similes, Veale and Hao (2007) describe how two English similes “as A as N1” and “as A as N2” will be mutually disambiguating if N1 and N2 are synonyms in WordNet, or if some sense of N1 is a hypernym or hyponym of some sense of N2 in WordNet . |
Tagging and Mapping of Similes | For instance, though HowNet has a much shallower hierarchical organization than WordNet , it compensates by encapsulating the meaning of different word senses using simple logical formulae of semantic primitives, or sememes, that are derived from the meaning of common Chinese characters. |
Tagging and Mapping of Similes | WordNet and HowNet thus offer two complementary levels or granularities of generalization that can be exploited as the context demands. |
Abstract | Manually constructing a Wordnet is a difficult task, needing years of experts’ time. |
Abstract | As a first step to automatically construct full Wordnets, we propose approaches to generate Wordnet synsets for languages both resource-rich and resource-poor, using publicly available Wordnets , a machine translator and/or a single bilingual dictionary. |
Abstract | Our algorithms translate synsets of existing Wordnets to a target language T, then apply a ranking method on the translation candidates to find best translations in T. Our approaches are applicable to any language which has at least one existing bilingual dictionary translating from English to it. |
Introduction | Wordnets are intricate and substantive repositories of lexical knowledge and have become important resources for computational processing of natural languages and for information retrieval. |
Introduction | Good quality Wordnets are available only for a few "resource-rich" languages such as English and Japanese. |
Introduction | Published approaches to automatically build new Wordnets are manual or semiautomatic and can be used only for languages that already possess some lexical resources. |
Abstract | The key aspect of our method is that it is the first unified approach that assigns the polarity of both word- and sense-level connotations, exploiting the innate bipartite graph structure encoded in WordNet . |
Conclusion | We have introduced a novel formulation of lexicon induction operating over both words and senses, by exploiting the innate structure between the words and senses as encoded in WordNet . |
Evaluation 11: Human Evaluation on ConnotationWordNet | 7Because senses in WordNet can be tricky to understand, care should be taken in designing the task so that the Turkers will focus only on the corresponding sense of a word. |
Evaluation 11: Human Evaluation on ConnotationWordNet | Therefore, we provided the part of speech tag, the WordNet gloss of the selected sense, and a few examples as given in WordNet . |
Introduction | We introduce ConnotationWordNet, a connotation lexicon over the network of words in conjunction with senses, as defined in WordNet . |
Introduction | For example, consider “abound”, for which lexicographers of WordNet prescribe two different senses: |
Introduction | Especially if we look up the WordNet entry for “bristle”, there are noticeably more negatively connotative words involved in its gloss and examples. |
Network of Words and Senses | Another benefit of our approach is that for various WordNet relations (e.g., antonym relations), which are defined over synsets (not over words), we can add edges directly between corresponding synsets, rather than projecting (i.e., approximating) those relations over words. |
Abstract | This paper examines the case for a graded notion of word meaning in two experiments, one which uses WordNet senses in a graded fashion, contrasted with the “winner takes all” annotation, and one which asks annotators to judge the similarity of two usages. |
Analyses | In the WSsim experiment, annotators rated the applicability of each WordNet 3.0 sense for a given target word occurrence. |
Analyses | In WordNet , they have 5, 7, and 4 senses, respectively. |
Annotation | WSsim is a word sense annotation task using WordNet senses.5 Unlike previous word sense annotation projects, we asked annotators to provide judgments on the applicability of every WordNet sense of the target lemma with the instruction: 6 |
Annotation | 3The SemCor dataset was produced alongside WordNet, so it can be expected to support the WordNet sense distinctions. |
Annotation | 5WordNet 1.7.1 was used in the annotation of both SE-3 and SemCor; we used the more current WordNet 3.0 after verifying that the lemmas included in this experiment had the same senses listed in both versions. |
Introduction | In the first one, referred to as WSsim (Word Sense Similarity), annotators give graded ratings on the applicability of WordNet senses. |
Introduction | The first study additionally tests to what extent the judgments on WordNet senses fall into clearcut clusters, while the second study allows us to explore meaning similarity independently of any lexicon resource. |
Related Work | Reported inter-annotator agreement (ITA) for fine-grained word sense assignment tasks has ranged between 69% (Kilgarriff and Rosenzweig, 2000) for a lexical sample using the HECTOR dictionary and 78.6% using WordNet (Landes et al., 1998) in all-words annotation. |
Related Work | Although we use WordNet for the annotation, our study is not a study of WordNet per se. |
Related Work | We choose WordNet because it is sufficiently fine- grained to examine subtle differences in usage, and because traditionally annotated datasets exist to which we can compare our results. |
Comparative Evaluation | As regards recall, we note that in two cases (i.e., DBpedia returning page super-types from its upper taxonomy, YAGO linking categories to WordNet synsets) the generalizations are neither pages nor categories and that MENTA returns heterogeneous hypernyms as mixed sets of WordNet synsets, Wikipedia pages and categories. |
Comparative Evaluation | MENTA seems to be the closest resource to ours, however, we remark that the hypernyms output by MENTA are very heterogeneous: 48% of answers are represented by a WordNet synset, 37% by Wikipedia categories and 15% are Wikipedia pages. |
Introduction | However, unlike the case with smaller manually-curated resources such as WordNet (Fellbaum, 1998), in many large automatically-created resources the taxonomical information is either missing, mixed across resources, e.g., linking Wikipedia categories to WordNet synsets as in YAGO, or coarse-grained, as in DBpedia whose hypernyms link to a small upper taxonomy. |
Introduction | (2005) provide a general vector-based method which, however, is incapable of linking pages which do not have a WordNet counterpart. |
Introduction | Higher coverage is provided by de Melo and Weikum (2010) thanks to the use of a set of effective heuristics, however, the approach also draws on WordNet and sense frequency information. |
Related Work | However, these methods do not link terms to existing knowledge resources such as WordNet , whereas those that explicitly link do so by adding new leaves to the existing taxonomy instead of acquiring wide-coverage taxonomies from scratch (Pan-tel and Ravichandran, 2004; Snow et al., 2006). |
Related Work | Other approaches, such as YAGO (Suchanek et al., 2008; Hoffart et al., 2013), yield a taxonomical backbone by linking Wikipedia categories to WordNet . |
Related Work | However, the categories are linked to the first, i.e., most frequent, sense of the category head in WordNet , involving only leaf categories in the linking. |
Abstract | This paper presents experiments with WordNet semantic classes to improve dependency parsing. |
Experimental Framework | Base WordNet WordNet Clusters |
Experimental Framework | WordNet . |
Experimental Framework | (2011), based on WordNet 2.1. |
Introduction | Broadly speaking, we can classify the methods to incorporate semantic information into parsers in two: systems using static lexical semantic repositories, such as WordNet or similar ontologies (Agirre et al., 2008; Agirre et al., 2011; Fujita et al., 2010), and systems using dynamic semantic clusters automatically acquired from corpora (Koo et al., 2008; Suzuki et al., 2009). |
Introduction | 0 Does semantic information in WordNet help |
Introduction | 0 How does WordNet compare to automatically obtained information? |
Related work | Broadly speaking, we can classify the attempts to add external knowledge to a parser in two sets: using large semantic repositories such as WordNet and approaches that use information automatically acquired from corpora. |
Related work | The results showed a signi-cant improvement, giving the first results over both WordNet and the Penn Treebank (PTB) to show that semantics helps parsing. |
Related work | (201 1) successfully introduced WordNet classes in a dependency parser, obtaining improvements on the full PTB using gold POS tags, trying different combinations of semantic classes. |
Abstract | To train the system, we extract substructures of WordNet and dis-criminatively learn to reproduce them, using adaptive subgradient stochastic optimization. |
Abstract | On the task of reproducing sub-hierarchies of WordNet , our approach achieves a 51% error reduction over a chance baseline, including a 15% error reduction due to the non-hypernym—factored sibling features. |
Experiments | We considered two distinct experimental setups, one that illustrates the general performance of our model by reproducing various medium-sized WordNet domains, and another that facilitates comparison to previous work by reproducing the much larger animal subtree provided by Kozareva and Hovy (2010). |
Experiments | General setup: In order to test the accuracy of structured prediction on medium-sized full-domain taxonomies, we extracted from WordNet 3.0 all bottomed-out full subtrees which had a tree-height of 3 (i.e., 4 nodes from root to leaf), and contained (10, 50] terms.11 This gives us 761 non-overlapping trees, which we partition into |
Experiments | To project WordNet synsets to terms, we used the first (most frequent) term in each synset. |
Introduction | However, currently available taxonomies such as WordNet are incomplete in coverage (Pennacchiotti and Pantel, 2006; Hovy et al., 2009), unavailable in many domains and languages, and |
Introduction | Figure 1: An excerpt of WordNet’s vertebrates taxonomy. |
Introduction | First, on the task of recreating fragments of WordNet , we achieve a 51% error reduction on ancestor-based F1 over a chance baseline, including a 15% error reduction due to the non-hypernym-factored sibling features. |
Parallel Datasets | In order to obtain parallel training data for the translation models, we collected three different datasets: manually-tagged question reformulations and question-answer pairs from the WikiAnswers social Q&A site (Section 3.1), and glosses from WordNet , Wiktionary, Wikipedia and Simple Wikipedia (Section 3.2). |
Parallel Datasets | 0 Wordnet (sense I): the natural satellite of the Earth. |
Parallel Datasets | o WordNet (Fellbaum, 1998). |
Related Work | Murdock and Croft (2005) created a first parallel corpus of synonym pairs extracted from WordNet , and an additional parallel corpus of English words translating to the same Arabic term in a parallel English-Arabic corpus. |
Related Work | Knowledge-based measures rely on lexical semantic resources such as WordNet and comprise path length based measures (Rada et al., 1989) and concept vector based measures (Qiu and Frei, 1993). |
Semantic Relatedness Experiments | The method consists in representing words as a concept vector, where concepts correspond to WordNet synsets, Wikipedia article titles or Wiktionary entry names. |
Semantic Relatedness Experiments | glosses in WordNet , the full article or the first paragraph of the article in Wikipedia or the full contents of a Wiktionary entry. |
Semantic Relatedness Experiments | WordNet .26 .46 Wikipedia .27 .03 WikipediaFirst .30 .38 Wiktionary .39 .58 Translation probabilities |
Experiments and evaluation | Table 2 shows the results of the evaluation of our initial thesaurus, achieved by comparing the selected semantic neighbors with two complementary reference resources: WordNet 3.0 synonyms (Miller, 1990) [W], which characterize a semantic similarity based on paradigmatic relations, and the Moby thesaurus (Ward, 1996) [M], which gathers a larger set of types of relations and is more representative of semantic relatedness3. |
Experiments and evaluation | WordNet provides a restricted number of synonyms for each noun while the Moby thesaurus contains for each entry a large number of synonyms and similar words. |
Experiments and evaluation | As a consequence, the precisions at different cutoffs have a significantly higher value with Moby as reference than with WordNet as reference. |
Introduction | The second approach makes use of a less structured source of knowledge about words such as the definitions of classical dictionaries or the glosses of WordNet . |
Introduction | WordNet’s glosses were used to support Lesk-like measures in (Banerjee and Pedersen, 2003) and more recently, measures were also defined from Wikipedia or Wiktionaries (Gabrilovich and |
Method | where: infll and inflg are inflected variants of nounl and noung generated using the Java WordNet Libraryl; THAT is a complementizer and can be that, which, or who; and * stands for 0 or more (up to 8) instances of Google’s star operator. |
Method | Finally, we lemmatize the main verb using WordNet’s morphological analyzer Morphy (Fellbaum, 1998). |
Related Work | (2005) apply both classic (SVM and decision trees) and novel supervised models (semantic scattering and iterative semantic specialization), using WordNet , word sense disambiguation, and a set of linguistic features. |
Related Work | Their approach is highly resource intensive (uses WordNet , CoreLex and Moby’s thesaurus), and is quite sensitive to the seed set of verbs: on a collection of 453 examples and 19 relations, they achieved 52.6% accuracy with 84 seed verbs, but only 46.7% with 57 seed verbs. |
Relational Similarity Experiments | We further experimented with the SemEval’07 task 4 dataset (Girju et al., 2007), where each example consists of a sentence, a target semantic relation, two nominals to be judged on whether they are in that relation, manually annotated WordNet senses, and the Web query used to obtain the sentence: |
Relational Similarity Experiments | WordNet(el) = "vessel%l:06:OO::", WordNet(e2) = "tool%l:O6:OO::", Content—Container(e2, el) = "true", Query = "contents of the * were a" |
Relational Similarity Experiments | The SemEval competition defines four types of systems, depending on whether the manually annotated WordNet senses and the Google query are used: A (WordNet=no, Query=no), B (WordNet=yes, Query=no), C (WordNet=no, Query=yes), and D (WordNet=yes, Query=yes). |
Background | (2007) compute the semantic similarity using WordNet . |
Background | Its definition is much larger than that of WordNet . |
Experiments | We ap—p1y standard DCL in Run 2, where concepts are determined according to their definitions in WordNet (Ye et a1., 2007). |
Experiments | When we merge diverse words having similar semantic according to WordNet concepts , we obtain 873 concepts per article on average in Run 2. |
Experiments | Even though the reduction of total concepts is limited, these new wiki concepts will group the terms that cannot be detected by WordNet . |
Introduction | Its results are better than other external resources such as WordNet , Gazetteers and Google’s define operator, especially for definition QA (Lita et al., 2004). |
Introduction | Most of these concepts are multi—word terms, whereas WordNet has only 50,000 plus multi—word terms. |
Introduction | Any term could appear in the definition of a concept if necessary, while the total vocabulary existing in WordNet’s glossary definition is less than 2000. |
Experiments | Finally, we report on the quality of a large database of Wordnet-based preferences obtained after manually associating our topics with Wordnet classes (Section 4.4). |
Experiments | We then automatically filtered out any rules which contained a negation, or for which the antecedent and consequent contained a pair of antonyms found in WordNet (this left us with 85 rules). |
Experiments | Table 3: Top 10 and Bottom 10 ranked inference rules ranked by LDA-SPafter automatically filtering out negations and antonyms (using WordNet ). |
Introduction | Resnik (1996) presented the earliest work in this area, describing an information-theoretic approach that inferred selectional preferences based on the WordNet hypernym hierarchy. |
Introduction | This avoids problems like WordNet’s poor coverage of proper nouns and is shown to improve performance. |
Previous Work | WordNet ), or automatically generated (Pantel, 2003). |
Automatic Metaphor Interpretation | Talking Points are a set of characteristics of concepts belonging to source and target domains and related facts about the world which the authors acquire automatically from WordNet and from the web. |
Automatic Metaphor Recognition | They mine WordNet (Fellbaum, 1998) for the examples of systematic polysemy, which allows to capture metonymic and metaphorical relations. |
Automatic Metaphor Recognition | The authors search for nodes that are relatively high up in the WordNet hierarchy and that share a set of common word forms among their descendants. |
Automatic Metaphor Recognition | They use hyponymy relation in WordNet and word bigram counts to predict metaphors at a sentence level. |
Metaphor Annotation in Corpora | A number of metaphorical senses are included in WordNet , however without any accompanying semantic annotation. |
Metaphor Resources | 5EuroWordNet is a multilingual database with wordnets for several European languages (Dutch, Italian, Spanish, German, French, Czech and Estonian). |
Metaphor Resources | The wordnets are structured in the same way as the Princeton WordNet for English. |
Metaphor Resources | WordNet (Lonneker and Eilts, 2004), would undoubtedly provide a new platform for experiments and enable researchers to directly compare their results. |
Methodology | mantic categories originating in WordNet . |
Methodology | 7Supersenses are called “lexicographer classes” in WordNet documentation (Fellbaum, 1998), http: / /worolnet . |
Methodology | English adjectives do not, as yet, have a similar high-level semantic partitioning in WordNet , thus we use a 13-class taxonomy of adjective supersenses constructed by Tsvetkov et al. |
Model and Feature Extraction | WordNet lacks coarse-grained semantic categories for adjectives. |
Model and Feature Extraction | For example, the top-level classes in GermaNet include: adj.feeling (e.g., willing, pleasant, cheerful); adj.sabstance (e.g., dry, ripe, creamy); adj.spatial (e.g., adjacent, gigantic).12 For each adjective type in WordNet , they produce a vector with a classifier posterior probabilities corresponding to degrees of membership of this word in one of the 13 semantic classes,13 similar to the feature vectors we build for nouns and verbs. |
Model and Feature Extraction | Consider an example related to projection of WordNet supersenses. |
Related Work | (2013) describe a Concrete Category Overlap algorithm, where co-occurrence statistics and Turney’s abstractness scores are used to determine WordNet supersenses that correspond to literal usage of a given adjective or verb. |
Related Work | To implement this idea, they extend MRC imageability scores to all dictionary words using links among WordNet supersenses (mostly hypernym and hyponym relations). |
Related Work | Because they heavily rely on WordNet and availability of imageability scores, their approach may not be applicable to low-resource languages. |
Abstract | This study presents a novel approach to the problem of system portability across different domains: a sentiment annotation system that integrates a corpus-based classifier trained on a small set of annotated in-domain data and a lexicon-based system trained on WordNet . |
Introduction | In this paper, we present a novel approach to the problem of system portability across different domains by developing a sentiment annotation system that integrates a corpus-based classifier with a lexicon-based system trained on WordNet . |
Introduction | The information contained in lexicographical sources, such as WordNet , reflects a lay person’s general knowledge about the world, while domain-specific knowledge can be acquired through classifier training on a small set of in-domain data. |
Introduction | The final, third part of the paper presents our system, composed of an ensemble of two classifiers —one trained on WordNet glosses and synsets and the other trained on a small in-domain training set. |
Lexicon-Based Approach | A lexicon-based approach capitalizes on the fact that dictionaries, such as WordNet (Fellbaum, 1998), contain a comprehensive and domain-independent set of sentiment clues that exist in general English. |
Lexicon-Based Approach | One of the limitations of general lexicons and dictionaries, such as WordNet (Fellbaum, 1998), as training sets for sentiment tagging systems is that they contain only definitions of individual words and, hence, only unigrams could be effectively learned from dictionary entries. |
Lexicon-Based Approach | Since the structure of WordNet glosses is fairly different from that of other types of corpora, we developed a system that used the list of human-annotated adjectives from (Hatzivassiloglou and McKeown, 1997) as a seed list and then learned additional unigrams |
Conclusion and Future Work | For verb classes, in particular, we devised a method for classifying all the verbs and verb phrases in WordNet into the activity and state classes. |
Conclusion and Future Work | The experimental results show that verb and verb phrase classification method is reasonably accurate with 91% precision and 78% recall with manually constructed gold standard consisting of 80 verbs and 82% accuracy for a random sample of all the WordNet entries. |
Experience Detection | We collected all hyponyms of words “do” and “act”, from WordNet (Fellbaum, 1998). |
Experience Detection | Lastly, we removed all the verbs that are under the hierarchy of “move” from WordNet . |
Experience Detection | In addition, the lexicon we constructed for the baseline (i.e., using the WordNet ) contains more errors than our activity lexicon for activity verbs. |
Lexicon Construction | We consider all the verbs and verb phrases in WordNet (Fellbaum, 1998) which is the largest electronic lexical database. |
Lexicon Construction | Based on the query matrix in table 2, we issued queries for all the verbs and verb phrases from WordNet to a search engine. |
Lexicon Construction | We finally trained our model with the top 10 features and classified all WordNet verbs and verb phrases. |
Related Work | Since our work is specifically geared toward domain-independent experience detection, we attempted to maximize the coverage by using all the verbs in WordNet , as opposed to the verbs appearing in a particular domain-specific corpus (e.g., medicine domain) as done in the previous work. |
Abstract | Frequently, such resources are constructed through automatic mergers of complementary resources, such as WordNet and Wikipedia. |
Introduction | Semantic knowledge bases such as WordNet (Fellbaum, 1998), YAGO (Suchanek et al., 2007), and BabelNet (Navigli and Ponzetto, 2010) provide ontological structure that enables a wide range of tasks, such as measuring semantic relatedness (Budanitsky and Hirst, 2006) and similarity (Pilehvar et al., 2013), paraphrasing (Kauchak and Barzilay, 2006), and word sense disambiguation (Navigli and Ponzetto, 2012; Moro et al., 2014). |
Introduction | extend WordNet using distributional or structural features to identify novel semantic connections between concepts. |
Introduction | The recent advent of large semistructured resources has enabled the creation of new semantic knowledge bases (Medelyan et al., 2009; Hovy et al., 2013) through automatically merging WordNet and Wikipedia (Suchanek et al., 2007; Navigli and Ponzetto, 2010; Nie-mann and Gurevych, 2011). |
Related Work | Rzeniewicz and Szymanski (2013) extend WordNet with commonsense knowledge using a 20 Questions-like game. |
Video Game with a Purpose Design | Knowledge base As the reference knowledge base, we chose BabelNet2 (Navigli and Ponzetto, 2010), a large-scale multilingual semantic ontology created by automatically merging WordNet with other collaboratively-constructed resources such as Wikipedia and OmegaWiki. |
Video Game with a Purpose Design | First, by connecting WordNet synsets to Wikipedia pages, most synsets are associated with a set of pictures; while often noisy, these pictures sometimes illustrate the target concept and are an ideal case for validation. |
Video Game with a Purpose Design | Second, BabelNet contains the semantic relations from both WordNet and hyperlinks in Wikipedia; these relations are again an ideal case of validation, as not all hyperlinks connect semantically-related pages in Wikipedia. |
Experiments | Many of these templates utilize information from WordNet (Fell-baum, 1998). |
Experiments | 0 WordNet link types (link type list) (e.g., attribute, hypernym, entailment) |
Experiments | 0 Lexicographer filenames (lexnames)—top level categories used in WordNet (e.g., noun.body, verb.cognition) |
Abstract | The features make extensive use of WordNet . |
Assessment of Lexical Resources | Since the PDEP system enables exploration of features from WordNet , FrameNet, and VerbNet, we are able to make some assessment of these resources. |
Assessment of Lexical Resources | WordNet played a statistically significant role in the systems developed by Tratz (2011) and Srikumar and Roth (2013). |
Assessment of Lexical Resources | This includes the WordNet lexicographer’s file name (e.g., noun.time), synsets, and hypernyms. |
Class Analyses | We are examining the WordNet detour to FrameNet, as described in Burchardt et al. |
Introduction | Section 4 describes how we are able to investigate the relationship of WordNet , FrameNet, and VerbNet to this effort and how this examination of preposition behavior can be used in working with these resources. |
See http://clg.wlv.ac.uk/proiects/DVC | The feature extraction rules are (1) word class (we), (2) part of speech (pos), (3) lemma (1), (4) word (w), (5) WordNet lexical name (In), (6) WordNet synonyms (s), (7) WordNet hypernyms (h), (8) whether the word is capitalized (c), and (9) affixes (af). |
See http://clg.wlv.ac.uk/proiects/DVC | For features such as the WordNet lexical name, synonyms and hypernyms, the number of values may be much larger. |
Comparison to Prior Work | Snow (Snow et al., 2006) has extended the WordNet 2.1 by adding thousands of entries (synsets) at a relatively high precision. |
Comparison to Prior Work | They have made several versions of extended WordNet available4. |
Comparison to Prior Work | For the experimental comparison, we focused on leaf semantic classes from the extended WordNet that have many hypernyms, so that a meaningful comparison could be made: specifically, we selected nouns that have at least three hypernyms, such that the hypernyms are the leaf nodes in the hypernym hierarchy of WordNet . |
Introduction | We also compare ASIA on twelve additional benchmarks to the extended Wordnet 2.1 produced by Snow et al (Snow et al., 2006), and show that for these twelve sets, ASIA produces more than five times as many set instances with much higher precision (98% versus 70%). |
Introduction | (Pennacchiotti and Pantel, 2006) proposed an algorithm for automatically ontologizing semantic relations into WordNet . |
Introduction | However, despite its high precision entries, WordNet’s limited coverage makes it impossible for relations whose arguments are not present in WordNet to be incorporated. |
Related Work | They mapped each argument of the relation into WordNet and identified the senses for which the relation holds. |
Related Work | Unfortunately, despite its very high precision entries, WordNet is known to have limited coverage, which makes it impossible for algorithms to map the content of a relation whose arguments are not present in WordNet . |
Related Work | To surmount this limitation, we do not use WordNet , but employ a different method of obtaining superclasses of a filler term: the inverse doubly-anchored patterns DAP‘1 (Hovy et al., 2009), which, given two arguments, harvests its supertypes from the source corpus. |
Results | Since our problem definition differs from available related work, and WordNet does not contain all harvested arguments as shown in (Hovy et al., 2009), it is not possible to make a direct comparison. |
Experimental Evaluation | html#wn) to WordNet (Miller, 1995) for lemmatization information. |
Experimental Evaluation | We also tried ablating the WordNet relations, and observed that the “identical-word” feature hurt the model the most. |
Experimental Evaluation | 10This is accomplished by eliminating lines 12 and 13 from the definition of pm and redefining pword to be the unigram word distribution estimated from the Gigaword corpus, as in G0, without the help of WordNet . |
Introduction | Because dependency syntax is still only a crude approximation to semantic structure, we augment the model with a lexical semantics component, based on WordNet (Miller, 1995), that models how words are probabilistically altered in generating a paraphrase. |
QG for Paraphrase Modeling | WordNet relation(s) The model next chooses a lexical semantics relation between 3360-) and the yet-to-be-chosen word ti (line 12). |
QG for Paraphrase Modeling | (2007),6 we employ a 14-feature log-linear model over all logically possible combinations of the 14 WordNet relations (Miller, 1995).7 Similarly to Eq. |
QG for Paraphrase Modeling | 14, we normalize this log-linear model based on the set of relations that are nonempty in WordNet for the word 3360-). |
Abstract | This paper presents a set of Bayesian methods for automatically extending the WORDNET ontology with new concepts and annotating existing concepts with generic property fields, or attributes. |
Abstract | We base our approach on Latent Dirichlet Allocation and evaluate along two dimensions: (l) the precision of the ranked lists of attributes, and (2) the quality of the attribute assignments to WORDNET concepts. |
Conclusion | This paper introduced a set of methods based on Latent Dirichlet Allocation (LDA) for jointly extending the WORDNET ontology and annotating its concepts with attributes (see Figure 4 for the end result). |
Introduction | We present a Bayesian approach for simultaneously extending IsA hierarchies such as those found in WORDNET (WN) (Fellbaum, 1998) with additional concepts, and annotating the resulting concept graph with attributes, i.e., generic property fields shared by instances of that concept. |
Introduction | We use WORDNET 3.0 as the specific test ontology for our annotation procedure, and evalu- |
Related Work | A large body of previous work exists on extending WORDNET with additional concepts and instances (Snow et al., 2006; Suchanek et al., 2007); these methods do not address attributes directly. |
Evaluation | Precisions are simply omitted because the difference to the recalls are always the number of failures on referring to WordNet by mislabeling of lemmata or POSs, which is always the same for the three methods. |
Metric Space Implementation | To calculate sense similarities, we used the WordNet similarity package by Pedersen et al. |
Metric Space Implementation | Those texts were parsed using RASP parser (Briscoe et al., 2006) version 3.1, to obtain grammatical relations for the distributional similarity, as well as to obtain lemmata and part-of-speech (POS) tags which are required to look up the sense inventory of WordNet . |
Related Work | (2010) used a WordNet pre-pruning. |
Related Work | Disambiguation is performed by considering only those candidate synsets that belong to the top-k largest connected components of the WordNet on domain corpus. |
Simultaneous Optimization of All-words WSD | The cluster centers are located at the means of hypotheses including miscellaneous alternatives not intended, thus the estimated probability distribution is, roughly speaking, offset toward the center of WordNet , which is not what we want. |
Abstract | We examine a number of variations of the method, including the addition of WordNet , partial matching, or removing relation labels from the dependencies. |
Dependency-based evaluation | This makes sense if we take into consideration that WordNet lists all possible synonyms for all possible senses of a word, and so, given a great number of cross-sentence comparisons in multi-sentence summaries, there is an increased risk of spurious matches between words which, despite being potentially synonymous in certain contexts, are not equivalent in the text. |
Discussion and future work | Three such variants are the baseline DEPEVAL(summ), the WordNet version DEPEVAL(summ) wn, and the version with removed relation labels DEPEVAL(summ) norel. |
Discussion and future work | The new implementation of BE presented at the TAC 2008 workshop (Tratz and Hovy, 2008) introduces transformations for dependencies in order to increase the number of matches among elements that are semantically similar yet differ in terms of syntactic structure and/or lexical choices, and adds WordNet for synonym matching. |
Discussion and future work | Since our method, presented in this paper, also uses the reranking parser, as well as WordNet , it would be interesting to compare both methods directly in terms of the performance of the dependency extraction procedure. |
Experimental results | The correlations are listed for the following versions of our method: pm - partial matching for dependencies; wn - WordNet ; pred - matching predicate-only dependencies; norel - ignoring dependency relation label; one - counting a match only once irrespective of how many instances of |
Related Work | These features are typically obtained from external resources such as Wordnet (Miller, 1990). |
Related Work | The first baseline searches for each head noun in WordNet and labels the noun as category Ck, if it has a hypernym synset corresponding to that category. |
Related Work | We manually identified the WordNet synsets that, to the best of our ability, seem to most closely correspond |
Experiments | from dependency parse tree) along with computing similarity in semantic spaces (using WordNet ) clearly produces an improvement in the summarization quality (+1.4 improvement in ROUGE-l F-score). |
Using the Framework | For each pair of nodes (u,v) in the graph, we compute the semantic similarity score (using WordNet ) between every pair of dependency relation (rel: a, b) in u and v as: s(u,v) = Z WN(a,-,aj) >< WN(b,-,bj), |
Using the Framework | WN(w,—, wj) is defined as the WordNet similarity score between words 212,- and to]? |
Using the Framework | 2There exists various semantic relatedness measures based on WordNet (Patwardhan and Pedersen, 2006). |
Learning Class Attributes | For these approaches, lists of instances are typically collected from publicly-available resources such as WordNet or Wikipedia (Pasca and Van Durme, 2007; |
Learning Class Attributes | 1Reisinger and Pasca (2009) considered the related problem of finding the most appropriate class for each attribute; they take an existing ontology of concepts ( WordNet ) as a class hierarchy and use a Bayesian approach to decide “the correct level of abstraction for each attribute.” |
Related Work | These efforts focused exclusively on the meronymy relation as used in WordNet (Miller et al., 1990). |
Related Work | Experts can manually specify the attributes of entities, as in the WordNet project (Miller et al., 1990). |
Related Work | In many ways WordNet can be regarded as a collection of commonsense relationships. |
Conclusion | We further showed that a weakly supervised heuristic, making use of WordNet sense ranks, can be significantly improved by incorporating information from our system. |
Experiment: Ranking Word Senses | WordNet (Fellbaum, 1998) senses of w apply to this occurrence of w. |
Experiment: Ranking Word Senses | The dataset contains ordinal judgments of the applicability of WordNet senses on a 5 point scale, ranging from completely difi‘er—ent to identical for eight different lemmas in 50 different sentential contexts. |
Experiment: Ranking Word Senses | We apply the same method as in Section 4.3: For each instance in the dataset, we compute the second-order vector of the target verb, contextually constrain it by the first-order vectors of the verb’s arguments, and compare the resulting vector to the vectors that represent the different WordNet senses of the verb. |
Background | A widely-used resource is WordNet (Fellbaum, 1998), where relations such as synonymy and hyponymy can be used to generate rules. |
Experimental Evaluation | single pair of words that are WordNet antonyms (2) Predicates differing by a single word of negation (3) Predicates p(t1, t2) and p(t2, 751) where p is a transitive verb (e.g., beat) in VerbNet (Kipper-Schuler et al., 2000). |
Learning Typed Entailment Graphs | Given a lexicographic resource ( WordNet ) and a set of predicates with their instances, we perform the following three steps (see Table 1): |
Learning Typed Entailment Graphs | 1) Training set generation We use WordNet to generate positive and negative examples, where each example is a pair of predicates. |
Learning Typed Entailment Graphs | For every predicate p(t1, t2) 6 P such that p is a single word, we extract from WordNet the set S of synonyms and direct hy-pernyms of p. For every p’ E S, if p’ (t1, t2) 6 P then p(t1, 752) —> p’ (t1, 752) is taken as a positive example. |
Background | Some have established concept hierarchies based on manually-built semantic resources such as WordNet (Miller, 1995). |
Background | Such hierarchies have good structures and high accuracy, but their coverage is limited to fine-grained concepts (e.g., “Ranunculaceae” is not included in WordNet .). |
Background | (2008) link the categories in Wikipedia onto WordNet . |
Introduction | In the WordNet hierarchy, senses are organized according to the “isa” relations. |
Related Work | (2006) provides a global optimization scheme for extending WordNet , which is different from the above-mentioned pairwise relationships identification methods. |
Conclusions and Future Work | We have empirically shown how automatically generated selectional preferences, using WordNet and distributional similarity measures, are able to effectively generalize lexical features and, thus, improve classification performance in a large-scale argument classification task on the CoNLL-2005 dataset. |
Experimental Setting | This happens more often with WordNet based models, which have a lower word coverage compared to distributional similarity—based models. |
Related Work | Resnik (1993) proposed to model selectional preferences using semantic classes from WordNet in order to tackle ambiguity issues in syntax (noun-compounds, coordination, PP-attachment). |
Results and Discussion | The performance loss on recall can be explained by the worse lexical coverage of WordNet when compared to automatically generated thesauri. |
Results and Discussion | Examples of words missing in WordNet include abbreviations (e.g., Inc., Corp.) and brand names (e.g., Texaco, Sony). |
Abstract | The evaluation set is derived from WordNet in a semi-supervised way. |
Conclusion and Future Work | We proposed a semi-supervised way to extract non-compositional MWEs from WordNet . |
Introduction and related work | Thirdly, we propose a semi-supervised approach for extracting non-compositional MWEs from WordNet , to decrease annotation cost. |
Proposed approach | Given a MWE, a set of queries is created: All synonyms of the M WE extracted from WordNet are collectedl. |
Test set of M WEs | For each of the 52, 217 MWEs of WordNet 3.0 (Miller, 1995) we collected: |
Related Work 2.1 WordNet-based Approach | For a given predicate q, the system firstly computes its distribution of argument semantic classes based on WordNet . |
Related Work 2.1 WordNet-based Approach | Clark and Weir (2002) suggest a hypothesis testing method by ascending the noun hierarchy of WordNet . |
Related Work 2.1 WordNet-based Approach | Cia-ramita and Johnson (2000) model WordNet as a Bayesian network to solve the “explain away” ambiguity. |
Empirical Evaluation | In Section 3.3, we developed three ways to compute the weight of an edge in the sentence quotation graph, i.e., clue words, semantic similarity based on WordNet and cosine similarity. |
Empirical Evaluation | The above experiments show that the widely used cosine similarity and the more sophisticated semantic similarity in WordNet are less accurate than the basic CWS in the summarization framework. |
Extracting Conversations from Multiple Emails | We explore three types of cohesion measures: (1) clue words that are based on stems, (2) semantic distance based on WordNet |
Extracting Conversations from Multiple Emails | 3.3.2 Semantic Similarity Based on WordNet |
Extracting Conversations from Multiple Emails | We use the well-known lexical database WordNet to get the semantic similarity of two words. |
Approach | Wherever applicable, we explore different syntactic and semantic representations of the textual content, e. g., extracting the dependency-based representation of the text or generalizing words to their WordNet supersenses (WNSS) (Ciaramita and Altun, 2006). |
Approach | In all these representations we skip stop words and normalize all words to their WordNet lemmas. |
The Corpus | Each word was morphologically simplified using the morphological functions of the WordNet library8. |
The Corpus | These tags, defined by WordNet lexicographers, provide a broad semantic categorization for nouns and verbs and include labels for nouns such as food, animal, body and feeling, and for verbs labels such as communication, contact, and possession. |
Context and Answer Detection | here, we use the product of sim(xu, Qi) and sim(:cv, {mm 62,} to estimate the possibility of being a context-answer pair for (u, v) , where sim(-, is the semantic similarity calculated on WordNet as described in Section 3.5. |
Context and Answer Detection | The semantic similarity between words is computed based on Wu and Palmer’s measure (Wu and Palmer, 1994) using WordNet (Fellbaum, 1998).1 The similarity between contiguous sentences will be used to capture the dependency for CRFs. |
Context and Answer Detection | - Similarity with the question using WordNet |
Experiments | CCG parser, WordNet |
Experiments | hypernyms, WordNet |
Experiments | head word, parser SVM hypernyms, WordNet |
Conclusion | Such methods include, but by no means limited to, semantic similarities between word pairs using lexical resources such as WordNet (Miller, 1995) and data-driven methods with various topic-dependent term weighting schemes on labeled corpus with topics such as MPQA. |
Term Weighting and Sentiment Analysis | Also, the distance between words in the local context or in the thesaurus-like dictionaries such as WordNet may be approximated as such measure. |
Term Weighting and Sentiment Analysis | It consists of WordNet synsets, where each synset is assigned three probability scores that add up to 1: positive, negative, and objective. |
Term Weighting and Sentiment Analysis | These scores are assigned at sense level (synsets in WordNet ), and we use the following equations to assess the sentiment scores at the word level. |
Introduction | Most established resources (e.g., WordNet ) represent only the main and widely accepted relationships such as hyper-nymy and meronymy. |
Related Work | There is a large body of related work that deals with discovery of basic relationship types represented in useful resources such as WordNet , including hyper-nymy (Hearst, 1992; Pantel et al., 2004; Snow et al., 2006), synonymy (Davidov and Rappoport, 2006; Widdows and Dorow, 2002) and meronymy (Berland and Charniak, 1999; Girju et al., 2006). |
Related Work | Several algorithms use manually-prepared resources, including WordNet (Moldovan et al., 2004; Costello et al., 2006) and Wikipedia (Strube and Ponzetto, 2006). |
Related Work | Evaluation for hypemymy and synonymy usually uses WordNet (Lin and Pantel, 2002; Widdows and Dorow, 2002; Davidov and Rappoport, 2006). |
Automated Classification | We use a Maximum Entropy (Berger et al., 1996) classifier with a large number of boolean features, some of which are novel (e. g., the inclusion of words from WordNet definitions). |
Automated Classification | The WordNet gloss terms had a surprisingly strong influence. |
Automated Classification | As far as we know, this is the first time that WordNet definition words have been used as features for noun compound interpretation. |
Related Work | Kim and Baldwin (2005) and Turney (2006) use nearest neighbor approaches based upon WordNet (Fellbaum, 1998) and Tumey’s Latent Relational Analysis, respectively. |
Related Work | Voorhees (1993) used the hyponymy (“ISA”) relation in WordNet (Miller, 1990) to disambiguate the polysemous nouns in a text. |
Related Work | (2004) tagged words with 25 root senses of nouns in WordNet . |
Related Work | Some researchers achieved improvements by expanding the disambiguated query words with synonyms and some other information from WordNet (Voorhees, 1994; Liu et al., 2004; Liu et al., 2005; Fang, 2008). |
Connotation Induction Algorithms | Hard constrains for WordNet relations: |
Precision, Coverage, and Efficiency | In particular, we found that it becomes nearly impractical to run the ILP formulation including all words in WordNet plus all words in the argument position in Google Web IT. |
Precision, Coverage, and Efficiency | Therefore we revise those hard constraints to encode various semantic relations ( WordNet and semantic coordination) more directly. |
Baselines | One is word co-occurrence (if word w and word wj occur in the same sentence or in the adjacent sentences, Sim(wi,wj) increases 1), and the other is WordNet (Miller, 1995) based similarity. |
Baselines | 3 Total weight of words in the focus candidate using the WordNet similarity. |
Baselines | 4 Max weight of words in the focus candidate using the WordNet similarity. |
Related work | In (Brown et al., 2011), the so called dynamic dependency neighborhoods (DDN), i.e., the set of verbs that are typically collocated with a direct object, are shown to be more helpful than lexical information (e. g., WordNet ). |
Related work | In supervised language learning, when few examples are available, DMs support cost-effective lexical generalizations, often outperforming knowledge based resources (such as WordNet , as in (Pantel et al., 2007)). |
Structural Similarity Functions | The contribution of (ii) is proportional to the lexical similarity of the tree lexical nodes, where the latter can be evaluated according to distributional models or also lexical resources, e. g., WordNet . |
Evaluation for SS | The SS algorithm has access to all the definitions in WordNet (WN). |
Experiments and Results | sourceforgenet, WordNet : :QueryData |
Limitations of Topic Models and LSA for Modeling Sentences | as WordNet (Fellbaum, 1998) (WN). |
Introduction | If the node is a preterminal node, we capture its lexical semantic by adding features indicating its WordNet sense information. |
Introduction | Specifically, the first WordNet sense of the terminal word, and all this sense’s hyponym senses will be added as features. |
Introduction | For example, WordNet senses {New Y0rk#], city#], district#], |
Experiments | In step 1, in order to make sure we select a diverse list of words, we consider three attributes of a word: frequency in a corpus, number of parts of speech, and number of synsets according to WordNet . |
Experiments | (2010), we use WordNet to first randomly select one synset of the first word, we then construct a set of words in various relations to the first word’s chosen synset, including hypemyms, hy-ponyms, holonyms, meronyms and attributes. |
Introduction | To capture interesting word pairs, we sample different senses of words using WordNet (Miller, 1995). |
Conclusion | In addition, we introduced the new and useful WordNet , Aflect, Length and Negation feature categories. |
Evaluation of Word Pairs | 1 WordNet 20.07 34.07 52.96 11.58 2 Verb Class 14.24 24.84 49.6 10.04 3 MPN 23.84 38.58 49.97 13.16 4 Modality 17.49 28.92 13.84 10.72 5 Polarity 16.46 26.36 65.15 11.58 6 Affect 18.62 31.59 59.8 13.37 7 8 9 |
Other Features | WordNet Features: We define four features based on WordNet (Fellbaum, 1998) - Synonyms, Antonyms, Hypernyms and Hyponyms. |
Related work | Many approaches to the problem use lexical taxonomies such as WordNet to identify the semantic classes that typically fill a particular argument slot for a predicate (Resnik, 1993; Clark and Weir, 2002; Schulte im Walde et al., 2008). |
Related work | (2007) integrate a model of random walks on the WordNet graph into an LDA topic model to build an unsupervised word sense disambiguation system. |
Related work | Reisinger and Pasca (2009) use LDA-like models to map automatically acquired attribute sets onto the WordNet hierarchy. |
Feature Extraction for Entailment | The NER module is based on a combination of user defined rules based on Lesk word disambiguation (Lesk, 1988), WordNet (Miller, 1995) lookups, and many user-defined dictionary lookups, e.g. |
Feature Extraction for Entailment | —When words don’t match we attempt matching synonyms in WordNet for most common senses. |
Feature Extraction for Entailment | —Verb match statistics using WordNet’s cause and entailment relations. |
PARMA | WordNet WordNet (Miller, 1995) is a database of information (synonyms, hypernyms, etc.) |
PARMA | For each entry, WordNet provides a set of synonyms, hypernyms, etc. |
PARMA | Given two spans, we use WordNet to determine semantic similarity by measuring how many synonym (or other) edges are needed to link two |
Analysis and discussion | (2008) use WordNet to develop sentiment lexicons in which each word has a positive or negative value associated with it, representing its strength. |
Analysis and discussion | The algorithm begins with seed sets of positive, negative, and neutral terms, and then uses the synonym and antonym structure of WordNet to expand those initial sets and refine the relative strength values. |
Methods | We also replace the negation and the adjective by the antonyms given in WordNet (using the first sense). |
ImpAr algorithm | named-entities and WordNet Super-Senses4. |
ImpAr algorithm | 4Lexicographic files according to WordNet terminology. |
Related Work | VENSES++ (Tonelli and Delmonte, 2010) applied a rule based anaphora resolution procedure and semantic similarity between candidates and thematic roles using WordNet (Fellbaum, 1998). |
Related Work | In the first phase, it extracts all synonyms from a thesaurus, such as WordNet , for the words to be substituted. |
Results and Analysis | They paraphrase a sentence s by replacing its words with WordNet synonyms, so that s can be more similar in wording to another sentence s’. |
Results and Analysis | has also been proposed in (Zhou et al., 2006), which uses paraphrase phrases like our PT—l instead of WordNet synonyms. |
Automatic Alignments | We use WordNet to generate candidate lemmas, and we also use a fuzzy match of a concept, defined to be a word in the sentence that has the longest string prefix match with that concept’s label, if the match length is Z 4. |
Automatic Alignments | WordNet lemmas and fuzzy matches are only used if the rule explicitly uses them. |
Training | Strips off trailing ‘-[0-9]+’ from the concept (for example run—01 —> run), and matches any exact matching word or WordNet lemma. |
Experimental Design | This algorithm generated a set of synonyms from WordNet and then used the SUBTLEX frequencies to find the most frequent synonym. |
Experimental Design | This measure is taken from WordNet (Fellbaum, 1998). |
Experimental Design | Synonym Count Also taken from WordNet , this is the number of potential synonyms with which a word could be replaced. |
Introduction | Distributional models that integrate the visual modality have been learned from texts and images (Feng and Lapata, 2010; Bruni et al., 2012b) or from ImageNet (Deng et al., 2009), e.g., by exploiting the fact that images in this database are hierarchically organized according to WordNet synsets (Leong and Mihalcea, 2011). |
The Attribute Dataset | Images for the concepts in McRae et al.’s (2005) production norms were harvested from ImageNet (Deng et al., 2009), an ontology of images based on the nominal hierarchy of WordNet (Fellbaum, 1998). |
The Attribute Dataset | ImageNet has more than 14 million images spanning 21K WordNet synsets. |
EXpt. 1: Predicting Absolute Scores | The first is that the lack of alignments for two function words is unproblematic; the second is that the alignment between fact and reality, which is established on the basis of WordNet similarity, is indeed licensed in the current context. |
Related Work | Banerjee and Lavie (2005) and Chan and Ng (2008) use WordNet , and Zhou et a1. |
Textual Entailment vs. MT Evaluation | The computation of these scores make extensive use of about ten lexical similarity resources, including WordNet , InfoMap, and Dekang Lin’s thesaurus. |
Introduction | processing tools, e.g., syntactic parsers (Wiebe, 2000), information extraction (IE) tools (Riloff and Wiebe, 2003) or rich lexical resources such as WordNet (Esuli and Sebastiani, 2006). |
Related Work | Many researchers have explored using relations in WordNet (Miller, 1995), e. g., Esuli and Sabastiani (2006), Andreevskaia and Bergler (2006) for English, Rao and Ravichandran (2009) for Hindi and French, and Perez-Rosas et al. |
Related Work | There is also a mismatch between the formality of many language resources, such as WordNet , and the extremely informal language of social media. |