Consensus Decoding Algorithms | For MBR decoding, we instead leverage a similarity measure 8(6; 6’) to choose a translation using the model’s probability distribution P(e| f), which has support over a set of possible translations E. The Viterbi derivation 6* is the mode of this distribution. |
Consensus Decoding Algorithms | Given any similarity measure 8 and a k-best list E, the minimum Bayes risk translation can be found by computing the similarity between all pairs of sentences in E, as in Algorithm 1. |
Consensus Decoding Algorithms | An example of a linear similarity measure is bag-of-words precision, which can be written as: |
Introduction | The Bayes optimal decoding objective is to minimize risk based on the similarity measure used for evaluation. |
Introduction | Unfortunately, with a nonlinear similarity measure like BLEU, we must resort to approximating the expected loss using a k-best list, which accounts for only a tiny fraction of a model’s full posterior distribution. |
Introduction | We show that if the similarity measure is linear in features of a sentence, then computing expected similarity for all k sentences requires only k similarity evaluations. |
Abstract | We present CoSimRank, a graph-theoretic similarity measure that is efficient because it can compute a single node similarity without having to compute the similarities of the entire graph. |
Abstract | Another advantage of CoSimRank is that it can be flexibly extended from basic node-node similarity to several other graph-theoretic similarity measures . |
Introduction | }raph-The0retic Similarity Measure |
Related Work | Apart from SimRank, many other similarity measures have been proposed. |
Related Work | (2006) introduce a similarity measure that is also based on the idea that nodes are similar when their neighbors are, but that is designed for bipartite graphs. |
Related Work | Another important similarity measure is cosine similarity of Personalized PageRank (PPR) vectors. |
Collocational Lexicon Induction | For each of these paraphrases, a DP is constructed and compared to that of the oov word using a similarity measure (Section 2.2). |
Collocational Lexicon Induction | where t is a phrase on the target side, 0 is the oov word or phrase, and s is a paraphrase of 0. p(s|0) is estimated using a similarity measure over DPs and p(t|s) is coming from the phrase-table. |
Collocational Lexicon Induction | 2.3 Similarity Measures |
Experiments & Results 4.1 Experimental Setup | In Section 2.2 and 2.3, different types of association measures and similarity measures have been explained to build and compare distributional profiles. |
Experiments & Results 4.1 Experimental Setup | As the results show, the combination of PMI as association measure and cosine as DP similarity measure outperforms the other possible combinations. |
Graph-based Lexicon Induction | Each phrase type represents a vertex in the graph and is connected to other vertices with a weight defined by a similarity measure between the two profiles (Section 2.3). |
Graph-based Lexicon Induction | However based on the definition of the similarity measures using context, it is quite possible that an oov node and a labeled node which are connected to the same unlabeled node do not share any context words and hence are not directly connected. |
Graph-based Lexicon Induction | In such a graph, the similarity of each pair of nodes is computed using one of the similarity measures discussed above. |
A Unified Semantic Representation | In order to compare semantic signatures, we adopt the Cosine similarity measure as a baseline method. |
Experiment 1: Textual Similarity | The top-ranking participating systems in the SemEval-2012 task were generally supervised systems utilizing a variety of lexical resources and similarity measurement techniques. |
Experiment 1: Textual Similarity | 3.3 Similarity Measure Analysis |
Experiment 1: Textual Similarity | In addition, we present in the table correlation scores for four other similarity measures reported by B'ar et al. |
Experiment 2: Word Similarity | Different evaluation methods exist in the literature for evaluating the performance of a word-level semantic similarity measure ; we adopted two well-established benchmarks: synonym recognition and correlating word similarity judgments with those from human annotators. |
Experiment 3: Sense Similarity | We adopt this task as a way of evaluating our similarity measure at the sense level. |
Abstract | Within this context, we propose a new methodology that adapts the classical K -means algorithm to a third-order similarity measure initially developed for NLP tasks. |
Conclusions | In this paper, we proposed a new PRC approach which (1) is based on the adaptation of the K -means algorithm to third-order similarity measures and (2) proposes a coherent stopping criterion. |
Evaluation | In particular, p is the size of the word feature vectors representing both Web snippets and centroids (p = 2.5), K is the number of clusters to be found (K = 2..10) and 8(Wik, 14/31) is the collocation measure integrated in the InfoSimba similarity measure . |
Introduction | feature vectors are hard to define in small collections of short text fragments (Timonen, 2013), (2) existing second-order similarity measures such as the cosine are unadapted to capture the semantic similarity between small texts, (3) Latent Semantic Analysis has evidenced inconclusive results (Osinski and Weiss, 2005) and (4) the labeling process is a surprisingly hard extra task (Carpineto et al., 2009). |
Introduction | For that purpose, we propose a new methodology that adapts the classical K -means algorithm to a third-order similarity measure initially developed for Topic Segmentation (Dias et al., 2007). |
Polythetic Post-Retrieval Clustering | Within the context of PRC, the K -means algorithm needs to be adapted to integrate third-order similarity measures (Mihalcea et al., 2006; Dias et al., 2007). |
Polythetic Post-Retrieval Clustering | Third-order similarity measures, also called weighted second-order similarity measures, do not rely on exact matches of word features as classical second-order similarity measures (6. g. the cosine metric), but rather evaluate similarity based on related matches. |
Polythetic Post-Retrieval Clustering | In this paper, we propose to use the third-order similarity measure called InfoSimba introduced in (Dias et al., 2007) for Topic Segmentation and implement its simplified version 83 in Equation 2. |
Evaluation and Results | To apply our PSSS on IQAPS inference task, we use it as the sentiment similarity measure in the algorithm explained in Figure 4. |
Evaluation and Results | The second row of Table 4 show the results of using a popular semantic similarity measure , PMI, as the sentiment similarity (SS) measure in Figure 4. |
Evaluation and Results | Table 4 shows the effectiveness of our sentiment similarity measure . |
Introduction | Semantic similarity measures such as Latent Semantic Analysis (LSA) (Landauer et al., 1998) can effectively capture the similarity between semantically related words like "car" and "automobile", but they are less effective in relating words with similar sentiment orientation like "excellent" and "superior". |
Introduction | We show that our approach effectively outperforms the semantic similarity measures in two NLP tasks: Indirect yes/no Question Answer Pairs (IQAPs) Inference and Sentiment Orientation (S0) prediction that are described as follows: |
Introduction | Previous research utilized the semantic relations between words obtained from WordNet (Hassan and Radev, 2010) and semantic similarity measures (e.g. |
Related Works | They also utilized Latent Semantic Analysis (LSA) (Landauer et al., 1998) as another semantic similarity measure . |
Related Works | However, both PM and LSA are semantic similarity measure . |
Related Works | (2012), we used two semantic similarity measures (PMI and LSA) for the IQAP inference task. |
Sentiment Similarity through Hidden Emotions | As we discussed above, semantic similarity measures are less effective to infer sentiment similarity between word pairs. |
Background and Model Setting | where sim(v, v’) is a vector similarity measure . |
Background and Model Setting | We note that the general DIRT scheme may be used while employing other “base” vector similarity measures . |
Background and Model Setting | This issue has been addressed in a separate line of research which introduced directional similarity measures suitable for inference relations (Bhagat et al., 2007; Szpektor and Dagan, 2008; Kotlerman et al., 2010). |
Introduction | Our scheme can be applied on top of any context-insensitive “base” similarity measure for rule learning, which operates at the word level, such as Cosine or Lin (Lin, 1998). |
Introduction | We apply our two-level scheme over three state-of-the-art context-insensitive similarity measures . |
Abstract | Then we propose various novel similarity measurements including surface features, meta-path based semantic features and social correlation features and combine them in a learning-to-rank framework. |
Introduction | 0 We propose two new similarity measures , as well as integrating temporal information into |
Introduction | the similarity measures to generate global semantic features. |
Target Candidate Ranking | We first extract surface features between the morph and the candidate based on measuring orthographic similarity measures which were commonly used in entity coreference resolution (e.g. |
Target Candidate Ranking | 4.2.3 Meta-Path-Based Semantic Similarity Measurements |
Target Candidate Ranking | We then adopt meta-path-based similarity measures (Sun et al., 2011a; Sun et al., 2011b), which are defined over heterogeneous networks to extract semantic features. |
Abstract | Existing word similarity measures are not robust to data sparseness since they rely only on the point estimation of words’ context profiles obtained from a limited amount of data. |
Abstract | The method uses a distribution of context profiles obtained by Bayesian estimation and takes the expectation of a base similarity measure under that distribution. |
Abstract | For the task of word similarity estimation using a large amount of Web data in Japanese, we show that the proposed measure gives better accuracies than other well-known similarity measures . |
Background | The BC is also a similarity measure on probability distributions and is suitable for our purposes as we describe in the next section. |
Background | Although BC has not been explored well in the literature on distributional word similarities, it is also a good similarity measure as the experiments show. |
Introduction | A number of semantic similarity measures have been proposed based on this hypothesis (Hindle, 1990; Grefenstette, 1994; Dagan et al., 1994; Dagan et al., 1995; Lin, 1998; Dagan et al., 1999). |
Introduction | In general, most semantic similarity measures have the following form: |
Introduction | Our technical contribution in this paper is to show that in the case where the context profiles are multinomial distributions, the priors are Dirichlet, and the base similarity measure is the Bhattacharyya coefficient (Bhattacharyya, 1943), we can derive an analytical form for Eq. |
Method | In this section, we show that if our base similarity measure is BC and the distributions under which we take the expectation are Dirichlet distributions, then Eq. |
Method | To put it all together, we can obtain a new Bayesian similarity measure on words, which can be calculated only from the hyperparameters for the Dirichlet prior, 04 and 6, and the observed counts C(wi, fk). |
Problem Formulation | For each term t in the dictionary and each SMS token 3,, we define a similarity measure a(t, 3,) that measures how closely the term 75 matches the SMS token 3,. |
Problem Formulation | Combining the similarity measure and the inverse document frequency (idf) of t in the corpus, we define a weight function to (t, 3,). |
Problem Formulation | The similarity measure and the weight function are discussed in detail in Section 5.1. |
System Implementation | The weight function is a combination of similarity measure between t and Si and Inverse Document Frequency (idf) of t. The next two subsections explain the calculation of the similarity measure and the idf in detail. |
System Implementation | 5.1.1 Similarity Measure |
System Implementation | For term t E D and token 3%- of the SMS, the similarity measure a(t, 81) between them is |
Abstract | Our approach leverages a similarity measure that enables the structural comparison of senses across lexical resources, achieving state-of-the-art performance on the task of aligning WordNet to three different collaborative resources: Wikipedia, Wiktionary and OmegaWiki. |
Conclusions | Our method leverages a novel similarity measure which enables a direct structural comparison of concepts across different lexical resources. |
Conclusions | In future work, we plan to extend our concept similarity measure across different natural languages. |
Experiments | 4.3 Similarity Measure Analysis |
Experiments | We explained in Section 2.1 that our concept similarity measure consists of two components: the definitional and the structural similarities. |
Experiments | structural similarity measure in comparison to the Dijkstra-WSA method, we carried out an experiment where our alignment system used only the structural similarity component, a variant of our system we refer to as SemAlignStr. |
Lexical Resource Ontologization | To do this, we apply our definitional similarity measure introduced in Section 2.1. |
Resource Alignment | Figure 1 illustrates the procedure underlying our cross-resource concept similarity measurement technique. |
Resource Alignment | The structural similarity component, instead, is a novel graph-based similarity measurement technique which calculates the similarity between a pair of concepts across the semantic networks of the two resources by leveraging the semantic |
A Statistical Inclusion Measure | Our research goal was to develop a directional similarity measure suitable for learning asymmetric relations, focusing empirically on lexical expansion. |
Abstract | This paper investigates the nature of directional (asymmetric) similarity measures , which aim to quantify distributional feature inclusion. |
Background | Then, word vectors are compared by some vector similarity measure . |
Conclusions and Future work | This paper advocates the use of directional similarity measures for lexical expansion, and potentially for other tasks, based on distributional inclusion of feature vectors. |
Evaluation and Results | We tested our similarity measure by evaluating its utility for lexical expansion, compared with baselines of the LIN, WeedsPrec and balPrec measures |
Evaluation and Results | Next, for each similarity measure , the terms found similar to any of the event’s seeds (‘u —> seed’) were taken as expansion terms. |
Introduction | Often, distributional similarity measures are used to identify expanding terms (e.g. |
Introduction | More generally, directional relations are abundant in NLP settings, making symmetric similarity measures less suitable for their identification. |
Introduction | Despite the need for directional similarity measures , their investigation counts, to the best of our knowledge, only few works (Weeds and Weir, 2003; Geffet and Dagan, 2005; Bhagat et al., 2007; Szpektor and Dagan, 2008; Michelbacher et al., 2007) and is utterly lacking. |
Abstract | As a consequence, improving such thesaurus is an important issue that is mainly tackled indirectly through the improvement of semantic similarity measures . |
Improving a distributional thesaurus | As in (Lin, 1998) or (Curran and Moens, 2002a), this building is based on the definition of a semantic similarity measure from a corpus. |
Improving a distributional thesaurus | For the extraction of distributional data and the characteristics of the distributional similarity measure , we adopted the options of (Ferret, 2010), resulting from a kind of grid search procedure performed with the extended TOEFL test proposed in (Freitag et al., 2005) as an optimization objective. |
Improving a distributional thesaurus | o similarity measure between contexts, for evaluating the semantic similarity of two words = Cosine measure. |
Introduction | Following work such as (Grefenstette, 1994), a widespread way to build a thesaurus from a corpus is to use a semantic similarity measure for extracting the semantic neighbors of the entries of the thesaurus. |
Introduction | Work based on WordNet-like lexical networks for building semantic similarity measures such as (Budanitsky and Hirst, 2006) or (Pedersen et al., 2004) falls into this category. |
Introduction | A part of these proposals focus on the weighting of the elements that are part of the contexts of words such as (Broda et al., 2009), in which the weights of context elements are turned into ranks, or (Zhitomirsky-Geffet and Dagan, 2009), followed and extended by (Yamamoto and Asakura, 2010), that proposes a bootstrapping method for modifying the weights of context elements according to the semantic neighbors found by an initial distributional similarity measure . |
Principles | For instance, features such as ngrams of words or ngrams of parts of speech are not considered whereas they are widely used in tasks such as word sense disambiguation (WSD) for instance, probably because they would lead to very large models and because similarity measures such as the Cosine measure are not necessarily suitable for heterogeneous representations (Alexandrescu and Kirchhoff, 2007). |
Related work | As a consequence, the improvement of such thesaurus is generally not directly addressed but is a possible consequence of the improvement of semantic similarity measures . |
Conclusion and Future Work | In this paper, we only tried Dice coefficient of n-grams and symmetrical sentence level BLEU as similarity measures . |
Conclusion and Future Work | In the future, we will explore other consensus features and other similarity measures , which may take document level information, or syntactic and semantic information into consideration. |
Experiments and Results | Instead of using graph-based consensus confidence as features in the log-linear model, we perform structured label propagation (Struct-LP) to re-rank the n-best list directly, and the similarity measures for source sentences and translation candidates are symmetrical sentence level BLEU (equation (10)). |
Features and Training | Tl(e,e') is the propagating probability in equation (8), with the similarity measure Sim(e,e') defined as the Dice coefficient over the set of all n-grams in e and those in e'. |
Features and Training | defined in equation (3), takes symmetrical sentence level BLEU as similarity measure ]: |
Features and Training | In theory we could use other similarity measures such as edit distance, string kernel. |
Graph-based Structured Learning | wilj defines the weight of the edge, which is a similarity measure between nodes i and j. |
Graph-based Structured Learning | Propagation probability TS (f, f ') is as defined in equation (3), and Tl(e,e') is defined given some similarity measure sim(e, 6') between labels 6 and |
Abstract | We show also how Thesaurus Rex supports a novel, generative similarity measure for WordNet. |
Related Work and Ideas | more rounded similarity measures . |
Related Work and Ideas | A similarity measure can draw on other |
Related Work and Ideas | Their best similarity measure achieves a remarkable 0.93 correlation with human judgments on the Miller & Charles word-pair set. |
Seeing is Believing (and Creating) | Using WordNet, for instance, a similarity measure can vertically converge on a common superordinate category of both inputs, and generate a single numeric result based on their distance to, and the information content of, this common generalization. |
Seeing is Believing (and Creating) | To be as useful for creative tasks as they are for conventional tasks, we need to re-imagine our computational similarity measures as generative rather than selective, expansive rather than reductive, divergent as well as convergent and lateral as well as vertical. |
Seeing is Believing (and Creating) | Section 2 provides a brief overview of past work in the area of similarity measurement , before section 3 describes a simple bootstrapping loop for acquiring richly diverse perspectives from the web for a wide variety of familiar ideas. |
Introduction | In this work, we analyze various Japanese corpora using a number of collocation and word similarity measures to deduce and suggest the best collocations for Japanese second language learners. |
Introduction | In order to build a system that is more sensitive to constructions that are difficult for learners, we use word similarity measures that generate collocation candidates using a large Japanese language learner corpus. |
Related Work | similarity measures are used. |
Related Work | Our work follows the general approach, that is, uses similarity measures for generating the confusion set and association measures for ranking the best candidates. |
Related Work | Similarity measures are used to generate the collocation candidates that are later ranked using association measures. |
Abstract | Instead of utilizing simple similarity measures and their disjoint combinations, our method directly optimizes document and entity representations for a given similarity measure . |
Abstract | A supervised fine-tuning stage follows to optimize the representation towards the similarity measure . |
Conclusion | We propose a deep learning approach that automatically learns context-entity similarity measure for entity disambiguation. |
Experiments and Analysis | When embedding our similarity measure sim(d, 6) into (Han et al., 2011), we achieve the best results on AIDA. |
Introduction | ument and entity representations for a fixed similarity measure . |
Introduction | In fact, the underlying representations for computing similarity measure add internal structure to the given similarity measure . |
Introduction | The learned similarity measure can be readily incorporated into any existing collective approaches, which further boosts performance. |
Abstract | The best results are obtained with a novel second-order distributional similarity measure , and the positive effect is specially relevant for out-of-domain data. |
Conclusions and Future Work | We have empirically shown how automatically generated selectional preferences, using WordNet and distributional similarity measures , are able to effectively generalize lexical features and, thus, improve classification performance in a large-scale argument classification task on the CoNLL-2005 dataset. |
Related Work | Pantel and Lin (2000) obtained very good results using the distributional similarity measure defined by Lin (1998). |
Results and Discussion | The second-order distributional similarity measures perform best overall, both in precision and recall. |
Selectional Preference Models | We will refer to this similarity measure as simg‘n. |
Selectional Preference Models | We will refer to these similarity measures as simE-ZE and simi’gg hereinafter. |
Abstract | A number of word similarity measures are proposed for clustering words for the Named Entity Recognition task. |
Conclusion | A number of word similarity measures are used for clustering. |
Evaluation of NE Recognition | Among the various similarity measures of clustering, improved results are obtained using the clus- |
Evaluation of NE Recognition | ters which uses the similarity measurement based on proximity of the words to NE categories (defined in Section 3.3). |
Word Clustering | The Euclidean distance is used to find the similarity between the above word vectors as a similarity measure . |
Word Clustering | Using the above similarity measures we have used the k-means algorithm. |
Abstract | We combine several graph alignment features with lexical semantic similarity measures using machine learning techniques and show that the student answers can be more accurately graded than if the semantic measures were used in isolation. |
Answer Grading System | In the final stage (Section 3.4), we produce an overall grade based upon the alignment scores found in the previous stage as well as the results of several semantic BOW similarity measures (Section 3.3). |
Answer Grading System | All eight WordNet-based similarity measures listed in Section 3.3 plus the LSA model are used to produce these features. |
Answer Grading System | In order to address this, we combine the graph alignment scores, which encode syntactic knowledge, with the scores obtained from semantic similarity measures . |
Results | One surprise while building this system was the consistency with which the novel technique of question demoting improved scores for the BOW similarity measures . |
Evaluation setting and results | Our method was evaluated for each (P1, P2, P3) combination and similarity measures J0 and 197,, separately. |
Introduction and related work | In this paper, we propose a novel unsupervised approach that compares the major senses of a MWE and its semantic head using distributional similarity measures to test the compositionality of the MWE. |
Proposed approach | Lee (1999) shows that J performs better than other symmetric similarity measures such as cosine, Jensen-Shannon divergence, etc. |
Proposed approach | Given the major uses of a MWE and its semantic head, the MWE is considered as compositional, when the corresponding distributional similarity measure (Jc or 197,) value is above a parameter threshold, sim. |
Unsupervised parameter tuning | The best performing distributional similarity measure is an. |
Discussion | One of the most effective similarity measures is the cosine similarity, which is a normalized dot product. |
Discussion | In order to appreciate the effect of these advantages, we perform an experiment that takes H to be the set of all LCs of size 1, and uses a single similarity measure . |
Our Proposal: A Latent LC Approach | where sim is some vector similarity measure . |
Our Proposal: A Latent LC Approach | We use two common similarity measures: the vector cosine metric, and the BInc (Szpektor and Dagan, 2008) similarity measure . |
Our Proposal: A Latent LC Approach | To do so, we use point-wise mutual information, and the conditional probabilities P(hf|hf) and POLE Similar measures have often been used for the unsupervised detection of MWEs (Villavicencio et al., 2007; Fazly and Stevenson, 2006). |
Experimental Evaluation | Local algorithms We described 12 distributional similarity measures computed over our corpus (Section 5.1). |
Experimental Evaluation | In addition, we obtained similarity lists learned by Lin and Pantel (2001), and replicated 3 similarity measures learned by Szpektor and Dagan (2008), over the RCV1 corpus7. |
Experimental Evaluation | For each distributional similarity measure (altogether 16 measures), we learned a graph by inserting any edge (u, v) , when u is in the top K templates most similar to 2). |
Learning Entailment Graph Edges | Similarity function We consider two similarity functions: The Lin (2001) similarity measure, and the Balanced Inclusion (BInc) similarity measure (Szpektor and Dagan, 2008). |
FrameNet — Wiktionary Alignment | They align senses in WordNet to Wikipedia entries in a supervised setting using semantic similarity measures . |
FrameNet — Wiktionary Alignment | Niemann and Gurevych (2011) combine two different types of similarity (i) cosine similarity on bag-of-words vectors (COS) and (ii) a personalized PageRank—based similarity measure (PPR). |
FrameNet — Wiktionary Alignment | For each similarity measure , Niemann and Gurevych (2011) determine a threshold (tppr and |
Related Work | The similarity measure is based on stem overlap of the candidates’ glosses expanded by WordNet domains, the WordNet synset, and the set of senses for a FrameNet frame. |
Named Entity Disambiguation by Leveraging Semantic Knowledge | Because the key problem of named entity disambiguation is to measure the similarity between name observations, we integrate the structural semantic relatedness in the similarity measure , so that it can better reflect the actual similarity between name observations. |
The Structural Semantic Relatedness Measure | The lexical relatedness lr between two WordNet concepts are measured using the Lin (l998)’s WordNet semantic similarity measure . |
The Structural Semantic Relatedness Measure | The problem of quantifying the relatedness between nodes in a graph is not a new problem, e.g., the structural equivalence and structural similarity (the SimRank in Jeh and Widom (2002) and the similarity measure in Leicht et al. |
The Structural Semantic Relatedness Measure | However, these similarity measures are not suitable for our task, because all of them assume that the edges are uniform so that they cannot take edge weight into consideration. |
Domain Adaptation | For both POS tagging and sentiment classification, we experimented with several alternative approaches for feature weighting, representation, and similarity measures using development data, which we randomly selected from the training instances from the datasets described in Section 5. |
Domain Adaptation | With respect to similarity measures, we experimented with cosine similarity and the similarity measure proposed by Lin (1998); cosine similarity performed consistently well over all the experimental settings. |
Domain Adaptation | The feature representation was held fixed during these similarity measure comparisons. |
O \ | As an example of the distribution prediction method, in Table 3 we show the top 3 similar distributional features u in the books (source) domain, predicted for the electronics (target) domain word 21) = lightweight, by different similarity measures . |
Conclusion | Moreover, our approach can combine two similarity measures in a hybrid hashing scheme, which is beneficial to comprehensively modeling the document similarity. |
Document Retrieval with Hashing | Given a query document vector q, we use the Cosine similarity measure to evaluate the similarity between q and a document a: in a dataset: |
Document Retrieval with Hashing | Enable a hybrid hashing scheme combining two similarity measures . |
Introduction | Furthermore, we make the hashing framework applicable to combine different similarity measures in NNS. |
Experiments: predicting relevance in context | Figure 3: Precision and recall on relevant links with respect to a threshold on the similarity measure (Lin’s score) |
Experiments: predicting relevance in context | A straightforward parameter to include to predict the relevance of a link is of course the similarity measure itself, here Lin’s information measure. |
Experiments: predicting relevance in context | This is already a big improvement on the use of the similarity measure alone (24%). |
Introduction | A distributional thesaurus is a lexical network that lists semantic neighbours, computed from a corpus and a similarity measure between lexical items, which generally captures the similarity of contexts in which the items occur. |
Algorithm | sim(h, h’) is a similarity measure between argument heads. |
Algorithm | The similarity measure we use is based on the slot distributions of the arguments. |
Algorithm | The similarity measure between two head words is then defined as the cosine measure of their vectors. |
Experiments | Siml and Sim2 respectively mean Formula 3.1 and Formula 3.2 are used in postprocessing as the similarity measure between |
Experiments | Sim2 achieves more performance improvement than Siml, which demonstrates the effectiveness of the similarity measure in Formula 3.2. |
Our Approach | One simple and straightforward similarity measure is the J accard |
Discussion and Future Work | 5.2 Fractional Similarity Measures |
Discussion and Future Work | In contrast, the linear-programming based TESLA metric allows fractional similarity measures between 0 (completely unrelated) and l (exact synonyms). |
Discussion and Future Work | Supporting fractional similarity measures is nontrivial in the TESLA-CELAB framework. |
ConceptResolver | We use several string similarity measures as features, including SoftTFIDF (Cohen et al., 2003), Level 2 JaroWinkler (Cohen et al., 2003), Fellegi-Sunter (Fellegi and Sunter, 1969), and Monge-Elkan (Monge and Elkan, 1996). |
Prior Work | Like other approaches (Basu et al., 2004; Xing et al., 2003; Klein et al., 2002), we learn a similarity measure for clustering based on a set of must-link and cannot-link constraints. |
Prior Work | Unlike prior work, our algorithm exploits multiple views of the data to improve the similarity measure . |
Genre Distance Measures | We can derive such distance measures from the genre hierarchy in a way similar to word similarity measures that were invented for lexical hierarchies such as WordNet (see (Pedersen et al., 2007) for an overview). |
Genre Distance Measures | The Leaeoek & Chodorow similarity measure (Leacock and Chodorow, 1998) normalizes the path length measure (6) by the maximum number of nodes D when traversing down from the root. |
Genre Distance Measures | Several other similarity measures have been proposed based on the Resnik similarity such as the one by (Lin, 1998): |
Introduction | They are used as matching sequences to locate corresponding candidate entries in the KB, and then to disambiguate those candidates using similarity measures . |
Introduction | This is usually done using similarity measures (such as cosine similarity, weighted J accard distance, KL divergence...) that evaluate the distance between a bag of words related to a candidate annotation, and the words surrounding the entity to annotate in the text. |
Related Work | It proposes a disambiguation method that combines popularity-based priors, similarity measures , and coherence. |
Evaluation | The Levenshtein similarity measure , on the other hand, is too restrictive and thus results in comparatively high precisions, but very low recall. |
Temporal Script Graphs | On the basis of this pseudo-parse, we compute the similarity measure sim: |
Temporal Script Graphs | The semantic constraints check whether the event descriptions of the merged node would be sufficiently consistent according to the similarity measure from Section 5.2. |