Index of papers in Proc. ACL that mention
  • similarity measure
DeNero, John and Chiang, David and Knight, Kevin
Consensus Decoding Algorithms
For MBR decoding, we instead leverage a similarity measure 8(6; 6’) to choose a translation using the model’s probability distribution P(e| f), which has support over a set of possible translations E. The Viterbi derivation 6* is the mode of this distribution.
Consensus Decoding Algorithms
Given any similarity measure 8 and a k-best list E, the minimum Bayes risk translation can be found by computing the similarity between all pairs of sentences in E, as in Algorithm 1.
Consensus Decoding Algorithms
An example of a linear similarity measure is bag-of-words precision, which can be written as:
Introduction
The Bayes optimal decoding objective is to minimize risk based on the similarity measure used for evaluation.
Introduction
Unfortunately, with a nonlinear similarity measure like BLEU, we must resort to approximating the expected loss using a k-best list, which accounts for only a tiny fraction of a model’s full posterior distribution.
Introduction
We show that if the similarity measure is linear in features of a sentence, then computing expected similarity for all k sentences requires only k similarity evaluations.
similarity measure is mentioned in 33 sentences in this paper.
Topics mentioned in this paper:
Rothe, Sascha and Schütze, Hinrich
Abstract
We present CoSimRank, a graph-theoretic similarity measure that is efficient because it can compute a single node similarity without having to compute the similarities of the entire graph.
Abstract
Another advantage of CoSimRank is that it can be flexibly extended from basic node-node similarity to several other graph-theoretic similarity measures .
Introduction
}raph-The0retic Similarity Measure
Related Work
Apart from SimRank, many other similarity measures have been proposed.
Related Work
(2006) introduce a similarity measure that is also based on the idea that nodes are similar when their neighbors are, but that is designed for bipartite graphs.
Related Work
Another important similarity measure is cosine similarity of Personalized PageRank (PPR) vectors.
similarity measure is mentioned in 20 sentences in this paper.
Topics mentioned in this paper:
Razmara, Majid and Siahbani, Maryam and Haffari, Reza and Sarkar, Anoop
Collocational Lexicon Induction
For each of these paraphrases, a DP is constructed and compared to that of the oov word using a similarity measure (Section 2.2).
Collocational Lexicon Induction
where t is a phrase on the target side, 0 is the oov word or phrase, and s is a paraphrase of 0. p(s|0) is estimated using a similarity measure over DPs and p(t|s) is coming from the phrase-table.
Collocational Lexicon Induction
2.3 Similarity Measures
Experiments & Results 4.1 Experimental Setup
In Section 2.2 and 2.3, different types of association measures and similarity measures have been explained to build and compare distributional profiles.
Experiments & Results 4.1 Experimental Setup
As the results show, the combination of PMI as association measure and cosine as DP similarity measure outperforms the other possible combinations.
Graph-based Lexicon Induction
Each phrase type represents a vertex in the graph and is connected to other vertices with a weight defined by a similarity measure between the two profiles (Section 2.3).
Graph-based Lexicon Induction
However based on the definition of the similarity measures using context, it is quite possible that an oov node and a labeled node which are connected to the same unlabeled node do not share any context words and hence are not directly connected.
Graph-based Lexicon Induction
In such a graph, the similarity of each pair of nodes is computed using one of the similarity measures discussed above.
similarity measure is mentioned in 12 sentences in this paper.
Topics mentioned in this paper:
Pilehvar, Mohammad Taher and Jurgens, David and Navigli, Roberto
A Unified Semantic Representation
In order to compare semantic signatures, we adopt the Cosine similarity measure as a baseline method.
Experiment 1: Textual Similarity
The top-ranking participating systems in the SemEval-2012 task were generally supervised systems utilizing a variety of lexical resources and similarity measurement techniques.
Experiment 1: Textual Similarity
3.3 Similarity Measure Analysis
Experiment 1: Textual Similarity
In addition, we present in the table correlation scores for four other similarity measures reported by B'ar et al.
Experiment 2: Word Similarity
Different evaluation methods exist in the literature for evaluating the performance of a word-level semantic similarity measure ; we adopted two well-established benchmarks: synonym recognition and correlating word similarity judgments with those from human annotators.
Experiment 3: Sense Similarity
We adopt this task as a way of evaluating our similarity measure at the sense level.
similarity measure is mentioned in 13 sentences in this paper.
Topics mentioned in this paper:
Moreno, Jose G. and Dias, Gaël and Cleuziou, Guillaume
Abstract
Within this context, we propose a new methodology that adapts the classical K -means algorithm to a third-order similarity measure initially developed for NLP tasks.
Conclusions
In this paper, we proposed a new PRC approach which (1) is based on the adaptation of the K -means algorithm to third-order similarity measures and (2) proposes a coherent stopping criterion.
Evaluation
In particular, p is the size of the word feature vectors representing both Web snippets and centroids (p = 2.5), K is the number of clusters to be found (K = 2..10) and 8(Wik, 14/31) is the collocation measure integrated in the InfoSimba similarity measure .
Introduction
feature vectors are hard to define in small collections of short text fragments (Timonen, 2013), (2) existing second-order similarity measures such as the cosine are unadapted to capture the semantic similarity between small texts, (3) Latent Semantic Analysis has evidenced inconclusive results (Osinski and Weiss, 2005) and (4) the labeling process is a surprisingly hard extra task (Carpineto et al., 2009).
Introduction
For that purpose, we propose a new methodology that adapts the classical K -means algorithm to a third-order similarity measure initially developed for Topic Segmentation (Dias et al., 2007).
Polythetic Post-Retrieval Clustering
Within the context of PRC, the K -means algorithm needs to be adapted to integrate third-order similarity measures (Mihalcea et al., 2006; Dias et al., 2007).
Polythetic Post-Retrieval Clustering
Third-order similarity measures, also called weighted second-order similarity measures, do not rely on exact matches of word features as classical second-order similarity measures (6. g. the cosine metric), but rather evaluate similarity based on related matches.
Polythetic Post-Retrieval Clustering
In this paper, we propose to use the third-order similarity measure called InfoSimba introduced in (Dias et al., 2007) for Topic Segmentation and implement its simplified version 83 in Equation 2.
similarity measure is mentioned in 12 sentences in this paper.
Topics mentioned in this paper:
Mohtarami, Mitra and Lan, Man and Tan, Chew Lim
Evaluation and Results
To apply our PSSS on IQAPS inference task, we use it as the sentiment similarity measure in the algorithm explained in Figure 4.
Evaluation and Results
The second row of Table 4 show the results of using a popular semantic similarity measure , PMI, as the sentiment similarity (SS) measure in Figure 4.
Evaluation and Results
Table 4 shows the effectiveness of our sentiment similarity measure .
Introduction
Semantic similarity measures such as Latent Semantic Analysis (LSA) (Landauer et al., 1998) can effectively capture the similarity between semantically related words like "car" and "automobile", but they are less effective in relating words with similar sentiment orientation like "excellent" and "superior".
Introduction
We show that our approach effectively outperforms the semantic similarity measures in two NLP tasks: Indirect yes/no Question Answer Pairs (IQAPs) Inference and Sentiment Orientation (S0) prediction that are described as follows:
Introduction
Previous research utilized the semantic relations between words obtained from WordNet (Hassan and Radev, 2010) and semantic similarity measures (e.g.
Related Works
They also utilized Latent Semantic Analysis (LSA) (Landauer et al., 1998) as another semantic similarity measure .
Related Works
However, both PM and LSA are semantic similarity measure .
Related Works
(2012), we used two semantic similarity measures (PMI and LSA) for the IQAP inference task.
Sentiment Similarity through Hidden Emotions
As we discussed above, semantic similarity measures are less effective to infer sentiment similarity between word pairs.
similarity measure is mentioned in 12 sentences in this paper.
Topics mentioned in this paper:
Melamud, Oren and Berant, Jonathan and Dagan, Ido and Goldberger, Jacob and Szpektor, Idan
Background and Model Setting
where sim(v, v’) is a vector similarity measure .
Background and Model Setting
We note that the general DIRT scheme may be used while employing other “base” vector similarity measures .
Background and Model Setting
This issue has been addressed in a separate line of research which introduced directional similarity measures suitable for inference relations (Bhagat et al., 2007; Szpektor and Dagan, 2008; Kotlerman et al., 2010).
Introduction
Our scheme can be applied on top of any context-insensitive “base” similarity measure for rule learning, which operates at the word level, such as Cosine or Lin (Lin, 1998).
Introduction
We apply our two-level scheme over three state-of-the-art context-insensitive similarity measures .
similarity measure is mentioned in 30 sentences in this paper.
Topics mentioned in this paper:
Huang, Hongzhao and Wen, Zhen and Yu, Dian and Ji, Heng and Sun, Yizhou and Han, Jiawei and Li, He
Abstract
Then we propose various novel similarity measurements including surface features, meta-path based semantic features and social correlation features and combine them in a learning-to-rank framework.
Introduction
0 We propose two new similarity measures , as well as integrating temporal information into
Introduction
the similarity measures to generate global semantic features.
Target Candidate Ranking
We first extract surface features between the morph and the candidate based on measuring orthographic similarity measures which were commonly used in entity coreference resolution (e.g.
Target Candidate Ranking
4.2.3 Meta-Path-Based Semantic Similarity Measurements
Target Candidate Ranking
We then adopt meta-path-based similarity measures (Sun et al., 2011a; Sun et al., 2011b), which are defined over heterogeneous networks to extract semantic features.
similarity measure is mentioned in 14 sentences in this paper.
Topics mentioned in this paper:
Kazama, Jun'ichi and De Saeger, Stijn and Kuroda, Kow and Murata, Masaki and Torisawa, Kentaro
Abstract
Existing word similarity measures are not robust to data sparseness since they rely only on the point estimation of words’ context profiles obtained from a limited amount of data.
Abstract
The method uses a distribution of context profiles obtained by Bayesian estimation and takes the expectation of a base similarity measure under that distribution.
Abstract
For the task of word similarity estimation using a large amount of Web data in Japanese, we show that the proposed measure gives better accuracies than other well-known similarity measures .
Background
The BC is also a similarity measure on probability distributions and is suitable for our purposes as we describe in the next section.
Background
Although BC has not been explored well in the literature on distributional word similarities, it is also a good similarity measure as the experiments show.
Introduction
A number of semantic similarity measures have been proposed based on this hypothesis (Hindle, 1990; Grefenstette, 1994; Dagan et al., 1994; Dagan et al., 1995; Lin, 1998; Dagan et al., 1999).
Introduction
In general, most semantic similarity measures have the following form:
Introduction
Our technical contribution in this paper is to show that in the case where the context profiles are multinomial distributions, the priors are Dirichlet, and the base similarity measure is the Bhattacharyya coefficient (Bhattacharyya, 1943), we can derive an analytical form for Eq.
Method
In this section, we show that if our base similarity measure is BC and the distributions under which we take the expectation are Dirichlet distributions, then Eq.
Method
To put it all together, we can obtain a new Bayesian similarity measure on words, which can be calculated only from the hyperparameters for the Dirichlet prior, 04 and 6, and the observed counts C(wi, fk).
similarity measure is mentioned in 25 sentences in this paper.
Topics mentioned in this paper:
Kothari, Govind and Negi, Sumit and Faruquie, Tanveer A. and Chakaravarthy, Venkatesan T. and Subramaniam, L. Venkata
Problem Formulation
For each term t in the dictionary and each SMS token 3,, we define a similarity measure a(t, 3,) that measures how closely the term 75 matches the SMS token 3,.
Problem Formulation
Combining the similarity measure and the inverse document frequency (idf) of t in the corpus, we define a weight function to (t, 3,).
Problem Formulation
The similarity measure and the weight function are discussed in detail in Section 5.1.
System Implementation
The weight function is a combination of similarity measure between t and Si and Inverse Document Frequency (idf) of t. The next two subsections explain the calculation of the similarity measure and the idf in detail.
System Implementation
5.1.1 Similarity Measure
System Implementation
For term t E D and token 3%- of the SMS, the similarity measure a(t, 81) between them is
similarity measure is mentioned in 15 sentences in this paper.
Topics mentioned in this paper:
Pilehvar, Mohammad Taher and Navigli, Roberto
Abstract
Our approach leverages a similarity measure that enables the structural comparison of senses across lexical resources, achieving state-of-the-art performance on the task of aligning WordNet to three different collaborative resources: Wikipedia, Wiktionary and OmegaWiki.
Conclusions
Our method leverages a novel similarity measure which enables a direct structural comparison of concepts across different lexical resources.
Conclusions
In future work, we plan to extend our concept similarity measure across different natural languages.
Experiments
4.3 Similarity Measure Analysis
Experiments
We explained in Section 2.1 that our concept similarity measure consists of two components: the definitional and the structural similarities.
Experiments
structural similarity measure in comparison to the Dijkstra-WSA method, we carried out an experiment where our alignment system used only the structural similarity component, a variant of our system we refer to as SemAlignStr.
Lexical Resource Ontologization
To do this, we apply our definitional similarity measure introduced in Section 2.1.
Resource Alignment
Figure 1 illustrates the procedure underlying our cross-resource concept similarity measurement technique.
Resource Alignment
The structural similarity component, instead, is a novel graph-based similarity measurement technique which calculates the similarity between a pair of concepts across the semantic networks of the two resources by leveraging the semantic
similarity measure is mentioned in 10 sentences in this paper.
Topics mentioned in this paper:
Kotlerman, Lili and Dagan, Ido and Szpektor, Idan and Zhitomirsky-Geffet, Maayan
A Statistical Inclusion Measure
Our research goal was to develop a directional similarity measure suitable for learning asymmetric relations, focusing empirically on lexical expansion.
Abstract
This paper investigates the nature of directional (asymmetric) similarity measures , which aim to quantify distributional feature inclusion.
Background
Then, word vectors are compared by some vector similarity measure .
Conclusions and Future work
This paper advocates the use of directional similarity measures for lexical expansion, and potentially for other tasks, based on distributional inclusion of feature vectors.
Evaluation and Results
We tested our similarity measure by evaluating its utility for lexical expansion, compared with baselines of the LIN, WeedsPrec and balPrec measures
Evaluation and Results
Next, for each similarity measure , the terms found similar to any of the event’s seeds (‘u —> seed’) were taken as expansion terms.
Introduction
Often, distributional similarity measures are used to identify expanding terms (e.g.
Introduction
More generally, directional relations are abundant in NLP settings, making symmetric similarity measures less suitable for their identification.
Introduction
Despite the need for directional similarity measures , their investigation counts, to the best of our knowledge, only few works (Weeds and Weir, 2003; Geffet and Dagan, 2005; Bhagat et al., 2007; Szpektor and Dagan, 2008; Michelbacher et al., 2007) and is utterly lacking.
similarity measure is mentioned in 10 sentences in this paper.
Topics mentioned in this paper:
Ferret, Olivier
Abstract
As a consequence, improving such thesaurus is an important issue that is mainly tackled indirectly through the improvement of semantic similarity measures .
Improving a distributional thesaurus
As in (Lin, 1998) or (Curran and Moens, 2002a), this building is based on the definition of a semantic similarity measure from a corpus.
Improving a distributional thesaurus
For the extraction of distributional data and the characteristics of the distributional similarity measure , we adopted the options of (Ferret, 2010), resulting from a kind of grid search procedure performed with the extended TOEFL test proposed in (Freitag et al., 2005) as an optimization objective.
Improving a distributional thesaurus
o similarity measure between contexts, for evaluating the semantic similarity of two words = Cosine measure.
Introduction
Following work such as (Grefenstette, 1994), a widespread way to build a thesaurus from a corpus is to use a semantic similarity measure for extracting the semantic neighbors of the entries of the thesaurus.
Introduction
Work based on WordNet-like lexical networks for building semantic similarity measures such as (Budanitsky and Hirst, 2006) or (Pedersen et al., 2004) falls into this category.
Introduction
A part of these proposals focus on the weighting of the elements that are part of the contexts of words such as (Broda et al., 2009), in which the weights of context elements are turned into ranks, or (Zhitomirsky-Geffet and Dagan, 2009), followed and extended by (Yamamoto and Asakura, 2010), that proposes a bootstrapping method for modifying the weights of context elements according to the semantic neighbors found by an initial distributional similarity measure .
Principles
For instance, features such as ngrams of words or ngrams of parts of speech are not considered whereas they are widely used in tasks such as word sense disambiguation (WSD) for instance, probably because they would lead to very large models and because similarity measures such as the Cosine measure are not necessarily suitable for heterogeneous representations (Alexandrescu and Kirchhoff, 2007).
Related work
As a consequence, the improvement of such thesaurus is generally not directly addressed but is a possible consequence of the improvement of semantic similarity measures .
similarity measure is mentioned in 10 sentences in this paper.
Topics mentioned in this paper:
Liu, Shujie and Li, Chi-Ho and Li, Mu and Zhou, Ming
Conclusion and Future Work
In this paper, we only tried Dice coefficient of n-grams and symmetrical sentence level BLEU as similarity measures .
Conclusion and Future Work
In the future, we will explore other consensus features and other similarity measures , which may take document level information, or syntactic and semantic information into consideration.
Experiments and Results
Instead of using graph-based consensus confidence as features in the log-linear model, we perform structured label propagation (Struct-LP) to re-rank the n-best list directly, and the similarity measures for source sentences and translation candidates are symmetrical sentence level BLEU (equation (10)).
Features and Training
Tl(e,e') is the propagating probability in equation (8), with the similarity measure Sim(e,e') defined as the Dice coefficient over the set of all n-grams in e and those in e'.
Features and Training
defined in equation (3), takes symmetrical sentence level BLEU as similarity measure ]:
Features and Training
In theory we could use other similarity measures such as edit distance, string kernel.
Graph-based Structured Learning
wilj defines the weight of the edge, which is a similarity measure between nodes i and j.
Graph-based Structured Learning
Propagation probability TS (f, f ') is as defined in equation (3), and Tl(e,e') is defined given some similarity measure sim(e, 6') between labels 6 and
similarity measure is mentioned in 9 sentences in this paper.
Topics mentioned in this paper:
Veale, Tony and Li, Guofu
Abstract
We show also how Thesaurus Rex supports a novel, generative similarity measure for WordNet.
Related Work and Ideas
more rounded similarity measures .
Related Work and Ideas
A similarity measure can draw on other
Related Work and Ideas
Their best similarity measure achieves a remarkable 0.93 correlation with human judgments on the Miller & Charles word-pair set.
Seeing is Believing (and Creating)
Using WordNet, for instance, a similarity measure can vertically converge on a common superordinate category of both inputs, and generate a single numeric result based on their distance to, and the information content of, this common generalization.
Seeing is Believing (and Creating)
To be as useful for creative tasks as they are for conventional tasks, we need to re-imagine our computational similarity measures as generative rather than selective, expansive rather than reductive, divergent as well as convergent and lateral as well as vertical.
Seeing is Believing (and Creating)
Section 2 provides a brief overview of past work in the area of similarity measurement , before section 3 describes a simple bootstrapping loop for acquiring richly diverse perspectives from the web for a wide variety of familiar ideas.
similarity measure is mentioned in 8 sentences in this paper.
Topics mentioned in this paper:
Pereira, Lis and Manguilimotan, Erlyn and Matsumoto, Yuji
Introduction
In this work, we analyze various Japanese corpora using a number of collocation and word similarity measures to deduce and suggest the best collocations for Japanese second language learners.
Introduction
In order to build a system that is more sensitive to constructions that are difficult for learners, we use word similarity measures that generate collocation candidates using a large Japanese language learner corpus.
Related Work
similarity measures are used.
Related Work
Our work follows the general approach, that is, uses similarity measures for generating the confusion set and association measures for ranking the best candidates.
Related Work
Similarity measures are used to generate the collocation candidates that are later ranked using association measures.
similarity measure is mentioned in 7 sentences in this paper.
Topics mentioned in this paper:
He, Zhengyan and Liu, Shujie and Li, Mu and Zhou, Ming and Zhang, Longkai and Wang, Houfeng
Abstract
Instead of utilizing simple similarity measures and their disjoint combinations, our method directly optimizes document and entity representations for a given similarity measure .
Abstract
A supervised fine-tuning stage follows to optimize the representation towards the similarity measure .
Conclusion
We propose a deep learning approach that automatically learns context-entity similarity measure for entity disambiguation.
Experiments and Analysis
When embedding our similarity measure sim(d, 6) into (Han et al., 2011), we achieve the best results on AIDA.
Introduction
ument and entity representations for a fixed similarity measure .
Introduction
In fact, the underlying representations for computing similarity measure add internal structure to the given similarity measure .
Introduction
The learned similarity measure can be readily incorporated into any existing collective approaches, which further boosts performance.
similarity measure is mentioned in 7 sentences in this paper.
Topics mentioned in this paper:
Zapirain, Beñat and Agirre, Eneko and Màrquez, Llu'is
Abstract
The best results are obtained with a novel second-order distributional similarity measure , and the positive effect is specially relevant for out-of-domain data.
Conclusions and Future Work
We have empirically shown how automatically generated selectional preferences, using WordNet and distributional similarity measures , are able to effectively generalize lexical features and, thus, improve classification performance in a large-scale argument classification task on the CoNLL-2005 dataset.
Related Work
Pantel and Lin (2000) obtained very good results using the distributional similarity measure defined by Lin (1998).
Results and Discussion
The second-order distributional similarity measures perform best overall, both in precision and recall.
Selectional Preference Models
We will refer to this similarity measure as simg‘n.
Selectional Preference Models
We will refer to these similarity measures as simE-ZE and simi’gg hereinafter.
similarity measure is mentioned in 6 sentences in this paper.
Topics mentioned in this paper:
Saha, Sujan Kumar and Mitra, Pabitra and Sarkar, Sudeshna
Abstract
A number of word similarity measures are proposed for clustering words for the Named Entity Recognition task.
Conclusion
A number of word similarity measures are used for clustering.
Evaluation of NE Recognition
Among the various similarity measures of clustering, improved results are obtained using the clus-
Evaluation of NE Recognition
ters which uses the similarity measurement based on proximity of the words to NE categories (defined in Section 3.3).
Word Clustering
The Euclidean distance is used to find the similarity between the above word vectors as a similarity measure .
Word Clustering
Using the above similarity measures we have used the k-means algorithm.
similarity measure is mentioned in 6 sentences in this paper.
Topics mentioned in this paper:
Mohler, Michael and Bunescu, Razvan and Mihalcea, Rada
Abstract
We combine several graph alignment features with lexical semantic similarity measures using machine learning techniques and show that the student answers can be more accurately graded than if the semantic measures were used in isolation.
Answer Grading System
In the final stage (Section 3.4), we produce an overall grade based upon the alignment scores found in the previous stage as well as the results of several semantic BOW similarity measures (Section 3.3).
Answer Grading System
All eight WordNet-based similarity measures listed in Section 3.3 plus the LSA model are used to produce these features.
Answer Grading System
In order to address this, we combine the graph alignment scores, which encode syntactic knowledge, with the scores obtained from semantic similarity measures .
Results
One surprise while building this system was the consistency with which the novel technique of question demoting improved scores for the BOW similarity measures .
similarity measure is mentioned in 6 sentences in this paper.
Topics mentioned in this paper:
Korkontzelos, Ioannis and Manandhar, Suresh
Evaluation setting and results
Our method was evaluated for each (P1, P2, P3) combination and similarity measures J0 and 197,, separately.
Introduction and related work
In this paper, we propose a novel unsupervised approach that compares the major senses of a MWE and its semantic head using distributional similarity measures to test the compositionality of the MWE.
Proposed approach
Lee (1999) shows that J performs better than other symmetric similarity measures such as cosine, Jensen-Shannon divergence, etc.
Proposed approach
Given the major uses of a MWE and its semantic head, the MWE is considered as compositional, when the corresponding distributional similarity measure (Jc or 197,) value is above a parameter threshold, sim.
Unsupervised parameter tuning
The best performing distributional similarity measure is an.
similarity measure is mentioned in 5 sentences in this paper.
Topics mentioned in this paper:
Abend, Omri and Cohen, Shay B. and Steedman, Mark
Discussion
One of the most effective similarity measures is the cosine similarity, which is a normalized dot product.
Discussion
In order to appreciate the effect of these advantages, we perform an experiment that takes H to be the set of all LCs of size 1, and uses a single similarity measure .
Our Proposal: A Latent LC Approach
where sim is some vector similarity measure .
Our Proposal: A Latent LC Approach
We use two common similarity measures: the vector cosine metric, and the BInc (Szpektor and Dagan, 2008) similarity measure .
Our Proposal: A Latent LC Approach
To do so, we use point-wise mutual information, and the conditional probabilities P(hf|hf) and POLE Similar measures have often been used for the unsupervised detection of MWEs (Villavicencio et al., 2007; Fazly and Stevenson, 2006).
similarity measure is mentioned in 5 sentences in this paper.
Topics mentioned in this paper:
Berant, Jonathan and Dagan, Ido and Goldberger, Jacob
Experimental Evaluation
Local algorithms We described 12 distributional similarity measures computed over our corpus (Section 5.1).
Experimental Evaluation
In addition, we obtained similarity lists learned by Lin and Pantel (2001), and replicated 3 similarity measures learned by Szpektor and Dagan (2008), over the RCV1 corpus7.
Experimental Evaluation
For each distributional similarity measure (altogether 16 measures), we learned a graph by inserting any edge (u, v) , when u is in the top K templates most similar to 2).
Learning Entailment Graph Edges
Similarity function We consider two similarity functions: The Lin (2001) similarity measure, and the Balanced Inclusion (BInc) similarity measure (Szpektor and Dagan, 2008).
similarity measure is mentioned in 5 sentences in this paper.
Topics mentioned in this paper:
Hartmann, Silvana and Gurevych, Iryna
FrameNet — Wiktionary Alignment
They align senses in WordNet to Wikipedia entries in a supervised setting using semantic similarity measures .
FrameNet — Wiktionary Alignment
Niemann and Gurevych (2011) combine two different types of similarity (i) cosine similarity on bag-of-words vectors (COS) and (ii) a personalized PageRank—based similarity measure (PPR).
FrameNet — Wiktionary Alignment
For each similarity measure , Niemann and Gurevych (2011) determine a threshold (tppr and
Related Work
The similarity measure is based on stem overlap of the candidates’ glosses expanded by WordNet domains, the WordNet synset, and the set of senses for a FrameNet frame.
similarity measure is mentioned in 4 sentences in this paper.
Topics mentioned in this paper:
Han, Xianpei and Zhao, Jun
Named Entity Disambiguation by Leveraging Semantic Knowledge
Because the key problem of named entity disambiguation is to measure the similarity between name observations, we integrate the structural semantic relatedness in the similarity measure , so that it can better reflect the actual similarity between name observations.
The Structural Semantic Relatedness Measure
The lexical relatedness lr between two WordNet concepts are measured using the Lin (l998)’s WordNet semantic similarity measure .
The Structural Semantic Relatedness Measure
The problem of quantifying the relatedness between nodes in a graph is not a new problem, e.g., the structural equivalence and structural similarity (the SimRank in Jeh and Widom (2002) and the similarity measure in Leicht et al.
The Structural Semantic Relatedness Measure
However, these similarity measures are not suitable for our task, because all of them assume that the edges are uniform so that they cannot take edge weight into consideration.
similarity measure is mentioned in 4 sentences in this paper.
Topics mentioned in this paper:
Bollegala, Danushka and Weir, David and Carroll, John
Domain Adaptation
For both POS tagging and sentiment classification, we experimented with several alternative approaches for feature weighting, representation, and similarity measures using development data, which we randomly selected from the training instances from the datasets described in Section 5.
Domain Adaptation
With respect to similarity measures, we experimented with cosine similarity and the similarity measure proposed by Lin (1998); cosine similarity performed consistently well over all the experimental settings.
Domain Adaptation
The feature representation was held fixed during these similarity measure comparisons.
O \
As an example of the distribution prediction method, in Table 3 we show the top 3 similar distributional features u in the books (source) domain, predicted for the electronics (target) domain word 21) = lightweight, by different similarity measures .
similarity measure is mentioned in 4 sentences in this paper.
Topics mentioned in this paper:
Li, Hao and Liu, Wei and Ji, Heng
Conclusion
Moreover, our approach can combine two similarity measures in a hybrid hashing scheme, which is beneficial to comprehensively modeling the document similarity.
Document Retrieval with Hashing
Given a query document vector q, we use the Cosine similarity measure to evaluate the similarity between q and a document a: in a dataset:
Document Retrieval with Hashing
Enable a hybrid hashing scheme combining two similarity measures .
Introduction
Furthermore, we make the hashing framework applicable to combine different similarity measures in NNS.
similarity measure is mentioned in 4 sentences in this paper.
Topics mentioned in this paper:
Muller, Philippe and Fabre, Cécile and Adam, Clémentine
Experiments: predicting relevance in context
Figure 3: Precision and recall on relevant links with respect to a threshold on the similarity measure (Lin’s score)
Experiments: predicting relevance in context
A straightforward parameter to include to predict the relevance of a link is of course the similarity measure itself, here Lin’s information measure.
Experiments: predicting relevance in context
This is already a big improvement on the use of the similarity measure alone (24%).
Introduction
A distributional thesaurus is a lexical network that lists semantic neighbours, computed from a corpus and a similarity measure between lexical items, which generally captures the similarity of contexts in which the items occur.
similarity measure is mentioned in 4 sentences in this paper.
Topics mentioned in this paper:
Abend, Omri and Rappoport, Ari
Algorithm
sim(h, h’) is a similarity measure between argument heads.
Algorithm
The similarity measure we use is based on the slot distributions of the arguments.
Algorithm
The similarity measure between two head words is then defined as the cosine measure of their vectors.
similarity measure is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Zhang, Huibin and Zhu, Mingjie and Shi, Shuming and Wen, Ji-Rong
Experiments
Siml and Sim2 respectively mean Formula 3.1 and Formula 3.2 are used in postprocessing as the similarity measure between
Experiments
Sim2 achieves more performance improvement than Siml, which demonstrates the effectiveness of the similarity measure in Formula 3.2.
Our Approach
One simple and straightforward similarity measure is the J accard
similarity measure is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Liu, Chang and Ng, Hwee Tou
Discussion and Future Work
5.2 Fractional Similarity Measures
Discussion and Future Work
In contrast, the linear-programming based TESLA metric allows fractional similarity measures between 0 (completely unrelated) and l (exact synonyms).
Discussion and Future Work
Supporting fractional similarity measures is nontrivial in the TESLA-CELAB framework.
similarity measure is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Krishnamurthy, Jayant and Mitchell, Tom
ConceptResolver
We use several string similarity measures as features, including SoftTFIDF (Cohen et al., 2003), Level 2 JaroWinkler (Cohen et al., 2003), Fellegi-Sunter (Fellegi and Sunter, 1969), and Monge-Elkan (Monge and Elkan, 1996).
Prior Work
Like other approaches (Basu et al., 2004; Xing et al., 2003; Klein et al., 2002), we learn a similarity measure for clustering based on a set of must-link and cannot-link constraints.
Prior Work
Unlike prior work, our algorithm exploits multiple views of the data to improve the similarity measure .
similarity measure is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Wu, Zhili and Markert, Katja and Sharoff, Serge
Genre Distance Measures
We can derive such distance measures from the genre hierarchy in a way similar to word similarity measures that were invented for lexical hierarchies such as WordNet (see (Pedersen et al., 2007) for an overview).
Genre Distance Measures
The Leaeoek & Chodorow similarity measure (Leacock and Chodorow, 1998) normalizes the path length measure (6) by the maximum number of nodes D when traversing down from the root.
Genre Distance Measures
Several other similarity measures have been proposed based on the Resnik similarity such as the one by (Lin, 1998):
similarity measure is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Charton, Eric and Meurs, Marie-Jean and Jean-Louis, Ludovic and Gagnon, Michel
Introduction
They are used as matching sequences to locate corresponding candidate entries in the KB, and then to disambiguate those candidates using similarity measures .
Introduction
This is usually done using similarity measures (such as cosine similarity, weighted J accard distance, KL divergence...) that evaluate the distance between a bag of words related to a candidate annotation, and the words surrounding the entity to annotate in the text.
Related Work
It proposes a disambiguation method that combines popularity-based priors, similarity measures , and coherence.
similarity measure is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Regneri, Michaela and Koller, Alexander and Pinkal, Manfred
Evaluation
The Levenshtein similarity measure , on the other hand, is too restrictive and thus results in comparatively high precisions, but very low recall.
Temporal Script Graphs
On the basis of this pseudo-parse, we compute the similarity measure sim:
Temporal Script Graphs
The semantic constraints check whether the event descriptions of the merged node would be sufficiently consistent according to the similarity measure from Section 5.2.
similarity measure is mentioned in 3 sentences in this paper.
Topics mentioned in this paper: