Index of papers in Proc. ACL 2013 that mention
  • similarity measure
Huang, Hongzhao and Wen, Zhen and Yu, Dian and Ji, Heng and Sun, Yizhou and Han, Jiawei and Li, He
Abstract
Then we propose various novel similarity measurements including surface features, meta-path based semantic features and social correlation features and combine them in a learning-to-rank framework.
Introduction
0 We propose two new similarity measures , as well as integrating temporal information into
Introduction
the similarity measures to generate global semantic features.
Target Candidate Ranking
We first extract surface features between the morph and the candidate based on measuring orthographic similarity measures which were commonly used in entity coreference resolution (e.g.
Target Candidate Ranking
4.2.3 Meta-Path-Based Semantic Similarity Measurements
Target Candidate Ranking
We then adopt meta-path-based similarity measures (Sun et al., 2011a; Sun et al., 2011b), which are defined over heterogeneous networks to extract semantic features.
similarity measure is mentioned in 14 sentences in this paper.
Topics mentioned in this paper:
Melamud, Oren and Berant, Jonathan and Dagan, Ido and Goldberger, Jacob and Szpektor, Idan
Background and Model Setting
where sim(v, v’) is a vector similarity measure .
Background and Model Setting
We note that the general DIRT scheme may be used while employing other “base” vector similarity measures .
Background and Model Setting
This issue has been addressed in a separate line of research which introduced directional similarity measures suitable for inference relations (Bhagat et al., 2007; Szpektor and Dagan, 2008; Kotlerman et al., 2010).
Introduction
Our scheme can be applied on top of any context-insensitive “base” similarity measure for rule learning, which operates at the word level, such as Cosine or Lin (Lin, 1998).
Introduction
We apply our two-level scheme over three state-of-the-art context-insensitive similarity measures .
similarity measure is mentioned in 30 sentences in this paper.
Topics mentioned in this paper:
Mohtarami, Mitra and Lan, Man and Tan, Chew Lim
Evaluation and Results
To apply our PSSS on IQAPS inference task, we use it as the sentiment similarity measure in the algorithm explained in Figure 4.
Evaluation and Results
The second row of Table 4 show the results of using a popular semantic similarity measure , PMI, as the sentiment similarity (SS) measure in Figure 4.
Evaluation and Results
Table 4 shows the effectiveness of our sentiment similarity measure .
Introduction
Semantic similarity measures such as Latent Semantic Analysis (LSA) (Landauer et al., 1998) can effectively capture the similarity between semantically related words like "car" and "automobile", but they are less effective in relating words with similar sentiment orientation like "excellent" and "superior".
Introduction
We show that our approach effectively outperforms the semantic similarity measures in two NLP tasks: Indirect yes/no Question Answer Pairs (IQAPs) Inference and Sentiment Orientation (S0) prediction that are described as follows:
Introduction
Previous research utilized the semantic relations between words obtained from WordNet (Hassan and Radev, 2010) and semantic similarity measures (e.g.
Related Works
They also utilized Latent Semantic Analysis (LSA) (Landauer et al., 1998) as another semantic similarity measure .
Related Works
However, both PM and LSA are semantic similarity measure .
Related Works
(2012), we used two semantic similarity measures (PMI and LSA) for the IQAP inference task.
Sentiment Similarity through Hidden Emotions
As we discussed above, semantic similarity measures are less effective to infer sentiment similarity between word pairs.
similarity measure is mentioned in 12 sentences in this paper.
Topics mentioned in this paper:
Moreno, Jose G. and Dias, Gaël and Cleuziou, Guillaume
Abstract
Within this context, we propose a new methodology that adapts the classical K -means algorithm to a third-order similarity measure initially developed for NLP tasks.
Conclusions
In this paper, we proposed a new PRC approach which (1) is based on the adaptation of the K -means algorithm to third-order similarity measures and (2) proposes a coherent stopping criterion.
Evaluation
In particular, p is the size of the word feature vectors representing both Web snippets and centroids (p = 2.5), K is the number of clusters to be found (K = 2..10) and 8(Wik, 14/31) is the collocation measure integrated in the InfoSimba similarity measure .
Introduction
feature vectors are hard to define in small collections of short text fragments (Timonen, 2013), (2) existing second-order similarity measures such as the cosine are unadapted to capture the semantic similarity between small texts, (3) Latent Semantic Analysis has evidenced inconclusive results (Osinski and Weiss, 2005) and (4) the labeling process is a surprisingly hard extra task (Carpineto et al., 2009).
Introduction
For that purpose, we propose a new methodology that adapts the classical K -means algorithm to a third-order similarity measure initially developed for Topic Segmentation (Dias et al., 2007).
Polythetic Post-Retrieval Clustering
Within the context of PRC, the K -means algorithm needs to be adapted to integrate third-order similarity measures (Mihalcea et al., 2006; Dias et al., 2007).
Polythetic Post-Retrieval Clustering
Third-order similarity measures, also called weighted second-order similarity measures, do not rely on exact matches of word features as classical second-order similarity measures (6. g. the cosine metric), but rather evaluate similarity based on related matches.
Polythetic Post-Retrieval Clustering
In this paper, we propose to use the third-order similarity measure called InfoSimba introduced in (Dias et al., 2007) for Topic Segmentation and implement its simplified version 83 in Equation 2.
similarity measure is mentioned in 12 sentences in this paper.
Topics mentioned in this paper:
Pilehvar, Mohammad Taher and Jurgens, David and Navigli, Roberto
A Unified Semantic Representation
In order to compare semantic signatures, we adopt the Cosine similarity measure as a baseline method.
Experiment 1: Textual Similarity
The top-ranking participating systems in the SemEval-2012 task were generally supervised systems utilizing a variety of lexical resources and similarity measurement techniques.
Experiment 1: Textual Similarity
3.3 Similarity Measure Analysis
Experiment 1: Textual Similarity
In addition, we present in the table correlation scores for four other similarity measures reported by B'ar et al.
Experiment 2: Word Similarity
Different evaluation methods exist in the literature for evaluating the performance of a word-level semantic similarity measure ; we adopted two well-established benchmarks: synonym recognition and correlating word similarity judgments with those from human annotators.
Experiment 3: Sense Similarity
We adopt this task as a way of evaluating our similarity measure at the sense level.
similarity measure is mentioned in 13 sentences in this paper.
Topics mentioned in this paper:
Razmara, Majid and Siahbani, Maryam and Haffari, Reza and Sarkar, Anoop
Collocational Lexicon Induction
For each of these paraphrases, a DP is constructed and compared to that of the oov word using a similarity measure (Section 2.2).
Collocational Lexicon Induction
where t is a phrase on the target side, 0 is the oov word or phrase, and s is a paraphrase of 0. p(s|0) is estimated using a similarity measure over DPs and p(t|s) is coming from the phrase-table.
Collocational Lexicon Induction
2.3 Similarity Measures
Experiments & Results 4.1 Experimental Setup
In Section 2.2 and 2.3, different types of association measures and similarity measures have been explained to build and compare distributional profiles.
Experiments & Results 4.1 Experimental Setup
As the results show, the combination of PMI as association measure and cosine as DP similarity measure outperforms the other possible combinations.
Graph-based Lexicon Induction
Each phrase type represents a vertex in the graph and is connected to other vertices with a weight defined by a similarity measure between the two profiles (Section 2.3).
Graph-based Lexicon Induction
However based on the definition of the similarity measures using context, it is quite possible that an oov node and a labeled node which are connected to the same unlabeled node do not share any context words and hence are not directly connected.
Graph-based Lexicon Induction
In such a graph, the similarity of each pair of nodes is computed using one of the similarity measures discussed above.
similarity measure is mentioned in 12 sentences in this paper.
Topics mentioned in this paper:
Ferret, Olivier
Abstract
As a consequence, improving such thesaurus is an important issue that is mainly tackled indirectly through the improvement of semantic similarity measures .
Improving a distributional thesaurus
As in (Lin, 1998) or (Curran and Moens, 2002a), this building is based on the definition of a semantic similarity measure from a corpus.
Improving a distributional thesaurus
For the extraction of distributional data and the characteristics of the distributional similarity measure , we adopted the options of (Ferret, 2010), resulting from a kind of grid search procedure performed with the extended TOEFL test proposed in (Freitag et al., 2005) as an optimization objective.
Improving a distributional thesaurus
o similarity measure between contexts, for evaluating the semantic similarity of two words = Cosine measure.
Introduction
Following work such as (Grefenstette, 1994), a widespread way to build a thesaurus from a corpus is to use a semantic similarity measure for extracting the semantic neighbors of the entries of the thesaurus.
Introduction
Work based on WordNet-like lexical networks for building semantic similarity measures such as (Budanitsky and Hirst, 2006) or (Pedersen et al., 2004) falls into this category.
Introduction
A part of these proposals focus on the weighting of the elements that are part of the contexts of words such as (Broda et al., 2009), in which the weights of context elements are turned into ranks, or (Zhitomirsky-Geffet and Dagan, 2009), followed and extended by (Yamamoto and Asakura, 2010), that proposes a bootstrapping method for modifying the weights of context elements according to the semantic neighbors found by an initial distributional similarity measure .
Principles
For instance, features such as ngrams of words or ngrams of parts of speech are not considered whereas they are widely used in tasks such as word sense disambiguation (WSD) for instance, probably because they would lead to very large models and because similarity measures such as the Cosine measure are not necessarily suitable for heterogeneous representations (Alexandrescu and Kirchhoff, 2007).
Related work
As a consequence, the improvement of such thesaurus is generally not directly addressed but is a possible consequence of the improvement of semantic similarity measures .
similarity measure is mentioned in 10 sentences in this paper.
Topics mentioned in this paper:
Veale, Tony and Li, Guofu
Abstract
We show also how Thesaurus Rex supports a novel, generative similarity measure for WordNet.
Related Work and Ideas
more rounded similarity measures .
Related Work and Ideas
A similarity measure can draw on other
Related Work and Ideas
Their best similarity measure achieves a remarkable 0.93 correlation with human judgments on the Miller & Charles word-pair set.
Seeing is Believing (and Creating)
Using WordNet, for instance, a similarity measure can vertically converge on a common superordinate category of both inputs, and generate a single numeric result based on their distance to, and the information content of, this common generalization.
Seeing is Believing (and Creating)
To be as useful for creative tasks as they are for conventional tasks, we need to re-imagine our computational similarity measures as generative rather than selective, expansive rather than reductive, divergent as well as convergent and lateral as well as vertical.
Seeing is Believing (and Creating)
Section 2 provides a brief overview of past work in the area of similarity measurement , before section 3 describes a simple bootstrapping loop for acquiring richly diverse perspectives from the web for a wide variety of familiar ideas.
similarity measure is mentioned in 8 sentences in this paper.
Topics mentioned in this paper:
He, Zhengyan and Liu, Shujie and Li, Mu and Zhou, Ming and Zhang, Longkai and Wang, Houfeng
Abstract
Instead of utilizing simple similarity measures and their disjoint combinations, our method directly optimizes document and entity representations for a given similarity measure .
Abstract
A supervised fine-tuning stage follows to optimize the representation towards the similarity measure .
Conclusion
We propose a deep learning approach that automatically learns context-entity similarity measure for entity disambiguation.
Experiments and Analysis
When embedding our similarity measure sim(d, 6) into (Han et al., 2011), we achieve the best results on AIDA.
Introduction
ument and entity representations for a fixed similarity measure .
Introduction
In fact, the underlying representations for computing similarity measure add internal structure to the given similarity measure .
Introduction
The learned similarity measure can be readily incorporated into any existing collective approaches, which further boosts performance.
similarity measure is mentioned in 7 sentences in this paper.
Topics mentioned in this paper:
Pereira, Lis and Manguilimotan, Erlyn and Matsumoto, Yuji
Introduction
In this work, we analyze various Japanese corpora using a number of collocation and word similarity measures to deduce and suggest the best collocations for Japanese second language learners.
Introduction
In order to build a system that is more sensitive to constructions that are difficult for learners, we use word similarity measures that generate collocation candidates using a large Japanese language learner corpus.
Related Work
similarity measures are used.
Related Work
Our work follows the general approach, that is, uses similarity measures for generating the confusion set and association measures for ranking the best candidates.
Related Work
Similarity measures are used to generate the collocation candidates that are later ranked using association measures.
similarity measure is mentioned in 7 sentences in this paper.
Topics mentioned in this paper:
Hartmann, Silvana and Gurevych, Iryna
FrameNet — Wiktionary Alignment
They align senses in WordNet to Wikipedia entries in a supervised setting using semantic similarity measures .
FrameNet — Wiktionary Alignment
Niemann and Gurevych (2011) combine two different types of similarity (i) cosine similarity on bag-of-words vectors (COS) and (ii) a personalized PageRank—based similarity measure (PPR).
FrameNet — Wiktionary Alignment
For each similarity measure , Niemann and Gurevych (2011) determine a threshold (tppr and
Related Work
The similarity measure is based on stem overlap of the candidates’ glosses expanded by WordNet domains, the WordNet synset, and the set of senses for a FrameNet frame.
similarity measure is mentioned in 4 sentences in this paper.
Topics mentioned in this paper: