Index of papers in Proc. ACL that mention
  • cosine similarity
Carenini, Giuseppe and Ng, Raymond T. and Zhou, Xiaodong
Abstract
We adopt three cohesion measures: clue words, semantic similarity and cosine similarity as the weight of the edges.
Conclusions
We adopt three cohesion metrics, clue words, semantic similarity and cosine similarity , to measure the weight of the edges.
Empirical Evaluation
In Section 3.3, we developed three ways to compute the weight of an edge in the sentence quotation graph, i.e., clue words, semantic similarity based on WordNet and cosine similarity .
Empirical Evaluation
The widely used cosine similarity does not perform well.
Empirical Evaluation
The above experiments show that the widely used cosine similarity and the more sophisticated semantic similarity in WordNet are less accurate than the basic CWS in the summarization framework.
Extracting Conversations from Multiple Emails
and (3) cosine similarity that is based on the word TFIDF vector.
Extracting Conversations from Multiple Emails
3.3.3 Cosine Similarity
Extracting Conversations from Multiple Emails
Cosine similarity is a popular metric to compute the similarity of two text units.
Introduction
(Carenini et al., 2007), semantic similarity and cosine similarity .
Summarization Based on the Sentence Quotation Graph
In the rest of this paper, let CWS denote the Generalized ClueWordSummarizer when the edge weight is based on clue words, and let CWS-Cosine and CWS-Semantic denote the summarizer when the edge weight is cosine similarity and semantic similarity respectively.
cosine similarity is mentioned in 11 sentences in this paper.
Topics mentioned in this paper:
Jansen, Peter and Surdeanu, Mihai and Clark, Peter
Approach
This is a commonly used setup in the CQA community (Wang et al., 2009).4 Thus, for a given question, all its answers are fetched from the answer collection, and an initial ranking is constructed based on the cosine similarity between theirs and the question’s lemma vector representations, with lemmas weighted using tfidf (Ch.
Approach
The candidate answers are scored using a linear interpolation of two cosine similarity scores: one between the entire parent document and question (to model global context), and a second between the answer candidate and question (for local context).6 Because the number of answer candidates is typically large (e.g., equal to the number of paragraphs in the textbook), we return the N top candidates with the highest scores.
Experiments
The following hyper parameters were tuned using grid search to maximize P@1 on each development partition: (a) the segment matching thresholds that determine the minimum cosine similarity between an answer segment and a question for the segment to be labeled QSEG; and (b)
Models and Features
6We empirically observed that this combination of scores performs better than using solely the cosine similarity between the answer and question.
Models and Features
If text before or after a marker out to a given sentence range matches the entire text of the question (with a cosine similarity score larger than a threshold), that argument takes on the label QSEG, or OTHER otherwise.
Models and Features
The values of the discourse features are the mean of the similarity scores (e. g., cosine similarity using tfidf weighting) of the two marker arguments and the corresponding question.
Related Work
(2011) extracted 47 cue phrases such as because from a small collection of web documents, and used the cosine similarity between an answer candidate and a bag of words containing these cue phrases as a single feature in their reranking model for non-factoid why QA.
cosine similarity is mentioned in 12 sentences in this paper.
Topics mentioned in this paper:
Wang, Baoxun and Wang, Xiaolong and Sun, Chengjie and Liu, Bingquan and Sun, Lin
Experiments
Cosine Similarity .
Experiments
Given a question q and its :andidate answer 3, their cosine similarity can be :omputed as follows:
Experiments
Method P@1(%) MRR (%) Nearest Answer 21.25 38.72 Cosine Similarity 23.15 43.50 HowNet 22.55 41.63 KL divergence 25 .30 51.40 DBN (without FT) 41.45 59.64 DBN (with FT) 45.00 62.03
Introduction
Because of this situation, the traditional relevance computing methods based on word co-occurrence, such as Cosine similarity and KL—divergence, are not effective for question-
cosine similarity is mentioned in 8 sentences in this paper.
Topics mentioned in this paper:
Rothe, Sascha and Schütze, Hinrich
CoSimRank
This is similar to cosine similarity except that the l-norm is used instead of the 2-norm.
Extensions
We are not including this method in our experiments, but we will give the equation here, as traditional document similarity measures (e.g., cosine similarity ) perform poorly on this task although there also are known alternatives with good results (Sahami and Heilman, 2006).
Extensions
To calculate PPR+cos, we computed 20 iterations with a decay factor of 0.8 and used the cosine similarity with the 2-norm in the denominator to compare two vectors.
Extensions
We compute 20 iterations of PPR+cos to reach convergence and then calculate a single cosine similarity .
Related Work
Another important similarity measure is cosine similarity of Personalized PageRank (PPR) vectors.
Related Work
These approaches use at least one of cosine similarity , PageRank and SimRank.
cosine similarity is mentioned in 6 sentences in this paper.
Topics mentioned in this paper:
Li, Hao and Liu, Wei and Ji, Heng
Document Retrieval with Hashing
Given a query document vector q, we use the Cosine similarity measure to evaluate the similarity between q and a document a: in a dataset:
Document Retrieval with Hashing
However, such a brute-force search does not scale to massive datasets since the search time complexity for each query is 0(n); additionally, the computational cost spent on Cosine similarity calculation is also nontrivial.
Document Retrieval with Hashing
(1) It tries to preserve the Cosine similarity of the original data with a probabilistic guarantee (Charikar, 2002).
Experiments
We first evaluate the quality of term vectors and ITQ binary codes by conducting the whole list Cosine similarity ranking and hamming distance ranking, respectively.
Experiments
For each query document, the top-K candidate documents with highest Cosine similarity scores and shortest hamming distances are returned, then we calculate the average precision for each K. Fig.
Introduction
for example, one can store 250 million documents with 1.9G memory using only 64 bits for each document while a large news corpus such as the English Gigaword fifth edition1 stores 10 million documents in a 26G hard drive; 2) the time efficiency of manipulating binary codes, for example, computing the hamming distance between a pair of binary codes is several orders of magnitude faster than computing the real-valued cosine similarity over a pair of document vectors.
cosine similarity is mentioned in 6 sentences in this paper.
Topics mentioned in this paper:
Bamman, David and Dyer, Chris and Smith, Noah A.
Evaluation
To illustrate how the model described above can learn geographically-informed semantic representations of words, table 1 displays the terms with the highest cosine similarity to wicked in Kansas and Massachusetts after running our joint model on the full 1.1 billion words of Twitter data; while wicked in Kansas is close to other evaluative terms like evil and pure and religious terms like gods and spirit, in Massachusetts it is most similar to other intensifiers like super, ridiculously and insanely.
Evaluation
Table 2 likewise presents the terms with the highest cosine similarity to city in both California and New York; while the terms most evoked by city in California include regional locations like Chinatown, Los Angeles’ South Bay and San Francisco’s East Bay, in New York the most similar terms include hamptons, upstate and borough
Evaluation
Table 1: Terms with the highest cosine similarity to wicked in Kansas and Massachusetts.
cosine similarity is mentioned in 6 sentences in this paper.
Topics mentioned in this paper:
Qazvinian, Vahed and Radev, Dragomir R.
Diversity-based Ranking
The edges between corresponding nodes (dz) represent the cosine similarity between them is above a threshold (0.10 following (Erkan and Radev, 2004)).
Diversity-based Ranking
Maximal Marginal Relevance (MMR) (Carbonell and Goldstein, 1998) uses the pairwise cosine similarity matrix and greedily chooses sentences that are the least similar to those already in the summary.
Diversity-based Ranking
C-LexRank is a clustering-based model in which the cosine similarities of document pairs are used to build a network of documents.
Prior Work
Once a lexical similarity graph is built, they modify the graph based on cluster information and perform LexRank on the modified cosine similarity graph.
cosine similarity is mentioned in 6 sentences in this paper.
Topics mentioned in this paper:
Pitler, Emily and Louis, Annie and Nenkova, Ani
Abstract
Focus, coherence and referential clarity are best evaluated by a class of features measuring local coherence on the basis of cosine similarity between sentences, coreference information, and summarization specific features.
Indicators of linguistic quality
Cosine similarity We use cosine similarity to compute the overlap of words in adjacent sentences s,- and 3H1 as a measure of continuity.
Indicators of linguistic quality
We compute the min, max, and average value of cosine similarity over the entire summary.
Indicators of linguistic quality
Cosine similarity is thus indicative of both continuity and redundancy.
Results and discussion
For all four other questions, the best feature set is Continuity, which is a combination of summarization specific features, coreference features and cosine similarity of adjacent sentences.
Results and discussion
We now investigate to what extent each of its components—summary-specific features, coreference, and cosine similarity between adjacent sentences—contribute to performance.
cosine similarity is mentioned in 6 sentences in this paper.
Topics mentioned in this paper:
Ding, Shilin and Cong, Gao and Lin, Chin-Yew and Zhu, Xiaoyan
Context and Answer Detection
The word similarity is based on cosine similarity of TF/IDF weighted vectors.
Context and Answer Detection
- Cosine similarity with the question
Context and Answer Detection
- Cosine similarity between contiguous sentences
Related Work
(2006a) used cosine similarity to match students’ query with reply posts for discussion-bot.
cosine similarity is mentioned in 6 sentences in this paper.
Topics mentioned in this paper:
Saha, Sujan Kumar and Mitra, Pabitra and Sarkar, Sudeshna
Introduction
For clustering we use a number of word similarities like cosine similarity among words and co-occurrence, along with the k-means clustering algorithm.
Word Clustering
3.1 Cosine Similarity based on Sentence Level Co-occurrence
Word Clustering
Then we measure cosine similarity between the word vectors.
Word Clustering
The cosine similarity between two word vectors (fl and E”) with dimension d is measured as:
cosine similarity is mentioned in 6 sentences in this paper.
Topics mentioned in this paper:
Pereira, Lis and Manguilimotan, Erlyn and Matsumoto, Yuji
Related Work
We computed the similarity between co-occurrence vectors using different metrics: Cosine Similarity , Dice coefficient (Curran, 2004), Kullback—Leibler divergence or KL divergence or relative entropy (Kullback and Leibler, 1951) and the J enson-Shannon divergence (Lee, 1999).
Related Work
One year data (1991) were used to extract the “noun wo verb” tuples to compute word similarity (using cosine similarity metric) and collocation scores.
Related Work
These data are necessary to compute the word similarity (using cosine similarity metric) and collocation scores.
cosine similarity is mentioned in 5 sentences in this paper.
Topics mentioned in this paper:
Ogura, Yukari and Kobayashi, Ichiro
Experiment
To measure the latent similarity among documents, we construct topic vectors with the topic probabilistic distribution, and then adopt the Jensen-Shannon divergence to measures it, on the other hand, in the case of using document vectors we adopt cosine similarity .
Experiment
Table 1: Extracting important sentences Methods Measure Accuracy F-value PageRank J enshen-Shannon 0.567 0.485 Cosine similarity 0.287 0.291 tf.
Experiment
idf J enshen-Shannon 0.550 0.43 5 Cosine similarity 0.275 0.270
cosine similarity is mentioned in 5 sentences in this paper.
Topics mentioned in this paper:
Sauper, Christina and Barzilay, Regina
Method
As a similarity function, we use cosine similarity weighted with TF*IDF.
Method
We define sim(efl, ej/lx) as the cosine similarity between excerpts 63-; from topic 253- and ej/l/ from topic if.
Method
If excerpts ejl and ej/l/ have cosine similarity
Rank(eij1...€ij7~,Wj)
For each topic present in the human-authored article, the Oracle selects the excerpt from our full model’s candidate excerpts with the highest cosine similarity to the human-authored text.
cosine similarity is mentioned in 5 sentences in this paper.
Topics mentioned in this paper:
Wang, Dong and Liu, Yang
Conclusion and Future Work
Our experiments show that both methods are able to improve the baseline approach, and we find that the cosine similarity between utterances or between an utterance and the whole document is not as useful as in other document summarization tasks.
Opinion Summarization Methods
0 sim(s, D) is the cosine similarity between DA 3 and all the utterances in the dialogue from the same speaker, D. It measures the relevancy of s to the entire dialogue from the target speaker.
Opinion Summarization Methods
For cosine similarity measure, we use TF*IDF (term frequency, inverse document frequency) term weighting.
Opinion Summarization Methods
is modeled as an adjacency matrix, where each node represents a sentence, and the weight of the edge between each pair of sentences is their similarity ( cosine similarity is typically used).
cosine similarity is mentioned in 5 sentences in this paper.
Topics mentioned in this paper:
Zweig, Geoffrey and Platt, John C. and Meek, Christopher and Burges, Christopher J.C. and Yessenalina, Ainur and Liu, Qiang
Experimental Results 5.1 Data Resources
Of all the methods in isolation, the simple approach of Section 4.1 — to use the total cosine similarity between a potential answer and the other words in the sentence — has performed best.
Experimental Results 5.1 Data Resources
For the LSA model, the linear combination has three inputs: the total word similarity, the cosine similarity between the sum of the answer word vectors and the sum of the rest of sentence’s word vectors, and the number of out—of—vocabulary terms in the answer.
Sentence Completion via Latent Semantic Analysis
An important property of SVD is that the rows of US — which represents the words — behave similarly to the original rows of W, in the sense that the cosine similarity between two rows in US approximates the cosine similarity between the corre—
Sentence Completion via Latent Semantic Analysis
sponding rows in W. Cosine similarity is defined as
Sentence Completion via Latent Semantic Analysis
Let m be the smallest cosine similarity between h and any word in the vocabulary V: m = minwev sim(h, w).
cosine similarity is mentioned in 5 sentences in this paper.
Topics mentioned in this paper:
Bhagat, Rahul and Ravichandran, Deepak
Acquiring Paraphrases
We use cosine similarity , which
Acquiring Paraphrases
As described in Section 3.2, we find paraphrases of a phrase p,- by finding its nearest neighbors based on cosine similarity between the feature vector of pi and other phrases.
Acquiring Paraphrases
If n is the number of vectors and d is the dimensionality of the vector space, finding cosine similarity between each pair of vectors has time complexity 0(n2 d).
cosine similarity is mentioned in 4 sentences in this paper.
Topics mentioned in this paper:
Wolfe, Travis and Van Durme, Benjamin and Dredze, Mark and Andrews, Nicholas and Beller, Charley and Callison-Burch, Chris and DeYoung, Jay and Snyder, Justin and Weese, Jonathan and Xu, Tan and Yao, Xuchen
Evaluation
The data was generated by clustering similar news stories from Gigaword using TF-IDF cosine similarity of their headlines.
Evaluation
Doc-pair Cosine Similarity
Evaluation
The x-axis represents the cosine similarity between the document pairs.
Results
Additionally, there is more data in the MTC dataset which has low cosine similarity than in RF.
cosine similarity is mentioned in 4 sentences in this paper.
Topics mentioned in this paper:
Alhelbawy, Ayman and Gaizauskas, Robert
Solution Graph
cos: The cosine similarity between the named entity textual mention and the KB entry title.
Solution Graph
ijim: While the cosine similarity between a textual mention in the document and the candidate
Solution Graph
The cosine similarity between “Essex” and “Danbury, Essex” is higher than that between “Essex” and “Essex County Cricket Club”, which is not helpful in the NED setting.
cosine similarity is mentioned in 4 sentences in this paper.
Topics mentioned in this paper:
Nenkova, Ani and Louis, Annie
Features
Cosine similarity between the document vector representations is probably the easiest and most commonly used among the various similarity measures.
Features
The cosine similarity between two (document representation) vectors v1 and 212 is given by 0036 = W. A value of 0 indicates that the vectors are orthogonal and dissimilar, a value of 1 indicates perfectly similar documents in terms of the words con-
Features
To compute the cosine overlap features, we find the pairwise cosine similarity between each two documents in an input set and compute their average.
cosine similarity is mentioned in 4 sentences in this paper.
Topics mentioned in this paper:
Qu, Zhonghua and Liu, Yang
Thread Structure Tagging
Cosine similarity with previous sentence.
Thread Structure Tagging
Here we use the cosine similarity between sentences, where each sentence is represented as a vector of words, with term weight calculated using TD-IDF (term frequency times inverse document frequency).
Thread Structure Tagging
* Cosine similarity with previous sentence.
cosine similarity is mentioned in 4 sentences in this paper.
Topics mentioned in this paper:
Kulkarni, Anagha and Callan, Jamie
Finding the Homographs in a Lexicon
Cohesiveness Score: Mean of the cosine similarities between each pair of definitions of w.
Finding the Homographs in a Lexicon
Average Number of Null Similarities: The number of definition pairs that have zero cosine similarity score (no word overlap).
Finding the Homographs in a Lexicon
The last feature sorts the pairwise cosine similarity scores in ascending order, prunes the top n% of the scores, and uses the maximum remaining score as the feature value.
cosine similarity is mentioned in 4 sentences in this paper.
Topics mentioned in this paper:
Mehdad, Yashar and Carenini, Giuseppe and Ng, Raymond T.
Experimental Setup
1) Cosine- 1 st: we rank the utterances in the chat log based on the cosine similarity between the utterance and query.
Experimental Setup
2) Cosine-all: we rank the utterances in the chat log based on the cosine similarity between the utterance and query and then select the utterances with a cosine similarity greater than 0;
Experimental Setup
Query Relevance: another interesting observation is that relying only on the cosine similarity (i.e., cosine-all) to measure the query relevance presents a quite strong baseline.
Phrasal Query Abstraction Framework
We use the K-mean clustering algorithm by cosine similarity as a distance function between sentence vectors composed of tfidf scores.
cosine similarity is mentioned in 4 sentences in this paper.
Topics mentioned in this paper:
Bhat, Suma and Xue, Huichao and Yoon, Su-Youn
Models for Measuring Grammatical Competence
The similarity between a test response and a score-specific vector is then calculated by a cosine similarity metric.
Models for Measuring Grammatical Competence
Although a total of 4 cosine similarity scores (one per score group) were generated, only 0034from among the four similarity scores, and cosmazc,
Models for Measuring Grammatical Competence
0 0034: the cosine similarity score between the test response and the vector of POS bigrams for the highest score class (level 4); and,
cosine similarity is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Daxenberger, Johannes and Gurevych, Iryna
Machine Learning with Edit-Turn-Pairs
We used the cosine similarity , longest common subsequence, and word n- gram similarity measures.
Machine Learning with Edit-Turn-Pairs
Cosine similarity was applied on binary weighted term vectors (L2 norm).
Machine Learning with Edit-Turn-Pairs
Cosine similarity , longest common subsequence, and word n-gram similarity were also applied to measure the similarity between the edit comment and the turn text as well as the similarity between the edit comment and the turn topic name.
cosine similarity is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Abend, Omri and Cohen, Shay B. and Steedman, Mark
Discussion
One of the most effective similarity measures is the cosine similarity , which is a normalized dot product.
Discussion
Indeed, taking Hp as above, and cosine similarity as the only feature (i.e., w E R), yields the distribution
Our Proposal: A Latent LC Approach
These measures give complementary perspectives on the similarity between the predicates, as the cosine similarity is symmetric between the LHS and RHS predicates, while BInc takes into account the directionality of the inference relation.
cosine similarity is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Guo, Weiwei and Li, Hao and Ji, Heng and Diab, Mona
Experiments
Evaluation: The similarity between a tweet and a news article is measured by cosine similarity .
Experiments
4 The cosine similarity
WTMF on Graphs
In the WTMF model, we would like the latent vectors of two text nodes Q.,j1,Q.,j-2 to be as similar as possible, namely that their cosine similarity to be close to 1.
cosine similarity is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Liu, Xiaohua and Li, Yitong and Wu, Haocheng and Zhou, Ming and Wei, Furu and Lu, Yi
Our Method
o 31(mi, mj): The cosine similarity of 75071;) and t(mj); and tweets are represented as TF-IDF vectors;
Our Method
0 32(mi, mj): The cosine similarity of 75071;) and t(mj); and tweets are represented as topic distribution vectors;
Related Work
SemTag uses the TAP knowledge base5, and employs the cosine similarity with TF-IDF weighting scheme to compute the match degree between a mention and an entity, achieving an accuracy of around 82%.
cosine similarity is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Feng, Yansong and Lapata, Mirella
Extractive Caption Generation
Cosine Similarity Word overlap is admittedly a naive measure of similarity, based on lexical identity.
Results
We compare four extractive models based on word overlap, cosine similarity , and two probabilistic similarity measures, namely KL and JS divergence and two abstractive models based on words (see equation (8)) and phrases (see equation (15)).
Results
As can be seen the probabilistic models (KL and J S divergence) outperform word overlap and cosine similarity (all differences are statistically significant, p < 0.01).6 They make use of the same topic model as the image annotation model, and are thus able to select sentences that cover common content.
cosine similarity is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Gao, Dehong and Li, Wenjie and Zhang, Renxian
Experiments and Evaluations
Borrowing this idea, for each sub-summary in a human-generated summary, we find its most matched sub-summary (judged by the cosine similarity measure) in the corresponding system-generated summary and then define the correlation according to the concordance between the two
Experiments and Evaluations
For the semantic-based approach, we compare three different approaches to defining the subtopic number K: (1) Semantic-based 1: Following the approach proposed in (Li et al., 2007), we first derive the matrix of tweet cosine similarity .
Sequential Summarization
For a tweet in a peak area, the linear combination of two measures is considered to evaluate its significance to be a sub-summary: (l) subtopic representativeness measured by the cosine similarity between the tweet and the centroid of all the tweets in the same peak area; (2) crowding endorsement measured by the times that the tweet is re-tweeted normalized by the total number of re-tweeting.
cosine similarity is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Ferschke, Oliver and Gurevych, Iryna and Rittberger, Marc
Selection of Reliable Training Instances
We can then estimate the topical similarity of two article sets by calculating the cosine similarity of their category frequency vectors C712=Aand6722= Bas
Selection of Reliable Training Instances
Cosine Similarity
Selection of Reliable Training Instances
Table 3: Cosine similarity scores between the category frequency vectors of the flawed article sets and the respective random or reliable negatives
cosine similarity is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Huang, Eric and Socher, Richard and Manning, Christopher and Ng, Andrew
Experiments
The nearest neighbors of a word are computed by comparing the cosine similarity between the center word and all other words in the dictionary.
Experiments
Table 1: Nearest neighbors of words based on cosine similarity .
Experiments
Table 2: Nearest neighbors of word embeddings learned by our model using the multi-prototype approach based on cosine similarity .
cosine similarity is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Chambers, Nathanael and Jurafsky, Dan
Learning Templates from Raw Text
Vector-based approaches are often adopted to represent words as feature vectors and compute their distance with cosine similarity .
Learning Templates from Raw Text
Distance is the cosine similarity between bag-of-words vector representations.
Learning Templates from Raw Text
We measure similarity using cosine similarity between the vectors in both approaches.
cosine similarity is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Wu, Zhili and Markert, Katja and Sharoff, Serge
Discussion
An alternative to structural distance measures would be distance measures between the genres based on pairwise cosine similarities between them.
Discussion
To assess this, we aggregated all character 4-gram training vectors of each genre and calculated standard cosine similarities .
Discussion
Inspecting the distance matrix visually, we determined that the cosine similarity could clearly distinguish between Fiction and NonFiction texts but not between any other genres.
cosine similarity is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Thater, Stefan and Fürstenau, Hagen and Pinkal, Manfred
Experiment: Ranking Word Senses
The WordNet senses are then ranked according to the cosine similarity between their sense vector and the contextually constrained target verb vector.
Experiments: Ranking Paraphrases
Therefore the choice of which word is contextualized does not strongly influence their cosine similarity , and contextualizing both should not add any useful information.
Experiments: Ranking Paraphrases
we compute [[swapOBmeadfl and compare it to the lifted first-order vectors of all paraphrase candidates, LOBJ([hint]) and LOBJ([star]), using cosine similarity .
cosine similarity is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Qazvinian, Vahed and Radev, Dragomir R.
Impact on Survey Generation
LexRank is a multidocument summarization system, which first builds a cosine similarity graph of all the candidate sentences.
Proposed Method
To formalize this assumption we use the sigmoid of the cosine similarity of two sentences to build it.
Proposed Method
Intuitively, if a sentence has higher similarity with the reference paper, it should have a higher potential of being in class 1 or C. The flag of each sentence here is a value between 0 and l and is determined by its cosine similarity to the reference.
cosine similarity is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Shnarch, Eyal and Barak, Libby and Dagan, Ido
Application Oriented Evaluations
During classification cosine similarity is measured between the feature vector of the classified document and the expanded vectors of all categories.
Application Oriented Evaluations
The first avoids any expansion, classifying documents based on cosine similarity with category names only.
The asterisk denotes an incorrect rule
We also examined another filtering score, the cosine similarity between the vectors representing the two rule sides in LSA (Latent Semantic Analysis) space (Deerwester et al., 1990).
cosine similarity is mentioned in 3 sentences in this paper.
Topics mentioned in this paper: