Index of papers in Proc. ACL that mention
  • tree kernel
Sun, Jun and Zhang, Min and Tan, Chew Lim
Abstract
We propose Bilingual Tree Kernels (BTKs) to capture the structural similarities across a pair of syntactic translational equivalences and apply BTKs to subtree alignment along with some plain features.
Abstract
Our study reveals that the structural features embedded in a bilingual parse tree pair are very effective for subtree alignment and the bilingual tree kernels can well capture such features.
Bilingual Tree Kernels
2.1 Independent Bilingual Tree Kernel (iBTK)
Bilingual Tree Kernels
In order to compute the dot product of the feature vectors in the exponentially high dimensional feature space, we introduce the tree kernel functions as follows:
Bilingual Tree Kernels
The iBTK is defined as a composite kernel consisting of a source tree kernel and a target tree kernel which measures the source and the target structural similarity respectively.
Introduction
Alternatively, convolution parse tree kernels (Collins and Duffy, 2001), which implicitly explore the tree structure information, have been successfully applied in many NLP tasks, such as Semantic parsing (Moschitti, 2004) and Relation Extraction (Zhang et al.
Introduction
In multilingual tasks such as machine translation, tree kernels are seldom applied.
Introduction
In this paper, we propose Bilingual Tree Kernels (BTKs) to model the bilingual translational equivalences, in our case, to conduct subtree alignment.
tree kernel is mentioned in 22 sentences in this paper.
Topics mentioned in this paper:
Wang, WenTing and Su, Jian and Tan, Chew Lim
Abstract
In this paper we propose using tree kernel based approach to automatically mine the syntactic information from the parse trees for discourse analysis, applying kernel function to the tree structures directly.
Abstract
The experiment shows tree kernel approach is able to give statistical significant improvements over flat syntactic path feature.
Abstract
We also illustrate that tree kernel approach covers more structure information than the production rules, which allows tree kernel to further incorporate information from a higher dimension space for possible better discrimination.
Introduction
In this paper we propose using tree kernel based method to automatically mine the syntactic
Introduction
The experiment shows that tree kernel is able to effectively incorporate syntactic structural information and produce statistical significant improvements over flat syntactic path feature for the recognition of both explicit and implicit relation in Penn Discourse Treebank (PDTB; Prasad et al., 2008).
Introduction
We also illustrate that tree kernel approach covers more structure information than the production rules, which allows tree kernel to further work on a higher dimensional space for possible better discrimination.
Related Work
Tree Kernel based Approach in NLP.
Related Work
While the feature based approach may not be able to fully utilize the syntactic information in a parse tree, an alternative to the feature-based methods, tree kernel methods (Haussler, 1999) have been proposed to implicitly explore features in a high dimensional space by employing a kernel function to calculate the similarity between two objects directly.
Related Work
Other sub-trees beyond 2-level (e. g. Tf- 7}) are only captured in the tree kernel, which allows tree kernel to further leverage on information from higher dimension space for possible better discrimination.
tree kernel is mentioned in 33 sentences in this paper.
Topics mentioned in this paper:
Plank, Barbara and Moschitti, Alessandro
Computational Structures for RE
In fact, lexical information is highly affected by data-sparseness, thus tree kernels combined with semantic information created from additional resources should provide a way to obtain a more robust system.
Computational Structures for RE
The idea is to (i) learn semantic similarity between words on the pivot corpus and (ii) use tree kernels embedding such a similarity to learn a RE system on the source, which allows to generalize to the new target domain.
Introduction
We encode word clusters or similarity in tree kernels , which, in turn, produce spaces of tree fragments.
Introduction
Rather than only matching the surface string of words, lexical similarity enables soft matches between similar words in convolution tree kernels .
Introduction
In the empirical evaluation on Automatic Content Extraction (ACE) data, we evaluate the impact of convolution tree kernels embedding lexical semantic similarities.
Related Work
Semantic syntactic tree kernels have been previously used for question classification (Bloehdorn and Moschitti, 2007a; Bloehdorn and Moschitti, 2007b; Croce et al., 2011).
Related Work
Thus, we present a novel application of semantic syntactic tree kernels and Brown clusters for domain adaptation of tree-kernel based relation extraction.
Semantic Syntactic Tree Kernels
Commonly used kernels in NLP are string kernels (Lodhi et al., 2002) and tree kernels (Moschitti, 2006; Moschitti, 2008).
Semantic Syntactic Tree Kernels
Figure l: Syntactic tree kernel (STK).
Semantic Syntactic Tree Kernels
Syntactic tree kernels (Collins and Duffy, 2001) compute the similarity between two trees T1 and T2 by counting common sub-trees (cf.
tree kernel is mentioned in 22 sentences in this paper.
Topics mentioned in this paper:
Lin, Chen and Miller, Timothy and Kho, Alvin and Bethard, Steven and Dligach, Dmitriy and Pradhan, Sameer and Savova, Guergana
Abstract
Convolution tree kernels are an efficient and effective method for comparing syntactic structures in NLP methods.
Abstract
However, current kernel methods such as subset tree kernel and partial tree kernel understate the similarity of very similar tree structures.
Background
2.1 Syntax-based Tree Kernels
Background
Syntax-based tree kernels quantify the similarity between two constituent parses by counting their common substructures.
Background
The partial tree kernel (PTK) relaxes the definition of subtrees to allow partial production rule
Introduction
Convolution kernels over syntactic trees ( tree kernels ) offer a potential solution to this problem by providing relatively efficient algorithms for computing similarities between entire discrete structures.
Introduction
However, conventional tree kernels are sensitive to pattern variations.
Introduction
For example, two trees in Figure 1(a) sharing the same structure except for one terminal symbol are deemed at most 67% similar by the conventional tree kernel (PTK) (Moschitti, 2006).
Methods
Compared with the previous tree kernels , our descending path kernel has the following advantages: l) the substructures are simplified so that they are more likely to be shared among trees, and therefore the sparse feature issues of previous kernels could be alleviated by this representation; 2) soft matching between two similar structures (e.g., NP—>DT JJ NN versus NP—>DT NN) have high similarity without reference to any corpus or grammar rules;
tree kernel is mentioned in 18 sentences in this paper.
Topics mentioned in this paper:
Sun, Le and Han, Xianpei
Abstract
Tree kernel is an effective technique for relation extraction.
Abstract
In this paper, we propose a new tree kernel, called feature-enriched tree kernel (F TK ), which can enhance the traditional tree kernel by: 1) refining the syntactic tree representation by annotating each tree node with a set of discriminant features; and 2) proposing a new tree kernel which can better measure the syntactic tree similarity by taking all features into consideration.
Abstract
Experimental results show that our method can achieve a 5.4% F—measure improvement over the traditional convolution tree kernel .
Introduction
An effective technique is the tree kernel (Zelenko et al., 2003; Zhou et al., 2007; Zhang et al., 2006; Qian et al., 2008), which can exploit syntactic parse tree information for relation extraction.
Introduction
Then the similarity between two trees are computed using a tree kernel, e. g., the convolution tree kernel proposed by Collins and Duffy (2001).
Introduction
Unfortunately, one main shortcoming of the traditional tree kernel is that the syntactic tree representation usually cannot accurately capture the
tree kernel is mentioned in 28 sentences in this paper.
Topics mentioned in this paper:
Croce, Danilo and Moschitti, Alessandro and Basili, Roberto and Palmer, Martha
Abstract
Then, we design advanced similarity functions between such structures, i.e., semantic tree kernel functions, for exploiting distributional and grammatical information in Support Vector Machines.
Model Analysis and Discussion
available; and (ii) in 76% of the errors only 2 or less argument heads are included in the extracted tree, therefore tree kernels cannot exploit enough lexical information to disambiguate verb senses.
Model Analysis and Discussion
the capability of tree kernels to implicitly trigger useful linguistic inductions for complex semantic tasks.
Related work
Recently, DMs have been also proposed in integrated syntacticsemantic structures that feed advanced learning functions, such as the semantic tree kernels discussed in (Bloehdorn and Moschitti, 2007a; Bloehdorn and Moschitti, 2007b; Mehdad et al., 2010; Croce et al., 2011).
Structural Similarity Functions
In particular, we design new models for verb classification by adopting algorithms for structural similarity, known as Smoothed Partial Tree Kernels (SPTKs) (Croce et al., 2011).
Structural Similarity Functions
3.2 Tree Kernels driven by Semantic Similarity To our knowledge, two main types of tree kernels exploit lexical similarity: the syntactic semantic tree kernel defined in (Bloehdom and Moschitti, 2007a) applied to constituency trees and the smoothed partial tree kernels (SPTKs) defined in (Croce et al., 2011), which generalizes the former.
Structural Similarity Functions
The A function determines the richness of the kernel space and thus induces different tree kernels, for example, the syntactic tree kernel (STK) (Collins and Duffy, 2002) or the partial tree kernel (PTK) (Moschitti, 2006).
Verb Classification Models
Here, we apply tree pruning to reduce the computational complexity of tree kernels as it is proportional to the number of nodes in the input trees.
Verb Classification Models
To encode dependency structure information in a tree (so that we can use it in tree kernels ), we use (i) lexemes as nodes of our tree, (ii) their dependencies as edges between the nodes and (iii) the dependency labels, e. g., grammatical functions (GR), and POS-Tags, again as tree nodes.
tree kernel is mentioned in 9 sentences in this paper.
Topics mentioned in this paper:
Severyn, Aliaksei and Moschitti, Alessandro and Uryupina, Olga and Plank, Barbara and Filippova, Katja
Abstract
We rely on the tree kernel technology to automatically extract and learn features with better generalization power than bag-of-words.
Introduction
In particular, we define an efficient tree kernel derived from the Partial Tree Kernel , (Moschitti, 2006a), suitable for encoding structural representation of comments into Support Vector Machines (SVMs).
Representations and models
These trees are input to tree kernel functions for generating structural features.
Representations and models
The latter are automatically generated and learned by SVMs with expressive tree kernels .
Representations and models
In other words, the tree fragment: [S [negative—VP [negative—V [destroy] ] [PRODUCT-NP [PRODUCT-N [xoom] ] ] ] is a strong feature (induced by tree kernels ) to help the classifier to discriminate such hard cases.
tree kernel is mentioned in 9 sentences in this paper.
Topics mentioned in this paper:
Xie, Boyi and Passonneau, Rebecca J. and Wu, Leon and Creamer, Germán G.
Abstract
We introduce a novel tree representation, and use it to train predictive models with tree kernels using support vector machines.
Conclusion
Our semantic frame-based model benefits from tree kernel learning using support vector machines.
Discussion
To analyze which were the best performing features within sectors, we extracted the best performing frame fragments for the polarity task using a tree kernel feature engineering method presented in Pighin and Moschitti (2009).
Experiments
3'SVM-light: http://svmlight.joachims.org and Tree Kernels in SVM-light: http://disi.unitn.it/moschitti/Tree-Kernel.htm.
Methods
The second is a tree representation that encodes semantic frame features, and depends on tree kernel measures for support vector machine classification.
Methods
The tree kernel (Moschitti, 2006; Collins and Duffy, 2002) is a function of tree similarity, based on common substructures (tree fragments).
tree kernel is mentioned in 6 sentences in this paper.
Topics mentioned in this paper:
Kim, Seokhwan and Banchs, Rafael E. and Li, Haizhou
Evaluation
the domain context tree kernel contributed to produce more precise outputs.
Introduction
Our composite kernel consists of a history sequence and a domain context tree kernels , both of which are composed based on similar textual units in Wikipedia articles to a given dialog context.
Wikipedia-based Composite Kernel for Dialog Topic Tracking
Our composite kernel consists of two different kernels: a history sequence kernel and a domain context tree kernel .
Wikipedia-based Composite Kernel for Dialog Topic Tracking
3.2 Domain Context Tree Kernel
Wikipedia-based Composite Kernel for Dialog Topic Tracking
Since this constructed tree structure represents semantic, discourse, and structural information extracted from the similar Wikipedia paragraphs to each given instance, we can explore these more enriched features to build the topic tracking model using a subset tree kernel (Collins and Duffy, 2002) which computes the similarity between each pair of trees in the feature space as follows:
tree kernel is mentioned in 6 sentences in this paper.
Topics mentioned in this paper:
Wang, Chang and Fan, James
Background
Many of them focus on using tree kernels to learn parse tree structure related features (Collins and Duffy, 2001; Culotta and Sorensen, 2004; Bunescu and Mooney, 2005).
Background
For example, by combining tree kernels and convolution string kernels, (Zhang et al., 2006) achieved the state of the art performance on ACE data (ACE, 2004).
Experiments
We compare our approaches to three state-of-the-art approaches including SVM with convolution tree kernels (Collins and Duffy, 2001), linear regression and SVM with linear kernels (Scholkopf and Smola, 2002).
Experiments
To adapt the tree kernel to medical domain, we followed the approach in (Nguyen et al., 2009) to take the syntactic structures into consideration.
Experiments
We also added the argument types as features to the tree kernel .
tree kernel is mentioned in 6 sentences in this paper.
Topics mentioned in this paper:
Chen, Yanping and Zheng, Qinghua and Zhang, Wei
Feature Construction
In this field, the tree kernel based method commonly uses the parse tree to capture the structure information (Zelenko et al., 2003; Culotta and Sorensen, 2004).
Related Work
(2012) proposed a convolution tree kernel .
Related Work
(2010) employed a model, combining both the feature based and the tree kernel based methods.
Related Work
(2008; 2010) also pointed out that, due to the inaccuracy of Chinese word segmentation and parsing, the tree kernel based approach is inappropriate for Chinese relation extraction.
tree kernel is mentioned in 5 sentences in this paper.
Topics mentioned in this paper:
Srivastava, Shashank and Hovy, Eduard
Introduction
2.2 Tree kernels
Introduction
Tree Kernel methods have gained popularity in the last decade for capturing syntactic information in the structure of parse trees (Collins and Duffy, 2002; Moschitti, 2006).
Introduction
(2013) have attempted to provide formulations to incorporate semantics into tree kernels through the use of distributional word vectors at the individual word-nodes.
tree kernel is mentioned in 5 sentences in this paper.
Topics mentioned in this paper:
Guzmán, Francisco and Joty, Shafiq and Màrquez, Llu'is and Nakov, Preslav
Our Discourse-Based Measures
A number of metrics have been proposed to measure the similarity between two labeled trees, e. g., Tree Edit Distance (Tai, 1979) and Tree Kernels (Collins and Duffy, 2001; Moschitti and Basili, 2006).
Our Discourse-Based Measures
Tree kernels (TKs) provide an effective way to integrate arbitrary tree structures in kernel-based machine learning algorithms like SVMs.
Our Discourse-Based Measures
the nuclearity and the relations, in order to allow the tree kernel to give partial credit to subtrees that differ in labels but match in their skeletons.
tree kernel is mentioned in 4 sentences in this paper.
Topics mentioned in this paper:
Surdeanu, Mihai and Ciaramita, Massimiliano and Zaragoza, Hugo
Approach
Counting the number of matched dependencies is essentially a simplified tree kernel for QA (e.g., see (Moschitti et al., 2007)) matching only trees of depth 2.
Approach
Experiments with full dependency tree kernels based on several variants of the convolution kernels of Collins and Duffy (2001) did not yield improvements.
Approach
We conjecture that the mistakes of the syntactic parser may be amplified in tree kernels , which consider an exponential number of sub-trees.
tree kernel is mentioned in 3 sentences in this paper.
Topics mentioned in this paper: