Index of papers in Proc. ACL that mention
  • grammar induction
Liu, Kai and Lü, Yajuan and Jiang, Wenbin and Liu, Qun
Bilingual Projection of Dependency Grammar
So we extract projected discrete dependency arc instances instead of treebank as training set for the projected grammar induction model.
Bilingually-Guided Dependency Grammar Induction
This section presents our bilingually-guided grammar induction model, which incorporates unsuperVised framework and bilingual projection model through a joint approach.
Bilingually-Guided Dependency Grammar Induction
According to following observation: unsuperVised induction model mines underlying syntactic structure of the monolingual language, however, it is hard to find good grammar induction in the exponential parsing space; bilingual projection obtains relatively reliable syntactic knowledge of the
Bilingually-Guided Dependency Grammar Induction
Based on the idea, we propose a novel strategy for training monolingual grammar induction model with the guidance of unsupervised and bilingually-projected dependency information.
Conclusion and Future Work
This paper presents a bilingually-guided strategy for automatic dependency grammar induction , which adopts an unsupervised skeleton and leverages the bilingually-projected dependency information during optimization.
Experiments
The results in Figure 3 prove that our unsupervised framework 04 = 1 can promote the grammar induction if it has a good start (well initialization), and it will be better once we incorporate the information from the projection side (04 = 0.9).
Introduction
Dependency Grammar Induction
Introduction
In dependency grammar induction , unsupervised methods achieve continuous improvements in recent years (Klein and Manning, 2004; Smith and Eisner, 2005; Bod, 2006; William et al., 2009; Spitkovsky et al., 2010).
Introduction
In the rest of the paper, we first describe the unsupervised dependency grammar induction framework in section 2 (where the unsupervised optimization objective is given), and introduce the bilingual projection method for dependency parsing in section 3 (where the projected optimization objective is given); Then in section 4 we present the bilingually-guided induction strategy for dependency grammar (where the two objectives above are jointly optimized, as shown in Figure 1).
grammar induction is mentioned in 13 sentences in this paper.
Topics mentioned in this paper:
Gormley, Matthew R. and Mitchell, Margaret and Van Durme, Benjamin and Dredze, Mark
Abstract
This work highlights a new application of unsupervised grammar induction and demonstrates several approaches to SRL in the absence of supervised syntax.
Introduction
To better understand the effect of the low-resource grammars and features used in these models, we further include comparisons with (1) models that use higher-resource versions of the same features; (2) state-of-the-art high resource models; and (3) previous work on low-resource grammar induction .
Introduction
0 New application of unsupervised grammar induction : low-resource SRL.
Introduction
o Constrained grammar induction using SRL for distant-supervision.
Related Work
Our work builds upon research in both semantic role labeling and unsupervised grammar induction (Klein and Manning, 2004; Spitkovsky et a1., 2010a).
Related Work
There has not yet been a comparison of techniques for SRL that do not rely on a syntactic treebank, and no exploration of probabilistic models for unsupervised grammar induction within an SRL pipeline that we have been able to find.
Related Work
Train Time, Constrained Grammar Induction : Observed Constraints
grammar induction is mentioned in 27 sentences in this paper.
Topics mentioned in this paper:
Gormley, Matthew R. and Eisner, Jason
Discussion
In principle, our branch-and-bound method can approach e-optimal solutions to Viterbi training of locally normalized generative models, including the NP-hard case of grammar induction with the DMV.
Introduction
dency grammar induction ).
Introduction
This is evident for grammar induction .
Introduction
(2010b) present evidence that hard EM can outperform soft EM for grammar induction in a hill-climbing setting.
Related Work
Alas, there are not yet any spectral learning methods that recover latent tree structure, as in grammar induction .
Related Work
Several integer linear programming (ILP) formulations of dependency parsing (Riedel and Clarke, 2006; Martins et al., 2009; Riedel et al., 2012) inspired our definition of grammar induction as a MP.
Relaxations
In this section, we will show how the RLT can be applied to our grammar induction problem and contrast it with the concave envelope relaxation presented in section 4.2.
grammar induction is mentioned in 7 sentences in this paper.
Topics mentioned in this paper:
Swanson, Ben and Yamangil, Elif and Charniak, Eugene and Shieber, Stuart
Conclusion
The most important direction of future work for OSTAG is the development of a principled grammar induction model, perhaps using the same techniques that have been successfully applied to TSG and TIG.
Experiments
A compact TSG can be obtained automatically using the MCMC grammar induction technique of Cohn and Blunsom (2010), retaining all TSG rules that appear in at least one derivation in after 1000 iterations of sampling.
Experiments
While our grammar induction method is a crude (but effective) heuristic, we can still highlight the qualities of the more important auxiliary trees by examining aggregate statistics over the MPD parses, shown in Figure 6.
TAG and Variants
This model has proven effective for grammar induction via Markov Chain Monte Carlo (MCMC), in which TSG derivations of the training set are repeatedly sampled to find frequently occurring elementary trees.
TAG and Variants
However, a large effort in non-probabilistic grammar induction has been performed through manual annotation with the XTAG project(Doran et al., 1994).
TAG and Variants
Later approaches (Shindo et al., 2011; Yamangil and Shieber, 2012) were able to extend the nonparametric modeling of TSGs to TIG, providing methods for both modeling and grammar induction .
grammar induction is mentioned in 7 sentences in this paper.
Topics mentioned in this paper:
Snyder, Benjamin and Naseem, Tahira and Barzilay, Regina
Conclusion
We have presented a probabilistic model for bilingual grammar induction which uses raw parallel text to learn tree pairs and their alignments.
Experimental setup
During preprocessing of the corpora we remove all punctuation marks and special symbols, following the setup in previous grammar induction work (Klein and Manning, 2002).
Introduction
We test the effectiveness of our bilingual grammar induction model on three corpora of parallel text: English-Korean, English-Urdu and English-Chinese.
Related Work
The unsupervised grammar induction task has been studied extensively, mostly in a monolingual setting (Charniak and Carroll, 1992; Stolcke and Omohundro, 1994; Klein and Manning, 2002; Seginer, 2007).
Related Work
We know of only one study which evaluates these bilingual grammar formalisms on the task of grammar induction itself (Smith and Smith, 2004).
Related Work
In contrast to this work, our goal is to explore the benefits of multilingual grammar induction in a fully unsupervised setting.
grammar induction is mentioned in 6 sentences in this paper.
Topics mentioned in this paper:
Gyawali, Bikash and Gardent, Claire
Abstract
A key feature of this approach is that grammar induction is driven by the extended domain of locality principle of TAG (Tree Adjoining Grammar); and that it takes into account both syntactic and semantic information.
Experimental Setup
We compare the results obtained with those obtained by two other systems participating in the KBGen challenge, namely the UDEL system, a symbolic rule based system developed by a group of students at the University of Delaware; and the IMS system, a statistical system using a probabilistic grammar induced from the training data.
Introduction
A key feature of this approach is that grammar induction is driven by the extended domain of locality principle of TAG (Tree Adjoining Grammar) and takes into account both syntactic and semantic information.
Related Work
They induce a probabilistic Tree Adjoining Grammar from a training set aligning frames and sentences using the grammar induction technique of (Chiang, 2000) and use a beam search that uses weighted features learned from the training data to rank alternative expansions at each step.
Related Work
We use a simple mainly symbolic approach whereas they use a generative approach for grammar induction and a discriminative approach for sentence generation.
grammar induction is mentioned in 5 sentences in this paper.
Topics mentioned in this paper:
Salway, Andrew and Touileb, Samia
Abstract
We report the first steps of a novel investigation into how a grammar induction algorithm can be modified and used to identify salient information structures in a corpus.
Approach
Thus, we propose to use a grammar induction algorithm to identify the most salient information structures in a corpus and take these as representations of important semantic content.
Closing Remarks
the use of grammar induction to elucidate semantic content for text mining purposes shows promise.
Implementation
Here we report our first attempt to apply grammar induction to text mining.
Introduction
Our approach is to induce information structures from an unannotated corpus by modifying and applying the ADIOS grammar induction algorithm (Solan et al., 2005): the modifications serve to focus the algorithm on what is typically written about key-terms (Section 3).
grammar induction is mentioned in 5 sentences in this paper.
Topics mentioned in this paper:
Cohen, Shay and Smith, Noah A
Introduction
For example, Smith and Eisner (2006) have penalized the approximate posterior over dependency structures in a natural language grammar induction task to avoid long range dependencies between words.
Introduction
We show that empirically, injecting prior knowledge improves performance on an unsupervised Chinese grammar induction task.
Variational Mixtures with Constraints
This is a strict model reminiscent of the successful application of structural bias to grammar induction (Smith and Eisner, 2006).
Variational Mixtures with Constraints
We demonstrated the effectiveness of the algorithm on a dependency grammar induction task.
grammar induction is mentioned in 4 sentences in this paper.
Topics mentioned in this paper:
Cohen, Shay B. and Johnson, Mark
Bayesian inference for PCFGs
9 Empirical effects of the three approaches in unsupervised grammar induction
Bayesian inference for PCFGs
In this section we present experiments using the three samplers just described in an unsupervised grammar induction problem.
Bayesian inference for PCFGs
Our goal here is not to improve the state-of-the-art in unsupervised grammar induction , but to try to measure empirical differences in the estimates produced by the three different approaches to tightness just described.
grammar induction is mentioned in 4 sentences in this paper.
Topics mentioned in this paper:
Dethlefs, Nina and Hastie, Helen and Cuayáhuitl, Heriberto and Lemon, Oliver
Cohesion across Utterances
3.4 Grammar Induction
Cohesion across Utterances
The grammar g of surface realisation candidates is obtained through an automatic grammar induction algorithm which can be run on unlabelled data and requires only minimal human intervention.
Cohesion across Utterances
We provide the human corpus of restaurant recommendations from Section 3.3 as input to grammar induction .
Evaluation
Algorithm 1 Grammar Induction .
grammar induction is mentioned in 4 sentences in this paper.
Topics mentioned in this paper:
Shindo, Hiroyuki and Miyao, Yusuke and Fujino, Akinori and Nagata, Masaaki
Background and Related Work
morphology analysis, word segmentation (Johnson and Goldwater, 2009), and dependency grammar induction (Cohen et al., 2010), rather than constituent syntax parsing.
Conclusion
Future work will involve examining the SR-TSG model for different languages and for unsupervised grammar induction .
Inference
The sentence-level MH sampler is a recently proposed algorithm for grammar induction (Johnson et al., 2007b; Cohn et al., 2010).
grammar induction is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Mareċek, David and Straka, Milan
Experiments
The first one (spiIZ) is the DMV-based grammar inducer by Spitkovsky et al.
Introduction
The task of unsupervised dependency parsing (which strongly relates to the grammar induction task) has become popular in the last decade, and its quality has been greatly increasing during this period.
Related Work
Such approaches achieve better results; however, they are useless for grammar induction from plain text.
grammar induction is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Ma, Xuezhe and Xia, Fei
Experiments
“PGI” is the phylogenetic grammar induction model of Berg-Kirkpatrick and Klein (2010).
Introduction
This led to a vast amount of research on unsupervised grammar induction (Carroll and Chamiak, 1992; Klein and Manning, 2004; Smith and Eisner, 2005; Cohen and Smith, 2009; Spitkovsky et al., 2010; Blun-som and Cohn, 2010; Marecek and Straka, 2013; Spitkovsky et al., 2013), which appears to be a natural solution to this problem, as unsupervised methods require only unannotated text for training parsers.
Introduction
Unfortunately, the unsupervised grammar induction systems’ parsing accuracies often significantly fall behind those of supervised systems (McDonald et al., 2011).
grammar induction is mentioned in 3 sentences in this paper.
Topics mentioned in this paper: