Index of papers in Proc. ACL 2011 that mention
  • iteratively
Lang, Joel and Lapata, Mirella
Abstract
We present an algorithm that iteratively splits and merges clusters representing semantic roles, thereby leading from an initial clustering to a final clustering of better quality.
Conclusions
We proposed a split-merge algorithm that iteratively manipulates clusters representing semantic roles whilst trading off cluster purity with collocation.
Conclusions
itive and requires no manual effort for training.
Related Work
Swier and Stevenson (2004) induce role labels with a bootstrapping scheme where the set of labeled instances is iteratively expanded using a classifier trained on previously labeled instances.
Related Work
We formulate the induction of semantic roles as a clustering problem and propose a split-merge algorithm which iteratively manipulates clusters representing semantic roles.
Split-Merge Role Induction
Our algorithm works by iteratively splitting and merging clusters of argument instances in order to arrive at increasingly accurate representations of semantic roles.
Split-Merge Role Induction
Then [3 is iteratively decreased again until it becomes zero, after which 7 is decreased by another 0.05.
iteratively is mentioned in 7 sentences in this paper.
Topics mentioned in this paper:
DeNero, John and Macherey, Klaus
Introduction
In this approach, we iteratively apply the same efficient sequence algorithms for the underlying directional models, and thereby optimize a dual bound on the model objective.
Model Inference
In particular, we can iteratively apply exact inference to the subgraph problems, adjusting their potentials to reflect the constraints of the full problem.
Model Inference
We can iteratively search for such a 11 via sub-gradient descent.
iteratively is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Ravi, Sujith and Knight, Kevin
Machine Translation as a Decipherment Task
4For Iterative EM, we start with a channel of size 101x101 (K =100) and in every pass we iteratively increase the vocabulary sizes by 50, repeating the training procedure until the channel size becomes 351x351.
Word Substitution Decipherment
Instead of instantiating the entire channel model (with all its parameters), we iteratively train the model in small steps.
Word Substitution Decipherment
Goto Step 2 and repeat the procedure, extending the channel size iteratively in each stage.
iteratively is mentioned in 3 sentences in this paper.
Topics mentioned in this paper: