Index of papers in Proc. ACL 2014 that mention
  • Viterbi
Chang, Yin-Wen and Rush, Alexander M. and DeNero, John and Collins, Michael
Abstract
The relaxation can be solved with a modified version of the Viterbi algorithm.
Adding Back Constraints
In order to compute the arg max of this optimization problem, we need to satisfy the constraints within the Viterbi algorithm.
Adding Back Constraints
We augment the Viterbi chart with a count vector d E D where D C leplll and d(i,j) is a count for the (i,j)’th constraint, i.e.
Adding Back Constraints
Figure 6 shows this constrained Viterbi relaxation approach.
Adjacent Agreement
The key algorithmic idea is to extend the Viterbi algorithm in order to consider possible adjacent links in the reverse direction.
Adjacent Agreement
Despite the new constraints, we can still compute L()\) in 0(I J (I + J time using a variant of the Viterbi algorithm.
Adjacent Agreement
Figure 5: Modified Viterbi algorithm for computing the adj a-cent agreement L()\).
Background
Note that for both of these models we can solve the optimization problem exactly using the standard Viterbi algorithm for HMM decoding.
Bidirectional Alignment
We then use the standard Viterbi update for computing the at variables, adding in the score of the y(i, j) necessary to satisfy the constraints.
Introduction
Our decoder uses a simple variant of the Viterbi algorithm for solving a relaxed version of this model.
Introduction
0 It is based on a novel relaxation for the model of DeNero and Macherey (2011), solvable with a variant of the Viterbi algorithm.
Viterbi is mentioned in 14 sentences in this paper.
Topics mentioned in this paper:
Hall, David and Berg-Kirkpatrick, Taylor and Klein, Dan
Abstract
(2013) presented an approach to GPU parsing that sacrifices traditional sparsity in exchange for raw computational power, obtaining a system that can compute Viterbi parses for a high-quality grammar at about 164 sentences per second on a midrange GPU.
Abstract
The resulting system is capable of computing over 404 Viterbi parses per second—more than a 2x speedup—on the same hardware.
Anatomy of a Dense GPU Parser
Table 1: Performance numbers for computing Viterbi inside charts on 20,000 sentences of length $40 from the Penn Treebank.
Anatomy of a Dense GPU Parser
The Viterbi inside loop for the grammar is inlined into a kernel.
Anatomy of a Dense GPU Parser
(2013)’s system is able to compute Viterbi charts at 164 sentences per second, for sentences up to length 40.
Grammar Clustering
The unpruned Viterbi computations in a fine grammar using the clustering method of Canny et al.
Introduction
Together these kernels implement the Viterbi inside algorithm.
Introduction
On a midrange GPU, their system can compute Viterbi derivations at 164 sentences per second on sentences of length 40 or less (see timing details below).
Introduction
(2013) is that it only computes Viterbi parses.
Minimum Bayes risk parsing
The Viterbi algorithm is a reasonably effective method for parsing.
Viterbi is mentioned in 27 sentences in this paper.
Topics mentioned in this paper:
Srivastava, Shashank and Hovy, Eduard
Introduction
in linear time (in the number of tokens) following the generalized Viterbi algorithm.
Introduction
A slightly modified version of Viterbi could also be used to find segmentations that are constrained to agree with some given motif boundaries, but can segment other parts of the sentence optimally under these constraints.
Introduction
Here y’ = Decode(x(k),w) is the optimal Viterbi decoding using the current estimates of the weights.
Viterbi is mentioned in 7 sentences in this paper.
Topics mentioned in this paper:
Tamura, Akihiro and Watanabe, Taro and Sumita, Eiichiro
RNN-based Alignment Model
The Viterbi alignment is determined using the Viterbi algorithm, similar to the FFNN-based model, where the model is sequentially applied from f1 to f J5.
RNN-based Alignment Model
5 Strictly speaking, we cannot apply the dynamic programming forward-backward algorithm (i.e., the Viterbi algorithm) due to the long alignment history of yi.
RNN-based Alignment Model
Thus, the Viterbi alignment is computed approximately using heuristic beam search.
Related Work
Given a specific model, the best alignment ( Viterbi alignment) of the sentence pair (ff, (if) can be found as
Related Work
a1 For example, the HMM model identifies the Viterbi alignment using the Viterbi algorithm.
Related Work
The model finds the Viterbi alignment using the Viterbi algorithm, similar to the classic HMM model.
Viterbi is mentioned in 6 sentences in this paper.
Topics mentioned in this paper:
Feng, Vanessa Wei and Hirst, Graeme
Abstract
To enhance the accuracy of the pipeline, we add additional constraints in the Viterbi decoding of the first CRF.
Bottom-up tree-building
Therefore, to improve its accuracy, we enforce additional commonsense constraints in its Viterbi decoding.
Bottom-up tree-building
Similar to Mfgfi‘cf’, for , we also include additional constraints in the Viterbi decoding, by disallowing transitions between two ones, and disallowing the sequence to be all zeros if it contains all the remaining constituents in the document.
Conclusions
Our approach was to adopt a greedy bottom-up tree-building, with two linear-chain CRFs as local probabilistic models, and enforce reasonable constraints in the first CRF’s Viterbi decoding.
Introduction
Specifically, in the Viterbi decoding of the first CRF, we include additional constraints elicited from common sense, to make more effective local decisions.
Viterbi is mentioned in 5 sentences in this paper.
Topics mentioned in this paper:
Gormley, Matthew R. and Mitchell, Margaret and Van Durme, Benjamin and Dredze, Mark
Approaches
Unsupervised Grammar Induction Our first method for grammar induction is fully unsupervised Viterbi EM training of the Dependency Model with Valence (DMV) (Klein and Manning, 2004), with uniform initialization of the model parameters.
Approaches
Constrained Grammar Induction Our second method, which we will refer to as DMV+C, induces grammar in a distantly supervised fashion by using a constrained parser in the E-step of Viterbi EM.
Experiments
We contrast the DMV trained with Viterbi EM+uniform initialization (DMV), our constrained DMV (DMV+C), and our model’s MBR decoding of latent syntax (Marginalized) with other recent work: Spitkovsky et al.
Related Work
(2010a) show that Viterbi (hard) EM training of the DMV with simple uniform initialization of the model parameters yields higher accuracy models than standard soft-EM
Related Work
In Viterbi EM, the E—step finds the maximum likelihood corpus parse given the current model parameters.
Viterbi is mentioned in 5 sentences in this paper.
Topics mentioned in this paper:
Jia, Zhongye and Zhao, Hai
Pinyin Input Method Model
The Viterbi algorithm ( Viterbi , 1967) is used for the decoding.
Pinyin Input Method Model
The shortest path algorithm for typo correction and Viterbi algorithm for PTC conversion are very closely related.
Related Works
The idea of “statistical input method” was proposed by modeling PTC conversion as a hidden Markov model (HMM), and using Viterbi ( Viterbi , 1967) algorithm to decode the sequence.
Viterbi is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Shi, Xing and Knight, Kevin and Ji, Heng
Training
To do this, we first take our phrasebook triples and construct sample string pairs <Epron, Pinyin-split> by pronouncing the phrasebook English with FST A, and by pronouncing the phrasebook Chinglish with FSTs D and C. Then we run the EM algorithm to learn FST B parameters (Table 3) and Viterbi alignments, such as:
Training
First, we obtain Viterbi alignments using the phoneme-based model, e. g.:
Training
EM learns values for parameters like P(nai te|night), plus Viterbi alignments such as:
Viterbi is mentioned in 3 sentences in this paper.
Topics mentioned in this paper: