Index of papers in Proc. ACL 2014 that mention
  • Gibbs sampling
Hingmire, Swapnil and Chakraborti, Sutanu
Background 3.1 LDA
In this paper we estimate approximate posterior inference using collapsed Gibbs sampling (Griffiths and Steyvers, 2004).
Background 3.1 LDA
The Gibbs sampling equation used to update the assignment of a topic I to the word 21) E W at the position n in document d, conditioned on at, flu, is:
Background 3.1 LDA
We use a subscript d, fin to denote the current token, zdm is ignored in the Gibbs sampling update.
Experimental Evaluation
l. Infer T number of topics on D for LDA using collapsed Gibbs sampling .
Experimental Evaluation
Update M D using collapsed Gibbs sampling update in Equation 1.
Experimental Evaluation
Infer ‘0‘ number of topics on the sprinkled document corpus D using collapsed Gibbs sampling update.
Topic Sprinkling in LDA
We then update the new LDA model using collapsed Gibbs sampling .
Topic Sprinkling in LDA
We then infer a set of |C | number of topics on the sprinkled dataset using collapsed Gibbs sampling , where C is the set of class labels of the training documents.
Topic Sprinkling in LDA
We modify collapsed Gibbs sampling update in Equation 1 to carry class label information while inferring topics.
Gibbs sampling is mentioned in 10 sentences in this paper.
Topics mentioned in this paper:
Chen, Zhiyuan and Mukherjee, Arjun and Liu, Bing
AKL: Using the Learned Knowledge
Most importantly, due to the use of the new form of knowledge, AKL’s inference mechanism ( Gibbs sampler ) is entirely different from that of MC-LDA (Section 5.2), which results in superior performances (Section 6).
AKL: Using the Learned Knowledge
In short, our modeling contributions are (1) the capability of handling more expressive knowledge in the form of clusters, (2) a novel Gibbs sampler to deal with inappropriate knowledge.
AKL: Using the Learned Knowledge
5.2 The Gibbs Sampler
Gibbs sampling is mentioned in 8 sentences in this paper.
Topics mentioned in this paper:
Zhang, Yuan and Lei, Tao and Barzilay, Regina and Jaakkola, Tommi and Globerson, Amir
Experimental Setup
Therefore, the first-order distribution is not well-defined and we only employ Gibbs sampling for simplicity.
Introduction
Our first strategy is akin to Gibbs sampling and samples a new head for each word in the sentence, modifying one arc at a time.
Results
iteration of this sampler makes multiple changes to the tree, in contrast to a single-edge change of Gibbs sampler .
Sampling-Based Dependency Parsing with Global Features
3.2.1 Gibbs Sampling
Sampling-Based Dependency Parsing with Global Features
One shortcoming of the Gibbs sampler is that it only changes one variable (arc) at a time.
Sampling-Based Dependency Parsing with Global Features
Note that blocked Gibbs sampling would be exponential in K, and is thus very slow already at K = 4.
Gibbs sampling is mentioned in 7 sentences in this paper.
Topics mentioned in this paper:
Andrews, Nicholas and Eisner, Jason and Dredze, Mark
Abstract
We present a block Gibbs sampler for posterior inference and an empirical evaluation on several datasets.
Inference by Block Gibbs Sampling
We use a block Gibbs sampler , which from an initial state (190,21), zo) repeats these steps: 1.
Inference by Block Gibbs Sampling
The topics of context words are assumed exchangeable, and so we re-sample them using Gibbs sampling (Griffiths and Steyvers, 2004).
Inference by Block Gibbs Sampling
Unfortunately, this is prohibitively expensive for the (nonexchangeable) topics of the named mentions c. A Gibbs sampler would have to choose a new value for cc.z with probability proportional to the resulting joint probability of the full sample.
Gibbs sampling is mentioned in 6 sentences in this paper.
Topics mentioned in this paper:
Hu, Yuening and Zhai, Ke and Eidelman, Vladimir and Boyd-Graber, Jordan
Experiments
3For Gibbs sampling , we use implementations available in Hu and Boyd-Graber (2012) for tLDA; and Mallet (McCallum, 2002) for LDA and pLDA.
Inference
We use a collapsed Gibbs sampler for tree-based topic models to sample the path ydn and topic assignment zdn for word wdn,
Inference
For topic .2 and path y, instead of variational updates, we use a Gibbs sampler within a document.
Inference
This equation embodies how this is a hybrid algorithm: the first term resembles the Gibbs sampling term encoding how much a document prefers a topic, while the second term encodes the expectation under the variational distribution of how much a path is preferred by this topic,
Gibbs sampling is mentioned in 4 sentences in this paper.
Topics mentioned in this paper:
Zou, Bowei and Zhou, Guodong and Zhu, Qiaoming
Baselines
Here, the topics are extracted from all the documents in the *SEM 2012 shared task using the LDA Gibbs Sampling algorithm (Griffiths, 2002).
Baselines
where Rel(w,, rm) is the weight of word w in topic rm calculated by the LDA Gibbs Sampling algorithm.
Baselines
> Topic Modeler: For estimating transition probability Pt(i,m), we employ GibbsLDA++6, an LDA model using Gibbs Sampling technique for parameter estimation and inference.
Gibbs sampling is mentioned in 4 sentences in this paper.
Topics mentioned in this paper:
Bamman, David and Underwood, Ted and Smith, Noah A.
Experiments
All experiments are run with 50 iterations of Gibbs sampling to collect samples for the personas p, alternating with maximization steps for 77.
Model
Rather than adopting a fully Bayesian approach (e.g., sampling all variables), we infer these values using stochastic EM, alternating between collapsed Gibbs sampling for each p and maximizing with respect to 77.
Model
8We assume the reader is familiar with collapsed Gibbs sampling as used in latent-variable NLP models.
Gibbs sampling is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Wang, Xiaolin and Utiyama, Masao and Finch, Andrew and Sumita, Eiichiro
Introduction
(2010) used the local best alignment to increase the speed of the Gibbs sampling in training but the impact on accuracy was not explored.
Introduction
To this end, we model bilingual UWS under a similar framework with monolingual UWS in order to improve efficiency, and replace Gibbs sampling with expectation maximization (EM) in training.
Methods
EF/{;}(P(.7-"k/|.7-")) = P(J:k'|f, M) in a similar manner to the marginalization in the Gibbs sampling process which we are replacing;
Gibbs sampling is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Yang, Bishan and Cardie, Claire
Approach
For constraints with higher-order structures, we use Gibbs Sampling (Geman and Geman, 1984) to approximate the expectations.
Approach
For documents where the higher-order constraints apply, we use the same Gibbs sampler as described above to infer the most likely label assignment, otherwise, we use the Viterbi algorithm.
Experiments
For approximation inference with higher-order constraints, we perform 2000 Gibbs sampling iterations where the first 1000 iterations are bum-in iterations.
Gibbs sampling is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Zhai, Ke and Williams, Jason D
Experiments
We run the Gibbs samplers for 1000 iterations and update all hyper-parameters using slice sampling (Neal, 2003; Wallach, 2008) every 10 iterations.
Latent Structure in Dialogues
We also assume symmetric Dirichlet priors on all multinomial distributions and apply collapsed Gibbs sampling .
Latent Structure in Dialogues
All probabilities can be computed using collapsed Gibbs sampler for LDA (Griffiths
Gibbs sampling is mentioned in 3 sentences in this paper.
Topics mentioned in this paper: