Index of papers in Proc. ACL 2011 that mention
  • Gibbs sampling
Hu, Yuening and Boyd-Graber, Jordan and Satinoff, Brianna
Constraints Shape Topics
3.1 Gibbs Sampling for Topic Models
Constraints Shape Topics
In topic modeling, collapsed Gibbs sampling (Griffiths and Steyvers, 2004) is a standard procedure for obtaining a Markov chain over the latent variables in the model.
Constraints Shape Topics
Given M documents the state of a Gibbs sampler for LDA consists of topic assignments for each token in the corpus and is represented as Z : {21,1...21,N1,22,1,...2M,NM}.
Discussion
As presented here, the technique for incorporating constraints is closely tied to inference with Gibbs sampling .
Interactively adding constraints
In the implementation of a Gibbs sampler , unassignment is done by setting a token’s topic assignment to an invalid topic (e. g. -l, as we use here) and decrementing any counts associated with that word.
Simulation Experiment
Next, we perform one of the strategies for state ablation, add additional iterations of Gibbs sampling , use the newly obtained topic distribution of each document as the feature vector, and perform classification on the test / train split.
Simulation Experiment
Each is averaged over five different chains using 10 additional iterations of Gibbs sampling per round (other numbers of iterations are discussed in Section 6.4).
Simulation Experiment
Figure 4 shows the effect of using different numbers of Gibbs sampling iterations after changing a constraint.
Gibbs sampling is mentioned in 9 sentences in this paper.
Topics mentioned in this paper:
Blunsom, Phil and Cohn, Trevor
Background
However this work approximated the derivation of the Gibbs sampler (omitting the interdependence between events when sampling from a collapsed model), resulting in a model which underperformed Brown et al.
Experiments
We have omitted the results for the HMM-LM as experimentation showed that the local Gibbs sampler became hopelessly stuck, failing to
The PYP-HMM
In order to induce a tagging under this model we use Gibbs sampling , a Markov chain Monte Carlo (MCMC) technique for drawing samples from the posterior distribution over the tag sequences given observed word sequences.
The PYP-HMM
We present two different sampling strategies: First, a simple Gibbs sampler which randomly samples an update to a single tag given all other tags; and second, a type-level sampler which updates all tags for a given word under a
The PYP-HMM
Gibbs samplers Both our Gibbs samplers perform the same calculation of conditional tag distributions, and involve first decrementing all trigrams and emissions affected by a sampling action, and then reintroducing the trigrams one at a time, conditioning their probabilities on the updated counts and table configurations as we progress.
Gibbs sampling is mentioned in 7 sentences in this paper.
Topics mentioned in this paper:
Celikyilmaz, Asli and Hakkani-Tur, Dilek
Final Experiments
For our models, we ran Gibbs samplers for 2000 iterations for each configuration throwing out first 500 samples as burn-in.
Two-Tiered Topic Model - TTM
We use Gibbs sampling which allows a combination of estimates from several local maxima of the posterior distribution.
Two-Tiered Topic Model - TTM
We obtain DS during Gibbs sampling (in §4.l), which indicates a saliency score of each sentence sj E S,j = LSD:
Gibbs sampling is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
He, Yulan and Lin, Chenghua and Alani, Harith
Introduction
The previously proposed J ST model uses the sentiment prior information in the Gibbs sampling inference step that a sentiment label will only be sampled if the current word token has no prior sentiment as defined in a sentiment lexicon.
Joint Sentiment-Topic (J ST) Model
Gibbs sampling was used to estimate the posterior distribution by sequentially sampling each variable of interest, 2,; and It here, from the distribution over
Joint Sentiment-Topic (J ST) Model
In our experiment, a was updated every 25 iterations during the Gibbs sampling procedure.
Gibbs sampling is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Ravi, Sujith and Knight, Kevin
Machine Translation as a Decipherment Task
Sampling IBM Model 3: We use point-wise Gibbs sampling to estimate the IBM Model 3 parameters.
Word Substitution Decipherment
channel.1 We perform inference using point-wise Gibbs sampling (Geman and Geman, 1984).
Word Substitution Decipherment
Parallelized Gibbs sampling : Secondly, we parallelize our sampling step using a Map-Reduce framework.
Gibbs sampling is mentioned in 3 sentences in this paper.
Topics mentioned in this paper: