Index of papers in Proc. ACL 2014 that mention
  • CRF
Anzaroot, Sam and Passos, Alexandre and Belanger, David and McCallum, Andrew
Background
For this underlying model, we employ a chain-structured conditional random field ( CRF ), since CRFs have been shown to perform better than other simple unconstrained models like hidden markov models for citation extraction (Peng and McCallum, 2004).
Background
The MAP inference task in a CRF be can expressed as an optimization problem with a lin-
Background
Since the log probability of some 3/ in the CRF is proportional to sum of the scores of all the factors, we can concatenate the indicator variables as a vector y and the scores as a vector 21) and write the MAP problem as
Citation Extraction Data
Here, 3/], represents an output tag of the CRF , SO if = = 1, then we have that 3/], was given a label with index i.
Citation Extraction Data
We constrain the output labeling of the chain-structured CRF to be a valid BIO encoding.
Citation Extraction Data
Rather than enforcing these constraints using dual decomposition, they can be enforced directly when performing MAP inference in the CRF by modifying the dynamic program of the Viterbi algorithm to only allow valid pairs of adj a-cent labels.
Soft Constraints in Dual Decomposition
Since we truncate penalties at 0, this suggests that we will learn a penalty of 0 for constraints in three categories: constraints that do not hold in the ground truth, constraints that hold in the ground truth but are satisfied in practice by performing inference in the base CRF model, and constraints that are satisfied in practice as a side-effect of imposing nonzero penalties on some other constraints .
CRF is mentioned in 19 sentences in this paper.
Topics mentioned in this paper:
Feng, Vanessa Wei and Hirst, Graeme
Abstract
To enhance the accuracy of the pipeline, we add additional constraints in the Viterbi decoding of the first CRF .
Bottom-up tree-building
Secondly, as a joint model, it is mandatory to use a dynamic CRF , for which exact inference is usually intractable or slow.
Bottom-up tree-building
Figure 4a shows our intra-sentential structure model in the form of a linear-chain CRF .
Bottom-up tree-building
Thus, different CRF chains have to be formed for different pairs of constituents.
Features
In our local models, to encode two adjacent units, U j and U j+1, within a CRF chain, we use the following 10 sets of features, some of which are modified from J oty et al.’s model.
Introduction
Specifically, in the Viterbi decoding of the first CRF , we include additional constraints elicited from common sense, to make more effective local decisions.
Related work
(2013) approach the problem of text-level discourse parsing using a model trained by Conditional Random Fields ( CRF ).
CRF is mentioned in 14 sentences in this paper.
Topics mentioned in this paper:
Yang, Bishan and Cardie, Claire
Abstract
The context-aware constraints provide additional power to the CRF model and can guide semi-supervised learning when labeled data is limited.
Approach
The CRF model the following conditional probabilities:
Approach
The objective function for a standard CRF is to maximize the log-likelihood over a collection of labeled doc-
Approach
Most of our constraints can be factorized in the same way as factorizing the model features in the first-order CRF model, and we can compute the expectations under q very efficiently using the forward-backward algorithm.
Experiments
We trained our model using a CRF incorporated with the proposed posterior constraints.
Experiments
For the CRF features, we include the tokens, the part-of-speech tags, the prior polarities of lexical patterns indicated by the opinion lexicon and the negator lexicon, the number of positive and negative tokens and the output of the voteflip algorithm (Choi and Cardie, 2009).
Experiments
We set the CRF regularization parameter a = l and set the posterior regularization parameter 6 and y (a tradeoff parameter we introduce to balance the supervised objective and the posterior regularizer in 2) by using grid search 8.
Introduction
Specifically, we use the Conditional Random Field (CRF) model as the learner for sentence-level sentiment classification, and incorporate rich discourse and lexical knowledge as soft constraints into the learning of CRF parameters via Posterior Regularization (PR) (Ganchev et al., 2010).
Introduction
Unlike most previous work, we explore a rich set of structural constraints that cannot be naturally encoded in the feature-label form, and show that such constraints can improve the performance of the CRF model.
CRF is mentioned in 22 sentences in this paper.
Topics mentioned in this paper:
Gormley, Matthew R. and Mitchell, Margaret and Van Durme, Benjamin and Dredze, Mark
Approaches
We define a conditional random field ( CRF ) (Lafferty et al., 2001) for this task.
Approaches
This model extends the CRF model in Section 3.1 to include the projective syntactic dependency parse for a sentence.
Approaches
train our CRF models by maximizing conditional log-likelihood using stochastic gradient descent with an adaptive learning rate (AdaGrad) (Duchi et al., 2011) over mini-batches.
Experiments
Similarly, improving the non-convex optimization of our latent-variable CRF (Marginalized) may offer further gains.
Introduction
0 Simpler joint CRF for syntactic and semantic dependency parsing than previously reported.
Introduction
The joint models use a non-loopy conditional random field ( CRF ) with a global factor constraining latent syntactic edge variables to form a tree.
CRF is mentioned in 7 sentences in this paper.
Topics mentioned in this paper:
Pei, Wenzhe and Ge, Tao and Chang, Baobao
Experiment
P R F OOV CRF 87.8 85.7 86.7 57.1 NN 92.4 92.2 92.3 60.0 NN+Tag Embed 93.0 92.7 92.9 61.0 MMTNN 93.7 93.4 93.5 64.2
Experiment
We also compare our model with the CRF model (Lafferty et al., 2001), which is a widely used log-linear model for Chinese word segmentation.
Experiment
The input feature to the CRF model is simply the context characters (unigram feature) without any additional feature engineering.
CRF is mentioned in 6 sentences in this paper.
Topics mentioned in this paper:
Bollegala, Danushka and Weir, David and Carroll, John
Domain Adaptation
Next, we train a CRF model using all features (i.e.
Domain Adaptation
Finally, the trained CRF model is applied to a target domain test sentence.
Experiments and Results
The L-BFGS (Liu and Nocedal, 1989) method is used to train the CRF and logistic regression models.
Experiments and Results
Specifically, in POS tagging, a CRF trained on source domain labeled sentences is applied to target domain test sentences, whereas in sentiment classification, a logistic regression classifier trained using source domain labeled reviews is applied to the target domain test reviews.
Related Work
Huang and Yates (2009) train a Conditional Random Field ( CRF ) tagger with features retrieved from a smoothing model trained using both source and target domain unlabeled data.
CRF is mentioned in 5 sentences in this paper.
Topics mentioned in this paper:
Hall, David and Durrett, Greg and Klein, Dan
Introduction
Formally, our model is a CRF where the features factor over anchored rules of a small backbone grammar, as shown in Figure 1.
Parsing Model
All of these past CRF parsers do also exploit span features, as did the structured margin parser of Taskar et al.
Surface Feature Framework
Recall that our CRF factors over anchored rules 7“, where each 7“ has identity rule(7“) and anchoring span(r).
Surface Feature Framework
As far as we can tell, all past CRF parsers have used “positive” features only.
CRF is mentioned in 4 sentences in this paper.
Topics mentioned in this paper:
Monroe, Will and Green, Spence and Manning, Christopher D.
Arabic Word Segmentation Model
A CRF model (Lafferty et a1., 2001) defines a distri-butionp(Y|X; 6), where X = {$1, .
Arabic Word Segmentation Model
The model of Green and DeNero is a third-order (i.e., 4—gram) Markov CRF , employing the following indicator features:
Introduction
The model is an extension of the character-level conditional random field ( CRF ) model of Green and DeNero (2012).
CRF is mentioned in 3 sentences in this paper.
Topics mentioned in this paper: