Index of papers in Proc. ACL 2013 that mention
  • CRFs
Feng, Minwei and Peter, Jan-Thorsten and Ney, Hermann
Experiments
In this section, we describe the baseline setup, the CRFs training results, the RNN training results
Experiments
0 Wapiti toolkit (Lavergne et al., 2010) used for CRFs ; RNN is built by the RNNLIB toolkit.
Experiments
5.2 CRFs Training Results
Tagging-style Reordering Model
For this supervised learning task, we choose the approach conditional random fields ( CRFs ) (Lafferty et al., 2001; Sutton and Mccallum, 2006; Lavergne et al., 2010) and recurrent neural network (RNN) (Elman, 1990; Jordan, 1990; Lang et al., 1990).
Tagging-style Reordering Model
For the first method, we adopt the linear-chain CRFs .
Tagging-style Reordering Model
However, even for the simple linear-chain CRFs , the complexity of learning and inference grows quadratically with respect to the number of output labels and the amount of structural features which are with regard to adjacent pairs of labels.
CRFs is mentioned in 19 sentences in this paper.
Topics mentioned in this paper:
Zeng, Xiaodong and Wong, Derek F. and Chao, Lidia S. and Trancoso, Isabel
Abstract
The derived label distributions are regarded as virtual evidences to regularize the learning of linear conditional random fields ( CRFs ) on unlabeled data.
Abstract
Empirical results on Chinese tree bank (CTB-7) and Microsoft Research corpora (MSR) reveal that the proposed model can yield better results than the supervised baselines and other competitive semi-supervised CRFs in this task.
Background
The first-order CRFs model (Lafferty et al., 2001) has been the most common one in this task.
Background
The goal is to learn a CRFs model in the form,
Introduction
The derived label distributions are regarded as prior knowledge to regularize the learning of a sequential model, conditional random fields ( CRFs ) in this case, on both
Introduction
Section 3 reviews the background, including supervised character-based joint S&T model based on CRFs and graph-based label propagation.
Related Work
Sun and Xu (2011) enhanced a CWS model by interpolating statistical features of unlabeled data into the CRFs model.
Related Work
also differs from other semi-supervised CRFs algorithms.
Related Work
(2006), extended by Mann and McCallum (2007), reported a semi-supervised CRFs model which aims to guide the learning by minimizing the conditional entropy of unlabeled data.
CRFs is mentioned in 30 sentences in this paper.
Topics mentioned in this paper:
Dethlefs, Nina and Hastie, Helen and Cuayáhuitl, Heriberto and Lemon, Oliver
Abstract
We formulate surface realisation as a sequence labelling task and combine the use of conditional random fields ( CRFs ) with semantic trees.
Abstract
Due to their extended notion of context, CRFs are able to take the global utterance context into account and are less constrained by local features than other realisers.
Cohesion across Utterances
This grammar defines the surface realisation space for the CRFs .
Conclusion and Future Directions
We have argued that CRFs are well suited for this task because they are not restricted by independence assumptions.
Conclusion and Future Directions
In addition, we may compare different sequence labelling algorithms for surface realisation (Nguyen and Guo, 2007) or segmented CRFs (Sarawagi and Cohen, 2005) and apply our method to more complex surface realisation domains such as text generation or summarisation.
Evaluation
CRFs and other state-of-the-art methods, we also compare our system to two other baselines:
Incremental Surface Realisation
Since CRFs are not restricted by the Markov condition, they are less constrained by local context than other models and can take nonlocal dependencies into account.
Incremental Surface Realisation
While their extended context awareness can often make CRFs slow to train, they are fast at execution and therefore very applicable to the incremental scenario.
Introduction
to surface realisation within incremental systems, because CRFs are able to model context across full as well as partial generator inputs which may undergo modifications during generation.
Related Work
(2009) who also use CRFs to find the best surface realisation from a semantic tree.
CRFs is mentioned in 10 sentences in this paper.
Topics mentioned in this paper:
Zeng, Xiaodong and Wong, Derek F. and Chao, Lidia S. and Trancoso, Isabel
Experiment
The feature templates in (Zhao et al., 2006) and (Zhang and Clark, 2007) are used in training the CRFs model and Perceptrons model, respectively.
Introduction
Sun and Xu (2011) enhanced the segmentation results by interpolating the statistics-based features derived from unlabeled data to a CRFs model.
Segmentation Models
This section briefly reviews two supervised models in these categories, a character-based CRFs model, and a word-based Perceptrons model, which are used in our approach.
Segmentation Models
2.1 Character-based CRFs Model
Segmentation Models
Xue (2003) first proposed the use of CRFs model (Lafferty et al., 2001) in character-based CWS.
Semi-supervised Learning via Co-regularizing Both Models
The model induction process is described in Algorithm 1: given labeled dataset D; and unlabeled dataset Du, the first two steps are training a CRFs (character-based) and Perceptrons (word-based) model on the labeled data D; , respectively.
Semi-supervised Learning via Co-regularizing Both Models
Afterwards, the agreements A are used as a set of constraints to bias the learning of CRFs (§ 3.2) and Perceptron (§ 3.3) on the unlabeled data.
Semi-supervised Learning via Co-regularizing Both Models
3.2 CRFs with Constraints
CRFs is mentioned in 10 sentences in this paper.
Topics mentioned in this paper:
Oh, Jong-Hoon and Torisawa, Kentaro and Hashimoto, Chikara and Sano, Motoki and De Saeger, Stijn and Ohtake, Kiyonori
Causal Relations for Why-QA
We regard this task as a sequence labeling problem and use Conditional Random Fields ( CRFs ) (Laf-ferty et al., 2001) as a machine learning framework.
Causal Relations for Why-QA
In our task, CRFs take three sentences of a causal relation candidate as input and generate their cause-effect annotations with a set of possible cause-effect IOB labels, including Begin-Cause (BC), Inside-Cause (IC), Begin-Effect (BE), Inside-Effect (IE), and Outside (0).
Causal Relations for Why-QA
We used the three types of feature sets in Table 3 for training the CRFs , where j is in the range of z' — 4 g j g i + 4 for current position i in a causal relation candidate.
CRFs is mentioned in 4 sentences in this paper.
Topics mentioned in this paper:
Wang, Aobo and Kan, Min-Yen
Methodology
Given the general performance and discrimi-native framework, Conditional Random Fields ( CRFs ) (Lafferty et al., 2001) is a suitable framework for tackling sequence labeling problems.
Methodology
CRFs represent a basic, simple and well-understood framework for sequence labeling, making it a suitable framework for adapting to perform joint inference.
Methodology
Figure 2: Graphical representations of the two types of CRFs used in this work.
CRFs is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Yang, Bishan and Cardie, Claire
Experiments
We trained CRFs for opinion entity identification using the following features: indicators for words, POS tags, and lexicon features (the subjectivity strength of the word in the Subjectivity Lexicon).
Model
We formulate the task of opinion entity identification as a sequence labeling problem and employ conditional random fields ( CRFs ) (Lafferty et al., 2001) to learn the probability of a sequence assignment y for a given sentence x.
Model
We define potential function fig that gives the probability of assigning a span 2' with entity label 2, and the probability is estimated based on the learned parameters from CRFs .
CRFs is mentioned in 3 sentences in this paper.
Topics mentioned in this paper: