Experiments | In this section, we describe the baseline setup, the CRFs training results, the RNN training results |
Experiments | 0 Wapiti toolkit (Lavergne et al., 2010) used for CRFs ; RNN is built by the RNNLIB toolkit. |
Experiments | 5.2 CRFs Training Results |
Tagging-style Reordering Model | For this supervised learning task, we choose the approach conditional random fields ( CRFs ) (Lafferty et al., 2001; Sutton and Mccallum, 2006; Lavergne et al., 2010) and recurrent neural network (RNN) (Elman, 1990; Jordan, 1990; Lang et al., 1990). |
Tagging-style Reordering Model | For the first method, we adopt the linear-chain CRFs . |
Tagging-style Reordering Model | However, even for the simple linear-chain CRFs , the complexity of learning and inference grows quadratically with respect to the number of output labels and the amount of structural features which are with regard to adjacent pairs of labels. |
Abstract | The derived label distributions are regarded as virtual evidences to regularize the learning of linear conditional random fields ( CRFs ) on unlabeled data. |
Abstract | Empirical results on Chinese tree bank (CTB-7) and Microsoft Research corpora (MSR) reveal that the proposed model can yield better results than the supervised baselines and other competitive semi-supervised CRFs in this task. |
Background | The first-order CRFs model (Lafferty et al., 2001) has been the most common one in this task. |
Background | The goal is to learn a CRFs model in the form, |
Introduction | The derived label distributions are regarded as prior knowledge to regularize the learning of a sequential model, conditional random fields ( CRFs ) in this case, on both |
Introduction | Section 3 reviews the background, including supervised character-based joint S&T model based on CRFs and graph-based label propagation. |
Related Work | Sun and Xu (2011) enhanced a CWS model by interpolating statistical features of unlabeled data into the CRFs model. |
Related Work | also differs from other semi-supervised CRFs algorithms. |
Related Work | (2006), extended by Mann and McCallum (2007), reported a semi-supervised CRFs model which aims to guide the learning by minimizing the conditional entropy of unlabeled data. |
Abstract | We formulate surface realisation as a sequence labelling task and combine the use of conditional random fields ( CRFs ) with semantic trees. |
Abstract | Due to their extended notion of context, CRFs are able to take the global utterance context into account and are less constrained by local features than other realisers. |
Cohesion across Utterances | This grammar defines the surface realisation space for the CRFs . |
Conclusion and Future Directions | We have argued that CRFs are well suited for this task because they are not restricted by independence assumptions. |
Conclusion and Future Directions | In addition, we may compare different sequence labelling algorithms for surface realisation (Nguyen and Guo, 2007) or segmented CRFs (Sarawagi and Cohen, 2005) and apply our method to more complex surface realisation domains such as text generation or summarisation. |
Evaluation | CRFs and other state-of-the-art methods, we also compare our system to two other baselines: |
Incremental Surface Realisation | Since CRFs are not restricted by the Markov condition, they are less constrained by local context than other models and can take nonlocal dependencies into account. |
Incremental Surface Realisation | While their extended context awareness can often make CRFs slow to train, they are fast at execution and therefore very applicable to the incremental scenario. |
Introduction | to surface realisation within incremental systems, because CRFs are able to model context across full as well as partial generator inputs which may undergo modifications during generation. |
Related Work | (2009) who also use CRFs to find the best surface realisation from a semantic tree. |
Experiment | The feature templates in (Zhao et al., 2006) and (Zhang and Clark, 2007) are used in training the CRFs model and Perceptrons model, respectively. |
Introduction | Sun and Xu (2011) enhanced the segmentation results by interpolating the statistics-based features derived from unlabeled data to a CRFs model. |
Segmentation Models | This section briefly reviews two supervised models in these categories, a character-based CRFs model, and a word-based Perceptrons model, which are used in our approach. |
Segmentation Models | 2.1 Character-based CRFs Model |
Segmentation Models | Xue (2003) first proposed the use of CRFs model (Lafferty et al., 2001) in character-based CWS. |
Semi-supervised Learning via Co-regularizing Both Models | The model induction process is described in Algorithm 1: given labeled dataset D; and unlabeled dataset Du, the first two steps are training a CRFs (character-based) and Perceptrons (word-based) model on the labeled data D; , respectively. |
Semi-supervised Learning via Co-regularizing Both Models | Afterwards, the agreements A are used as a set of constraints to bias the learning of CRFs (§ 3.2) and Perceptron (§ 3.3) on the unlabeled data. |
Semi-supervised Learning via Co-regularizing Both Models | 3.2 CRFs with Constraints |
Causal Relations for Why-QA | We regard this task as a sequence labeling problem and use Conditional Random Fields ( CRFs ) (Laf-ferty et al., 2001) as a machine learning framework. |
Causal Relations for Why-QA | In our task, CRFs take three sentences of a causal relation candidate as input and generate their cause-effect annotations with a set of possible cause-effect IOB labels, including Begin-Cause (BC), Inside-Cause (IC), Begin-Effect (BE), Inside-Effect (IE), and Outside (0). |
Causal Relations for Why-QA | We used the three types of feature sets in Table 3 for training the CRFs , where j is in the range of z' — 4 g j g i + 4 for current position i in a causal relation candidate. |
Methodology | Given the general performance and discrimi-native framework, Conditional Random Fields ( CRFs ) (Lafferty et al., 2001) is a suitable framework for tackling sequence labeling problems. |
Methodology | CRFs represent a basic, simple and well-understood framework for sequence labeling, making it a suitable framework for adapting to perform joint inference. |
Methodology | Figure 2: Graphical representations of the two types of CRFs used in this work. |
Experiments | We trained CRFs for opinion entity identification using the following features: indicators for words, POS tags, and lexicon features (the subjectivity strength of the word in the Subjectivity Lexicon). |
Model | We formulate the task of opinion entity identification as a sequence labeling problem and employ conditional random fields ( CRFs ) (Lafferty et al., 2001) to learn the probability of a sequence assignment y for a given sentence x. |
Model | We define potential function fig that gives the probability of assigning a span 2' with entity label 2, and the probability is estimated based on the learned parameters from CRFs . |