Index of papers in Proc. ACL 2014 that mention
  • MaxEnt
Bhat, Suma and Xue, Huichao and Yoon, Su-Youn
Experimental Setup
Subsequently, the feature extraction stage (a VSM or a MaxEnt model as the case may be) generates the syntactic complexity feature which is then incorporated in a multiple linear regression model to generate a score.
Experimental Setup
We used the maximum entropy classifier implementation in the MaxEnt toolkit4.
Experimental Setup
The results that follow are based on MaxEnt classifier’s parameter settings initialized to zero.
Models for Measuring Grammatical Competence
The inductive classifier we use here is the maximum-entropy model ( MaxEnt ) which has been used to solve several statistical natural language processing problems with much success (Berger et al., 1996; Borthwick et al., 1998; Borthwick, 1999; Pang et al., 2002; Klein et al., 2003; Rosenfeld, 2005).
Models for Measuring Grammatical Competence
The productive feature engineering aspects of incorporating features into the discriminative MaxEnt classifier motivate the model choice for the problem at hand.
Models for Measuring Grammatical Competence
In particular, the ability of the MaxEnt model’s estimation routine to handle overlapping (correlated) features makes it directly applicable to address the first limitation of the VSM model.
MaxEnt is mentioned in 15 sentences in this paper.
Topics mentioned in this paper:
Xiong, Deyi and Zhang, Min
Decoding with Sense-Based Translation Model
MaxEnt classifiers
Decoding with Sense-Based Translation Model
Once we get sense clusters for word tokens in test sentences, we load pre-trained MaXEnt classifiers of the corresponding word types.
Experiments
We trained our MaxEnt classifiers with the off-the-shelf MaxEnt tool.4 We performed 100 iterations of the L-BFGS algorithm implemented in the training toolkit on the collected training events from the sense-annotated data as described in Section 3.2.
Experiments
It took an average of 57.5 seconds for training a Maxent classifier.
Sense-Based Translation Model
entropy ( MaxEnt ) based classifier that is used to predict the translation probability p(é|C(c)).
Sense-Based Translation Model
The MaxEnt classifier can be formulated as follows.
Sense-Based Translation Model
This is not a issue for the MaxEnt classifier as it can deal with arbitrary overlapping features (Berger et al., 1996).
MaxEnt is mentioned in 14 sentences in this paper.
Topics mentioned in this paper:
Chen, Liwei and Feng, Yansong and Huang, Songfang and Qin, Yong and Zhao, Dongyan
Conclusions
Furthermore, our framework is scalable for other local sentence level extractors in addition to the MaxEnt model.
Experiments
Our ILP model and its variants all outperform Mintz++ in precision in both datasets, indicating that our approach helps filter out incorrect predictions from the output of MaxEnt model.
Experiments
However, in the Riedel’s dataset, Mintz++, the MaxEnt relation extractor, does not perform well, and our framework cannot improve its performance.
Experiments
Hence, our framework does not perform well due to the poor performance of MaXEnt extractor and the lack of clues.
The Framework
By adopting ILP, we can combine the local information including MaXEnt confidence scores and the implicit relation backgrounds that are embedded into global consistencies of the entity tuples together.
MaxEnt is mentioned in 7 sentences in this paper.
Topics mentioned in this paper:
Doyle, Gabriel and Bicknell, Klinton and Levy, Roger
Experiment
To establish performance for the phonological standard, we use the IBPOT learner to find constraint weights but do not update M. The resultant learner is essentially MaxEnt OT with the weights estimated through Metropolis sampling instead of gradient ascent.
Introduction
We consider this question by examining the dominant framework in modern phonology, Optimality Theory (Prince and Smolensky, 1993, OT), implemented in a log-linear framework, MaXEnt OT (Goldwater and Johnson, 2003), with output forms’ probabilities based on a weighted sum of
Phonology and Optimality Theory 2.1 OT structure
In IBPOT, we use the log-linear EVAL developed by Goldwater and J ohn-son (2003) in their MaxEnt OT system.
Phonology and Optimality Theory 2.1 OT structure
MEOT also is motivated by the general MaxEnt framework, whereas most other OT formulations are ad hoc constructions specific to phonology.
Phonology and Optimality Theory 2.1 OT structure
In MaXEnt OT, each constraint has a weight, and the candidates’ scores are the sums of the weights of violated constraints.
MaxEnt is mentioned in 5 sentences in this paper.
Topics mentioned in this paper:
Huang, Fei and Xu, Jian-Ming and Ittycheriah, Abraham and Roukos, Salim
Document-specific MT System
ment (HMM (Vogel et al., 1996) and MaxEnt (Ittycheriah and Roukos, 2005) alignment models, phrase pair extraction, MT model training (Ittycheriah and Roukos, 2007) and LM model training.
Related Work
Target part-of-speech and null dependency link are exploited in a MaXEnt classifier to improve the MT quality estimation (Xiong et al., 2010).
Static MT Quality Estimation
0 17 decoding features, including phrase translation probabilities (source-to-target and target-to-source), word translation probabilities (also in both directions), maxent prob-abilitiesl, word count, phrase count, distor-
Static MT Quality Estimation
1The maxent probability is the translation probability
MaxEnt is mentioned in 4 sentences in this paper.
Topics mentioned in this paper:
Li, Junhui and Marton, Yuval and Resnik, Philip and Daumé III, Hal
Discussion
It shows that 1) as expected, our classifiers do worse on the harder semantic reordering prediction than syntactic reordering prediction; 2) thanks to the high accuracy obtained by the maxent classifiers, integrating either the syntactic or the semantic reordering constraints results in better reordering performance from both syntactic and semantic perspectives; 3) in terms of the mutual impact, the syntactic reordering models help improving semantic reordering more than the semantic reordering
Discussion
Syntactic Semantic l-m rm l-m rm MR08 75.0 78.0 66.3 68.5 +syn-reorder 78.4 80.9 69.0 70.2 +sem—reorder 76.0 78.8 70.7 72.7 +b0th 78.6 81.7 70.6 72.1 Maxent Classifier 80.7 85.6 70.9 73.5
Experiments
tactic parsing and semantic role labeling on the Chinese sentences, then train the models by using MaxEnt toolkit with L1 regularizer (Tsuruoka et al., 2009).3 Table 3 shows the reordering type distribution over the training data.
Related Work
Marton and Resnik (2008) employed soft syntactic constraints with weighted binary features and no MaXEnt model.
MaxEnt is mentioned in 4 sentences in this paper.
Topics mentioned in this paper:
Pershina, Maria and Min, Bonan and Xu, Wei and Grishman, Ralph
Available at http://nlp. stanford.edu/software/mimlre. shtml.
—l— Guided DS Semi—MIML —.— DS+upsampling —'— MaxEnt
Available at http://nlp. stanford.edu/software/mimlre. shtml.
Our baselines: 1) MaXEnt is a supervised maximum entropy baseline trained on a human-labeled data; 2) DS+upsamp|ing is an upsampling experiment, where MIML was trained on a mix of a distantly-labeled and human-labeled data; 3) Semi-MIML is a recent semi-supervised extension.
The Challenge
We experimentally tested alternative feature sets by building supervised Maximum Entropy ( MaxEnt ) models using the hand-labeled data (Table 3), and selected an effective combination of three features from the full feature set used by Surdeanu et al., (2011):
The Challenge
Table 3: Performance of a MaxEnt , trained on hand-labeled data using all features (Surdeanu et al., 2011) vs using a subset of two (types of entities, dependency path), or three (adding a span word) features, and evaluated on the test set.
MaxEnt is mentioned in 4 sentences in this paper.
Topics mentioned in this paper:
Joshi, Aditya and Mishra, Abhijit and Senthamilselvan, Nivvedan and Bhattacharyya, Pushpak
Discussion
We use three sentiment classification techniques: Na‘1've Bayes, MaxEnt and SVM with un-igrams, bigrams and trigrams as features.
Discussion
MaxEnt (Movie) -0.29 (72.17) MaxEnt (Twitter) -0.26 (71.68) SVM (Movie) -().24 (66.27) SVM (Twitter) -().19 (73.15)
Discussion
MaxEnt has the highest negative correlation of -().29 and -().26.
MaxEnt is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Kalchbrenner, Nal and Grefenstette, Edward and Blunsom, Phil
Experiments
unigram, bigram, trigram 92.6 MAXENT POS, chunks, NE, supertags
Experiments
unigram, bigram, trigram 93.6 MAxENT POS, wh-word, head word
Experiments
SVM 81.6 BINB 82.7 MAXENT 83.0 MAX-TDNN 78.8 NBOW 80.9 DCNN 87.4
MaxEnt is mentioned in 3 sentences in this paper.
Topics mentioned in this paper: