Index of papers in Proc. ACL 2010 that mention
  • log-linear
Wuebker, Joern and Mauser, Arne and Ney, Hermann
Experimental Evaluation
The features are combined in a log-linear way.
Experimental Evaluation
In Table 5 we can see that the performance of the heuristic phrase model can be increased by 0.6 BLEU on TEST by filtering the phrase table to contain the same phrases as the count model and reoptimizing the log-linear model weights.
Experimental Evaluation
Log-linear interpolation of the count model with the heuristic yields a further increase, showing an improvement of 1.3 BLEU on DEV and 1.4 BLEU on TEST over the baseline.
Introduction
The translation process is implemented as a weighted log-linear combination of several models hm(e{, sf , ff) including the logarithm of the phrase probability in source-to-target as well as in target-to-source direction.
Phrase Model Training
The log-linear interpolations pint(f|é) of the phrase translation probabilities are estimated as
Phrase Model Training
As a generalization of the fixed interpolation of the two phrase tables we also experimented with adding the two trained phrase probabilities as additional features to the log-linear framework.
Phrase Model Training
With good log-linear feature weights, feature-wise combination should perform at least as well as fixed interpolation.
log-linear is mentioned in 10 sentences in this paper.
Topics mentioned in this paper:
Branavan, S.R.K. and Zettlemoyer, Luke and Barzilay, Regina
Algorithm
Specifically, we modify the log-linear policy p(a|s; q, 6) by adding lookahead features gb(s, a, q) which complement the local features used in the previous model.
Background
A Log-Linear Parameterization The policy
Background
function used for action selection is defined as a log-linear distribution over actions: €9-¢(s,a)
log-linear is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Xiao, Tong and Zhu, Jingbo and Zhu, Muhua and Wang, Huizhen
Background
where Pr(e| f) is the probability that e is the translation of the given source string f. To model the posterior probability Pr(e| f) , most of the state-of-the-art SMT systems utilize the log-linear model proposed by Och and Ney (2002), as follows,
Background
In this paper, u denotes a log-linear model that has Mfixed features {h1(f,e), ..., hM(f,e)}, ,1 = {3.1, ..., AM} denotes the M parameters of u, and u(/1) denotes a SMT system based on u with parameters ,1.
Background
In this paper, we use the term training set to emphasize the training of log-linear model.
log-linear is mentioned in 3 sentences in this paper.
Topics mentioned in this paper: