Introduction | They use aligned sequences of words, called biphrases, as building blocks for translations, and score alternative candidate translations for the same source sentence based on a log-linear model of the conditional probability of target sentences given the source sentence: |
Introduction | alignment a, where the alignment is a representation of the sequence of biphrases that where used in order to build T from S; The Ak’s are weights and ZS is a normalization factor that guarantees that p is a proper conditional probability distribution over the pairs (T, A). |
Introduction | These typically include forward and reverse phrase conditional probability features log p(f|§) as well as logp(§ where 59' is the source side of the biphrase and f the target side, and the so-called “phrase penalty” and “word penalty” features, which count the number of phrases and words in the alignment. |
Phrase-based Decoding as TSP | For example, in the example of Figure 3, the cost of this - machine translation - is - strange, can only take into account the conditional probability of the word strange relative to the word is, but not relative to the words translation and is. |
Related work | its translation has larger conditional probability ) or because the associated heuristics is less tight, hence more optimistic. |
Experiments | The joint probability is expressed as a chain product of a series of conditional probabilities of token pairs P({ei}, {cj}) = P((ék,ck)|(ék_1,ck_1)), k = 1 . |
Experiments | The conditional probabilities for token pairs are estimated from the aligned training corpus. |
Transliteration alignment techniques | From the affinity matrix conditional probabilities P(ei|cj) can be estimated as |
Transliteration alignment techniques | We estimate conditional probability of Chinese phoneme cpk, after observing English character 6, as |
Active Learning for Sequence Labeling | We apply CRFs as our base learner throughout this paper and employ a utility function which is based on the conditional probability of the most likely label sequence 37* for an observation sequence :8 (cf. |
Conditional Random Fields for Sequence Labeling | See Equation (1) for the conditional probability PX(37 with ZX calculated as in Equation (6). |
Conditional Random Fields for Sequence Labeling | The marginal and conditional probabilities are used by our AL approaches as confidence estimators. |
Image Clustering with Annotated Auxiliary Data | Then we apply the EM algorithm (Dempster et al., 1977) to estimate the conditional probabilities P(f|z), and P(z|v) with respect to each dependence in Figure 3 as follows. |
Image Clustering with Annotated Auxiliary Data | o M-Step: re-estimates conditional probabilities P(zk|vi) and P(zk|wl): |
Image Clustering with Annotated Auxiliary Data | and conditional probability P(fj |zk), which is a mixture portion of posterior probability of latent variables |