Abstract | Based on these measures, we improve the alignment quality by selecting high confidence sentence alignments and alignment links from multiple word alignments of the same sentence pair . |
Alignment Link Confidence Measure | Similar to the sentence alignment confidence measure, the confidence of an alignment link aij in the sentence pair (S, T) is defined as |
Alignment Link Confidence Measure | From multiple alignments of the same sentence pair , we select high confidence links from different alignments based on their link confidence scores and alignment agreement ratio. |
Alignment Link Confidence Measure | Suppose the sentence pair (8, T) have alignments A1,. |
Improved MaXEnt Aligner with Confidence-based Link Filtering | 512 sentence pairs, and the A-E alignment test set is the 200 Arabic-English sentence pairs from NIST MT03 test set. |
Introduction | The example in Figure 1 shows the word alignment of the given Chinese and English sentence pair , where the English words following each Chinese word is its literal translation. |
Introduction | In this paper we introduce a confidence measure for word alignment, which is robust to extra or missing words in the bilingual sentence pairs , as well as word alignment errors. |
Sentence Alignment Confidence Measure | Given a bilingual sentence pair (S,T) where S={31,. . |
Sentence Alignment Confidence Measure | We randomly selected 512 Chinese-English (CE) sentence pairs and generated word alignment using the MaxEnt aligner (Ittycheriah and Roukos, 2005). |
Sentence Alignment Confidence Measure | For each sentence pair , we also calculate the sentence alignment confidence score — log 0 (A|S, T). |
Conclusions and Discussions | In this paper, we applied a graph-based SSL algorithm to improve the performance of QA task by exploiting unlabeled entailment relations between affirmed question and candidate sentence pairs . |
Experiments | We also used a set of 340 QA-type sentence pairs from RTE02-03 and 195 pairs from RTE04 by converting the hypothesis sentences into question form to create additional set of q/a pairs. |
Experiments | Each of these headline-candidate sentence pairs is used as additional unlabeled q/a pair. |
Experiments | This is due to the fact that we establish two seperate entailment models for copula and non-copula q/a sentence pairs that enables extracting useful information and better representation of the specific data. |
Feature Extraction for Entailment | (2) (QComp) Question component match features: The sentence component analysis is applied on both the affirmed question and the candidate sentence pairs to characterize their semantic components including subject(S), object(O), head (H) and modifiers(M). |
Graph Based Semi-Supervised Learning for Entailment Ranking | Let each data point in X = {531, ..., an}, xi E SW represents information about a question and candidate sentence pair and Y = {3/1, be their output labels. |
Consensus Decoding Algorithms | 2.1 Minimum Bayes Risk over Sentence Pairs |
Consensus Decoding Algorithms | Algorithm 1 MBR over Sentence Pairs 12 A <— —00 2: for e E E do 3: A6 <— 0 4 for e’ E E do 5: Ae<—A6+P(e’|f)-S(e;e’) 6 7 |
Consensus Decoding Algorithms | MBR over Sentence Pairs MBR over Features |
Experimental Results | A phrase discovery procedure over word-aligned sentence pairs provides rule frequency counts, which are normalized to estimate features on rules. |
Experimental Results | The synchronous grammar rules are extracted from word aligned sentence pairs where the target sentence is annotated with a syntactic parse (Galley et al., 2004). |
Approach | Figure 1(b) shows an aligned sentence pair example where dependencies are perfectly conserved across the alignment. |
Approach | For example, in some sentence pair we might find 10 edges that have both end points aligned and can be transferred. |
Experiments | Our basic model uses constraints of the form: the expected proportion of conserved edges in a sentence pair is at least 77 = 90%.1 |
Posterior Regularization | In grammar transfer, our basic constraint is of the form: the expected proportion of conserved edges in a sentence pair is at least 77 (the exact proportion we used was 0.9, which was determined using unlabeled data as described in Section 5). |
Conclusion | 9We used about 70k sentence pairs for CE model training, while Wang et a1. |
Conclusion | (2008) used about 100k sentence pairs , a CE translation dictionary and more monolingual corpora for model training. |
Experiments | For English-Spanish translation, we selected 400k sentence pairs from the Europarl corpus that are close to the English parts of both the BTEC CE corpus and the BTEC ES corpus. |
Using RBMT Systems for Pivot Translation | Another way to use the synthetic multilingual corpus is to add the source-pivot or pivot-target sentence pairs in this corpus to the training data to rebuild the source-pivot or pivot-target SMT model. |
Experiments | Table 4: Comparison of rule extraction time (seconds/ 1000 sentence pairs ) and decoding time (sec-ond/sentence) |
Experiments | Table 4 gives the rule extraction time (seconds/ 1000 sentence pairs ) and decoding time (sec-ond/sentence) with varying pruning thresholds. |
Experiments | the new training corpus contained about 260K sentence pairs with 7.39M Chinese words and 9.41M English words. |
EXpt. 1: Predicting Absolute Scores | Each language consists of 1500—2800 sentence pairs produced by 7—15 MT systems. |
EXpt. 1: Predicting Absolute Scores | RTER has a rather flat learning curve that climbs to within 2 points of the final correlation value for 20% of the training set (about 400 sentence pairs ). |
Textual Entailment vs. MT Evaluation | The average total runtime per sentence pair is 5 seconds on an AMD 2.6GHz Opteron core — efficient enough to perform regular evaluations on development and test sets. |
Experimental Setup | We trained the system on the NIST MT06 Eval corpus excluding the UN data (approximately 900K sentence pairs ). |
Experimental Setup | We trained the system on a subset of 950K sentence pairs from the NIST MT08 training data, selected by |
Experimental Setup | The subsampling algorithm selects sentence pairs from the training data in a way that seeks reasonable representation for all n-grams appearing in the test set. |