Index of papers in Proc. ACL 2009 that mention
  • sentence pairs
Huang, Fei
Abstract
Based on these measures, we improve the alignment quality by selecting high confidence sentence alignments and alignment links from multiple word alignments of the same sentence pair .
Alignment Link Confidence Measure
Similar to the sentence alignment confidence measure, the confidence of an alignment link aij in the sentence pair (S, T) is defined as
Alignment Link Confidence Measure
From multiple alignments of the same sentence pair , we select high confidence links from different alignments based on their link confidence scores and alignment agreement ratio.
Alignment Link Confidence Measure
Suppose the sentence pair (8, T) have alignments A1,.
Improved MaXEnt Aligner with Confidence-based Link Filtering
512 sentence pairs, and the A-E alignment test set is the 200 Arabic-English sentence pairs from NIST MT03 test set.
Introduction
The example in Figure 1 shows the word alignment of the given Chinese and English sentence pair , where the English words following each Chinese word is its literal translation.
Introduction
In this paper we introduce a confidence measure for word alignment, which is robust to extra or missing words in the bilingual sentence pairs , as well as word alignment errors.
Sentence Alignment Confidence Measure
Given a bilingual sentence pair (S,T) where S={31,. .
Sentence Alignment Confidence Measure
We randomly selected 512 Chinese-English (CE) sentence pairs and generated word alignment using the MaxEnt aligner (Ittycheriah and Roukos, 2005).
Sentence Alignment Confidence Measure
For each sentence pair , we also calculate the sentence alignment confidence score — log 0 (A|S, T).
sentence pairs is mentioned in 15 sentences in this paper.
Topics mentioned in this paper:
Celikyilmaz, Asli and Thint, Marcus and Huang, Zhiheng
Conclusions and Discussions
In this paper, we applied a graph-based SSL algorithm to improve the performance of QA task by exploiting unlabeled entailment relations between affirmed question and candidate sentence pairs .
Experiments
We also used a set of 340 QA-type sentence pairs from RTE02-03 and 195 pairs from RTE04 by converting the hypothesis sentences into question form to create additional set of q/a pairs.
Experiments
Each of these headline-candidate sentence pairs is used as additional unlabeled q/a pair.
Experiments
This is due to the fact that we establish two seperate entailment models for copula and non-copula q/a sentence pairs that enables extracting useful information and better representation of the specific data.
Feature Extraction for Entailment
(2) (QComp) Question component match features: The sentence component analysis is applied on both the affirmed question and the candidate sentence pairs to characterize their semantic components including subject(S), object(O), head (H) and modifiers(M).
Graph Based Semi-Supervised Learning for Entailment Ranking
Let each data point in X = {531, ..., an}, xi E SW represents information about a question and candidate sentence pair and Y = {3/1, be their output labels.
sentence pairs is mentioned in 6 sentences in this paper.
Topics mentioned in this paper:
DeNero, John and Chiang, David and Knight, Kevin
Consensus Decoding Algorithms
2.1 Minimum Bayes Risk over Sentence Pairs
Consensus Decoding Algorithms
Algorithm 1 MBR over Sentence Pairs 12 A <— —00 2: for e E E do 3: A6 <— 0 4 for e’ E E do 5: Ae<—A6+P(e’|f)-S(e;e’) 6 7
Consensus Decoding Algorithms
MBR over Sentence Pairs MBR over Features
Experimental Results
A phrase discovery procedure over word-aligned sentence pairs provides rule frequency counts, which are normalized to estimate features on rules.
Experimental Results
The synchronous grammar rules are extracted from word aligned sentence pairs where the target sentence is annotated with a syntactic parse (Galley et al., 2004).
sentence pairs is mentioned in 6 sentences in this paper.
Topics mentioned in this paper:
Ganchev, Kuzman and Gillenwater, Jennifer and Taskar, Ben
Approach
Figure 1(b) shows an aligned sentence pair example where dependencies are perfectly conserved across the alignment.
Approach
For example, in some sentence pair we might find 10 edges that have both end points aligned and can be transferred.
Experiments
Our basic model uses constraints of the form: the expected proportion of conserved edges in a sentence pair is at least 77 = 90%.1
Posterior Regularization
In grammar transfer, our basic constraint is of the form: the expected proportion of conserved edges in a sentence pair is at least 77 (the exact proportion we used was 0.9, which was determined using unlabeled data as described in Section 5).
sentence pairs is mentioned in 4 sentences in this paper.
Topics mentioned in this paper:
Wu, Hua and Wang, Haifeng
Conclusion
9We used about 70k sentence pairs for CE model training, while Wang et a1.
Conclusion
(2008) used about 100k sentence pairs , a CE translation dictionary and more monolingual corpora for model training.
Experiments
For English-Spanish translation, we selected 400k sentence pairs from the Europarl corpus that are close to the English parts of both the BTEC CE corpus and the BTEC ES corpus.
Using RBMT Systems for Pivot Translation
Another way to use the synthetic multilingual corpus is to add the source-pivot or pivot-target sentence pairs in this corpus to the training data to rebuild the source-pivot or pivot-target SMT model.
sentence pairs is mentioned in 4 sentences in this paper.
Topics mentioned in this paper:
Liu, Yang and Lü, Yajuan and Liu, Qun
Experiments
Table 4: Comparison of rule extraction time (seconds/ 1000 sentence pairs ) and decoding time (sec-ond/sentence)
Experiments
Table 4 gives the rule extraction time (seconds/ 1000 sentence pairs ) and decoding time (sec-ond/sentence) with varying pruning thresholds.
Experiments
the new training corpus contained about 260K sentence pairs with 7.39M Chinese words and 9.41M English words.
sentence pairs is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Pado, Sebastian and Galley, Michel and Jurafsky, Dan and Manning, Christopher D.
EXpt. 1: Predicting Absolute Scores
Each language consists of 1500—2800 sentence pairs produced by 7—15 MT systems.
EXpt. 1: Predicting Absolute Scores
RTER has a rather flat learning curve that climbs to within 2 points of the final correlation value for 20% of the training set (about 400 sentence pairs ).
Textual Entailment vs. MT Evaluation
The average total runtime per sentence pair is 5 seconds on an AMD 2.6GHz Opteron core — efficient enough to perform regular evaluations on development and test sets.
sentence pairs is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Setiawan, Hendra and Kan, Min Yen and Li, Haizhou and Resnik, Philip
Experimental Setup
We trained the system on the NIST MT06 Eval corpus excluding the UN data (approximately 900K sentence pairs ).
Experimental Setup
We trained the system on a subset of 950K sentence pairs from the NIST MT08 training data, selected by
Experimental Setup
The subsampling algorithm selects sentence pairs from the training data in a way that seeks reasonable representation for all n-grams appearing in the test set.
sentence pairs is mentioned in 3 sentences in this paper.
Topics mentioned in this paper: