Index of papers in Proc. ACL 2014 that mention
  • SMT system
Riezler, Stefan and Simianer, Patrick and Haas, Carolin
Experiments
The bilingual SMT system used in our experiments is the state-of-the-art SCFG decoder CDEC (Dyer et al., 2010)5.
Experiments
We trained the SMT system on the English-German parallel web data provided in the COMMON CRAWL6 (Smith et al., 2013) dataset.
Experiments
Method 1 is the baseline system, consisting of the CDEC SMT system trained on the COMMON CRAWL data as described above.
Grounding SMT in Semantic Parsing
Given a manual German translation of the English query as source sentence, the SMT system produces an English target translation.
Introduction
This avoids the problem of un-reachability of independently generated reference translations by the SMT system .
Introduction
in which SMT systems can be trained and evaluated.
Related Work
(2012) propose a setup where an SMT system feeds into cross-language information retrieval, and receives feedback from the performance of translated queries with respect to cross-language retrieval performance.
Related Work
However, despite offering direct and reliable prediction of translation quality, the cost and lack of reusability has confined task-based evaluations involving humans to testing scenarios, but prevented a use for interactive training of SMT systems as in our work.
Response-based Online Learning
Firstly, update rules that require to compute a feature representation for the reference translation are suboptimal in SMT, because often human-generated reference translations cannot be generated by the SMT system .
Response-based Online Learning
Such “un-reachable” gold-standard translations need to be replaced by “surrogate” gold-standard translations that are close to the human-generated translations and still lie within the reach of the SMT system .
SMT system is mentioned in 10 sentences in this paper.
Topics mentioned in this paper:
Cui, Lei and Zhang, Dongdong and Liu, Shujie and Chen, Qiming and Li, Mu and Zhou, Ming and Yang, Muyun
Conclusion and Future Work
Experimental results show that our approach is promising for SMT systems to learn a better translation model.
Experiments
This proves that bilingually induced topic representation with neural network helps the SMT system disambiguate translation candidates.
Introduction
For example, translation sense disambiguation approaches (Carpuat and Wu, 2005; Carpuat and Wu, 2007) are proposed for phrase-based SMT systems .
Introduction
Meanwhile, for hierarchical phrase-based or syntax-based SMT systems , there is also much work involving rich contexts to guide rule selection (He et al., 2008; Liu et al., 2008; Marton and Resnik, 2008; Xiong et al., 2009).
Introduction
Although these methods are effective and proven successful in many SMT systems , they only leverage within-
Topic Similarity Model with Neural Network
For the SMT system , the best translation candidate 6 is given by:
SMT system is mentioned in 8 sentences in this paper.
Topics mentioned in this paper:
Xiong, Deyi and Zhang, Min
Decoding with Sense-Based Translation Model
Figure 2: Architecture of SMT system with the sense-based translation model.
Decoding with Sense-Based Translation Model
Figure 2 shows the architecture of the SMT system enhanced with the sense-based translation model.
Experiments
Our baseline system is a state-of-the-art SMT system which adapts Bracketing Transduction Grammars (Wu, 1997) to phrasal translation and equips itself with a maximum entropy based reordering model (Xiong et al., 2006).
Introduction
These glosses, used as the sense predictions of their WSD system, are integrated into a word-based SMT system either to substitute for translation candidates of their translation model or to postedit the output of their SMT system .
Introduction
We integrate the proposed sense-based translation model into a state-of-the-art SMT system and conduct experiments on Chines-to-English translation using large-scale training data.
SMT system is mentioned in 5 sentences in this paper.
Topics mentioned in this paper:
Hu, Yuening and Zhai, Ke and Eidelman, Vladimir and Boyd-Graber, Jordan
Experiments
To optimize SMT system , we tune the parameters on NIST MT06, and report results on three test sets: MT02, MT03 and MT05.2
Introduction
In contrast, machine translation uses inherently multilingual data: an SMT system must translate a phrase or sentence from a source language to a different target language, so existing applications of topic models (Eidelman et al., 2012) are wilfully ignoring available information on the target side that could aid domain discovery.
Topic Models for Machine Translation
Cross-Domain SMT A SMT system is usually trained on documents with the same genre (e.g., sports, business) from a similar style (e.g., newswire, blog-posts).
Topic Models for Machine Translation
Domain Adaptation for SMT Training a SMT system using diverse data requires domain adaptation.
SMT system is mentioned in 4 sentences in this paper.
Topics mentioned in this paper:
Schamoni, Shigehiko and Hieber, Felix and Sokolov, Artem and Riezler, Stefan
Experiments
order to simulate pass-through behavior of out-of-vocabulary terms in SMT systems , additional features accounting for source and target term identity were added to DK and BM models.
Introduction
This approach is advantageous if large amounts of in-domain sentence-parallel data are available to train SMT systems , but relevance rankings to train retrieval models are not.
Related Work
In a direct translation approach (DT), a state-of-the-art SMT system is used to produce a single best translation that is used as search query in the target language.
Related Work
For example, Google’s CLIR approach combines their state-of-the-art SMT system with their proprietary search engine (Chin et al., 2008).
SMT system is mentioned in 4 sentences in this paper.
Topics mentioned in this paper:
Tu, Mei and Zhou, Yu and Zong, Chengqing
Conclusion
In the future, we will extend our methods to other translation models, such as the syntax-based model, to study how to further improve the performance of SMT systems .
Introduction
In Section 3, we present our models and show how to integrate the models into an SMT system .
Related Work
They added the labels assigned to connectives as an additional input to an SMT system , but their experimental results show that the improvements under the evaluation metric of BLEU were not significant.
Related Work
To the best of our knowledge, our work is the first attempt to exploit the source functional relationship to generate the target transitional expressions for grammatical cohesion, and we have successfully incorporated the proposed models into an SMT system with significant improvement of BLEU metrics.
SMT system is mentioned in 4 sentences in this paper.
Topics mentioned in this paper:
Guzmán, Francisco and Joty, Shafiq and Màrquez, Llu'is and Nakov, Preslav
Introduction
Although modern SMT systems have switched to a discriminative log-linear framework, which allows for additional sources as features, it is generally hard to incorporate dependencies beyond a small window of adjacent words, thus making it difficult to use linguistically-rich models.
Introduction
We believe that the semantic and pragmatic information captured in the form of DTs (i) can help develop discourse-aware SMT systems that produce coherent translations, and (ii) can yield better MT evaluation metrics.
Introduction
While in this work we focus on the latter, we think that the former is also within reach, and that SMT systems would benefit from preserving the coherence relations in the source language when generating target-language translations.
SMT system is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Salameh, Mohammad and Cherry, Colin and Kondrak, Grzegorz
Methods
We approach this problem by augmenting an SMT system built over target segments with features that reflect the desegmented target words.
Methods
In this section, we describe our various strategies for desegmenting the SMT system’s output space, along with the features that we add to take advantage of this desegmented view.
Related Work
Other approaches train an SMT system to predict lemmas instead of surface forms, and then inflect the SMT output as a postprocessing step (Minkov et al., 2007; Clifton and Sarkar, 2011; Fraser et al., 2012; El Kholy and Habash, 2012b).
SMT system is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Salloum, Wael and Elfardy, Heba and Alamir-Salloum, Linda and Habash, Nizar and Diab, Mona
Machine Translation Experiments
We used MT08 and EgyDevV3 to tune SMT systems while we divided the remaining sets among classifier training data (5,562 sentences), dev (1,802 sentences) and blind test (1,804 sentences) sets to ensure each of these new sets has a variety of dialects and genres (weblog and newswire).
Machine Translation Experiments
This MSA-pivoting system uses Salloum and Habash (2013)’s DA-MSA MT system followed by an Arabic-English SMT system which is trained on both corpora augmented with the DA-English where the DA side is preprocessed with the same DA-MSA MT system then tokenized with MADA-ARZ.
Machine Translation Experiments
Test sets are similarly preprocessed before decoding with the SMT system .
SMT system is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Tamura, Akihiro and Watanabe, Taro and Sumita, Eiichiro
Training
In the translation tasks, we used the Moses phrase-based SMT systems (Koehn et al., 2007).
Training
In addition, for a detailed comparison, we evaluated the SMT system where the IBM Model 4 was trained from all the training data (I BM 4a“).
Training
Consequently, the SMT system using RN Nu+c trained from a small part of training data can achieve comparable performance to that using I BM 4 trained from all training data, which is shown in Table 3.
SMT system is mentioned in 3 sentences in this paper.
Topics mentioned in this paper: