Index of papers in Proc. ACL that mention
  • context information
Zhai, Feifei and Zhang, Jiajun and Zhou, Yu and Zong, Chengqing
Abstract
In this way, we incorporate rich context information of PAS for disambiguation.
Conclusion and Future Work
The two methods successfully incorporate the rich context information into the translation process.
Experiment
Specifically, after integrating the inside context information of PAS into transformation, we can see that system IC-PASTR significantly outperforms system PASTR by 0.71 BLEU points.
Experiment
Conversely, by considering the context information , the PASTR+MEPD system chooses a correct rule for translation:
Introduction
In this paper, we propose two novel methods to incorporate rich context information to handle PAS ambiguities.
Maximum Entropy PAS Disambiguation (MEPD) Model
In order to handle the role ambiguities, in this section, we concentrate on utilizing a maximum entropy model to incorporate the context information for PAS disambiguation.
PAS-based Translation Framework
The target-side-like PAS is selected only according to the language model and translation probabilities, without considering any context information of PAS.
Related Work
They combine rich context information to do disambiguation for words or phrases, and achieve improved translation performance.
Related Work
By incorporating the rich context information as features, they chose better rules for translation and yielded stable improvements on translation quality.
Related Work
They also combine the context information in the model.
context information is mentioned in 11 sentences in this paper.
Topics mentioned in this paper:
Sun, Le and Han, Xianpei
Introduction
The feature we used includes characteristics of relation instance, phrase properties and context information (See Section 3 for details).
Introduction
3.3 Context Information Feature
Introduction
The context information of a phrase node is critical for identifying the role and the importance of a subtree in the whole relation instance.
context information is mentioned in 10 sentences in this paper.
Topics mentioned in this paper:
Zhou, Guangyou and Liu, Fang and Liu, Yang and He, Shizhu and Zhao, Jun
Experiments
Table 6: Impact of the contextual information .
Experiments
3.5 Impact of the Contextual Information
Experiments
In this paper, we translate the English questions into other four languages using Google Translate (GTrans), which takes into account contextual information during translation.
Introduction
The basic idea is to capture the contextual information in modeling the translation of phrases as a whole, thus the word ambiguity problem is somewhat alleviated.
Introduction
tions: (1) Contextual information is exploited during the translation from one language to another.
Introduction
For example in Table 1, English words “interest” and “bank” that have multiple meanings under different contexts are correctly addressed by using the state-of-the-art translation tool — —Google Translate.4 Thus, word ambiguity based on contextual information is naturally involved when questions are translated.
Our Approach
Statistical machine translation (e.g., Google Translate) can utilize contextual information during the question translation, so it can solve the word ambiguity and word mismatch problems to some extent.
context information is mentioned in 10 sentences in this paper.
Topics mentioned in this paper:
Qazvinian, Vahed and Radev, Dragomir R.
Abstract
In this paper, we propose a general framework based on probabilistic inference to extract such context information from scientific papers.
Abstract
Our experiments show greater pyramid scores for surveys generated using such context information rather than citation sentences alone.
Conclusion
Our experiments on generating surveys for Question Answering and Dependency Parsing show how surveys generated using such context information along with citation sentences have higher quality than those built using citations alone.
Conclusion
Our future goal is to combine summarization and bibliometric techniques towards building automatic surveys that employ context information as an important part of the generated surveys.
Introduction
We refer to such implicit citations that contain information about a specific secondary source but do not explicitly cite it, as sentences with context information or context sentences for short.
Proposed Method
In this section we propose our methodology that enables us to identify the context information of a cited paper.
Proposed Method
To find the sentences from a paper that form the context information of a given cited paper, we build an MRF in which a hidden node :13,- and an observed node y,- correspond to each sentence.
context information is mentioned in 7 sentences in this paper.
Topics mentioned in this paper:
Setiawan, Hendra and Zhou, Bowen and Xiang, Bing and Shen, Libin
Abstract
Long distance reordering remains one of the greatest challenges in statistical machine translation research as the key contextual information may well be beyond the confine of translation units.
Conclusion
We presented a novel approach to address a kind of long-distance reordering that requires global cross-boundary contextual information .
Conclusion
Empirical results confirm our intuition that incorporating cross-boundaries contextual information improves translation quality.
Experiments
As shown, the empirical results confirm our intuition that SMT can greatly benefit from reordering model that incorporate cross-unit contextual information .
Introduction
Often, such reordering decisions require contexts that span across multiple translation units.1 Unfortunately, previous approaches fall short in capturing such cross-unit contextual information that could be
Introduction
In this paper, we argue that reordering modeling would greatly benefit from richer cross-boundary contextual information
Introduction
We introduce a reordering model that incorporates such contextual information , named the Two-Neighbor Orientation (TNO) model.
context information is mentioned in 7 sentences in this paper.
Topics mentioned in this paper:
Zhang, Dongdong and Li, Mu and Duan, Nan and Li, Chi-Ho and Zhou, Ming
Introduction
We also compared the performance of our model based on different contextual information , and show that both large-scale monolingual data and parallel bilingual data can be helpful to generate correct measure words.
Model Training and Application 3.1 Training
Then, the collocation between measure words and head words and their surrounding contextual information are extracted to train the measure word selection models.
Model Training and Application 3.1 Training
Then, contextual information within the windows in the source and the target sentence is extracted and fed to the measure word selection model.
Model Training and Application 3.1 Training
We do not integrate our measure word generation module into the SMT decoder since there is only little target contextual information available during SMT decoding.
Our Method
Based on contextual information contained in both input source sentence and SMT system’s output translation, a measure word candidate set M is constructed.
Our Method
After obtaining the measure word candidate set M, a measure word selection model is employed to select the best one from M. Given the contextual information C in both source window and target
context information is mentioned in 6 sentences in this paper.
Topics mentioned in this paper:
Sun, Xu and Gao, Jianfeng and Micol, Daniel and Quirk, Chris
A Phrase-Based Error Model
Rather than replacing single words in isolation, this model replaces sequences of words with sequences of words, thus incorporating contextual information .
A Phrase-Based Error Model
Notice that when we set L=1, the phrase-based error model is reduced to a word-based error model which assumes that words are transformed independently from C to Q, without taking into account any contextual information .
Introduction
Comparing to traditional error models that account for transformation probabilities between single characters (Kernighan et al., 1990) or sub-word strings (Brill and Moore, 2000), the phrase-based model is more powerful in that it captures some contextual information by retaining inter-term dependencies.
Introduction
We show that this information is crucial to detect the correction of a query term, because unlike in regular written text, any query word can be a valid search term and in many cases the only way for a speller system to make the judgment is to explore its usage according to the contextual information .
Related Work
Typically, a language model (source model) is used to capture contextual information, while an error model (channel model) is considered to be context free in that it does not take into account any contextual information in modeling word transformation probabilities.
Related Work
In this study we argue that it is beneficial to capture contextual information in the error model.
context information is mentioned in 6 sentences in this paper.
Topics mentioned in this paper:
Ding, Shilin and Cong, Gao and Lin, Chin-Yew and Zhu, Xiaoyan
Experiments
Table 6: Contextual Information for Answer Detection.
Experiments
Linear CRFs with contextual information perform better than those without context.
Experiments
The results clearly shows that contextual information greatly improves the performance of answer detection.
Introduction
As shown in the example, a forum question usually requires contextual information to provide background or constraints.
Introduction
Moreover, it sometimes needs contextual information to provide explicit link to its answers.
Introduction
We call contextual information the context of a question in this paper.
context information is mentioned in 6 sentences in this paper.
Topics mentioned in this paper:
Yogatama, Dani and Sim, Yanchuan and Smith, Noah A.
Experiments
We also analyze how much each of our four main extensions (shape features, context information , noise, and first-order column dependencies) to EEA contributes to overall performance by ablating each in turn (also shown in Fig.
Experiments
Our model can do better, since it makes use of context information and features, and it can put a person and an organization in one row even though they do not share common words.
Experiments
It shows that in some cases context information is not adequate, and a possible improvement might be obtained by providing more context to the model.
Learning and Inference
We further incorporate context information and a notion of noise.
Learning and Inference
It is important to be able to se context information to determine which row mention should go into.
context information is mentioned in 5 sentences in this paper.
Topics mentioned in this paper:
Celikyilmaz, Asli and Hakkani-Tur, Dilek and Tur, Gokhan and Sarikaya, Ruhi
Conclusions
Our results show that encoding priors on words and context information contributes significantly to the performance of semantic clustering.
Conclusions
Rather than using single turn utterances, we hope to utilize the context information , e.g., information from previous turns for improving the performance of the semantic tagging of the current turns.
Experiments
To include contextual information , we add binary features for all possible tags.
Experiments
The results indicate that incorporating context information with MTR is an effective option for identifying semantic ambiguity.
context information is mentioned in 4 sentences in this paper.
Topics mentioned in this paper:
van Gompel, Maarten and van den Bosch, Antal
Baselines
A context-insensitive yet informed baseline was constructed to assess the impact of L2 context information in translating Ll fragments.
Data preparation
Nevertheless, we hope to show that our automated way of test set generation is sufficient to test the feasibility of our core hypothesis that L1 fragments can be translated to L2 using L2 context information .
Introduction
The main research question in this research is how to disambiguate an L1 word or phrase to its L2 translation based on an L2 context, and whether such cross-lingual contextual approaches provide added value compared to baseline models that are not context informed or compared to standard language models.
System
If so, we are done quickly and need not rely on context information .
context information is mentioned in 4 sentences in this paper.
Topics mentioned in this paper:
Thater, Stefan and Fürstenau, Hagen and Pinkal, Manfred
Conclusion
Another direction for further study will be the generalization of our model to larger syntactic contexts, including more than only the direct neighbors in the dependency graph, ultimately incorporating context information from the whole sentence in a recursive fashion.
Experiments: Ranking Paraphrases
3Note that the context information is the same for both words.
Experiments: Ranking Paraphrases
The main difference between verbs on the one hand, and nouns, adjectives, and adverbs on the other hand, is that verbs typically come with a rich context—subject, object, and so on—while non-verbs often have either no dependents at all or only closed class dependents such as determiners which provide only limited contextual informations , if any at all.
The model
A more flexible approach than simple filtering, however, is to re-weight those dimensions with context information .
context information is mentioned in 4 sentences in this paper.
Topics mentioned in this paper:
Manshadi, Mehdi and Li, Xiao
Abstract
In order to take contextual information into account, a discriminative model is used on top of the parser to re—rank the n—best parse trees generated by the parser.
Discriminative re-ranking
When there is enough labeled data, then a discriminative model can be trained on the labeled data to learn contextual information and to further enhance the tagging performance.
Introduction
Contextual information often plays a big role in resolving tagging ambiguities and is one of the key benefits of discriminative models such as CRFs.
Summary
to take contextual information into account.
context information is mentioned in 4 sentences in this paper.
Topics mentioned in this paper:
Tang, Duyu and Wei, Furu and Yang, Nan and Zhou, Ming and Liu, Ting and Qin, Bing
Related Work
SSWE outperforms MVSA by exploiting more contextual information in the sentiment predictor function.
Related Work
Among three sentiment-specific word embeddings, SSWEu captures more context information and yields best performance.
Related Work
SSWE outperforms MVSA and ReEmb by exploiting more context information of words and sentiment information of sentences, respectively.
context information is mentioned in 4 sentences in this paper.
Topics mentioned in this paper:
Xiong, Deyi and Zhang, Min
Related Work
Rather than predicting word senses for ambiguous words, the reformulated WSD directly predicts target translations for source words with context information .
Related Work
Lexical selection Our work is also related to lexical selection in SMT where appropriate target lexical items for source words are selected by a statistical model with context information (Bangalore et al., 2007; Mauser et al., 2009).
Sense-Based Translation Model
The sense-based translation model estimates the probability that a source word 0 is translated into a target phrase 6 given contextual information , including word senses that are obtained using the HDP-based WSI as described in the last section.
WSI-Based Broad-Coverage Sense Tagger
A pseudo document is composed of either a bag of neighboring words of a word token, or the Part-to-Speech tags of neighboring words, or other contextual information elements.
context information is mentioned in 4 sentences in this paper.
Topics mentioned in this paper:
Cirik, Volkan
Algorithm
To make use of both word identity and context information of a given type, we use S-CODE co-occurrence modeling (Maron et al., 2010) as (Yatbaz et al., 2012) does.
Experiments
In that experiment, POS induction is done by using word identities and context information represented by substitute words.
Introduction
of a target word, we separate occurrences of the word into different groups depending on the context information represented by substitute vectors.
context information is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Wintrode, Jonathan and Khudanpur, Sanjeev
Abstract
We aim to improve spoken term detection performance by incorporating contextual information beyond traditional N-gram language models.
Introduction
We will show that by focusing on contextual information in the form of word repetition within documents, we obtain consistent improvement across five languages in the so called Base Phase of the IARPA BABEL program.
Motivation
Clearly topic or context information is relevant to a retrieval type task, but we need a stable, consistent framework in which to apply it.
context information is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Muller, Philippe and Fabre, Cécile and Adam, Clémentine
Abstract
We first set up a human annotation of semantic links with or without contextual information to show the importance of the textual context in evaluating the relevance of semantic similarity, and to assess the prevalence of actual semantic relations between word tokens.
Evaluation of lexical similarity in context
To verify that this methodology is useful, we did a preliminary annotation to contrast judgment on lexical pairs with or without this contextual information .
Introduction
We present the experiments we set up to automatically filter semantic relations in context, with various groups of features that take into account information from the corpus used to build the thesaurus and contextual information related to occurrences of semantic neighbours 3).
context information is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Liu, Yang
Introduction
As the stack of a state keeps changing during the decoding process, the context information needed to calculate dependency language model and maximum entropy model probabilities (e. g., root word, leftmost child, etc.)
Introduction
As a result, the chance of risk-free hypothesis recombination (Koehn et al., 2003) significantly decreases because complicated contextual information is much less likely to be identical.
Introduction
In the future, we plan to include more contextual information (e.g., the uncovered source phrases) in the maximum entropy model to resolve conflicts.
context information is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Zhao, Shiqi and Wang, Haifeng and Liu, Ting and Li, Sheng
Conclusion
In addition, we will try to make better use of the context information when replacing paraphrase patterns in context sentences.
Experiments
In Section 4.1, we have evaluated the precision of the paraphrase patterns without considering context information .
Experiments
The context information was also considered by our judges.
context information is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Szpektor, Idan and Dagan, Ido and Bar-Haim, Roy and Goldberger, Jacob
Conclusions
CP enriches the representation of textual objects with typical contextual information that constrains or disambiguates their meaning, and provides matching functions that compare the preferences of objects involved in the inference.
Contextual Preferences
Overall, such incorrect inferences may be avoided by considering contextual information for t, h and 7“ during their matching process.
Contextual Preferences
In this framework, the representation of an object 2, where 2 may be a text, a template or an entailment rule, is enriched with contextual information denoted cp(z).
context information is mentioned in 3 sentences in this paper.
Topics mentioned in this paper: