Index of papers in Proc. ACL 2013 that mention
  • manually annotated
Lan, Man and Xu, Yu and Niu, Zhengyu
Experiments and Results
This is contrary to our original expectation that exp data which has been manually annotated for discourse connective disambiguation should outperform BLLIP which contains a lot of noise.
Experiments and Results
This finding indicates that under the multitask learning, it may not be worthy of using manually annotated corpus to generate auxiliary data.
Implementation Details of Multitask Learning Method
This is because the former data is manually annotated whether a word serves as discourse connective or not, while the latter does not manually disambiguate two types of ambiguity, i.e., whether a word serves as discourse connective or not, and the type of discourse relation if it is a discourse connective.
Introduction
To overcome the shortage of manually annotated training data, (Marcu and Echihabi, 2002) proposed a pattern-based approach to automatically generate training data from raw corpora.
Introduction
Later, with the release of manually annotated corpus, such as Penn Discourse Treebank 2.0 (PDTB) (Prasad et al., 2008), recent studies performed implicit discourse relation recognition on natural (i.e., genuine) implicit discourse data (Pitler et al., 2009) (Lin et al., 2009) (Wang et al., 2010) with the use of linguistically informed features and machine learning algorithms.
Introduction
2002) and (Sporleder and Lascarides, 2008), and (2) manually annotated explicit data with the removal of explicit discourse connectives.
Multitask Learning for Discourse Relation Prediction
Among them, it is manually annotated as an Expansion relation for 2, 938 times.
manually annotated is mentioned in 7 sentences in this paper.
Topics mentioned in this paper:
Wang, Mengqiu and Che, Wanxiang and Manning, Christopher D.
Bilingual NER Results
If a model only gives good performance with well-tuned hyper-parameters, then we must have manually annotated data for tuning, which would significantly reduce the applicability and portability of this method to other language pairs and tasks.
Introduction
Our method does not require any manual annotation of word alignments or named entities over the bilingual training data.
Joint Alignment and NER Decoding
On the other hand, the CRF—based NER models are trained on manually annotated data, and admit richer sequence and lexical features.
Related Work
proach is that training such a feature-rich model requires manually annotated bilingual NER data, which can be prohibitively expensive to generate.
Related Work
The model demonstrates performance improvements in both parsing and alignment, but shares the common limitations of other supervised work in that it requires manually annotated bilingual joint parsing and word alignment data.
manually annotated is mentioned in 5 sentences in this paper.
Topics mentioned in this paper:
Kozhevnikov, Mikhail and Titov, Ivan
Conclusion
It allows one to quickly construct an SRL model for a new language without manual annotation or language-specific heuristics, provided an accurate model is available for one of the related languages along with a certain amount of parallel data for the two languages.
Evaluation
For English-French, the English CoNLL-ST dataset was used as a source and the model was evaluated on the manually annotated dataset from van der Plas et al.
Evaluation
Note that in case of French we evaluate against the output of a supervised system, since manual annotation is not available for this dataset.
Setup
Several approaches have been proposed to obtain an SRL model for a new language with little or no manual annotation .
manually annotated is mentioned in 4 sentences in this paper.
Topics mentioned in this paper:
Laparra, Egoitz and Rigau, German
Conclusions and Future Work
That is, it can be applied where there is no available manual annotations to train.
Evaluation
Instead, the supervised approach would need a large amount of manual annotations for every predicate to be processed.
Introduction
However, current automatic systems require large amounts of manually annotated training data for each predicate.
Introduction
The effort required for this manual annotation explains the absence of generally applicable tools.
manually annotated is mentioned in 4 sentences in this paper.
Topics mentioned in this paper:
Zhang, Meishan and Zhang, Yue and Che, Wanxiang and Liu, Ting
Introduction
Manually annotated corpora, such as the Chinese Treebank (CTB) (Xue et al., 2005), usually have words as the basic syntactic elements
Introduction
We manually annotate the structures of 37,382 words, which cover the entire CTB5.
Introduction
Luo (2003) exploited this advantage by adding flat word structures without manually annotation to CTB trees, and building a generative character-based parser.
Related Work
In addition, instead of using flat structures, we manually annotate hierarchal tree structures of Chinese words for converting word-based constituent trees into character-based constituent trees.
manually annotated is mentioned in 4 sentences in this paper.
Topics mentioned in this paper:
Berg-Kirkpatrick, Taylor and Durrett, Greg and Klein, Dan
Experiments
To investigate the effects of using an in domain language model, we created a corpus composed of the manual annotations of all the documents in the Old Bailey proceedings, excluding those used in our test set.
Results and Analysis
We manually annotated the ink blotches (shown in red), and made them unobserved in the model.
Results and Analysis
We performed error analysis on our development set by randomly choosing 100 word errors from the WER alignment and manually annotating them with relevant features.
manually annotated is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Fader, Anthony and Zettlemoyer, Luke and Etzioni, Oren
Abstract
Given a large, community-authored, question-paraphrase corpus, we demonstrate that it is possible to learn a semantic lexicon and linear ranking function without manually annotating questions.
Introduction
These algorithms require no manual annotation and can be applied to large, noisy databases of relational triples.
Overview of the Approach
The final result is a scalable learning algorithm that requires no manual annotation of questions.
manually annotated is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Hartmann, Silvana and Gurevych, Iryna
FrameNet — Wiktionary Alignment
teas) independently on a manually annotated gold standard.
Related Work
In Tonelli and Pighin (2009), they use these features to train an SVM-classifier to identify valid alignments and report an Fl-score of 0.66 on a manually annotated gold standard.
Resource FNWKde
We compare FNWKde to two German frame-semantic resources, the manually annotated SALSA corpus (Burchardt et al., 2006) and a resource from Pado and Lapata (2005), henceforth P&L05.
manually annotated is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
McDonald, Ryan and Nivre, Joakim and Quirmbach-Brundage, Yvonne and Goldberg, Yoav and Das, Dipanjan and Ganchev, Kuzman and Hall, Keith and Petrov, Slav and Zhang, Hao and Täckström, Oscar and Bedini, Claudia and Bertomeu Castelló, Núria and Lee, Jungmee
Towards A Universal Treebank
The first is traditional manual annotation , as previously used by Helmreich et al.
Towards A Universal Treebank
2.2 Manual Annotation
Towards A Universal Treebank
Such a reduction may ultimately be necessary also in the case of dependency relations, but since most of our data sets were created through manual annotation , we could afford to retain a fine-grained analysis, knowing that it is always possible to map from finer to coarser distinctions, but not vice versa.4
manually annotated is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Yu, Haonan and Siskind, Jeffrey Mark
Experiment
We manually annotated each video with several sentences that describe what occurs in that video.
Experiment
To evaluate our results, we manually annotated the correctness of each such pair.
Experiment
The scores are thresholded to decide hits, which together with the manual annotations , can generate TP, TN, FF, and FN counts.
manually annotated is mentioned in 3 sentences in this paper.
Topics mentioned in this paper: