Index of papers in Proc. ACL that mention
  • state of the art
Bhat, Suma and Sproat, Richard
Abstract
Finally, we compare our proposal with the state of the art estimators (both parametric and nonparametric) on large standard corpora; apart from showing the favorable performance of our estimator, we also see that the classical Good-Turing estimator consistently underestimates the vocabulary size.
Conclusion
We then compared the performance of the proposed estimator with that of the state of the art estimators on large corpora.
Experiments
In this study we consider state of the art parametric estimators, as surveyed by (Baroni and Evert, 2005).
Introduction
While compared with other vocabulary size estimates, we see that our estimator performs at least as well as some of the state of the art estimators.
Previous Work
A good survey of the state of the art is available in (Gandolfi and Sastri, 2004).
Results and Discussion
0 From the Figure 1, we see that our estimator compares quite favorably with the best of the state of the art estimators.
Results and Discussion
The best of the state of the art estimator is a parametric one (ZM), while ours is a nonparametric estimator.
Results and Discussion
Further, it compares very favorably to the state of the art estimators (both parametric and nonparametric).
state of the art is mentioned in 10 sentences in this paper.
Topics mentioned in this paper:
Hermann, Karl Moritz and Das, Dipanjan and Weston, Jason and Ganchev, Kuzman
Conclusion
We have presented a simple model that outperforms the prior state of the art on FrameNet-style frame-semantic parsing, and performs at par with one of the previous-best single-parser systems on PropBank SRL.
Discussion
For FrameNet, the WSABIE EMBEDDING model we propose strongly outperforms the baselines on all metrics, and sets a new state of the art .
Discussion
In comparison to prior work on FrameNet, even our baseline models outperform the previous state of the art .
Experiments
This would be a standard NLP approach for the frame identification problem, but is surprisingly competitive with the state of the art .
Experiments
(2014) describe the state of the art
Experiments
While comparing with prior state of the art on the same corpus, we noted that Das et al.
Introduction
First, we show that for frame identification on the FrameNet corpus (Baker et al., 1998; Fillmore et al., 2003), we outperform the prior state of the art (Das et al., 2014).
Introduction
Second, we present results on PropBank-style semantic role labeling (Palmer et al., 2005; Meyers et al., 2004; Marquez et al., 2008), that approach strong baselines, and are on par with prior state of the art (Punyakanok et al., 2008).
Overview
(2010) improved performance, and later set the current state of the art on this task (Das et al., 2014).
state of the art is mentioned in 10 sentences in this paper.
Topics mentioned in this paper:
Hermann, Karl Moritz and Blunsom, Phil
Abstract
We evaluate these models on two cross-lingual document classification tasks, outperforming the prior state of the art .
Conclusion
Coupled with very simple composition functions, vectors learned with this method outperform the state of the art on the task of cross-lingual document classification.
Experiments
Our models outperform the prior state of the art , with the BI models performing slightly better than the ADD models.
Experiments
We compare our embeddings with the SENNA embeddings, which achieve state of the art performance on a number of tasks (Collobert et al., 2011).
Introduction
First, we show that for cross-lingual document classification on the Reuters RCVIRCV2 corpora (Lewis et al., 2004), we outperform the prior state of the art (Klementiev et al., 2012).
Related Work
They have received a lot of attention in recent years (Collobert and Weston, 2008; Mnih and Hinton, 2009; Mikolov et al., 2010, inter alia) and have achieved state of the art performance in language modelling.
state of the art is mentioned in 6 sentences in this paper.
Topics mentioned in this paper:
Hasan, Kazi Saidul and Ng, Vincent
Abstract
We present a survey of the state of the art in automatic keyphrase extraction, examining the major sources of errors made by existing systems and discussing the challenges ahead.
Conclusion and Future Directions
We have presented a survey of the state of the art in automatic keyphrase extraction.
Corpora
n: A Survey of the State of the Art
Evaluation
4.2 The State of the Art
Introduction
Our goal in this paper is to survey the state of the art in keyphrase extraction, examining the major sources of errors made by existing systems and discussing the challenges ahead.
state of the art is mentioned in 5 sentences in this paper.
Topics mentioned in this paper:
Baroni, Marco and Dinu, Georgiana and Kruszewski, Germán
Evaluation materials
State of the art performance on this set has been reported by Hassan and Mi-halcea (2011) using a technique that exploits the Wikipedia linking structure and word sense disambiguation techniques.
Evaluation materials
The current state of the art is reached by Halawi et al.
Evaluation materials
Current state of the art was reached by the window-based count model of Baroni and Lenci (2010).
Results
Indeed, the predictive models achieve an impressive overall performance, beating the current state of the art in several cases, and approaching it in many more.
state of the art is mentioned in 4 sentences in this paper.
Topics mentioned in this paper:
Escalante, Hugo Jair and Solorio, Thamar and Montes-y-Gomez, Manuel
Abstract
We report experimental results in AA data sets that confirm that LHs over character n-grams are more helpful for AA than the usual global histograms, yielding results far superior to state of the art approaches.
Conclusions
Our experimental results showed that LHs outperform traditional bag-of-words formulations and state of the art techniques in balanced, imbalanced, and reduced data sets.
Experiments and Results
The BOLH formulation outperforms state of the art approaches by a considerable margin that ranges from 10% to 27%.
Introduction
c We report experimental results that are superior to state of the art approaches (Plakias and Stamatatos, 2008b; Plakias and Stamatatos, 2008a), with improvements ranging from 2% — 6% in balanced data sets and from 14% — 30% in imbalanced data sets.
state of the art is mentioned in 4 sentences in this paper.
Topics mentioned in this paper:
Pantel, Patrick and Lin, Thomas and Gamon, Michael
Abstract
We show that modeling user intent significantly improves entity type resolution for head queries over the state of the art , on several metrics, without degradation in tail query performance.
Conclusion
Jointly modeling the interplay between the underlying user intents and entity types in web search queries shows significant improvements over the current state of the art on the task of resolving entity types in head queries.
Introduction
We show that jointly modeling user intent and entity type significantly outperforms the current state of the art on the task of entity type resolution in queries.
Introduction
0 We empirically show that our models outperform the state of the art and that modeling latent intent contributes significantly to these results.
state of the art is mentioned in 4 sentences in this paper.
Topics mentioned in this paper:
Wolfe, Travis and Van Durme, Benjamin and Dredze, Mark and Andrews, Nicholas and Beller, Charley and Callison-Burch, Chris and DeYoung, Jay and Snyder, Justin and Weese, Jonathan and Xu, Tan and Yao, Xuchen
Abstract
PARMA achieves state of the art results on an existing and a new dataset.
Conclusion
PARMA achieves state of the art performance on three datasets for predicate argument alignment.
Introduction
We demonstrate the effectiveness of this approach by achieving state of the art performance on the data of Roth and Frank despite having little relevant training data.
Results
On RF, compared to Roth and Frank, the best published method for this task, we also improve, making PARMA the state of the art system for this task.
state of the art is mentioned in 4 sentences in this paper.
Topics mentioned in this paper:
Plank, Barbara and Moschitti, Alessandro
Computational Structures for RE
We will use this gold information also in Section 6.1 to show that our system aligns well to the state of the art on the ACE 2004 benchmark.
Experimental Setup
To compare our model against the state of the art we use the ACE 2004 data.
Introduction
We first show that our system aligns well with the state of the art on the ACE 2004 benchmark.
Results
Table 3 shows that our system (bottom) aligns well with the state of the art .
state of the art is mentioned in 4 sentences in this paper.
Topics mentioned in this paper:
Rastrow, Ariya and Dredze, Mark and Khudanpur, Sanjeev
Experiments
Our ASR system is based on the 2007 IBM Speech transcription system for the GALE Distillation Go/No-go Evaluation (Chen et al., 2006) with state of the art discriminative acoustic models.
Experiments
(2010), a state of the art broadcast news (BN) parser, with phrase structures converted to labeled dependencies by the Stanford converter.
Introduction
Discriminative models, which directly distinguish correct from incorrect hypothesis, are particularly attractive because they allow the inclusion of arbitrary features (Kuo et al., 2002; Roark et al., 2007; Collins et al., 2005); these models with syntactic information have obtained state of the art results.
state of the art is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Paperno, Denis and Pham, Nghia The and Baroni, Marco
Evaluation
Table 5 summarizes the performance of our models on the chosen tasks, and compares it to the state of the art reported in previous work, as well as to various strong baselines.
Evaluation
For anvanl, plf is just below the state of the art , which is based on disambiguating the verb vector in context (Kartsaklis and Sadrzadeh, 2013), and If outperforms the baseline, which consists in using the verb vector only as a proxy to sentence similarity.5 On anvan2, plf outperforms the best model
Evaluation
5 We report state of the art from Kartsaklis and Sadrzadeh
state of the art is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Narayan, Shashi and Gardent, Claire
Abstract
When compared against current state of the art methods, our model yields significantly simpler output that is both grammatical and meaning preserving.
Experiments
To evaluate performance, we compare our approach with three other state of the art systems using the test set provided by Zhu et al.
Introduction
When compared against current state of the art methods (Zhu et al., 2010; Woodsend and Lapata, 2011; Wubben et al., 2012), our model yields significantly simpler output that is both grammatical and meaning preserving.
state of the art is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Hermann, Karl Moritz and Blunsom, Phil
Experiments
— Overall, our models compare favourably with the state of the art .
Experiments
With an additional, unsupervised training step we achieved results beyond the current state of the art on this task, too.
Introduction
We show that this combination of state of the art machine learning and an advanced linguistic formalism translates into concise models with competitive performance on a variety of tasks.
state of the art is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Biran, Or and McKeown, Kathleen
Abstract
In addition, we present results for a full system using additional features which achieves close to state of the art performance without resorting to gold syntactic parses or to context outside the relation.
Conclusion
(2009), who used mostly similar features, for comparison and temporal and is competitive with the most recent state of the art systems for contingency and expansion without using any syntactic or context features.
Introduction
In addition, we present a system which combines these word pairs with additional features to achieve near state of the art performance without the use of syntactic parse features and of context outside the arguments of the relation.
state of the art is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Ganchev, Kuzman and Graça, João V. and Taskar, Ben
Introduction
Section 5 explores how the new alignments lead to consistent and significant improvement in a state of the art phrase base machine translation by using posterior decoding rather than Viterbi decoding.
Phrase-based machine translation
In particular we fix a state of the art machine translation system1 and measure its performance when we vary the supplied word alignments.
Word alignment results
These values are competitive with other state of the art systems (Liang et al., 2006).
state of the art is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Clifton, Ann and Sarkar, Anoop
Conclusion and Future Work
Using our proposed approach we obtain better scores than the state of the art on the English-Finnish translation task (Luong et al., 2010): from 14.82% BLEU to 15.09%, while using a
Translation and Morphology
Both of these approaches beat the state of the art on the English-Finnish translation task.
Translation and Morphology
Our proposed approaches are significantly better than the state of the art , achieving the highest reported BLEU scores on the English-Finnish Europarl version 3 dataset.
state of the art is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Ritter, Alan and Mausam and Etzioni, Oren
Experiments
We use this experiment to compare the various topic models as well as the best model with the known state of the art approaches to selectional preferences.
Experiments
We first compare the three LDA-based approaches to each other and two state of the art similarity based systems (Erk, 2007) (using mutual information and J accard similarity respectively).
Introduction
Our experiments demonstrate that LDA-SP significantly outperforms state of the art approaches obtaining an 85% increase in recall at precision 0.9 on the standard pseudo-disambiguation task.
state of the art is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Li, Zhifei and Eisner, Jason and Khudanpur, Sanjeev
Abstract
Experiments show that our approach improves the state of the art .
Conclusions and Future Work
Our empirical results improve the state of the art .
Introduction
Experiments show that our approach improves the state of the art .
state of the art is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Wang, Qin Iris and Schuurmans, Dale and Lin, Dekang
Introduction
Supervised learning algorithms still represent the state of the art approach for inferring dependency parsers from data (McDonald et al., 2005a; McDonald and Pereira, 2006; Wang et al., 2007).
Introduction
Consequently, most previous work that has attempted semi-supervised or unsupervised approaches to parsing have not produced results beyond the state of the art supervised results (Klein and Manning, 2002; Klein and Manning, 2004).
Introduction
In recent years, SVMs have demonstrated state of the art results in many supervised learning tasks.
state of the art is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Penn, Gerald and Zhu, Xiaodan
Abstract
We assess the current state of the art in speech summarization, by comparing a typical summarizer on two different domains: lecture data and the SWITCHBOARD corpus.
Problem definition and related literature
The purpose of this paper is not so much to introduce a new way of summarizing speech, as to critically reappraise how well the current state of the art really works.
Problem definition and related literature
These four results provide us with valuable insight into the current state of the art in speech summarization: it is not summarization, the aspiration to measure the relative merits of knowledge sources has masked the prominence of some very simple baselines, and the Zechner & Waibel pipe-ASR-output—into-text-summarizer model is still very competitive — what seems to matter more than having access to the raw spoken data is simply knowing that it is spoken data, so that the most relevant, still textually available features can be used.
state of the art is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Nivre, Joakim and McDonald, Ryan
Abstract
By letting one model generate features for the other, we consistently improve accuracy for both models, resulting in a significant improvement of the state of the art when evaluated on data sets from the CoNLL-X shared task.
Conclusion
Our experimental results show that both models consistently improve their accuracy when given access to features generated by the other model, which leads to a significant advancement of the state of the art in data-driven dependency parsing.
Experiments
Finally, given that the two base models had the previously best performance for these data sets, the guided models achieve a substantial improvement of the state of the art .
state of the art is mentioned in 3 sentences in this paper.
Topics mentioned in this paper: