Index of papers in Proc. ACL 2011 that mention
  • semi-supervised
Krishnamurthy, Jayant and Mitchell, Tom
Abstract
Synonym detection exploits redundant information to train several domain-specific synonym classifiers in a semi-supervised fashion.
Background: Never-Ending Language Learner
NELL is an information extraction system that has been running 24x7 for over a year, using coupled semi-supervised learning to populate an ontology from unstructured text found on the web.
ConceptResolver
After mapping each noun phrase to one or more senses (each with a distinct category type), ConceptResolver performs semi-supervised clustering to find synonymous senses.
ConceptResolver
For each category, ConceptResolver trains a semi-supervised synonym classifier then uses its predictions to cluster word senses.
Introduction
Train a semi-supervised classifier to predict synonymy.
Introduction
used to train a semi-supervised classifier.
Prior Work
However, our evaluation shows that ConceptResolver has higher synonym resolution precision than Resolver, which we attribute to our semi-supervised approach and the known relation schema.
Prior Work
ConceptResolver’s approach lies between these two extremes: we label a small number of synonyms (10 pairs), then use semi-supervised training to learn a similarity function.
Prior Work
ConceptResolver uses a novel algorithm for semi-supervised clustering which is conceptually similar to other work in the area.
semi-supervised is mentioned in 12 sentences in this paper.
Topics mentioned in this paper:
LIU, Xiaohua and ZHANG, Shaodian and WEI, Furu and ZHOU, Ming
Abstract
We propose to combine a K-Nearest Neighbors (KNN) classifier with a linear Conditional Random Fields (CRF) model under a semi-supervised learning framework to tackle these challenges.
Abstract
The semi-supervised learning plus the gazetteers alleviate the lack of training data.
Abstract
Extensive experiments show the advantages of our method over the baselines as well as the effectiveness of KNN and semi-supervised learning.
Introduction
(2010), which introduces a high-level rule language, called NERL, to build the general and domain specific NER systems; and 2) semi-supervised learning, which aims to use the abundant unlabeled data to compensate for the lack of annotated data.
Introduction
Indeed, it is the combination of KNN and CRF under a semi-supervised learning framework that differentiates ours from the existing.
Introduction
It is also demonstrated that integrating KNN classified results into the CRF model and semi-supervised learning considerably boost the performance.
Related Work
Related work can be roughly divided into three categories: NER on tweets, NER on non-tweets (e.g., news, biological medicine, and clinical notes), and semi-supervised learning for NER.
Related Work
To achieve this, a KNN classifier with a CRF model is combined to leverage cross tweets information, and the semi-supervised learning is adopted to leverage unlabeled tweets.
Related Work
2.3 Semi-supervised Learning for NER
semi-supervised is mentioned in 24 sentences in this paper.
Topics mentioned in this paper:
Titov, Ivan
Abstract
We consider a semi-supervised setting for domain adaptation where only unlabeled data is available for the target domain.
Constraints on Inter-Domain Variability
As we discussed in the introduction, our goal is to provide a method for domain adaptation based on semi-supervised learning of models with distributed representations.
Constraints on Inter-Domain Variability
In this section, we first discuss the shortcomings of domain adaptation with the above-described semi-supervised approach and motivate constraints on inter-domain variability of
Empirical Evaluation
For every pair, the semi-supervised methods use labeled data from the source domain and unlabeled data from both domains.
Empirical Evaluation
All the methods, supervised and semi-supervised , are based on the model described in Section 2.
Empirical Evaluation
This does not seem to have an adverse effect on the accuracy but makes learning very efficient: the average training time for the semi-supervised methods was about 20 minutes on a standard PC.
Introduction
The danger of this semi-supervised approach in the domain-adaptation setting is that some of the latent variables will correspond to clusters of features specific only to the source domain, and consequently, the classifier relying on this latent variable will be badly affected when tested on the target domain.
Related Work
Various semi-supervised techniques for domain-adaptation have also been considered, one example being self-training (McClosky et al., 2006).
Related Work
Semi-supervised leam-ing with distributed representations and its application to domain adaptation has previously been considered in (Huang and Yates, 2009), but no attempt has been made to address problems specific to the domain-adaptation setting.
semi-supervised is mentioned in 13 sentences in this paper.
Topics mentioned in this paper:
Jiang, Qixia and Sun, Maosong
Abstract
This paper proposes a novel (semi-)supervised hashing method named Semi-Supervised SimHash (83H) for high—dimensional data similarity search.
Background and Related Works
2.3 Semi-Supervised Hashing
Background and Related Works
Semi-Supervised Hashing (SSH) (Wang et al., 2010a) is recently proposed to incorporate prior knowledge for better hashing.
Introduction
vated by this, some supervised methods are proposed to derive effective hash functions from prior knowledge, i.e., Spectral Hashing (Weiss et al., 2009) and Semi-Supervised Hashing (SSH) (Wang et al., 2010a).
Introduction
This paper proposes a novel (semi-)supervised hashing method, Semi-Supervised SimHash (S3H), for high-dimensional data similarity search.
Introduction
In Section 3, we describe our proposed Semi-Supervised SimHash (S3H).
Semi-Supervised SimHash
In this section, we present our hashing method, named Semi-Supervised SimHash (83H).
The direction is determined by concatenating w L times.
We have proposed a novel supervised hashing method named Semi-Supervised Simhash (83H) for high-dimensional data similarity search.
semi-supervised is mentioned in 8 sentences in this paper.
Topics mentioned in this paper:
Sajjad, Hassan and Fraser, Alexander and Schmid, Helmut
Abstract
We conduct experiments on data sets from the NEWS 2010 shared task on transliteration mining and achieve an F-measure of up to 92%, outperforming most of the semi-supervised systems that were submitted.
Conclusion
We evaluated it against the semi-supervised systems of NEWS10 and achieved high F-measure and performed better than most of the semi-supervised systems.
Experiments
On the WIL data sets, we compare our fully unsupervised system with the semi-supervised systems presented at the NEWSlO (Kumaran et al., 2010).
Experiments
For English/Arabic, English/Hindi and English/Tamil, our system is better than most of the semi-supervised systems presented at the NEWS 2010 shared task for transliteration mining.
Introduction
We compare our unsupervised transliteration mining method with the semi-supervised systems presented at the NEWS 2010 shared task on transliteration mining (Kumaran et al., 2010) using four language pairs.
Introduction
These systems used a manually labelled set of data for initial supervised training, which means that they are semi-supervised systems.
Introduction
We achieve an F-measure of up to 92% outperforming most of the semi-supervised systems.
Previous Research
Our unsupervised method seems robust as its performance is similar to or better than many of the semi-supervised systems on three language pairs.
semi-supervised is mentioned in 8 sentences in this paper.
Topics mentioned in this paper:
Sun, Ang and Grishman, Ralph and Sekine, Satoshi
Abstract
We present a simple semi-supervised relation extraction system with large-scale word clustering.
Abstract
When training on different sizes of data, our semi-supervised approach consistently outperformed a state-of-the-art supervised baseline system.
Cluster Feature Selection
The cluster based semi-supervised system works by adding an additional layer of lexical features that incorporate word clusters as shown in column 4 of Table 4.
Conclusion and Future Work
We have described a semi-supervised relation extraction system with large-scale word clustering.
Experiments
For the semi-supervised system, 70 percent of the rest of the documents were randomly selected as training data and 30 percent as development data.
Experiments
For the semi-supervised system, each test fold was the same one used in the baseline and the other 4 folds were further split into a training set and a development set in a ratio of 7:3 for selecting clusters.
semi-supervised is mentioned in 6 sentences in this paper.
Topics mentioned in this paper:
Chen, Harr and Benson, Edward and Naseem, Tahira and Barzilay, Regina
Related Work
(2007) propose an objective function for semi-supervised extraction that balances likelihood of labeled instances and constraint violation on unlabeled instances.
Results
Comparison against Supervised CRF Our final set of experiments compares a semi-supervised version of our model against a conditional random field (CRF) model.
Results
For finance, it takes at least 10 annotated documents (corresponding to roughly 130 annotated relation instances) for the CRF to match the semi-supervised model’s performance.
semi-supervised is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Kobdani, Hamidreza and Schuetze, Hinrich and Schiehlen, Michael and Kamp, Hans
Related Work
There are three main approaches to CoRe: supervised, semi-supervised (or weakly supervised) and unsupervised.
Related Work
We use the term semi-supervised for approaches that use some amount of human-labeled coreference pairs.
Related Work
(2002) used co-training for coreference resolution, a semi-supervised method.
semi-supervised is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Lu, Bin and Tan, Chenhao and Cardie, Claire and K. Tsou, Benjamin
Related Work
Semi-supervised Learning.
Related Work
Another line of related work is semi-supervised learning, which combines labeled and unlabeled data to improve the performance of the task of interest (Zhu and Goldberg, 2009).
Related Work
Among the popular semi-supervised methods (e. g. EM on Nai've Bayes (Nigam et al., 2000), co-training (Blum and Mitchell, 1998), transductive SVMs (Joachims, 1999b), and co-regularization (Sindhwani et al., 2005; Amini et al., 2010)), our approach employs the EM algorithm, extending it to the bilingual case based on maximum entropy.
semi-supervised is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Zhou, Guangyou and Zhao, Jun and Liu, Kang and Cai, Li
Experiments
Type D, C and S denote discriminative, combined and semi-supervised systems, respectively.
Experiments
We also compare our method with the semi-supervised approaches, the semi-supervised approaches achieved very high accuracies by leveraging on large unlabeled data directly into the systems for joint learning and decoding, while in our method, we only explore the N-gram features to further improve supervised dependency parsing performance.
Introduction
(2008) proposed a semi-supervised dependency parsing by introducing lexical intermediaries at a coarser level than words themselves via a cluster method.
semi-supervised is mentioned in 3 sentences in this paper.
Topics mentioned in this paper: