Index of papers in Proc. ACL that mention
  • regression model
Cao, Guihong and Robertson, Stephen and Nie, Jian-Yun
Abstract
The selection is made according to the appropriateness of the alteration to the query context (using a bigram language model), or according to its expected impact on the retrieval effectiveness (using a regression model ).
Introduction
In this paper, we will use a regression model to predict the impact on retrieval effectiveness.
Regression Model for Alteration Selection
This method develops a regression model from a set of training data, and it is capable of predicting the expected change in performance when the original query is augmented by this alteration.
Regression Model for Alteration Selection
5.1 Linear Regression Model
Regression Model for Alteration Selection
The goal of the regression model is to predict the performance change when a query term is augmented with an alteration.
regression model is mentioned in 20 sentences in this paper.
Topics mentioned in this paper:
Wang, William Yang and Hua, Zhenhao
Copula Models for Text Regression
Our proposed semiparametric copula regression model takes a different perspective.
Copula Models for Text Regression
Then we describe the proposed semiparametric Gaussian copula text regression model .
Copula Models for Text Regression
We formulate the copula regression model as follows.
Experiments
In the first experiment, we compare the proposed semiparametric Gaussian copula regression model to three baselines on three datasets with all features.
Experiments
On the post—2009 dataset, none of results from the linear and nonlinear SVM models can match up with the linear regression model , but our proposed copula model still improves over all baselines by a large margin.
Experiments
To understand the learning curve of our proposed copula regression model , we use the 25%, 50%, 75% subsets from the training data, and evaluate all four models.
regression model is mentioned in 13 sentences in this paper.
Topics mentioned in this paper:
Morin, Emmanuel and Hazem, Amir
Abstract
Moreover, we have introduced a regression model that boosts the observations of word co-occurrences used in the context-based projection method.
Bilingual Lexicon Extraction
We then present an extension of this approach based on regression models .
Bilingual Lexicon Extraction
First, while they experienced the linear regression model, we propose to contrast different regression models .
Bilingual Lexicon Extraction
As most regression models have already been described in great detail (Christensen, 1997; Agresti, 2007), the derivation of most models is only briefly introduced in this work.
Experiments and Results
Table 6: Results (MAP %) of the standard approach using different regression models on the balanced breast cancer and diabetes corpora
Experiments and Results
4.2.1 Regression Models Comparison
Experiments and Results
We contrast the simple linear regression model (Lin) with the second and the third order polynomial regressions (Poly2 and P0ly3) and the logistic regression model (Logit).
Introduction
To make them more reliable, our second contribution is to contrast different regression models in order to boost the observations of word co-occurrences.
regression model is mentioned in 13 sentences in this paper.
Topics mentioned in this paper:
Li, Chen and Qian, Xian and Liu, Yang
Abstract
For each bigram, a regression model is used to estimate its frequency in the reference summary.
Abstract
The regression model uses a variety of indicative features and is trained discriminatively to minimize the distance between the estimated and the ground truth bigram frequency in the reference summary.
Experiment and Analysis
We used the estimated value from the regression model ; the ICSI system just uses the bigram’s document frequency in the original text as weight.
Experiment and Analysis
# bigrams used in our regression model 2140.7 (i.e., in selected sentences)
Experiments
In our method, we first extract all the bigrams from the selected sentences and then estimate each bigram’s N We f using the regression model .
Experiments
When training our bigram regression model , we use each of the 4 reference summaries separately, i.e., the bigram frequency is obtained from one reference summary.
Introduction
To estimate the bigram frequency in the summary, we propose to use a supervised regression model that is discriminatively trained using a variety of features.
Proposed Method 2.1 Bigram Gain Maximization by ILP
2.2 Regression Model for Bigram Frequency Estimation
Proposed Method 2.1 Bigram Gain Maximization by ILP
We propose to use a regression model for this.
Proposed Method 2.1 Bigram Gain Maximization by ILP
To train this regression model using the given reference abstractive summaries, rather than trying to minimize the squared error as typically done, we propose a new objective function.
regression model is mentioned in 16 sentences in this paper.
Topics mentioned in this paper:
Celikyilmaz, Asli and Hakkani-Tur, Dilek
Abstract
In this paper, we formulate extractive summarization as a two step learning problem building a generative model for pattern discovery and a regression model for inference.
Abstract
Then, using these scores, we train a regression model based on the lexical and structural characteristics of the sentences, and use the model to score sentences of new documents to form a summary.
Background and Motivation
Our approach differs from the early work, in that, we combine a generative hierarchical model and regression model to score sentences in new documents, eliminating the need for building a generative model for new document clusters.
Experiments and Discussions
Later, we build a regression model with the same features as our HybHSum to create a summary.
Experiments and Discussions
We keep the parameters and the features of the regression model of hierarchical HybHSum intact for consistency.
Introduction
In this paper, we present a novel approach that formulates MDS as a prediction problem based on a two-step hybrid model: a generative model for hierarchical topic discovery and a regression model for inference.
Introduction
We construct a hybrid learning algorithm by extracting salient features to characterize summary sentences, and implement a regression model for inference (Fig.3).
Introduction
Our aim is to find features that can best represent summary sentences as described in § 5, — implementation of a feasible inference method based on a regression model to enable scoring of sentences in test document clusters without retraining, (which has not been investigated in generative summarization models) described in § 5.2.
Regression Model
We build a regression model using sentence scores as output and selected salient features as input variables described below:
Regression Model
(4), we train a regression model .
Regression Model
Once the SVR model is trained, we use it to predict the scores of ntest number of sentences in test (unseen) document clusters, Otest = {01, “plowstl Our HybHSum captures the sentence characteristics with a regression model using sentences in different document clusters.
regression model is mentioned in 12 sentences in this paper.
Topics mentioned in this paper:
Kozareva, Zornitsa
Conclusion
Through experiments carried out on the developed datasets, we showed that the proposed polarity classification and valence regression models significantly improve baselines (from 11.90% to 39.69% depending on the language) and work well for all four languages.
Task B: Valence Prediction
5.2 Regression Model
Task B: Valence Prediction
Full details of the regression model and its implementation are beyond the scope of this paper; for more details see (Scho'lkopf and Smola, 2001; Smola et al., 2003).
Task B: Valence Prediction
Evaluation Measures: To evaluate the quality of the valence prediction model, we compare the actual valence score of the metaphor given by human annotators denoted with 3/ against those valence scores predicted by the regression model denoted with ac.
regression model is mentioned in 10 sentences in this paper.
Topics mentioned in this paper:
Pado, Sebastian and Galley, Michel and Jurafsky, Dan and Manning, Christopher D.
Conclusion and Outlook
We have used an off-the-shelf RTE system to compute these features, and demonstrated that a regression model over these features can outperform an ensemble of traditional MT metrics in two experiments on different datasets.
EXpt. 1: Predicting Absolute Scores
We optimize the weights of our regression models on two languages and then predict the human scores on the third language.
Experimental Evaluation
They are small regression models as described in Section 2 over component scores of four widely used MT metrics.
Experimental Evaluation
2The regression models can simulate the behaviour of each component by setting the weights appropriately, but are strictly more powerful.
Experimental Evaluation
We therefore verified that the three nontrivial “baseline” regression models indeed confer a benefit over the default component combination scores: BLEU—1 (which outperformed BLEU-4 in the MetricsMATR 2008 evaluation), NIST-4, and TER (with all costs set to 1).
Expt. 2: Predicting Pairwise Preferences
6We also experimented with a logistic regression model that predicts binary preferences directly.
Textual Entailment vs. MT Evaluation
This allows us to use an off-the-shelf RTE system to obtain features, and to combine them using a regression model as described in Section 2.
regression model is mentioned in 9 sentences in this paper.
Topics mentioned in this paper:
Chang, Kai-min K. and Cherkassky, Vladimir L. and Mitchell, Tom M. and Just, Marcel Adam
Brain Imaging Experiments on Adj ec-tive-Noun Comprehension
In this analysis, we train a regression model to fit the activation profile for the 12 phrase stimuli.
Brain Imaging Experiments on Adj ec-tive-Noun Comprehension
The regression model examined to what extent the semantic feature vectors (explanatory variables) can account for the variation in neural activity (response variable) across the 12 stimuli.
Brain Imaging Experiments on Adj ec-tive-Noun Comprehension
All explanatory variables were entered into the regression model simultaneously.
regression model is mentioned in 8 sentences in this paper.
Topics mentioned in this paper:
Ai, Hua and Litman, Diane J.
Abstract
When building prediction models of human judgments using previously proposed automatic measures, we find that we cannot reliably predict human ratings using a regression model , but we can predict human rankings by a ranking model.
Conclusion and Future Work
We would also want to include more automatic measures that may be available in the richer corpora to improve the ranking and the regression models .
Introduction
Similarly, when we use previously proposed automatic measures to predict human judgments, we cannot reliably predict human ratings using a regression model , but we can consistently mimic human judges’ rankings using a ranking model.
Related Work
Some studies (e.g., (Walker et al., 1997)) build regression models to predict user satisfaction scores from the system log as well as the user survey.
Related Work
In this study, we build both a regression model and a ranking model to evaluate user simulation.
Validating Automatic Measures
6.1 The Regression Model
regression model is mentioned in 6 sentences in this paper.
Topics mentioned in this paper:
Huang, Fei and Xu, Jian-Ming and Ittycheriah, Abraham and Roukos, Salim
Adaptive MT Quality Estimation
The above QE regression model is trained on a portion of the sentences from the input document, and evaluated on the remaining sentences from the same document.
Adaptive MT Quality Estimation
Therefore it is necessary to build a QB regression model that’s robust to different document-specific translation models.
Adaptive MT Quality Estimation
We compute the TER of Tq using Rq as the reference, and train a QB regression model with the 26 features proposed in section 4.1.
Related Work
Soricut and Echihabi (2010b) proposed various regression models to predict the expected BLEU score of a given sentence translation hypothesis.
Static MT Quality Estimation
We experiment with several classifiers: linear regression model, decision tree based regression model and SVM model.
Static MT Quality Estimation
Our experiments show that the decision tree-based regression model obtains the highest correlation coefficients (0.53) and lowest RMSE (0.23) in both the training and test sets.
regression model is mentioned in 6 sentences in this paper.
Topics mentioned in this paper:
Cai, Qingqing and Yates, Alexander
Experiments
Figure 3 shows a Precision-Recall (PR) curve for MATCHER and three baselines: a “Frequency” model that ranks candidate matches for TD by their frequency during the candidate identification step; a “Pattern” model that uses MATCHER’s linear regression model for ranking, but is restricted to only the pattern-based features; and an “Extractions” model that similarly restricts the ranking model to ReVerb features.
Experiments
All regression models for learning alignments outperform the Frequency ranking by a wide margin.
Extending a Semantic Parser Using a Schema Alignment
For W, we use a linear regression model whose features are the score from MATCHER, the probabilities from the Syn and Sem NBC models, and the average weight of all lexical entries in UBL with matching syntax and semantics.
Textual Schema Matching
3.5 Regression models for scoring candidates
Textual Schema Matching
MATCHER uses a regression model to combine these various statistics into a score for (77,719).
Textual Schema Matching
The regression model is a linear regression with least-squares parameter estimation; we experimented with support vector regression models with nonlinear kernels, with no significant improvements in accuracy.
regression model is mentioned in 6 sentences in this paper.
Topics mentioned in this paper:
Bamman, David and O'Connor, Brendan and Smith, Noah A.
Evaluation
The difference between the persona regression model and the Dirichlet persona model here is not
Evaluation
by the persona regression model , along with links fn
Evaluation
In practice, we find that while the Dirichlet model distinguishes between character personas in different movies, the persona regression model helps distinguish between different personas within the same movie.
Exploratory Data Analysis
To illustrate this, we present results from the persona regression model learned above, with 50 latent lexical classes and 100 latent personas.
Models
Distribution over topics for persona p in role 7“ 0d Movie d’s distribution over personas pe Character e’s persona (integer, p E {1..P}) j A specific (7“, w) tuple in the data Zj Word topic for tuple j 1113' Word for tuple j oz Concentration parameter for Dirichlet model 6 Feature weights for regression model [1,02 Gaussian mean and variance (for regularizing B) md Movie features (from movie metadata) me Entity features (from movie actor metadata) VT, 7 Dirichlet concentration parameters
Models
Figure 2: Above: Dirichlet persona model (left) and persona regression model (right).
regression model is mentioned in 6 sentences in this paper.
Topics mentioned in this paper:
Beigman Klebanov, Beata and Flor, Michael
Application to Essay Scoring
From this set, pl-p6 were used for feature selection, data visualization, and estimation of the regression models (training), while sets p7-p9 were reserved for a blind test.
Application to Essay Scoring
To evaluate the usefulness of WAP in improving automated scoring of essays, we estimate a linear regression model using the human score as a dependent variable (label) and e-rater score and the HAT as the two independent variables (features).
Application to Essay Scoring
We estimate a regression model on each of setA-pi, i E {1, .., 6}, and evaluate them on each of setA-pj, j E {7, .., 9}, and compare the performance with that of e-rater alone on setA-pj.
Related Work
11We also performed a cross-validation test on setA p1-p6, where we estimated a regression model on setA-pi and evaluate it on setA-pj, for all i,j E {1, ..,6},i 7E j, and compared the performance with that of e-rater alone on setA-pj, yielding 30 different train-test combinations.
regression model is mentioned in 5 sentences in this paper.
Topics mentioned in this paper:
Kolachina, Prasanth and Cancedda, Nicola and Dymetman, Marc and Venkatapathy, Sriram
Inferring a learning curve from mostly monolingual data
Regression model 10K 75K 500K Ridge 0.063 0.060 0.053
Inferring a learning curve from mostly monolingual data
Table 4: Root mean squared error of the linear regression models for each anchor size
Inferring a learning curve from mostly monolingual data
Table 4 shows these results for Ridge and Lasso regression models at the three anchor sizes.
Selecting a parametric family of curves
We consider such observations to be generated by a regression model of the form:
regression model is mentioned in 5 sentences in this paper.
Topics mentioned in this paper:
Recasens, Marta and Danescu-Niculescu-Mizil, Cristian and Jurafsky, Dan
Automatically Identifying Biased Language
We trained a logistic regression model on a feature vector for every word that appears in the NPOV sentences from the training set, with the bias-inducing words as the positive class, and all the other words as the negative class.
Automatically Identifying Biased Language
The types of features used in the logistic regression model are listed in Table 3, together with their value space.
Automatically Identifying Biased Language
Logistic regression model that only uses the features based on Liu et al.’s (2005) lexicons of positive and negative words (i.e., features 26—29).
Conclusions
However, our logistic regression model reveals that epistemological and other features can usefully augment the traditional sentiment and subjectivity features for addressing the difficult task of identifying the bias-inducing word in a biased sentence.
regression model is mentioned in 5 sentences in this paper.
Topics mentioned in this paper:
Chaturvedi, Snigdha and Goldwasser, Dan and Daumé III, Hal
Intervention Prediction Models
Our logistic regression model uses the following two types of features: Thread only features and Aggregated post features.
Intervention Prediction Models
p,- and h,- represent the posts of the thread and their latent categories respectively; 7“ represents the instructor’s intervention and gb(t) represent the nonstructural features used by the logistic regression model .
Intervention Prediction Models
The logistic regression model is good at exploiting the thread level features but not the content of individual posts.
Introduction
The first uses a logistic regression model that primarily incorporates high level information about threads and posts.
regression model is mentioned in 5 sentences in this paper.
Topics mentioned in this paper:
Wang, Ziqi and Xu, Gu and Li, Hang and Zhang, Ming
Experimental Results
Two representative methods were used as baselines: the generative model proposed by (Brill and Moore, 2000) referred to as generative and the logistic regression model proposed by (Okazaki et al., 2008)
Experimental Results
When using their method for ranking, we used outputs of the logistic regression model as rank scores.
Introduction
(2008) proposed using a logistic regression model for approximate dictionary matching.
Related Work
(2008) utilized substring substitution rules and incorporated the rules into a L1-regularized logistic regression model .
regression model is mentioned in 4 sentences in this paper.
Topics mentioned in this paper:
Mairesse, François and Walker, Marilyn
Parameter Estimation Models
Continuous parameters are modeled with a linear regression model (LR), an M5’ model tree (M5), and a model based on support vector machines with a linear kernel (SVM).
Parameter Estimation Models
As regression models can extrapolate beyond the [0, 1] interval, the output parameter values are truncated if needed—at generation time—before being sent to the base generator.
Parameter Estimation Models
Table 3: Pearson’s correlation between parameter model predictions and continuous parameter values, for different regression models .
regression model is mentioned in 4 sentences in this paper.
Topics mentioned in this paper:
Wang, Chang and Fan, James
Experiments
The SVM with linear kernels and the linear regression model used the same features as the manifold models.
Experiments
By integrating unlabeled data, the manifold model under setting (1) made a 15% improvement over linear regression model on F1 score, where the improvement was significant across all relations.
Introduction
Our model goes beyond regular regression models in that it applies constraints to those coefficients, such that the topology of the given data manifold will be respected.
Introduction
Computing the optimal weights in a regression model and preserving manifold topology are conflicting objectives, we
regression model is mentioned in 4 sentences in this paper.
Topics mentioned in this paper:
Mohler, Michael and Bunescu, Razvan and Mihalcea, Rada
Answer Grading System
We train the isotonic regression model on each type of system output (i.e., alignment scores, SVM output, BOW scores).
Discussion and Conclusions
This is likely due to the different objective function in the corresponding optimization formulations: while the ranking model attempts to ensure a correct ordering between the grades, the regression model seeks to minimize an error objective that is closer to the RMSE.
Results
For each fold, one additional fold is held out for later use in the development of an isotonic regression model (see Figure 3).
regression model is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Bamman, David and Underwood, Ted and Smith, Noah A.
Experiments
In contrast, the Persona Regression model of Bamman et al.
Experiments
The Persona Regression model of Bamman et al.
Experiments
As expected, the Persona Regression model performs best at hypothesis class B (correctly judging two characters from the same author to be more similar to each other than to a character from a different author); this behavior is encouraged in this model by allowing an author (as an external metadata variable) to directly influence
regression model is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Bhat, Suma and Xue, Huichao and Yoon, Su-Youn
Experimental Results
The results reported are averaged over a 5-fold cross validation of the multiple regression model , where 80% of the SM data
Experimental Setup
Subsequently, the feature extraction stage (a VSM or a MaxEnt model as the case may be) generates the syntactic complexity feature which is then incorporated in a multiple linear regression model to generate a score.
Experimental Setup
As in prior studies, here too the level of agreement is evaluated by means of the weighted kappa measure as well as unrounded and rounded Pearson’s correlations between machine and human scores (since the output of the regression model can either be rounded or regarded as is).
regression model is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
Tomanek, Katrin and Hahn, Udo and Lohmann, Steffen and Ziegler, Jürgen
Cognitively Grounded Cost Modeling
Therefore, we learn a linear regression model with time (an operationalization of annotation costs) as the dependent variable.
Cognitively Grounded Cost Modeling
We learned a simple linear regression model with the annotation time as dependent variable and the features described above as independent variables.
Summary and Conclusions
This optimization may include both exploration of additional features (such as domain-specific ones) as well as experimentation with other, presumably nonlinear, regression models .
regression model is mentioned in 3 sentences in this paper.
Topics mentioned in this paper:
de Marneffe, Marie-Catherine and Manning, Christopher D. and Potts, Christopher
Analysis and discussion
The fitted logistic regression model (black line) has a statistically significant coefficient for response entropy (p < 0.001).
Analysis and discussion
Figure 5 plots the relationship between the response entropy and the accuracy of our decision procedure, along with a fitted logistic regression model using response entropy to predict whether our system’s inference was correct.
Methods
A logistic regression model can capture these facts.
regression model is mentioned in 3 sentences in this paper.
Topics mentioned in this paper: