ϟ
 
DOI: 10.1017/s026988899900404x
¤ OpenAccess: Green
This work has “Green” OA status. This means it may cost money to access on the publisher landing page, but there is a free copy in an OA repository.

An overview of regression techniques for knowledge discovery

Ilhan Uysal,H. Altay Güvenir

Local regression
Artificial intelligence
Nonparametric regression
Predicting or learning numeric features is called regression in the statistical literature, and it is the subject of research in both machine learning and statistics. This paper reviews the important techniques and algorithms for regression developed by both communities. Regression is important for many applications, since lots of real life problems can be modeled as regression problems. The review includes Locally Weighted Regression (LWR), rule-based regression, Projection Pursuit Regression (PPR), instance-based regression, Multivariate Adaptive Regression Splines (MARS) and recursive partitioning regression methods that induce regression trees (CART, RETIS and M5).
    Cite this:
Generate Citation
Powered by Citationsy*
Referenced Papers:
DOI: 10.1016/b978-1-55860-307-3.50037-x
1993
Cited 399 times
Combining Instance-Based and Model-Based Learning
This paper concerns learning tasks that require the prediction of a continuous value rather than a discrete class. A general method is presented that allows predictions to use both instance-based and model-based learning. Results with three approaches to constructing models and with eight datasets demonstrate improvements due to the composite method. Keywords: learning with continuous classes, instance-based learning, model-based learning, empirical evaluation.
MAG: 148291325
1992
Cited 69 times
Employing linear regression in regression tree leaves
DOI: 10.1613/jair.199
¤ Open Access
1995
Cited 132 times
Rule-based Machine Learning Methods for Functional Prediction
We describe a machine learning method for predicting the value of a real-valued function, given the values of multiple input variables. The method induces solutions from samples in the form of ordered disjunctive normal form (DNF) decision rules. A central objective of the method and representation is the induction of compact, easily interpretable solutions. This rule-based decision model can be extended to search efficiently for similar cases prior to approximating function values. Experimental results on real-world data demonstrate that the new techniques are competitive with existing machine learning and statistical methods and can sometimes yield superior regression performance.
MAG: 153647241
1987
Cited 99 times
Learning decision rules in noisy domains
MAG: 1610836425
1991
Cited 753 times
Nearest neighbor (NN) norms: NN pattern classification techniques
DOI: 10.1023/a:1006559212014
1997
Cited 1,485 times
MAG: 177086481
1993
Cited 14 times
Rule-Based Regression.
DOI: 10.1109/64.248354
1993
Cited 71 times
Optimized rule induction
Swap-1, a state-of-the-art system for learning decision rules from data, is presented. The method embodied in Swap-1 generates reduced-complexity solutions by inducing compact solutions in larger dimensions where many rules might be needed to make accurate predictions. For many applications, such systems can automatically construct relatively compact rule sets with highly predictive performance. >
DOI: 10.1111/j.1467-8640.1989.tb00315.x
¤ Open Access
1989
Cited 139 times
Instance-based prediction of real-valued attributes
Instance-based representations have been applied to numerous classification tasks with some success. Most of these applications involved predicting a symbolic class based on observed attributes. This paper presents an instance-based method for predicting a numeric value based on observed attributes. We prove that, given enough instances, if the numeric values are generated by continuous functions with bounded slope, then the predicted values are accurate approximations of the actual values. We demonstrate the utility of this approach by comparing it with a standard approach for value prediction. The instance-based approach requires neither ad hoc parameters nor background knowledge.
DOI: 10.1093/comjnl/3.3.175
¤ Open Access
1960
Cited 2,458 times
An Automatic Method for Finding the Greatest or Least Value of a Function
DOI: 10.1145/240455.240464
1996
Cited 1,528 times
The KDD process for extracting useful knowledge from volumes of data
DOI: 10.1145/240455.240472
1996
Cited 447 times
A database perspective on knowledge discovery
DOI: 10.1109/t-c.1974.224051
¤ Open Access
1974
Cited 1,390 times
A Projection Pursuit Algorithm for Exploratory Data Analysis
An algorithm for the analysis of multivariate data is presented and is discussed in terms of specific examples. The algorithm seeks to find one-and two-dimensional linear projections of multivariate data that are relatively highly revealing.
DOI: 10.1080/01621459.1981.10477729
¤ Open Access
1981
Cited 1,679 times
Projection Pursuit Regression
Abstract A new method for nonparametric multiple regression is presented. The procedure models the regression surface as a sum of general smooth functions of linear combinations of the predictor variables in an iterative manner. It is more general than standard stepwise and stagewise regression procedures, does not require the definition of a metric in the predictor space, and lends itself to graphical interpretation.
DOI: 10.1214/aos/1176347126
¤ Open Access
1989
Cited 172 times
On Projection Pursuit Regression
We construct a tractable mathematical model for kernel-based projection pursuit regression approximation. The model permits computation of explicit formulae for bias and variance of estimators. It is shown that the bias of an orientation estimate dominates error about the mean--indeed, the latter is asymptotically negligible in comparison with bias. However, bias and error about the mean are of the same order in the case of projection pursuit curve estimates. Implications of our formulae for bias and variance are discussed.
DOI: 10.1214/aos/1176347963
¤ Open Access
1991
Cited 5,860 times
Multivariate Adaptive Regression Splines
A new method is presented for flexible regression modeling of high dimensional data. The model takes the form of an expansion in product spline basis functions, where the number of basis functions as well as the parameters associated with each one (product degree and knot locations) are automatically determined by the data. This procedure is motivated by the recursive partitioning approach to regression and shares its attractive properties. Unlike recursive partitioning, however, this method produces continuous models with continuous derivatives. It has more power and flexibility to model relationships that are nearly additive or involve interactions in at most a few variables. In addition, the model can be represented in a form that separately identifies the additive contributions and those associated with the different multivariable interactions.
DOI: 10.1007/978-1-4612-2660-4_33
1994
Cited 22 times
Learning to Catch: Applying Nearest Neighbor Algorithms to Dynamic Control Tasks
This paper examines the hypothesis that local weighted variants of k-nearest neighbor algorithms can support dynamic control tasks. We evaluated several k-nearest neighbor (k-NN) algorithms on the simulated learning task of catching a flying ball. Previously, local regression algorithms have been advocated for this class of problems. These algorithms, which are variants of k-NN, base their predictions on a (possibly weighted) regression computed from the k nearest neighbors. While they outperform simpler k-NN algorithms on many tasks, they have trouble on this ball-catching task. We hypothesize that the non-linearities in this task are the cause of this behavior, and that local regression algorithms may need to be modified to work well under similar conditions.
DOI: 10.1023/a:1022689900470
¤ Open Access
1991
Cited 1,996 times
Storing and using specific instances improves the performance of several supervised learning algorithms. These include algorithms that learn decision trees, classification rules, and distributed networks. However, no investigation has analyzed algorithms that use only specific instances to solve incremental learning tasks. In this paper, we describe a framework and methodology, called instance-based learning, that generates classification predictions using only specific instances. Instance-based learning algorithms do not maintain a set of abstractions derived from specific instances. This approach extends the nearest neighbor algorithm, which has large storage requirements. We describe how storage requirements can be significantly reduced with, at most, minor sacrifices in learning rate and classification accuracy. While the storage-reducing algorithm performs well on several real-world databases, its performance degrades rapidly with the level of attribute noise in training instances. Therefore, we extended it with a significance test to distinguish noisy instances. This extended algorithm's performance degrades gracefully with increasing noise levels and compares favorably with a noise-tolerant decision tree algorithm.
MAG: 2222298566
1990
Cited 4 times
Knowledge Engineering Vol I: Fundamentals
From the Publisher: This contributed volume,prepared by a team of internationally renowned experts,is an excellent sourcebook covering the fundamentals of expert systems. Comprehensive discussions of the topics of representation of knowledge and rule-based systems are included.
DOI: 10.5860/choice.28-4558
1991
Cited 4,648 times
Finding groups in data: an introduction to cluster analysis
1. Introduction. 2. Partitioning Around Medoids (Program PAM). 3. Clustering large Applications (Program CLARA). 4. Fuzzy Analysis. 5. Agglomerative Nesting (Program AGNES). 6. Divisive Analysis (Program DIANA). 7. Monothetic Analysis (Program MONA). Appendix 1. Implementation and Structure of the Programs. Appendix 2. Running the Programs. Appendix 3. Adapting the Programs to Your Needs. Appendix 4. The Program CLUSPLOT. References. Author Index. Subject Index.
MAG: 3017143921
1973
Cited 7,364 times
Pattern classification and scene analysis
Provides a unified, comprehensive and up-to-date treatment of both statistical and descriptive methods for pattern recognition. The topics treated include Bayesian decision theory, supervised and unsupervised learning, nonparametric techniques, discriminant analysis, clustering, preprosessing of pictorial data, spatial filtering, shape description techniques, perspective transformations, projective invariants, linguistic procedures, and artificial intelligence techniques for scene analysis.
MAG: 622270208
1990
Cited 3 times
Understanding Knowledge Engineering
An overview of regression techniques for knowledge discovery” is a paper by Ilhan Uysal H. Altay Güvenir published in the journal Knowledge Engineering Review in 1999. It was published by Cambridge University Press. It has an Open Access status of “green”. You can read and download a PDF Full Text of this paper here.