![]() |
Image source : becominghuman.ai |
Machine Learning, computational learning theory, and similar terms are often used in the context of Data Mining, to denote the application of generic model-fitting or classification algorithms for predictive data mining. Unlike traditional statistical data analysis, which is usually concerned with the estimation of population parameters by statistical inference, the emphasis in data mining (and machine learning) is usually on the accuracy of prediction (predicted classification), regardless of whether or not the "models" or techniques that are used to generate the prediction is interpretable or open to simple explanation.
Good examples of this type of technique often applied to
predictive data mining are neural networks or meta-learning techniques such as
boosting, etc. These methods usually involve the fitting of very
complex "generic" models, that are not related to any
reasoning or theoretical understanding of underlying causal processes; instead,
these techniques can be shown to generate accurate predictions or
classification in crossvalidation samples.
The concept of meta-learning applies to the area of
predictive data mining, to combine the predictions from multiple models. It is
particularly useful when the types of models included in the project are very
different. In this context, this procedure is also referred to as Stacking
(Stacked Generalization).
![]() |
Image source : lilianweng.github.io |
Suppose your data mining project includes tree classifiers,
such as C&RT and CHAID, linear discriminant analysis (e.g., see GDA), and
Neural Networks. Each computes predicted classifications for a crossvalidation
sample, from which overall goodness-of-fit statistics (e.g., misclassification
rates) can be computed. Experience has shown that combining the predictions
from multiple methods often yields more accurate predictions than can be
derived from any one method (e.g., see Witten
and Frank, 2000).
The predictions from different classifiers can be used as
input into a meta-learner, which will attempt to combine the predictions to
create a final best predicted classification. So, for example, the predicted
classifications from the tree classifiers, linear model, and the neural network
classifier(s) can be used as input variables into a neural network
meta-classifier, which will attempt to "learn" from the data
how to combine the predictions from the different models to yield maximum
classification accuracy.
We can apply meta-learners to the results from different meta-learners
to create "meta-meta"-learners, and so on; however, in practice such
exponential increase in the amount of data processing, in order to derive an
accurate prediction, will yield less and less marginal utility.
One of the preliminary stage in predictive data mining,
when the data set includes more variables than could be included (or would be
efficient to include) in the actual model building phase (or even in initial
exploratory operations), is to select predictors from a large list of
candidates. For example, when data are collected via automated (computerized)
methods, it is not uncommon that measurements are recorded for thousands or
hundreds of thousands (or more) of predictors.
The standard analytic methods
for predictive data mining, such as neural network analyses, classification and
regression trees, generalized linear models, or general linear models become
impractical when the number of predictors exceed more than a few hundred
variables.
![]() |
Image source : medium.com |
Feature selection selects a subset of predictors from a
large list of candidate predictors without assuming that the relationships
between the predictors and the dependent or outcome variables of interest are
linear, or even monotone. Therefore, this is used as a pre-processor for predictive
data mining, to select manageable sets of predictors that are likely related to
the dependent (outcome) variables of interest, for further analyses with any of
the other methods for regression and classification.
Reference source : documentation(dot)statsoft(dot)com
Machine Learning, Meta-Learning and Feature Selection
Reviewed by AIA
on
December 21, 2019
Rating:

No comments: