Lazy
Learners (or Learning from Your Neighbors)
Imagine
a contrasting lazy approach, in which the learner instead waits until the last minute
before doing any model construction in order to classify a given test tuple.
That is, when given a training tuple, a lazy learner simply
stores it (or does only a little minor processing) and waits until it is given
a test tuple. Only when it sees the test tuple does it perform generalization
in order to classify the tuple based on its similarity to the stored training
tuples.
K-Nearest-Neighbor
Classifiers
The
k-nearest-neighbor method was first
described in the early 1950s. The method is labor intensive when given large
training sets, and did not gain popularity until the 1960s when increased
computing power became available. It has since been widely used in the area of
pattern recognition. Nearest-neighbor classifiers are based on learning by
analogy, that is, by comparing a given test tuplewh training tuples that are
similar to it. The training tuples are described by n
attributes. Each tuple represents a point in an n-dimensional
space. In this way, all of the training tuples are stored in an n-dimensional
pattern space. When given an unknown tuple, a k-nearest-neighbor
classifier searches the pattern space for the k
training tuples that are closest to the unknown tuple. These k
training tuples are the k “nearest
Neighbors”
of the unknown tuple. “Closeness” is defined in terms of a distance metric,
such as Euclidean distance. The Euclidean distance between two points or
tuples, say, X1
= (x11,
x12, : : : ,
x1n) and
X2 = (x21, x22, : : : , x2n), is In other words, for each numeric
attribute, we take the difference between the corresponding values of that
attribute in tuple X1 and in tuple X2, square this
difference, and accumulate it. The square root is taken of the total
accumulated distance count. Typically, we normalize the values of each
attribute before using Equation (6.45). This helps prevent attributes with
initially large ranges (such as income) from outweighing attributes with initially smaller ranges
(such as binary attributes). Min-max normalization,
No comments:
Post a Comment