1. Trang chủ
  2. » Công Nghệ Thông Tin

Data Mining and Knowledge Discovery Handbook, 2 Edition part 8 potx

10 431 0
Tài liệu đã được kiểm tra trùng lặp

Đang tải... (xem toàn văn)

THÔNG TIN TÀI LIỆU

Thông tin cơ bản

Định dạng
Số trang 10
Dung lượng 390,99 KB

Các công cụ chuyển đổi và chỉnh sửa cho tài liệu này

Nội dung

Key words: Feature Extraction, Dimensional Reduction, Principal Components Analysis, Distortion Discriminant Analysis, Nystr¨om method, Projection Pursuit, Kernel PCA, Multidimensional S

Trang 1

50 Jerzy W Grzymala-Busse and Witold J Grzymala-Busse

Grzymala-Busse J.W., Grzymala-Busse W.J., and Goodwin L.K A comparison of three

clos-est fit approaches to missing attribute values in preterm birth data International Journal

of Intelligent Systems 17 (2002) 125–134.

Grzymala-Busse, J.W and Hu, M A comparison of several approaches to missing attribute values in Data Mining Proceedings of the Second In-ternational Conference on Rough Sets and Current Trends in Computing RSCTC’2000, Banff, Canada, October 16–19, 2000, 340–347

Grzymala-Busse, J.W and Wang A.Y Modified algorithms LEM1 and LEM2 for rule induc-tion from data with missing attribute values Proc of the Fifth Internainduc-tional Workshop

on Rough Sets and Soft Computing (RSSC’97) at the Third Joint Conference on Infor-mation Sciences (JCIS’97), Research Triangle Park, NC, March 2–5, 1997, 69–72 Grzymala-Busse J.W and Siddhaye S Rough set approaches to rule induction from incom-plete data Proceedings of the IPMU’2004, the 10th International Conference on In-formation Processing and Management of Uncertainty in Knowledge-Based Systems, Perugia, Italy, July 49, 2004, vol 2, 923930

Imielinski T and Lipski W Jr Incomplete information in relational databases, Journal of the ACM 31 (1984) 761–791.

Kononenko I., Bratko I., and Roskar E Experiments in automatic learning of medical diag-nostic rules Technical Report, Jozef Stefan Institute, Lljubljana, Yugoslavia, 1984 Kryszkiewicz M Rough set approach to incomplete information systems Proceedings of the Second Annual Joint Conference on Information Sciences, Wrightsville Beach, NC, September 28–October 1, 1995, 194–197

Kryszkiewicz M Rules in incomplete information systems Information Sciences 113 (1999)

271–292

Lakshminarayan K., Harp S.A., and Samad T Imputation of missing data in industrial

databases Applied Intelligence 11 (1999) 259 – 275.

Latkowski, R On decomposition for incomplete data Fundamenta Informaticae 54 (2003)

1-16

Latkowski R and Mikolajczyk M Data decomposition and decision rule join-ing for classification of data with missing values Proceedings of the RSCTC’2004, the Fourth International Conference on Rough Sets and Current Trends in Computing, Uppsala, Sweden, June 1–5, 2004 Lecture Notes in Artificial Intelligence 3066, Springer-Verlag 2004, 254–263

Lipski W Jr On semantic issues connected with incomplete information

databases ACM Transactions on Database Systems 4 (1979), 262–296.

Lipski W Jr On databases with incomplete information Journal of the ACM 28 (1981) 41–

70

Little R.J.A and Rubin D.B Statistical Analysis with Missing Data, Second Edition, J Wiley

& Sons, Inc., 2002

Pawlak Z Rough Sets International Journal of Computer and Information Sciences 11

(1982) 341–356

Pawlak Z Rough Sets Theoretical Aspects of Reasoning about Data Kluwer Academic

Publishers, Dordrecht, Boston, London, 1991

Pawlak Z., Grzymala-Busse J.W., Slowinski R., and Ziarko, W Rough sets Communications

of the ACM 38 (1995) 88–95.

Polkowski L and Skowron A (eds.) Rough Sets in Knowledge Discovery, 2, Applications, Case Studies and Software Systems, Appendix 2: Software Systems Physica Verlag,

Heidelberg New York (1998) 551–601

Trang 2

3 Handling Missing Attribute Values 51 Quinlan J.R Unknown attribute values in induction Proc of the 6-th Int Workshop on Ma-chine Learning, Ithaca, NY, 1989, 164 – 168

Quinlan J R C4.5: Programs for Machine Learning Morgan Kaufmann Publishers, San

Mateo CA (1993)

Schafer J.L Analysis of Incomplete Multivariate Data Chapman and Hall, London, 1997.

Slowinski R and Vanderpooten D A generalized definition of rough approximations based

on similarity IEEE Transactions on Knowledge and Data Engineering 12 (2000) 331–

336

Stefanowski J Algorithms of Decision Rule Induction in Data Mining Poznan University of Technology Press, Poznan, Poland (2001)

Stefanowski J and Tsoukias A On the extension of rough sets under incomplete information Proceedings of the 7th International Workshop on New Directions in Rough Sets, Data Mining, and Granular-Soft Computing,

RSFDGrC’1999, Ube, Yamaguchi, Japan, November 8–10, 1999, 73–81

Stefanowski J and Tsoukias A Incomplete information tables and rough classification Com-putational Intelligence 17 (2001) 545–566.

Weiss S and Kulikowski C.A Computer Systems That Learn: Classification and Prediction Methods from Statistics, Neural Nets, Machine Learning, and Expert Systems, chapter How to Estimate the True Performance of a Learning System, pp 17–49, San Mateo,

CA: Morgan Kaufmann Publishers, Inc., 1991

Wong K.C and Chiu K.Y Synthesizing statistical knowledge for incomplete mixed-mode data IEEE Transactions on Pattern Analysis and Machine Intelligence 9 (1987) 796805

Wu X and Barbara D Learning missing values from summary constraints ACM SIGKDD Explorations Newsletter 4 (2002) 21 – 30.

Wu X and Barbara D Modeling and imputation of large incomplete multidimensional datasets Proc of the 4-th Int Conference on Data Warehousing and Knowledge Dis-covery, Aix-en-Provence, France, 2002, 286 – 295

Yao Y.Y On the generalizing rough set theory Proc of the 9th Int Conference on Rough Sets, Fuzzy Sets, Data Mining and Granular Computing (RSFDGrC’2003), Chongqing, China, October 19–22, 2003, 44–51

Trang 4

Geometric Methods for Feature Extraction and

Dimensional Reduction - A Guided Tour

Christopher J.C Burges

Microsoft Research

Summary We give a tutorial overview of several geometric methods for feature extraction and dimensional reduction We divide the methods into projective methods and methods that model the manifold on which the data lies For projective methods, we review projection pursuit, principal component analysis (PCA), kernel PCA, probabilistic PCA, and oriented PCA; and for the manifold methods, we review multidimensional scaling (MDS), landmark MDS, Isomap, locally linear embedding, Laplacian eigenmaps and spectral clustering The Nystr¨om method, which links several of the algorithms, is also reviewed The goal is to provide

a self-contained review of the concepts and mathematics underlying these algorithms

Key words: Feature Extraction, Dimensional Reduction, Principal Components Analysis, Distortion Discriminant Analysis, Nystr¨om method, Projection Pursuit, Kernel PCA, Multidimensional Scaling, Landmark MDS, Locally Linear Embed-ding, Isomap

Introduction

Feature extraction can be viewed as a preprocessing step which removes distracting variance from a dataset, so that downstream classifiers or regression estimators per-form better The area where feature extraction ends and classification, or regression, begins is necessarily murky: an ideal feature extractor would simply map the data

to its class labels, for the classification task On the other hand, a character recog-nition neural net can take minimally preprocessed pixel values as input, in which case feature extraction is an inseparable part of the classification process (LeCun and Bengio, 1995) Dimensional reduction - the (usually non-invertible) mapping of data to a lower dimensional space - is closely related (often dimensional reduction

is used as a step in feature extraction), but the goals can differ Dimensional reduc-tion has a long history as a method for data visualizareduc-tion, and for extracting key low dimensional features (for example, the 2-dimensional orientation of an object, from its high dimensional image representation) The need for dimensionality reduction

O Maimon, L Rokach (eds.), Data Mining and Knowledge Discovery Handbook, 2nd ed.,

DOI 10.1007/978-0-387-09823-4_4, © Springer Science+Business Media, LLC 2010

Trang 5

54 Christopher J.C Burges

also arises for other pressing reasons (Stone, 1982) showed that, under certain reg-ularity assumptions, the optimal rate of convergence1for nonparametric regression

varies as m −p/(2p+d) , where m is the sample size, the data lies in R d, and where the

regression function is assumed to be p times differentiable Consider 10,000 sample points, for p = 2 and d = 10 If d is increased to 20, the number of sample points

must be increased to approximately 10 million in order to achieve the same optimal rate of convergence If our data lie (approximately) on a low dimensional manifold

L that happens to be embedded in a high dimensional manifold H , modeling the

projected data inL rather than in H may turn an infeasible problem into a feasible

one

The purpose of this review is to describe the mathematics and ideas underlying the algorithms Implementation details, although important, are not discussed Some notes on notation: vectors are denoted by boldface, whereas components are denoted

by x a, or by(xi)a for the a’th component of the i’th vector Following

(Horn and Johnson, 1985), the set of p by q matrices is denoted M pq, the set of

(square) p by p matrices by M p , and the set of symmetric p by p matrices by S p(all matrices considered are real) e with no subscript is used to denote the vector of all ones; on the other hand ea denotes the a’th eigenvector We denote sample size by m, and dimension usually by d or d , with typically d d.δi jis the Kronecker delta

(the i j’th component of the unit matrix) We generally reserve indices i, j, to index vectors and a, b to index dimension.

We place feature extraction and dimensional reduction techniques into two broad categories: methods that rely on projections (Section 4.1) and methods that attempt to model the manifold on which the data lies (Section 4.2) Section 4.1 gives a detailed description of principal component analysis; apart from its intrinsic usefulness, PCA

is interesting because it serves as a starting point for many modern algorithms, some

of which (kernel PCA, probabilistic PCA, and oriented PCA) are also described However it has clear limitations: it is easy to find even low dimensional examples where the PCA directions are far from optimal for feature extraction (Duda and Hart, 1973), and PCA ignores correlations in the data that are higher than second order Section 4.2 starts with an overview of the Nystr¨om method, which can be used

to extend, and link, several of the algorithms described in this chapter We then ex-amine some methods for dimensionality reduction which assume that the data lie

on a low dimensional manifold embedded in a high dimensional spaceH , namely

locally linear embedding, multidimensional scaling, Isomap, Laplacian eigenmaps, and spectral clustering

1For convenience we reproduce Stone’s definitions (Stone, 1982) Letθ be the unknown regression function, ˆT nan estimator ofθ using n samples, and {b n } a sequence of positive

constants Then{b n } is called a lower rate of convergence if there exists c > 0 such that

limninfTnˆ supθP n − n) = 1, and it is called an achievable rate of convergence if there is a sequence of estimators{ ˆT n } and c > 0 such that lim nsupθP n − n) = 0;

{b n } is called an optimal rate of convergence if it is both a lower rate of convergence and

an achievable rate of convergence

Trang 6

4 Geometric Methods for Feature Extraction and Dimensional Reduction 55

4.1 Projective Methods

If dimensional reduction is so desirable, how should we go about it? Perhaps the

simplest approach is to attempt to find low dimensional projections that extract

use-ful information from the data, by maximizing a suitable objective function This is the idea of projection pursuit (Friedman and Tukey, 1974) The name ’pursuit’ arises from the iterative version, where the currently optimal projection is found in light of previously found projections (in fact originally this was done manually2) Apart from handling high dimensional data, projection pursuit methods can be robust to noisy

or irrelevant features (Huber, 1985), and have been applied to regression (Friedman and Stuetzle, 1981), where the regression is expressed as a sum of ’ridge functions’ (functions of the one dimensional projections) and at each iteration the projection is chosen to minimize the residuals; to classification; and to density estimation (Fried-man et al., 1984) How are the interesting directions found? One approach is to search for projections such that the projected data departs from normality (Huber, 1985) One might think that, since a distribution is normal if and only if all of its one di-mensional projections are normal, if the least normal projection of some dataset is still approximately normal, then the dataset is also necessarily approximately nor-mal, but this is not true; Diaconis and Freedman have shown that most projections

of high dimensional data are approximately normal (Diaconis and Freedman, 1984) (see also below) Given this, finding projections along which the density departs from normality, if such projections exist, should be a good exploratory first step

The sword of Diaconis and Freedman cuts both ways, however If most jections of most high dimensional datasets are approximately normal, perhaps pro-jections are not always the best way to find low dimensional representations Let’s review their results in a little more detail The main result can be stated informally

as follows: consider a model where the data, the dimension d, and the sample size

and d Suppose that asν tends to infinity, the fraction of vectors which are not ap-proximately the same length tends to zero, and suppose further that under the same conditions, the fraction of pairs of vectors which are not approximately orthogonal

to each other also tends to zero3 Then ( (Diaconis and Freedman, 1984), theorem 1.1) the empirical distribution of the projections along any given unit direction tends

to N (0,σ2) weakly in probability However, if the conditions are not fulfilled, as for some long-tailed distributions, then the opposite result can hold - that is, most

pro-jections are not normal (for example, most propro-jections of Cauchy distributed data4 will be Cauchy (Diaconis and Freedman, 1984))

2See J.H Friedman’s interesting response to (Huber, 1985) in the same issue

3More formally, the conditions are: for σ2 positive and finite, and for any positive ε,

j 2σ2d| > εd} → 0 and (1/m2)card{1 ≤ j,k ≤ m : |x j · x k | > εd} → 0 (Diaconis and Freedman, 1984).

4The Cauchy distribution in one dimension has density c /(c2+ x2) for constant c.

Trang 7

56 Christopher J.C Burges

As a concrete example5, consider data uniformly distributed over the unit n+ 1-sphereS n+1for odd n Let’s compute the density projected along any line I passing

through the origin By symmetry, the result will be independent of

the direction we choose If the distance along the projection is parameterized by

on the sphere, then the density atξ is proportional to the volume of an n-sphere of

radius sinθ:ρ(ξ) = C(1−ξ2)n−12 Requiring that1

−1ρ(ξ)dξ= 1 gives the constant

C:

C= 21(n+1) n!!

(1

Let’s plot this density and compare against a one dimensional Gaussian density fitted using maximum likelihood For that we just need the variance, which can be com-puted analytically:σ2= 1

n+2, and the mean, which is zero Figure 4.1 shows the re-sult for the 20-sphere Although data uniformly distributed onS20is far from Gaus-sian, its projection along any direction is close to Gaussian for all such directions, and we cannot hope to uncover such structure using one dimensional projections

0

0.2

0.4

0.6

0.8

1

1.2

1.4

1.6

1.8

2

Fig 4.1 Dotted line: a Gaussian with zero mean and variance 1/21 Solid line: the density projected from data distributed uniformly over the 20-sphere, to any line passing through the origin

The notion of searching for non-normality, which is at the heart of projection pursuit (the goal of which is dimensional reduction), is also the key idea underly-ing independent component analysis (ICA) (the goal of which is source separation) ICA (Hyv¨arinen et al., 2001) searches for projections such that the probability distri-butions of the data along those projections are statistically independent: for example,

5The story for even n is similar but the formulae are slightly different

Trang 8

4 Geometric Methods for Feature Extraction and Dimensional Reduction 57 consider the problem of separating the source signals in a linear combinations of sig-nals, where the sources consist of speech from two speakers who are recorded using two microphones (and where each microphone captures sound from both speakers) The signal is the sum of two statistically independent signals, and so finding those independent signals is required in order to decompose the signal back into the two original source signals, and at any given time, the separated signal values are re-lated to the microphone signals by two (time independent) projections (forming an invertible 2 by 2 matrix) If the data is normally distributed, finding projections along which the data is uncorrelated is equivalent to finding projections along which it is independent, so although using principal component analysis (see below) will suf-fice to find independent projections, those projections will not be useful for the above task For most other distributions, finding projections along which the data is statis-tically independent is a much stronger (and for ICA, useful) condition than finding projections along which the data is uncorrelated Hence ICA concentrates on situa-tions where the distribution of the data departs from normality, and in fact, finding the maximally non-Gaussian component (under the constraint of constant variance) will give you an independent component (Hyv¨arinen et al., 2001)

4.1.1 Principal Component Analysis (PCA)

PCA: Finding an Informative Direction

Given data xi ∈ R d , i = 1,··· ,m, suppose you’d like to find a direction v ∈ R d for which the projection xi · v gives a good one dimensional representation of your

orig-inal data: that is, informally, the act of projecting loses as little information about your expensively-gathered data as possible (we will examine the information theo-retic view of this below) Suppose that unbeknownst to you, your data in fact lies along a lineI embedded in R d, that is, xi=μ+θin, whereμis the sample mean6,

is then

m

m

i=1

((xi −μ) · n)2= 1

m

m

i=1θ2

and that along some other unit direction n is

v n ≡ 1 m

m

i=1((xi −μ) · n )2= 1

m

m

i=1θ2

Since (n · n )2= cos2φ, whereφ is the angle between n and n , we see that the projected variance is maximized if and only if n= ±n Hence in this case, finding

the projection for which the projected variance is maximized gives you the direction

you are looking for, namely n, regardless of the distribution of the data along n, as

long as the data has finite variance You would then quickly find that the variance along all directions orthogonal to n is zero, and conclude that your data in fact lies

6Note that if all x lie along a given line then so doesμ

Trang 9

58 Christopher J.C Burges

along a one dimensional manifold embedded in R d This is one of several basic results of PCA that hold for arbitrary distributions, as we shall see

Even if the underlying physical process generates data that ideally lies alongI ,

noise will usually modify the data at various stages up to and including the mea-surements themselves, and so your data will very likely not lie exactly alongI If

the overall noise is much smaller than the signal, it makes sense to try to findI by

searching for that projection along which the projected data has maximum variance

If in addition your data lies in a two (or higher) dimensional subspace, the above argument can be repeated, picking off the highest variance directions in turn Let’s see how that works

PCA: Ordering by Variance

We’ve seen that directions of maximum variance can be interesting, but how can we find them? The variance along unit vector n (Eq (4.2)) is n Cn where C is the sample

covariance matrix Since C is positive semidefinite, its eigenvalues are positive or

zero; let’s choose the indexing such that the (unit normed) eigenvectors ea , a =

Since the{e a } span the space, we can expand n in terms of them: n = ∑ d

=1αaea, and we’d like to find theαathat maximize n Cn= n aαa Ce a= ∑aλaα2

a, subject

to∑aα2

a= 1 (to give unit normed n) This is just a convex combination of theλ’s, and since a convex combination of any set of numbers is maximized by taking the largest, the optimal n is just e1, the principal eigenvector (or any one of the set of such eigenvectors, if multiple eigenvectors share the same largest eigenvalue), and furthermore, the variance of the projection of the data along n is justλ1

The above construction captures the variance of the data along the direction n

To characterize the remaining variance of the data, let’s find that direction m which

is both orthogonal to n, and along which the projected data again has maximum

variance Since the eigenvectors of C form an orthonormal basis (or can be so

cho-sen), we can expand m in the subspaceR d−1 orthogonal to n as m= ∑d

a=2βaea Just as above, we wish to find theβathat maximize m Cm= ∑d

a=2λaβ2

a, subject to

d

a=2β2

a= 1, and by the same argument, the desired direction is given by the (or any) remaining eigenvector with largest eigenvalue, and the corresponding variance is just

that eigenvalue Repeating this argument gives d orthogonal directions, in order of monotonically decreasing projected variance Since the d directions are orthogonal, they also provide a complete basis Thus if one uses all d directions, no informa-tion is lost, and as we’ll see below, if one uses the d < d principal directions, then

the mean squared error introduced by representing the data in this manner is mini-mized Finally, PCA for feature extraction amounts to projecting the data to a lower dimensional space: given an input vector x, the mapping consists of computing the projections of x along the ea , a = 1, ,d , thereby constructing the components of the projected d -dimensional feature vectors

Trang 10

4 Geometric Methods for Feature Extraction and Dimensional Reduction 59 PCA Decorrelates the Samples

Now suppose we’ve performed PCA on our samples, and instead of using it to con-struct low dimensional features, we simply use the full set of orthonormal eigen-vectors as a choice of basis In the old basis, a given input vector x is expanded as

x= ∑d

a=1x auafor some orthonormal set{u a }, and in the new basis, the same vector

is expanded as x= ∑d

b=1˜x beb , so ˜x a ≡ x · e a= ea · ∑ b x bub The mean μ 1

mixi has components ˜μa· e ain the new basis The sample covariance matrix depends

on the choice of basis: if C is the covariance matrix in the old basis, then the

cor-responding covariance matrix in the new basis is ˜C ab ≡ 1

mi ( ˜x ia − ˜μa )( ˜x ib − ˜μb) = 1

mi {e a · (∑ p x ipup −μ)}{∑ q x iquq −μ) · e b } = e

a Ce bbδab Hence in the new basis the covariance matrix is diagonal and the samples are uncorrelated It’s worth emphasizing two points: first, although the covariance matrix can be viewed as a ge-ometric object in that it transforms as a tensor (since it is a summed outer product of vectors, which themselves have a meaning independent of coordinate system), nev-ertheless, the notion of correlation is basis-dependent (data can be correlated in one basis and uncorrelated in another) Second, PCA decorrelates the samples whatever their underlying distribution; it does not have to be Gaussian

PCA: Reconstruction with Minimum Squared Error

The basis provided by the eigenvectors of the covariance matrix is also optimal for di-mensional reduction in the following sense Again consider some arbitrary orthonor-mal basis{u a , a = 1, ,d}, and take the first d of these to perform the dimensional

reduction: ˜x≡ ∑ d

a=1(x · u a)ua The chosen uaform a basis forR d

, so we may take the components of the dimensionally reduced vectors to be x· u a , a = 1, ,d (al-though here we leave ˜x with dimension d) Define the reconstruction error summed

over the dataset as∑m

i=1 i − ˜x i 2 Again assuming that the eigenvectors{e a } of the

covariance matrix are ordered in order of non-increasing eigenvalues, choosing to use those eigenvectors as basis vectors will give minimal reconstruction error If the data is not centered, then the mean should be subtracted first, the dimensional reduc-tion performed, and the mean then added back7; thus in this case, the dimensionally reduced data will still lie in the subspaceR d

, but that subspace will be offset from the origin by the mean Bearing this caveat in mind, to prove the claim we can assume that the data is centered Expanding ua ≡ ∑ d

p=1βapep, we have 1

i

i − ˜x i 2= 1

i

i 21 m

d

a=1∑

i

(xi · u a)2

(4.4)

with the constraints∑d

p=1βapβbpab The second term on the right is

7The principal eigenvectors are not necessarily the directions that give minimal reconstruc-tion error if the data is not centered: imagine data whose mean is both orthogonal to the principal eigenvector and far from the origin The single direction that gives minimal re-construction error will be close to the mean

... are: for σ2< /small> positive and finite, and for any positive ε,

j 2< /sup>σ2< /small>d| > εd} → and (1/m2< /small>)card{1... was done manually2< /sup>) Apart from handling high dimensional data, projection pursuit methods can be robust to noisy

or irrelevant features (Huber, 1 985 ), and have been applied...

i=1θ2< /small>

Since (n · n )2< /small>= cos2< /small>φ, whereφ is the angle between n and

Ngày đăng: 04/07/2014, 05:21

TỪ KHÓA LIÊN QUAN