You are on page 1of 118

Concepts and

Techniques

— Chapter 2 —

Jiawei Han and Micheline Kamber


Department of Computer Science
University of Illinois at Urbana-Champaign
www.cs.uiuc.edu/~hanj
©2008 Jiawei Han. All rights reserved.
August 10, 2009 Data Mining: Concepts and Techniques 1
August 10, 2009 Data Mining: Concepts and Techniques 2
Chapter 2: Data
Preprocessing
 General data characteristics
 Basic data description and exploration
 Measuring data similarity
 Data cleaning
 Data integration and transformation
 Data reduction
 Summary

August 10, 2009 Data Mining: Concepts and Techniques 3


Types of Data Sets
 Record
 Relational records

timeout

season
coach

score

game
Data matrix, e.g., numerical

team

ball

lost
pla

wi
n
y
matrix, crosstabs
 Document data: text Document 1 3 0 5 0 2 6 0 2 0 2

documents: term-frequency Document 2 0 7 0 2 1 0 0 3 0 0

vector Document 3 0 1 0 0 1 2 2 0 3 0

 Transaction data
TID Items
 Graph
 World Wide Web 1 Bread, Coke, Milk
2 Beer, Bread
 Social or information
networks 3 Beer, Coke, Diaper, Milk
 Molecular Structures 4 Beer, Bread, Diaper, Milk
 Ordered 5 Coke, Diaper, Milk
 Spatial data: maps
 Temporal data: time-series
 Sequential Data: transaction
sequences
August 10, 2009 Data Mining: Concepts and Techniques 4
Important Characteristics of
Structured Data

 Dimensionality
 Curse of dimensionality

 Sparsity
 Only presence counts

 Resolution

Patterns depend on the scale
 Similarity
 Distance measure

August 10, 2009 Data Mining: Concepts and Techniques 5


Types of Attribute Values
 Nominal
 E.g., profession, ID numbers, eye color, zip

codes
 Ordinal
 E.g., rankings (e.g., army, professions), grades,

height in {tall, medium, short}


 Binary
 E.g., medical test (positive vs. negative)

 Interval
 E.g., calendar dates, body temperatures

 Ratio

E.g., temperature in Kelvin, length, time, counts
August 10, 2009 Data Mining: Concepts and Techniques 6
Discrete vs. Continuous
Attributes
 Discrete Attribute
 Has only a finite or countably infinite set of

values
 E.g., zip codes, profession, or the set of words in

a collection of documents
 Sometimes, represented as integer variables

 Note: Binary attributes are a special case of

discrete attributes
 Continuous Attribute

 Has real numbers as attribute values

 Examples: temperature, height, or weight

 Practically, real values can only be measured and

represented using a finite number of digits


 Continuous attributes are typically represented

as floating-point
August 10, 2009 variables
Data Mining: Concepts and Techniques 7
Chapter 2: Data
Preprocessing
 General data characteristics
 Basic data description and exploration
 Measuring data similarity
 Data cleaning
 Data integration and transformation
 Data reduction
 Summary

August 10, 2009 Data Mining: Concepts and Techniques 8


Mining Data Descriptive
Characteristics

 Motivation
 To better understand the data: central tendency,
variation and spread
 Data dispersion characteristics
 median, max, min, quantiles, outliers, variance, etc.
 Numerical dimensions correspond to sorted intervals
 Data dispersion: analyzed with multiple granularities
of precision
 Boxplot or quantile analysis on sorted intervals
 Dispersion analysis on computed measures
 Folding measures into numerical dimensions
 Boxplot or quantile analysis on the transformed cube
August 10, 2009 Data Mining: Concepts and Techniques 9
Measuring the Central Tendency
1 n
x = ∑xi µ= ∑x
 Mean (algebraic measure) (sample vs. population):
n i =1 N
n
 Weighted arithmetic mean: ∑w x i i
x= i =1
 Trimmed mean: chopping extreme values n

∑w
i =1
i
 Median: A holistic measure
 Middle value if odd number of values, or average of the
middle two values otherwise
Estimated by interpolation (for grouped data):
N / 2 − (∑ freq)l

 Mode median = L1 + ( ) width


freqmedian
 Value that occurs most frequently in the data
 Unimodal, bimodal, trimodal
 Empirical formula: mean − mode = 3 × (mean − median)
August 10, 2009 Data Mining: Concepts and Techniques 10
Symmetric vs. Skewed
Data
 Median, mean and mode of symmetric

symmetric, positively and


negatively skewed data

positively negatively
skewed skewed

August 10, 2009 Data Mining: Concepts and Techniques 11


Measuring the Dispersion of Data

 Quartiles, outliers and boxplots


 Quartiles: Q1 (25th percentile), Q3 (75th percentile)
 Inter-quartile range: IQR = Q3 – Q1
 Five number summary: min, Q1, M, Q3, max
 Boxplot: ends of the box are the quartiles, median is marked,
whiskers, and plot outlier individually
 Outlier: usually, a value higher/lower than 1.5 x IQR
 Variance and standard deviation (sample: s, population: σ)
 Variance: (algebraic, scalable computation)
1 n 1 n 2 1 n 1 n 1 n
s =
2

n − 1 i =1
2

( xi − x ) =
n − 1 i =1

[ xi − ( xi ) ]
n i =1
2
σ = ∑ ( xi − µ ) 2 =
2

N i =1 N
∑ xi − µ 2
i =1
2

 Standard deviation s (or σ) is the square root of variance s2 (or σ2)

August 10, 2009 Data Mining: Concepts and Techniques 12


Boxplot Analysis

 Five-number summary of a distribution:


Minimum, Q1, M, Q3, Maximum
 Boxplot
 Data is represented with a box
 The ends of the box are at the first and
third quartiles, i.e., the height of the box is
IQR
 The median is marked by a line within the
box
 Whiskers: two lines outside the box extend
to Minimum and
August 10, 2009
Maximum
Data Mining: Concepts and Techniques 13
Visualization of Data Dispersion: 3-D
Boxplots

August 10, 2009 Data Mining: Concepts and Techniques 14


Properties of Normal Distribution
Curve

 The normal (distribution) curve


 From μ–σ to μ+σ: contains about 68% of the

measurements (μ: mean, σ: standard deviation)


 From μ–2σ to μ+2σ: contains about 95% of it
 From μ–3σ to μ+3σ: contains about 99.7% of it

August 10, 2009 Data Mining: Concepts and Techniques 15


Graphic Displays of Basic Statistical
Descriptions
 Boxplot: graphic display of five-number summary
 Histogram: x-axis are values, y-axis repres.
frequencies
 Quantile plot: each value xi is paired with fi
indicating that approximately 100 fi % of data are
≤ xi
 Quantile-quantile (q-q) plot: graphs the quantiles
of one univariant distribution against the
corresponding quantiles of another
 Scatter plot: each pair of values is a pair of
coordinates and plotted as points in the plane
 Loess (local regression) curve: add a smooth curve
to a scatter plot to provide better perception of the
August 10, 2009 Data Mining: Concepts and Techniques 16
Histogram Analysis
 Graph displays of basic statistical class
descriptions
 Frequency histograms


A univariate graphical method

Consists of a set of rectangles that reflect the counts
or frequencies of the classes present in the given
data

August 10, 2009 Data Mining: Concepts and Techniques 17


Histograms Often Tells More than
Boxplots

 The two histograms


shown in the left
may have the same
boxplot
representation
 The same values
for: min, Q1,
median, Q3, max
 But they have
rather different data
distributions

August 10, 2009 Data Mining: Concepts and Techniques 18


Quantile Plot
 Displays all of the data (allowing the user to
assess both the overall behavior and unusual
occurrences)
 Plots quantile information
 For a data x data sorted in increasing order, f
i i
indicates that approximately 100 fi% of the data
are below or equal to the value xi

August 10, 2009 Data Mining: Concepts and Techniques 19


Quantile-Quantile (Q-Q) Plot
 Graphs the quantiles of one univariate distribution
against the corresponding quantiles of another
 Allows the user to view whether there is a shift in
going from one distribution to another

August 10, 2009 Data Mining: Concepts and Techniques 20


Scatter plot
 Provides a first look at bivariate data to see
clusters of points, outliers, etc
 Each pair of values is treated as a pair of
coordinates and plotted as points in the plane

August 10, 2009 Data Mining: Concepts and Techniques 21


Loess Curve
 Adds a smooth curve to a scatter plot in order to
provide better perception of the pattern of
dependence
 Loess curve is fitted by setting two parameters: a
smoothing parameter, and the degree of the
polynomials that are fitted by the regression

August 10, 2009 Data Mining: Concepts and Techniques 22


Positively and Negatively Correlated
Data

 The left half fragment is


positively correlated
 The right half is negative
correlated
August 10, 2009 Data Mining: Concepts and Techniques 23
Not Correlated Data

August 10, 2009 Data Mining: Concepts and Techniques 24


Data Visualization and Its
Methods
 Why data visualization?
 Gain insight into an information space by mapping data
onto graphical primitives
 Provide qualitative overview of large data sets
 Search for patterns, trends, structure, irregularities,
relationships among data
 Help find interesting regions and suitable parameters for
further quantitative analysis
 Provide a visual proof of computer representations derived
 Typical visualization methods:
 Geometric techniques
 Icon-based techniques
 Hierarchical techniques

August 10, 2009 Data Mining: Concepts and Techniques 25


Ribbons with Twists Based on Vorticity
Direct Data Visualization

August 10, 2009 Data Mining: Concepts and Techniques 26


Geometric Techniques

 Visualization of geometric transformations and


projections of the data
 Methods
 Landscapes
 Projection pursuit technique
 Finding meaningful projections of
multidimensional data
 Scatterplot matrices
 Prosection views
 Hyperslice
 Parallel coordinates
August 10, 2009 Data Mining: Concepts and Techniques 27
Scatterplot Matrices

Used by permission of M. Ward, Worcester Polytechnic Institute

Matrix of scatterplots (x-y-diagrams) of the k-dim. data [total of C(k, 2) = (k2 ̶ k)/2
scatterplots]
August 10, 2009 Data Mining: Concepts and Techniques 28
Landscapes
Used by permission of B. Wright, Visible Decisions Inc.

news articles
visualized as
a landscape

 Visualization of the data as perspective landscape


 The data needs to be transformed into a (possibly artificial)
2D spatial representation which preserves the characteristics
of the data
August 10, 2009 Data Mining: Concepts and Techniques 29
Parallel Coordinates
 n equidistant axes which are parallel to one of the screen
axes and correspond to the attributes
 The axes are scaled to the [minimum, maximum]: range of
the corresponding attribute
 Every data item corresponds to a polygonal line which
intersects each of the axes at the point which corresponds to
the value for the attribute

•  •  •

Attr. 1 Attr. 2 Attr. 3 Attr. k


August 10, 2009 Data Mining: Concepts and Techniques 30
Parallel Coordinates of a Data Set

August 10, 2009 Data Mining: Concepts and Techniques 31


Icon-based Techniques

 Visualization of the data values as features of icons


 Methods:
 Chernoff Faces
 Stick Figures
 Shape Coding:
 Color Icons:
 TileBars: The use of small icons representing the
relevance feature vectors in document retrieval

August 10, 2009 Data Mining: Concepts and Techniques 32


Chernoff Faces
 A way to display variables on a two-dimensional surface, e.g.,
let x be eyebrow slant, y be eye size, z be nose length, etc.
 The figure shows faces produced using 10 characteristics--
head eccentricity, eye size, eye spacing, eye eccentricity,
pupil size, eyebrow slant, nose size, mouth shape, mouth
size, and mouth opening): Each assigned one of 10 possible
values, generated using Mathematica (S. Dickson)
 REFERENCE: Gonick, L. and Smith, W.
The Cartoon Guide to Statistics. New
York: Harper Perennial, p. 212, 1993
 Weisstein, Eric W. "Chernoff Face."
From MathWorld--A Wolfram Web
Resource.
mathworld.wolfram.com/ChernoffFace.html
August 10, 2009 Data Mining: Concepts and Techniques 33
Stick Figures
used by permission of G. Grinstein, University of Massachusettes at Lowell

census data
showing age,
income, sex,
education, etc.

August 10, 2009 Data Mining: Concepts and Techniques 34


Hierarchical Techniques

 Visualization of the data using a


hierarchical partitioning into subspaces.
 Methods
 Dimensional Stacking
 Worlds-within-Worlds
 Treemap
 Cone Trees
 InfoCube

August 10, 2009 Data Mining: Concepts and Techniques 35


Dimensional Stacking

attribute 4
attribute 2

attribute 3

attribute 1

 Partitioning of the n-dimensional attribute space in


2-D subspaces which are ‘stacked’ into each other
 Partitioning of the attribute value ranges into

classes the important attributes should be used on


the outer levels
 Adequate for data with ordinal attributes of low

cardinality
AugustBut,
10, 2009difficult to display moreandthan nine dimensions

Data Mining: Concepts Techniques 36
Dimensional Stacking
Used by permission of M. Ward, Worcester Polytechnic Institute

Visualization of oil mining data with longitude and latitude mapped to the
outer x-, y-axes and ore grade and depth mapped to the inner x-, y-axes
August 10, 2009 Data Mining: Concepts and Techniques 37
Tree-Map
 Screen-filling method which uses a hierarchical
partitioning of the screen into regions depending on
the attribute values
 The x- and y-dimension of the screen are partitioned
alternately according to the attribute values
(classes)

MSR Netscan Image

August 10, 2009 Data Mining: Concepts and Techniques 38


Tree-Map of a File System
(Schneiderman)

August 10, 2009 Data Mining: Concepts and Techniques 39


Chapter 2: Data
Preprocessing
 General data characteristics
 Basic data description and exploration
 Measuring data similarity (Sec. 7.2)
 Data cleaning
 Data integration and transformation
 Data reduction
 Summary

August 10, 2009 Data Mining: Concepts and Techniques 40


Similarity and Dissimilarity
 Similarity
 Numerical measure of how alike two data objects

are
 Value is higher when objects are more alike

 Often falls in the range [0,1]

 Dissimilarity (i.e., distance)


 Numerical measure of how different are two data

objects
 Lower when objects are more alike

 Minimum dissimilarity is often 0

 Upper limit varies

 Proximity refers to a similarity or dissimilarity


August 10, 2009 Data Mining: Concepts and Techniques 41
Data Matrix and Dissimilarity
Matrix
 Data matrix
 n data points with  x11 ... x1f ... x1p 
 
p dimensions  ... ... ... ... ... 
 Two modes x ... xif ... xip 
 i1 
 ... ... ... ... ... 
x ... xnf ... xnp 
 n1 
 Dissimilarity matrix
 n data points, but
 0 
 d(2,1) 0 
registers only the  
 d(3,1) d ( 3,2) 0 
distance  
 A triangular matrix  : : : 
d ( n,1) d ( n,2) ... ... 0
 Single mode

August 10, 2009 Data Mining: Concepts and Techniques 42


Example: Data Matrix and Distance
Matrix

3
point x y
2 p1
p1 0 2
p3 p4
1
p2 2 0
p2 p3 3 1
0 p4 5 1
0 1 2 3 4 5 6 Data Matrix

p1 p2 p3 p4
p1 0 2.828 3.162 5.099
p2 2.828 0 1.414 3.162
p3 3.162 1.414 0 2
p4 5.099 3.162 2 0

Distance Matrix (i.e., Dissimilarity Matrix) for Euclidean


Distance

August 10, 2009 Data Mining: Concepts and Techniques 43


Minkowski Distance
 Minkowski distance: A popular distance measure

d (i, j) = q (| x − x |q + | x − x |q +...+ | x − x |q )
i1 j1 i2 j2 ip jp
where i = (xi1, xi2, …, xip) and j = (xj1, xj2, …, xjp)
are two p-dimensional data objects, and q is the
order
 Properties
 d(i, j) > 0 if i ≠ j, and d(i, i) = 0 (Positive
definiteness)
 d(i, j) = d(j, i) (Symmetry)
 d(i, j) ≤ d(i, k)Data
August 10, 2009
+Mining:
d(k,Concepts
j) (Triangle
and Techniques
Inequality) 44
Special Cases of Minkowski Distance
 q = 1: Manhattan (city block, L1 norm) distance
 E.g., the Hamming distance: the number of bits that
are different between two binary vectors
d (i, j) =| x − x | + | x − x | +...+ | x − x |
i1 j1 i2 j 2 ip jp

 q= 2: (L2 norm) Euclidean distance


d (i, j) = (| x − x |2 + | x − x |2 +...+ | x − x |2 )
i1 j1 i2 j 2 ip jp

 q → ∞. “supremum” (Lmax norm, L∞ norm) distance.


This is the maximum difference between any
component of the vectors
 Do not confuse q with n, i.e., all these distances are
defined for all numbers of dimensions.
 Also, one can use weighted distance, parametric Pearson
product moment correlation, or other dissimilarity
August 10, 2009 Data Mining: Concepts and Techniques 45
Example: Minkowski Distance
L1 p1 p2 p3 p4
p1 0 4 4 6
p2 4 0 2 4
p3 4 2 0 2
p4 6 4 2 0
point x y
p1 0 2 L2 p1 p2 p3 p4
p2 2 0 p1 0 2.828 3.162 5.099
p3 3 1 p2 2.828 0 1.414 3.162
p4 5 1 p3 3.162 1.414 0 2
p4 5.099 3.162 2 0

L∞ p1 p2 p3 p4
p1 0 2 3 5
p2 2 0 1 3
p3 3 1 0 2
p4 5 3 2 0

Distance Matrix

August 10, 2009 Data Mining: Concepts and Techniques 46


Interval-valued variables

 Standardize data

 Calculate the mean absolute deviation:


s f = 1n (| x1 f − m f | + | x2 f − m f | +...+ | xnf − m f |)

where m f = 1n (x1 f + x2 f + ... + xnf )


.

 Calculate the standardized measurement (z-score)


xif − m f
zif = sf
 Using mean absolute deviation is more robust than using

standard deviation

 Then calculate the Enclidean distance of other Minkowski

distance
August 10, 2009 Data Mining: Concepts and Techniques 47
Binary Variables Object j
1 0 sum
1 a b a +b
 A contingency table for binary Object i
data
0 c d c+d
sum a + c b + d p
 Distance measure for d (i, j) = b +c
symmetric binary variables:
a +b +c +d
 Distance measure for d (i, j) = b +c
asymmetric binary variables: a +b +c
 Jaccard coefficient (similarity simJaccard (i, j) = a
measure for asymmetric binary a+b+c
 variables):
Note: Jaccard coefficient is the same as “coherence”:

coherence(i, j) = sup(i, j) = a
sup(i) +sup( j) −sup(i, j) (a +b) + (a + c) − a
August 10, 2009 Data Mining: Concepts and Techniques 48
Dissimilarity between Binary
Variables
 Example
Name Gender Fever Cough Test-1 Test-2 Test-3 Test-4
Jack M Y N P N N N
Mary F Y N P N P N
Jim M Y P N N N N
 gender is a symmetric attribute
 the remaining attributes are asymmetric binary
 let the values Y and P be set to 1, and the value N be set
to 0 d ( jack , mary ) = 0 +1 =0.33
2 +0 +1
1 +1
d ( jack , jim ) = =0.67
1 +1 +1
1 +2
d ( jim , mary ) = =0.75
1 +1 +2
August 10, 2009 Data Mining: Concepts and Techniques 49
Nominal Variables

 A generalization of the binary variable in that it can


take more than 2 states, e.g., red, yellow, blue,
green
 Method 1: Simple matching
 m: # of matches,
d (i, j)p: p −m# of variables
=total p

 Method 2: Use a large number of binary variables


 creating a new binary variable for each of the M
nominal states

August 10, 2009 Data Mining: Concepts and Techniques 50


Ordinal Variables

 An ordinal variable can be discrete or continuous


 Order is important, e.g., rank
 Can be treated like interval-scaled
 replace x
if by their rank
rif ∈{1,..., M f }
 map the range of each variable onto [0, 1] by
replacing i-th object in the f-th variable by
rif − 1
zif =
Mf − 1

 compute the dissimilarity using methods for


interval-scaled variables

August 10, 2009 Data Mining: Concepts and Techniques 51


Ratio-Scaled Variables
 Ratio-scaled variable: a positive measurement on a
nonlinear scale, approximately at exponential
scale, such as AeBt or Ae-Bt
 Methods:
 treat them like interval-scaled variables—not a
good choice! (why?—the scale can be distorted)
 apply logarithmic transformation
yif = log(xif)
 treat them as continuous ordinal data treat their
rank as interval-scaled
August 10, 2009 Data Mining: Concepts and Techniques 52
Variables of Mixed Types
 A database may contain all the six types of
variables
 symmetric binary, asymmetric binary, nominal,

ordinal, interval and ratio


One may use a weighted formula
( f ) ( f ) to combine their
Σ δ

effects
p
d
d (i, j) = f = 1 ij ij
Σ pf = 1δ ij( f )

 f is binary or nominal:
dij(f) = 0 if xif = xjf , or dij(f) = 1 otherwise
 f is interval-based: use the normalized distance
if
zif = r −1
 f is ordinal or ratio-scaled
f M −1
 Compute ranks r and
August 10, 2009 if Concepts and Techniques
Data Mining: 53
Vector Objects: Cosine
Similarity
 Vector objects: keywords in documents, gene features in micro-
arrays, …
 Applications: information retrieval, biologic taxonomy, ...
 Cosine measure: If d1 and d2 are two vectors, then
cos(d1, d2) = (d1 • d2) /||d1|| ||d2|| ,
where • indicates vector dot product, ||d||: the length of
vector d
 Example:
d1 = 3 2 0 5 0 0 0 2 0 0
d2 = 1 0 0 0 0 0 0 1 0 2
d1•d2 = 3*1+2*0+0*0+5*0+0*0+0*0+0*0+2*1+0*0+0*2 = 5
||d1||=
(3*3+2*2+0*0+5*5+0*0+0*0+0*0+2*2+0*0+0*0)0.5=(42)0.
5
= 6.481
August 10, 2009 Data Mining: Concepts and Techniques 54
Chapter 2: Data
Preprocessing
 General data characteristics
 Basic data description and exploration
 Measuring data similarity
 Data cleaning
 Data integration and transformation
 Data reduction
 Summary

August 10, 2009 Data Mining: Concepts and Techniques 55


Major Tasks in Data Preprocessing
 Data cleaning
 Fill in missing values, smooth noisy data,

identify or remove outliers, and resolve


inconsistencies
 Data integration

 Integration of multiple databases, data cubes,

or files
 Data transformation

 Normalization and aggregation

 Data reduction

 Obtains reduced representation in volume but

produces the same or similar analytical results


 Data discretization: part of data reduction, of
August 10, 2009 Data Mining: Concepts and Techniques 56
Data Cleaning
 No quality data, no quality mining results!
 Quality decisions must be based on quality data
 e.g., duplicate or missing data may cause incorrect or
even misleading statistics
 “Data cleaning is the number one problem in data
warehousing”—DCI survey
 Data extraction, cleaning, and transformation comprises
the majority of the work of building a data warehouse
 Data cleaning tasks
 Fill in missing values
 Identify outliers and smooth out noisy data
 Correct inconsistent data
 Resolve redundancy caused by data integration

August 10, 2009 Data Mining: Concepts and Techniques 57


Data in the Real World Is Dirty
 incomplete: lacking attribute values, lacking
certain attributes of interest, or containing only
aggregate data
 e.g., occupation=“ ” (missing data)

 noisy: containing noise, errors, or outliers


 e.g., Salary=“−10” (an error)

 inconsistent: containing discrepancies in codes or


names, e.g.,
 Age=“42” Birthday=“03/07/1997”

 Was rating “1,2,3”, now rating “A, B, C”

 discrepancy between duplicate records

August 10, 2009 Data Mining: Concepts and Techniques 58


Why Is Data Dirty?
 Incomplete data may come from
 “Not applicable” data value when collected
 Different considerations between the time when the data
was collected and when it is analyzed.
 Human/hardware/software problems
 Noisy data (incorrect values) may come from
 Faulty data collection instruments
 Human or computer error at data entry
 Errors in data transmission
 Inconsistent data may come from
 Different data sources
 Functional dependency violation (e.g., modify some
linked data)
 Duplicate records also need data cleaning
August 10, 2009 Data Mining: Concepts and Techniques 59
Multi-Dimensional Measure of Data
Quality
 A well-accepted multidimensional view:
 Accuracy

 Completeness

 Consistency

 Timeliness

 Believability

 Value added

 Interpretability

 Accessibility

 Broad categories:
 Intrinsic, contextual, representational, and

accessibility

August 10, 2009 Data Mining: Concepts and Techniques 60


Missing Data

 Data is not always available


 E.g., many tuples have no recorded value for

several attributes, such as customer income in


sales data
 Missing data may be due to
 equipment malfunction

 inconsistent with other recorded data and thus

deleted
 data not entered due to misunderstanding

 certain data may not be considered important

at the time of entry


 not register history or changes of the data

August
 10, 2009 Data Mining: Concepts and Techniques 61
How to Handle Missing
Data?

 Ignore the tuple: usually done when class label is


missing (when doing classification)—not effective
when the % of missing values per attribute varies
considerably
 Fill in the missing value manually: tedious +
infeasible?
 Fill in it automatically with
 a global constant : e.g., “unknown”, a new
class?!
 the attribute mean
 the attribute mean for all samples belonging to
the same class:
August 10, 2009
smarter
Data Mining: Concepts and Techniques 62
Noisy Data
 Noise: random error or variance in a measured
variable
 Incorrect attribute values may due to
 faulty data collection instruments

 data entry problems

 data transmission problems

 technology limitation

 inconsistency in naming convention

 Other data problems which requires data cleaning


 duplicate records

 incomplete data

 inconsistent data

August 10, 2009 Data Mining: Concepts and Techniques 63


How to Handle Noisy Data?
 Binning
 first sort data and partition into (equal-

frequency) bins
 then one can smooth by bin means, smooth by

bin median, smooth by bin boundaries, etc.


 Regression

 smooth by fitting the data into regression

functions
 Clustering

 detect and remove outliers

 Combined computer and human inspection

 detect suspicious values and check by human

(e.g., deal with


August 10, 2009 Datapossible
Mining: Conceptsoutliers)
and Techniques 64
Simple Discretization Methods: Binning

 Equal-width (distance) partitioning


 Divides the range into N intervals of equal size: uniform
grid
 if A and B are the lowest and highest values of the
attribute, the width of intervals will be: W = (B –A)/N.
 The most straightforward, but outliers may dominate
presentation
 Skewed data is not handled well
 Equal-depth (frequency) partitioning
 Divides the range into N intervals, each containing
approximately same number of samples
August 10,Good data scaling
 2009 Data Mining: Concepts and Techniques 65
Binning Methods for Data
Smoothing
 Sorted data for price (in dollars): 4, 8, 9, 15, 21, 21, 24,
25, 26, 28, 29, 34
* Partition into equal-frequency (equi-depth) bins:
- Bin 1: 4, 8, 9, 15
- Bin 2: 21, 21, 24, 25
- Bin 3: 26, 28, 29, 34
* Smoothing by bin means:
- Bin 1: 9, 9, 9, 9
- Bin 2: 23, 23, 23, 23
- Bin 3: 29, 29, 29, 29
* Smoothing by bin boundaries:
- Bin 1: 4, 4, 4, 15
- Bin 2: 21, 21, 25, 25
- Bin 3: 26, 26, 26, 34
August 10, 2009 Data Mining: Concepts and Techniques 66
Regression

Y1

Y1’ y=x+1

X1 x

August 10, 2009 Data Mining: Concepts and Techniques 67


Cluster Analysis

August 10, 2009 Data Mining: Concepts and Techniques 68


Data Cleaning as a Process
 Data discrepancy detection
 Use metadata (e.g., domain, range, dependency,

distribution)
 Check field overloading

 Check uniqueness rule, consecutive rule and null rule

 Use commercial tools


Data scrubbing: use simple domain knowledge (e.g.,
postal code, spell-check) to detect errors and make
corrections
 Data auditing: by analyzing data to discover rules and

relationship to detect violators (e.g., correlation and


clustering to find outliers)
 Data migration and integration
 Data migration tools: allow transformations to be specified

 ETL (Extraction/Transformation/Loading) tools: allow users

to specify transformations through a graphical user


interface
August
 10, 2009 Data Mining: Concepts and Techniques 69
Chapter 2: Data
Preprocessing
 General data characteristics
 Basic data description and exploration
 Measuring data similarity
 Data cleaning
 Data integration and transformation
 Data reduction
 Summary

August 10, 2009 Data Mining: Concepts and Techniques 70


Data Integration
 Data integration:
 Combines data from multiple sources into a

coherent store
 Schema integration: e.g., A.cust-id ≡ B.cust-#
 Integrate metadata from different sources

 Entity identification problem:


 Identify real world entities from multiple data

sources, e.g., Bill Clinton = William Clinton


 Detecting and resolving data value conflicts
 For the same real world entity, attribute values

from different sources are different


 Possible reasons: different representations,

different scales, e.g., metric vs. British units

August 10, 2009 Data Mining: Concepts and Techniques 71


Handling Redundancy in Data
Integration

 Redundant data occur often when integration of


multiple databases
 Object identification: The same attribute or
object may have different names in different
databases
 Derivable data: One attribute may be a
“derived” attribute in another table, e.g.,
annual revenue
 Redundant attributes may be able to be detected
by correlation analysis
 Careful integration of the data from multiple
August sources
10, 2009 may help
Data reduce/avoid redundancies and
Mining: Concepts and Techniques 72
Correlation Analysis (Numerical Data)

 Correlation coefficient (also called Pearson’s


product moment coefficient)

rp ,q =
∑ ( p − p )(q − q ) ∑ ( pq ) − n p q
=
(n − 1)σ pσ q (n − 1)σ pσ q

where n is the number of tuples,


p q
and are the
respective means of p and q, σp and σq are the respective
standard deviation of p and q, and Σ(pq) is the sum of the
pq cross-product.
 If rp,q > 0, p and q are positively correlated (p’s
values increase as q’s). The higher, the stronger
correlation.
August 10, 2009 Data Mining: Concepts and Techniques 73
Correlation (viewed as linear
relationship)
 Correlation measures the linear relationship
between objects
 To compute correlation, we standardize
data objects, p and q, and then take their
dot product
pk′ = ( pk − mean( p)) / std ( p)

qk′ = ( qk − mean( q)) / std ( q)

correlation( p, q) = p′ • q′

August 10, 2009 Data Mining: Concepts and Techniques 74


Visually Evaluating Correlation

Scatter plots
showing the
similarity from
–1 to 1.

August 10, 2009 Data Mining: Concepts and Techniques 75


Correlation Analysis (Categorical
Data)
 Χ2 (chi-square) test
(Observed − Expected ) 2
χ2 = ∑
Expected
 The larger the Χ2 value, the more likely the
variables are related
 The cells that contribute the most to the Χ2 value
are those whose actual count is very different from
the expected count
 Correlation does not imply causality
 # of hospitals and # of car-theft in a city are correlated
 Both are causally linked to the third variable: population

August 10, 2009 Data Mining: Concepts and Techniques 76


Chi-Square Calculation: An
Example

Play Not play Sum


Like science fiction chess
250(90) chess
200(360) (row)
450

Not like science 50(210) 1000(840) 1050


fiction
Sum(col.) 300 1200 1500

 Χ2 (chi-square) calculation (numbers in parenthesis


are expected counts calculated based on the data
distribution in the two categories)
(250 − 90) 2 (50 − 210) 2 (200 − 360) 2 (1000 − 840) 2
χ =2
+ + + = 507.93
90 210 360 840
 It shows that like_science_fiction and play_chess
are correlated in the group
August 10, 2009 Data Mining: Concepts and Techniques 77
Data Transformation
 A function that maps the entire set of values of a
given attribute to a new set of replacement values
s.t. each old value can be identified with one of
the new values
 Methods

 Smoothing: Remove noise from data

 Aggregation: Summarization, data cube

construction
 Generalization: Concept hierarchy climbing

 Normalization: Scaled to fall within a small,

specified range
 min-max normalization

 z-score normalization


normalization by decimal scaling
August 10, Attribute/feature
 2009 construction
Data Mining: Concepts and Techniques 78
Data Transformation:
Normalization
 Min-max normalization: to [new_minA, new_maxA]
v − minA
v' = (new _ maxA − new _ minA) + new _ minA
maxA − minA
 Ex. Let income range $12,000 to $98,000
73,600 − 12,000 normalized to
(1.0 − 0) + 0 = 0.716
98,000 − 12
[0.0, 1.0]. Then $73,000 is mapped to
,000

 Z-score normalization (μ: mean, σ: standard deviation):


v −µA
v' =
σ A

73,600 − 54,000
= 1.225
 Ex. Let μ = 54,000, σ = 16,000. Then16,000
 Normalization by decimal scaling
v
v' = j Where j is the smallest integer such that Max(|ν’|) < 1
10
August 10, 2009 Data Mining: Concepts and Techniques 79
Chapter 2: Data
Preprocessing
 General data characteristics
 Basic data description and exploration
 Measuring data similarity
 Data cleaning
 Data integration and transformation
 Data reduction
 Summary

August 10, 2009 Data Mining: Concepts and Techniques 80


Data Reduction Strategies
 Why data reduction?
 A database/data warehouse may store terabytes of data

 Complex data analysis/mining may take a very long time

to run on the complete data set


 Data reduction: Obtain a reduced representation of the

data set that is much smaller in volume but yet produce the
same (or almost the same) analytical results
 Data reduction strategies

 Dimensionality reduction — e.g., remove unimportant

attributes
 Numerosity reduction (some simply call it: Data

Reduction)
 Data cub aggregation

 Data compression

 Regression

 Discretization
August 10, 2009 (and
Data concept
Mining: hierarchy
Concepts and Techniques generation) 81
Dimensionality Reduction
 Curse of dimensionality
 When dimensionality increases, data becomes increasingly

sparse
 Density and distance between points, which is critical to

clustering, outlier analysis, becomes less meaningful


 The possible combinations of subspaces will grow

exponentially
 Dimensionality reduction
 Avoid the curse of dimensionality

 Help eliminate irrelevant features and reduce noise

 Reduce time and space required in data mining

 Allow easier visualization

 Dimensionality reduction techniques


 Principal component analysis

 Singular value decomposition

August 10, 2009 Data Mining: Concepts and Techniques 82


Dimensionality Reduction: Principal
Component Analysis (PCA)
 Find a projection that captures the largest amount
of variation in data
 Find the eigenvectors of the covariance matrix,
and these eigenvectors define the new space
x2

x1
August 10, 2009 Data Mining: Concepts and Techniques 83
Principal Component Analysis
(Steps)
 Given N data vectors from n-dimensions, find k ≤ n
orthogonal vectors (principal components) that can be best
used to represent data
 Normalize input data: Each attribute falls within the same
range
 Compute k orthonormal (unit) vectors, i.e., principal
components
 Each input data (vector) is a linear combination of the k
principal component vectors
 The principal components are sorted in order of decreasing
“significance” or strength
 Since the components are sorted, the size of the data can
be reduced by eliminating the weak components, i.e.,
those with low variance
August 10, 2009 (i.e., using
Data Mining: Concepts the strongest principal
and Techniques 84
Feature Subset Selection
 Another way to reduce dimensionality of data
 Redundant features
 duplicate much or all of the information
contained in one or more other attributes
 E.g., purchase price of a product and the
amount of sales tax paid
 Irrelevant features
 contain no information that is useful for the data
mining task at hand
 E.g., students' ID is often irrelevant to the task
of predicting students' GPA

August 10, 2009 Data Mining: Concepts and Techniques 85


Heuristic Search in Feature Selection
 There are 2d possible feature combinations of d
features
 Typical heuristic feature selection methods:

 Best single features under the feature

independence assumption: choose by


significance tests
 Best step-wise feature selection:

 The best single-feature is picked first

 Then next best feature condition to the first,

...
 Step-wise feature elimination:

 Repeatedly eliminate the worst feature

 Best combined feature selection and elimination


August 10, 2009 Data Mining: Concepts and Techniques 86
Feature Creation
 Create new attributes that can capture the
important information in a data set much more
efficiently than the original attributes
 Three general methodologies

 Feature extraction


domain-specific
 Mapping data to new space (see: data

reduction)
 E.g., Fourier transformation, wavelet

transformation
 Feature construction

 Combining features


August 10, 2009
Data discretization
Data Mining: Concepts and Techniques 87
Mapping Data to a New Space
 Fourier transform
 Wavelet transform

Two Sine Waves Two Sine Waves + Noise Frequency

August 10, 2009 Data Mining: Concepts and Techniques 88


Numerosity (Data)
Reduction
 Reduce data volume by choosing alternative,
smaller forms of data representation
 Parametric methods (e.g., regression)
 Assume the data fits some model, estimate

model parameters, store only the parameters,


and discard the data (except possible outliers)
 Example: Log-linear models—obtain value at a

point in m-D space as the product on


appropriate marginal subspaces
 Non-parametric methods
 Do not assume models

 Major families: histograms, clustering,

sampling
August 10, 2009 Data Mining: Concepts and Techniques 89
Parametric Data Reduction:
Regression and Log-Linear
Models
 Linear regression: Data are modeled to fit a
straight line
 Often uses the least-square method to fit the
line
 Multiple regression: allows a response variable Y
to be modeled as a linear function of
multidimensional feature vector
 Log-linear model: approximates discrete
multidimensional
August 10, 2009 Data probability distributions
Mining: Concepts and Techniques 90
Regress Analysis and Log-
Linear Models
 Linear regression: Y = w X + b
 Two regression coefficients, w and b, specify the

line and are to be estimated by using the data


at hand
 Using the least squares criterion to the known

values of Y1, Y2, …, X1, X2, ….


 Multiple regression: Y = b0 + b1 X1 + b2 X2.
 Many nonlinear functions can be transformed
into the above
 Log-linear models:
 The multi-way table of joint probabilities is

approximated by a product of lower-order tables


Probability: p(a,
 2009
August 10, b, c,
Data Mining: d) and
Concepts α β χ
= Techniques δ
Data Reduction:
Wavelet Transformation
Haar2 Daubechie4
 Discrete wavelet transform (DWT): linear signal
processing, multi-resolutional analysis
 Compressed approximation: store only a small
fraction of the strongest of the wavelet coefficients
 Similar to discrete Fourier transform (DFT), but
better lossy compression, localized in space
 Method:
 Length, L, must be an integer power of 2 (padding with
0’s, when necessary)
 Each transform has 2 functions: smoothing, difference
 Applies to pairs of data, resulting in two set of data of
length L/2
 Applies two functions
August 10, 2009 recursively,
Data Mining: until reaches the
Concepts and Techniques 92
DWT for Image Compression

 Image

Low Pass High Pass

Low Pass High Pass

Low Pass High Pass

August 10, 2009 Data Mining: Concepts and Techniques 93


Data Cube Aggregation

 The lowest level of a data cube (base cuboid)


 The aggregated data for an individual entity of
interest
 E.g., a customer in a phone calling data
warehouse
 Multiple levels of aggregation in data cubes
 Further reduce the size of data to deal with
 Reference appropriate levels
 Use the smallest representation which is enough
to solve the task
 Queries regarding
August 10, 2009 Data aggregated information should
Mining: Concepts and Techniques 94
Data Compression
 String compression
 There are extensive theories and well-tuned

algorithms
 Typically lossless

 But only limited manipulation is possible without

expansion
 Audio/video compression

 Typically lossy compression, with progressive

refinement
 Sometimes small fragments of signal can be

reconstructed without reconstructing the whole


 Time sequence is not audio

 Typically short and vary slowly with time


August 10, 2009 Data Mining: Concepts and Techniques 95
Data Compression

Original Data Compressed


Data
lossless

s s y
lo
Original Data
Approximated

August 10, 2009 Data Mining: Concepts and Techniques 96


Data Reduction: Histograms
 Divide data into buckets and 40
store average (sum) for each
35
bucket
 Partitioning rules: 30
 Equal-width: equal bucket 25
range 20
 Equal-frequency (or equal-
15
depth)
10
 V-optimal: with the least
histogram variance 5
(weighted sum of the 0
original values that each

90000
20000

40000

70000
50000
10000

30000

60000

80000

100000
bucket represents)
August MaxDiff: set bucketData Mining: Concepts and Techniques
 10, 2009 97
Data Reduction Method:
Clustering
 Partition data set into clusters based on similarity,
and store cluster representation (e.g., centroid
and diameter) only
 Can be very effective if data is clustered but not if
data is “smeared”
 Can have hierarchical clustering and be stored in
multi-dimensional index tree structures
 There are many choices of clustering definitions
and clustering algorithms
 Cluster analysis will be studied in depth in
Chapter 7
August 10, 2009 Data Mining: Concepts and Techniques 98
Data Reduction Method: Sampling

 Sampling: obtaining a small sample s to represent


the whole data set N
 Allow a mining algorithm to run in complexity that
is potentially sub-linear to the size of the data
 Key principle: Choose a representative subset of
the data
 Simple random sampling may have very poor
performance in the presence of skew
 Develop adaptive sampling methods, e.g.,
stratified sampling:
 Note: Sampling may not reduce database I/Os
August 10, 2009 Data Mining: Concepts and Techniques 99
Types of Sampling

 Simple random sampling


 There is an equal probability of selecting any

particular item
 Sampling without replacement

 Once an object is selected, it is removed from

the population
 Sampling with replacement

 A selected object is not removed from the

population
 Stratified sampling:

 Partition the data set, and draw samples from

each partition (proportionally, i.e., approximately


the same percentage of the data)
 Used in conjunction with skewed data
August 10, 2009 Data Mining: Concepts and Techniques 100
Sampling: With or without
Replacement

W O R
SRS le random
i m p h o u t
( s e wi t
l
samp ment)
p l a c e
re

SRSW
R

Raw Data
August 10, 2009 Data Mining: Concepts and Techniques 101
Sampling: Cluster or Stratified
Sampling

Raw Data Cluster/Stratified Sample

August 10, 2009 Data Mining: Concepts and Techniques 102


Data Reduction: Discretization
 Three types of attributes:
 Nominal — values from an unordered set, e.g., color,
profession
 Ordinal — values from an ordered set, e.g., military or
academic rank
 Continuous — real numbers, e.g., integer or real numbers
 Discretization:
 Divide the range of a continuous attribute into intervals
 Some classification algorithms only accept categorical
attributes.
 Reduce data size by discretization
August 10,

2009 Data Mining: Concepts and Techniques 103
Discretization and Concept
Hierarchy

 Discretization
 Reduce the number of values for a given continuous
attribute by dividing the range of the attribute into
intervals
 Interval labels can then be used to replace actual data
values
 Supervised vs. unsupervised
 Split (top-down) vs. merge (bottom-up)
 Discretization can be performed recursively on an attribute
 Concept hierarchy formation
 Recursively reduce the data by collecting and replacing
low level concepts (such as numeric values for age) by
August 10, 2009 Data Mining: Concepts and Techniques 104
Discretization and Concept Hierarchy
Generation for Numeric Data
 Typical methods: All the methods can be applied recursively
 Binning (covered above)

Top-down split, unsupervised,
 Histogram analysis (covered above)
 Top-down split, unsupervised
 Clustering analysis (covered above)

Either top-down split or bottom-up merge, unsupervised
 Entropy-based discretization: supervised, top-down split
 Interval merging by χ2 Analysis: unsupervised, bottom-up
merge
 Segmentation by natural partitioning: top-down split,
August 10, 2009 Data Mining: Concepts and Techniques 105
Labels

 Entropy based approach

3 categories for both x and y 5 categories for both x and y

August 10, 2009 Data Mining: Concepts and Techniques 106


Entropy-Based Discretization
 Given a set of samples S, if S is partitioned into two intervals
S1 and S2 using boundary T, the information gain after
| | | |
partitioning is I ( S , T ) = S 1 Entropy ( S 1) + S 2 Entropy ( S 2)
|S| |S|

 Entropy is calculated based on class


m
distribution of the
samples in the set. Given
Entropy 1) =
( Sm −∑pi log
classes, the
2 ( pentropy
i) of S1 is
i =1

where pi is the probability of class i in S1


 The boundary that minimizes the entropy function over all
possible boundaries is selected as a binary discretization
 The process is recursively applied to partitions obtained until
some stopping criterion is met
 Such a boundary may reduce data size and improve
August 10, 2009 Data Mining: Concepts and Techniques 107
Class Labels

Data Equal interval width

Equal frequency K-means

August 10, 2009 Data Mining: Concepts and Techniques 108


Interval Merge by χ 2 Analysis
 Merging-based (bottom-up) vs. splitting-based methods
 Merge: Find the best neighboring intervals and merge them to
form larger intervals recursively
 ChiMerge [Kerber AAAI 1992, See also Liu et al. DMKD 2002]
 Initially, each distinct value of a numerical attr. A is
considered to be one interval
 χ2 tests are performed for every pair of adjacent intervals
 Adjacent intervals with the least χ2 values are merged
together, since low χ2 values for a pair indicate similar class
distributions
 This merge process proceeds recursively until a predefined
stopping criterion is met (such as significance level, max-
interval, max inconsistency,
August 10, 2009 etc.)
Data Mining: Concepts and Techniques 109
Segmentation by Natural
Partitioning

 A simply 3-4-5 rule can be used to segment


numeric data into relatively uniform, “natural”
intervals.
 If an interval covers 3, 6, 7 or 9 distinct values
at the most significant digit, partition the range
into 3 equi-width intervals
 If it covers 2, 4, or 8 distinct values at the most
significant digit, partition the range into 4
intervals
 If it covers 1, 5, or 10 distinct values at the
August 10, 2009 Data Mining: Concepts and Techniques 110
Example of 3-4-5 Rule
count

Step 1: -$351 -$159 profit $1,838 $4,700


Min Low (i.e, 5%-tile) High(i.e, 95%-0 tile) Max
Step 2: msd=1,000 Low=-$1,000 High=$2,000

(-$1,000 - $2,000)
Step 3:

(-$1,000 - 0) (0 -$ 1,000) ($1,000 - $2,000)

(-$400 -$5,000)
Step 4:

(-$400 - 0) ($2,000 - $5, 000)


(0 - $1,000) ($1,000 - $2, 000)
(0 -
(-$400 - ($1,000 -
$200)
$1,200) ($2,000 -
-$300)
($200 - $3,000)
($1,200 -
(-$300 - $400)
$1,400)
-$200) ($3,000 -
($400 - ($1,400 - $4,000)
(-$200 - $600) $1,600) ($4,000 -
-$100) $5,000)
($600 - ($1,600 -
$800) ($800 - ($1,800 -
$1,800)
(-$100 - $1,000) $2,000)
0)
August 10, 2009 Data Mining: Concepts and Techniques 111
Concept Hierarchy Generation for
Categorical Data
 Specification of a partial/total ordering of attributes
explicitly at the schema level by users or experts
 street < city < state < country
 Specification of a hierarchy for a set of values by
explicit data grouping
 {Urbana, Champaign, Chicago} < Illinois
 Specification of only a partial set of attributes
 E.g., only street < city, not others
 Automatic generation of hierarchies (or attribute
levels) by the analysis of the number of distinct
values
 E.g., for a set Data
August 10, 2009 of Mining:
attributes: {street, city, state,
Concepts and Techniques 112
Automatic Concept Hierarchy
Generation
 Some hierarchies can be automatically
generated based on the analysis of the number
of distinct values per attribute in the data set
 The attribute with the most distinct values is

placed at the lowest level of the hierarchy


 Exceptions, e.g., weekday, month, quarter,

year
country 15 distinct values

province_or_ state 365 distinct values

city 3567 distinct values

street 674,339 distinct values


August 10, 2009 Data Mining: Concepts and Techniques 113
Chapter 2: Data
Preprocessing
 General data characteristics
 Basic data description and exploration
 Measuring data similarity
 Data cleaning
 Data integration and transformation
 Data reduction
 Summary

August 10, 2009 Data Mining: Concepts and Techniques 114


Summary
 Data preparation/preprocessing: A big issue for data
mining
 Data description, data exploration, and measure
data similarity set the base for quality data
preprocessing
 Data preparation includes
 Data cleaning
 Data integration and data transformation
 Data reduction (dimensionality and numerosity
reduction)
 A lot a methods have been developed but data
August 10, 2009 Data Mining: Concepts and Techniques 115
References
 D. P. Ballou and G. K. Tayi. Enhancing data quality in data warehouse
environments. Communications of ACM, 42:73-78, 1999
 W. Cleveland, Visualizing Data, Hobart Press, 1993
 T. Dasu and T. Johnson. Exploratory Data Mining and Data Cleaning. John
Wiley, 2003
 T. Dasu, T. Johnson, S. Muthukrishnan, V. Shkapenyuk. 
Mining Database Structure; Or, How to Build a Data Quality Browser.
SIGMOD’02
 U. Fayyad, G. Grinstein, and A. Wierse. Information Visualization in Data Mining
and Knowledge Discovery, Morgan Kaufmann, 2001
 H. V. Jagadish et al., Special Issue on Data Reduction Techniques. Bulletin of
the Technical Committee on Data Engineering, 20(4), Dec. 1997
 D. Pyle. Data Preparation for Data Mining. Morgan Kaufmann, 1999
 E. Rahm and H. H. Do. Data Cleaning: Problems and Current Approaches. IEEE
Bulletin of the Technical Committee on Data Engineering. Vol.23, No.4
 V. Raman and J. Hellerstein. Potters Wheel: An Interactive Framework for Data
Cleaning and Transformation, VLDB’2001
 T. Redman. Data Quality: Management and Technology. Bantam Books, 1992
 E. R. Tufte. The Visual Display of Quantitative Information, 2nd ed., Graphics
Press, 2001
August

10, 2009 Data Mining: Concepts and Techniques 116
August 10, 2009 Data Mining: Concepts and Techniques 117
Feature Subset Selection
Techniques

 Brute-force approach:
 Try all possible feature subsets as input to data

mining algorithm
 Embedded approaches:
 Feature selection occurs naturally as part of the

data mining algorithm


 Filter approaches:
 Features are selected before data mining

algorithm is run
 Wrapper approaches:
 Use the data mining algorithm as a black box to

find best subset of attributes


August 10, 2009 Data Mining: Concepts and Techniques 118

You might also like