TECHNIQUES FOR BIG DATA
FEATURE EXTRACTION USING
DISTANCE COVARIANCE
BASED PCA
Big Data
 Big Data' is a blanket term for any collection of data
sets so large and complex that it becomes difficult to
process using on-hand database management tools or
traditional data processing applications.
 Big data requires exceptional technologies to efficiently
process large quantities of data within tolerable
elapsed times. A 2011 McKinsey report suggests
suitable technologies include crowdsourcing, data fusion
and integration, genetic algorithms, machine learning,
natural language processing, signal processing,
simulation, time series analysis and visualization.
How Big is Big Data?
 Very large, distributed aggregations of loosely structured data – often
incomplete and inaccessible.
 Petabytes/exabytes of data Millions/billions of people Billions/trillions of
records.
 Loosely-structured and often distributed data.
 Flat schemas with few complex interrelationships
 Often involving time-stamped events
 Often made up of incomplete data
 Often including connections between data elements that must be
probabilistically inferred.
 Applications that involved Big-data can be: Transactional (e.g., Facebook,
PhotoBox), or, Analytic (e.g., ClickFox, Merced Applications).
 (Reference Wikibon.org)
Big Data
Big Data Can be of three types:
1. Large number of attributes (>16)
2. Large number of samples
3. Large number both of attributes and samples
I have tried to work on the first case.
What is Dimensionality Reduction?
 Dimensionality reduction or dimension reduction
is the process of reducing the number of random
variables under consideration (or attributes or
features or descriptors), and can be divided into
feature selection and feature extraction.
Feature Selection
 Filters: Pearson’s Correlation
 Wrappers: Run a classifier again and again, each
time with a new set of features selected using
backward selection or forward selection.
Feature Extraction
 Feature extraction transforms the data in the high-
dimensional space to a space of fewer dimensions.
The data transformation may be linear, as in
principal component analysis (PCA), but many
nonlinear dimensionality reduction techniques also
exist. For multidimensional data, tensor
representation can be used in dimensionality
reduction through multilinear subspace learning.
Feature Extraction
 The main linear technique for dimensionality
reduction, principal component analysis, performs a
linear mapping of the data to a lower-dimensional
space in such a way that the variance of the data in
the low-dimensional representation is maximized
What is Principal Component Analysis?
 Principal component analysis (PCA) is a statistical procedure that
uses an orthogonal transformation to convert a set of observations
of possibly correlated variables into a set of values of linearly
uncorrelated variables called principal components. The number of
principal components is less than or equal to the number of original
variables. This transformation is defined in such a way that the first
principal component has the largest possible variance (that is,
accounts for as much of the variability in the data as possible), and
each succeeding component in turn has the highest variance possible
under the constraint that it is orthogonal to (i.e., uncorrelated with)
the preceding components. Principal components are guaranteed to
be independent if the data set is jointly normally distributed. PCA is
sensitive to the relative scaling of the original variables.
That is fine, but show me the MATH!
 Online tutorial
(http://www.cs.otago.ac.nz/cosc453/student_tutori
als/principal_components.pdf)
PCA and BIG DATA
 BIG DATA containing thousands will require a lot of
computation time for an average computer.
 PCA becomes an important tool while drawing
inference from such large data sets.
What is Distance Correlation?
 Distance correlation is a measure of statistical
dependence between two random variables or two
random vectors of arbitrary, not necessarily equal
dimension. An important property is that this measure of
dependence is zero if and only if the random variables
are statistically independent. This measure is derived
from a number of other quantities that are used in its
specification, specifically: distance variance, distance
standard deviation and distance covariance. These
take the same roles as the ordinary moments with
corresponding names in the specification of the Pearson
product-moment correlation coefficient.
Distance Covariance Solved Example
Sample Data
Column 1 Column 2
1 1
2 0
-1 2
0 3
Mean 0.5 1.5
Distances
 For Column 1 (aij = pow((ai^2 – aj^2), 0.5))
0 1.73 0 1
1.73 0 1.73 2
0 1.73 0 1
1 2 1 0
Using Euclidean formula to calculate distances
Mean 0.62 1.365 0.62 1
0.62
1.365
0.62
1
Grand Mean : 0.932
Similarly
 Distances for column 2 (bij)
0 1 1.73 2.8
1 0 2 3
1.73 2 0 2.23
2.8 3 2.23 0
Mean
Mean 1.38 1.5 1.49
2.66
1.38
1.5
1.49
2.66
Grand Mean : 1.595
Centering both the columns
 Aij = aij – ~ai – ~aj + ~a;
 where
 ~ai = row mean of ai
 ~aj = column mean of aj
 ~a = grand mean of a
Aij
-0.308 0.677 -0.308 0.312
0.677 -1.668 0.677 0.567
-0.308 0.677 -0.308 0.312
0.312 0.567 0.312 -1.608
Similarly
 We can calculate Bij
 Distance Covariance = (Aij*Bij)/n^2
Distance Covariance Principal
Component Analysis
 After we have obtained distance covariance, we
can find the highest eigen vectors of the covariance
matrix and then use those eigen vectors to extract
new features
 These eigen vectors can be multiplied by the real
dataset to generate the reduced dataset.
PCA vs D-PCA
 The classical measure of dependence, the Pearson
correlation coefficient, is mainly sensitive to a linear
relationship between two variables. Distance correlation
was introduced in 2005 by Gabor J Szekely in several
lectures to address this deficiency of Pearson’s
correlation, namely that it can easily be zero for
dependent variables. Correlation = 0
(uncorrelatedness) does not imply independence while
distance correlation = 0 does imply independence. The
first results on distance correlation were published in
2007 and 2009.
Confusion Matrix
Modifications of D-PCA
 1. pow((ai^2 – aj^2),0.5)/ai+aj
 2. pow((ai^2 – aj^2),0.5)/ai
 These modification can be used to scale the data
which can then eliminate Normalization Step.
Results
Drawbacks
 Cannot handle time series data
 Cannot handle noisy data
 Assumes data distribution to be normal
 Sensitive to scaling of the data
Future work
 Rank correlation
 Distance based source separation

More Related Content

PDF
PCA (Principal component analysis)
PDF
Principal Component Analysis
PDF
Principal Component Analysis and Clustering
PPTX
Dimension Reduction: What? Why? and How?
PDF
Influence over the Dimensionality Reduction and Clustering for Air Quality Me...
PDF
Bank loan purchase modeling
PPTX
Data Reduction Stratergies
PCA (Principal component analysis)
Principal Component Analysis
Principal Component Analysis and Clustering
Dimension Reduction: What? Why? and How?
Influence over the Dimensionality Reduction and Clustering for Air Quality Me...
Bank loan purchase modeling
Data Reduction Stratergies

What's hot (18)

PDF
KNN and ARL Based Imputation to Estimate Missing Values
PPTX
Data reduction
PDF
Canonical correlation
PPT
Textmining Retrieval And Clustering
PDF
An intro to applied multi stat with r by everitt et al
PPTX
Cannonical Correlation
PDF
MPSKM Algorithm to Cluster Uneven Dimensional Time Series Subspace Data
PDF
Data Science - Part VII - Cluster Analysis
PPTX
Lect4 principal component analysis-I
PPTX
Datamining
PDF
IRJET- Supervised Learning Classification Algorithms Comparison
PPTX
Introduction to Linear Discriminant Analysis
PPTX
AI: Belief Networks
PDF
COMPARISION OF PERCENTAGE ERROR BY USING IMPUTATION METHOD ON MID TERM EXAMIN...
PPTX
03 Data Mining Techniques
PPTX
Lect 3 background mathematics
PPT
Cluster spss week7
PDF
IRJET- Performance Evaluation of Various Classification Algorithms
KNN and ARL Based Imputation to Estimate Missing Values
Data reduction
Canonical correlation
Textmining Retrieval And Clustering
An intro to applied multi stat with r by everitt et al
Cannonical Correlation
MPSKM Algorithm to Cluster Uneven Dimensional Time Series Subspace Data
Data Science - Part VII - Cluster Analysis
Lect4 principal component analysis-I
Datamining
IRJET- Supervised Learning Classification Algorithms Comparison
Introduction to Linear Discriminant Analysis
AI: Belief Networks
COMPARISION OF PERCENTAGE ERROR BY USING IMPUTATION METHOD ON MID TERM EXAMIN...
03 Data Mining Techniques
Lect 3 background mathematics
Cluster spss week7
IRJET- Performance Evaluation of Various Classification Algorithms
Ad

Viewers also liked (6)

PPTX
Theory and the review of related literature
PPTX
Image feature extraction
PPTX
Matlab Feature Extraction Using Segmentation And Edge Detection
PDF
Deep learning - Conceptual understanding and applications
PPTX
Introduction to Machine Learning
PDF
Deep Learning through Examples
Theory and the review of related literature
Image feature extraction
Matlab Feature Extraction Using Segmentation And Edge Detection
Deep learning - Conceptual understanding and applications
Introduction to Machine Learning
Deep Learning through Examples
Ad

Similar to Slides distancecovariance (20)

PPTX
M5.pptx
PPTX
Unit 3-L6.pptx introdction to machine learning
PDF
Machine Learning.pdf
PDF
Principal Component Analysis in Machine Learning.pdf
PPTX
Singular Value Decomposition (SVD).pptx
PPTX
EDAB Module 5 Singular Value Decomposition (SVD).pptx
PPTX
Principal Component Analysis in Machine learning.pptx
PDF
Survey on Feature Selection and Dimensionality Reduction Techniques
PDF
A study on rough set theory based
PDF
IRJET- Performance Evaluation of Various Classification Algorithms
PDF
A Novel Algorithm for Design Tree Classification with PCA
PDF
1376846406 14447221
PDF
High dimensionality reduction on graphical data
PPTX
Principal Component Analysis (PCA) machine Learning.
PDF
Density Based Clustering Approach for Solving the Software Component Restruct...
PDF
IRJET- Comparative Study of PCA, KPCA, KFA and LDA Algorithms for Face Re...
PDF
Building Azure Machine Learning Models
DOC
Bra a bidirectional routing abstraction for asymmetric mobile ad hoc networks...
PDF
Survey on Supervised Method for Face Image Retrieval Based on Euclidean Dist...
PDF
Working with the data for Machine Learning
M5.pptx
Unit 3-L6.pptx introdction to machine learning
Machine Learning.pdf
Principal Component Analysis in Machine Learning.pdf
Singular Value Decomposition (SVD).pptx
EDAB Module 5 Singular Value Decomposition (SVD).pptx
Principal Component Analysis in Machine learning.pptx
Survey on Feature Selection and Dimensionality Reduction Techniques
A study on rough set theory based
IRJET- Performance Evaluation of Various Classification Algorithms
A Novel Algorithm for Design Tree Classification with PCA
1376846406 14447221
High dimensionality reduction on graphical data
Principal Component Analysis (PCA) machine Learning.
Density Based Clustering Approach for Solving the Software Component Restruct...
IRJET- Comparative Study of PCA, KPCA, KFA and LDA Algorithms for Face Re...
Building Azure Machine Learning Models
Bra a bidirectional routing abstraction for asymmetric mobile ad hoc networks...
Survey on Supervised Method for Face Image Retrieval Based on Euclidean Dist...
Working with the data for Machine Learning

Recently uploaded (20)

PPTX
Unit IImachinemachinetoolopeartions.pptx
PDF
IAE-V2500 Engine for Airbus Family 319/320
PDF
Project_Mgmt_Institute_-Marc Marc Marc .pdf
PDF
Research on ultrasonic sensor for TTU.pdf
PPTX
Agentic Artificial Intelligence (Agentic AI).pptx
PDF
Unit1 - AIML Chapter 1 concept and ethics
PPTX
Principal presentation for NAAC (1).pptx
DOCX
An investigation of the use of recycled crumb rubber as a partial replacement...
PDF
UEFA_Carbon_Footprint_Calculator_Methology_2.0.pdf
PDF
IAE-V2500 Engine Airbus Family A319/320
PDF
Beginners-Guide-to-Artificial-Intelligence.pdf
PPTX
MAD Unit - 3 User Interface and Data Management (Diploma IT)
PDF
Micro 3 New.ppt.pdf tools the laboratory the method
PPT
UNIT-I Machine Learning Essentials for 2nd years
PDF
electrical machines course file-anna university
PPTX
Micro1New.ppt.pptx the main themes if micro
PDF
Lesson 3 .pdf
PPTX
CNS - Unit 1 (Introduction To Computer Networks) - PPT (2).pptx
PPTX
Environmental studies, Moudle 3-Environmental Pollution.pptx
PPTX
Chapter-8 Introduction to Quality Standards.pptx
Unit IImachinemachinetoolopeartions.pptx
IAE-V2500 Engine for Airbus Family 319/320
Project_Mgmt_Institute_-Marc Marc Marc .pdf
Research on ultrasonic sensor for TTU.pdf
Agentic Artificial Intelligence (Agentic AI).pptx
Unit1 - AIML Chapter 1 concept and ethics
Principal presentation for NAAC (1).pptx
An investigation of the use of recycled crumb rubber as a partial replacement...
UEFA_Carbon_Footprint_Calculator_Methology_2.0.pdf
IAE-V2500 Engine Airbus Family A319/320
Beginners-Guide-to-Artificial-Intelligence.pdf
MAD Unit - 3 User Interface and Data Management (Diploma IT)
Micro 3 New.ppt.pdf tools the laboratory the method
UNIT-I Machine Learning Essentials for 2nd years
electrical machines course file-anna university
Micro1New.ppt.pptx the main themes if micro
Lesson 3 .pdf
CNS - Unit 1 (Introduction To Computer Networks) - PPT (2).pptx
Environmental studies, Moudle 3-Environmental Pollution.pptx
Chapter-8 Introduction to Quality Standards.pptx

Slides distancecovariance

  • 1. TECHNIQUES FOR BIG DATA FEATURE EXTRACTION USING DISTANCE COVARIANCE BASED PCA
  • 2. Big Data  Big Data' is a blanket term for any collection of data sets so large and complex that it becomes difficult to process using on-hand database management tools or traditional data processing applications.  Big data requires exceptional technologies to efficiently process large quantities of data within tolerable elapsed times. A 2011 McKinsey report suggests suitable technologies include crowdsourcing, data fusion and integration, genetic algorithms, machine learning, natural language processing, signal processing, simulation, time series analysis and visualization.
  • 3. How Big is Big Data?  Very large, distributed aggregations of loosely structured data – often incomplete and inaccessible.  Petabytes/exabytes of data Millions/billions of people Billions/trillions of records.  Loosely-structured and often distributed data.  Flat schemas with few complex interrelationships  Often involving time-stamped events  Often made up of incomplete data  Often including connections between data elements that must be probabilistically inferred.  Applications that involved Big-data can be: Transactional (e.g., Facebook, PhotoBox), or, Analytic (e.g., ClickFox, Merced Applications).  (Reference Wikibon.org)
  • 4. Big Data Big Data Can be of three types: 1. Large number of attributes (>16) 2. Large number of samples 3. Large number both of attributes and samples I have tried to work on the first case.
  • 5. What is Dimensionality Reduction?  Dimensionality reduction or dimension reduction is the process of reducing the number of random variables under consideration (or attributes or features or descriptors), and can be divided into feature selection and feature extraction.
  • 6. Feature Selection  Filters: Pearson’s Correlation  Wrappers: Run a classifier again and again, each time with a new set of features selected using backward selection or forward selection.
  • 7. Feature Extraction  Feature extraction transforms the data in the high- dimensional space to a space of fewer dimensions. The data transformation may be linear, as in principal component analysis (PCA), but many nonlinear dimensionality reduction techniques also exist. For multidimensional data, tensor representation can be used in dimensionality reduction through multilinear subspace learning.
  • 8. Feature Extraction  The main linear technique for dimensionality reduction, principal component analysis, performs a linear mapping of the data to a lower-dimensional space in such a way that the variance of the data in the low-dimensional representation is maximized
  • 9. What is Principal Component Analysis?  Principal component analysis (PCA) is a statistical procedure that uses an orthogonal transformation to convert a set of observations of possibly correlated variables into a set of values of linearly uncorrelated variables called principal components. The number of principal components is less than or equal to the number of original variables. This transformation is defined in such a way that the first principal component has the largest possible variance (that is, accounts for as much of the variability in the data as possible), and each succeeding component in turn has the highest variance possible under the constraint that it is orthogonal to (i.e., uncorrelated with) the preceding components. Principal components are guaranteed to be independent if the data set is jointly normally distributed. PCA is sensitive to the relative scaling of the original variables.
  • 10. That is fine, but show me the MATH!  Online tutorial (http://www.cs.otago.ac.nz/cosc453/student_tutori als/principal_components.pdf)
  • 11. PCA and BIG DATA  BIG DATA containing thousands will require a lot of computation time for an average computer.  PCA becomes an important tool while drawing inference from such large data sets.
  • 12. What is Distance Correlation?  Distance correlation is a measure of statistical dependence between two random variables or two random vectors of arbitrary, not necessarily equal dimension. An important property is that this measure of dependence is zero if and only if the random variables are statistically independent. This measure is derived from a number of other quantities that are used in its specification, specifically: distance variance, distance standard deviation and distance covariance. These take the same roles as the ordinary moments with corresponding names in the specification of the Pearson product-moment correlation coefficient.
  • 13. Distance Covariance Solved Example Sample Data Column 1 Column 2 1 1 2 0 -1 2 0 3 Mean 0.5 1.5
  • 14. Distances  For Column 1 (aij = pow((ai^2 – aj^2), 0.5)) 0 1.73 0 1 1.73 0 1.73 2 0 1.73 0 1 1 2 1 0 Using Euclidean formula to calculate distances Mean 0.62 1.365 0.62 1 0.62 1.365 0.62 1 Grand Mean : 0.932
  • 15. Similarly  Distances for column 2 (bij) 0 1 1.73 2.8 1 0 2 3 1.73 2 0 2.23 2.8 3 2.23 0 Mean Mean 1.38 1.5 1.49 2.66 1.38 1.5 1.49 2.66 Grand Mean : 1.595
  • 16. Centering both the columns  Aij = aij – ~ai – ~aj + ~a;  where  ~ai = row mean of ai  ~aj = column mean of aj  ~a = grand mean of a
  • 17. Aij -0.308 0.677 -0.308 0.312 0.677 -1.668 0.677 0.567 -0.308 0.677 -0.308 0.312 0.312 0.567 0.312 -1.608
  • 18. Similarly  We can calculate Bij  Distance Covariance = (Aij*Bij)/n^2
  • 19. Distance Covariance Principal Component Analysis  After we have obtained distance covariance, we can find the highest eigen vectors of the covariance matrix and then use those eigen vectors to extract new features  These eigen vectors can be multiplied by the real dataset to generate the reduced dataset.
  • 20. PCA vs D-PCA  The classical measure of dependence, the Pearson correlation coefficient, is mainly sensitive to a linear relationship between two variables. Distance correlation was introduced in 2005 by Gabor J Szekely in several lectures to address this deficiency of Pearson’s correlation, namely that it can easily be zero for dependent variables. Correlation = 0 (uncorrelatedness) does not imply independence while distance correlation = 0 does imply independence. The first results on distance correlation were published in 2007 and 2009.
  • 22. Modifications of D-PCA  1. pow((ai^2 – aj^2),0.5)/ai+aj  2. pow((ai^2 – aj^2),0.5)/ai  These modification can be used to scale the data which can then eliminate Normalization Step.
  • 24. Drawbacks  Cannot handle time series data  Cannot handle noisy data  Assumes data distribution to be normal  Sensitive to scaling of the data
  • 25. Future work  Rank correlation  Distance based source separation