SlideShare a Scribd company logo
1 
Feature Selection in 
Hierarchical Feature Spaces 
10/12/2014 Petar Ristoski, Heiko Paulheim
Motivation: Linked Open Data as Background 
Knowledge 
• Linked Open Data is a method for publishing interlinked 
datasets using machine interpretable semantics 
• Started 2007 
• A collection of ~1,000 datasets 
– Various domains, e.g. general knowledge, government data, … 
– Using semantic web standards (HTTP, RDF, SPARQL) 
• Free of charge 
• Machine processable 
• Sophisticated tool stacks 
10/12/2014 Petar Ristoski, Heiko Paulheim 
2
Motivation: Linked Open Data as Background 
Knowledge 
10/12/2014 Petar Ristoski, Heiko Paulheim 
3
Example: the Auto MPG Dataset 
• A well-known UCI dataset 
– Goal: predict fuel consumption of cars 
• Hypothesis: background knowledge → more accurate predictions 
• Used background knowledge: 
– Entity types and categories from DBpedia (=Wikipedia) 
• Results: M5Rules down to almost half the prediction error 
– i.e. on average, we are wrong by 1.6 instead of 2.9 MPG 
Attribute set 
Linear Regression M5Rules 
RMSE RE RMSE RE 
original 3.359 0.118 2.859 0.088 
original + direct types 3.334 0.117 2.835 0.091 
original + categories 4.474 0.144 2.926 0.090 
original + direct types + categories 2.551 0.088 1.574 0.042 
10/12/2014 Petar Ristoski, Heiko Paulheim 4
Drawbacks 
• The generated feature sets are rather large 
– e.g. for dataset of 300 instances, it may generate up to 5,000 features 
from one source 
• Increase complexity and runtime 
• Overfitting for too specific features 
10/12/2014 Petar Ristoski, Heiko Paulheim 5
Linked Open Data is Backed by Ontologies 
LOD Graph Excerpt Ontology Excerpt 
10/12/2014 Petar Ristoski, Heiko Paulheim 6
HIERARCHICAL FEATURE 
SPACE 
10/12/2014 Petar Ristoski, Heiko Paulheim 7
Problem Statement 
• Each instance is an n-dimensional binary feature vector (v1,v2,…,vn), 
where vi ∈ {0,1} for all 1≤ vi ≤n 
• Feature space: V={v1,v2,…, vn} 
• Hierarchic relation between two features vi and vj can be denoted as 
vi < vj, where vi is more specific than vj 
• For all hierarchical features, the following implication holds: 
vi < vj → (vi = 1 → vj = 1) 
• Transitivity between hierarchical features exists: 
vi < vj ˄ vj < vk → vi < vk 
• The problem of feature selection can be defined as finding a 
projection of V to V’, where V’ ⊆ V and p(V’) ≥ p(V), where p is a 
performance function: 
푝: 푃 푉 → [0,1] 
10/12/2014 Petar Ristoski, Heiko Paulheim 8
Hierarchical Feature Space: Example 
Josh Donaldson is the best 3rd 
baseman in the American League. 
LeBron James NOT ranked #1 after 
newly released list of Top NBA players 
“Two things are infinite: the universe 
and human stupidity; and I'm not sure 
about the universe.”―Albert Einstein 
Nineteen-year-old figure skater Yuzuru 
Hanyu, who won a gold medal in the 
Sochi Olympics, is among the 684 
peo... http://bit.ly/1kb6W5y 
In his weekly address, President 
Barack Obama discusses expanding 
opportunity for hard-working 
Americans: http://ofa.bo/ccH 
Barack Obama cracks jokes at Vladimir 
Putin's expense http://dlvr.it/5Z7JCR 
I spotted the Lance Armstrong case in 
2006 when everyone thought he was 
God, and now this case catches my 
attention. 
10/12/2014 Petar Ristoski, Heiko Paulheim 9
Hierarchical Feature Space: Example 
dbpedia-owl: 
Basketball_Player 
dbpedia-owl: 
Baseball_Player 
dbpedia-owl: 
Athlete 
dbpedia:LeBron_James dbpedia:Josh_Donaldson 
Josh Donaldson is the best 3rd 
baseman in the American League. 
LeBron James NOT ranked #1 after 
newly released list of Top NBA players 
10/12/2014 Petar Ristoski, Heiko Paulheim 10
Hierarchical Feature Space: Example 
10/12/2014 Petar Ristoski, Heiko Paulheim 11
Hierarchical Feature Space 
• Linked Open Data 
– DBpedia, YAGO, Biperpedia, Google Knowledge Graph 
• Lexical Databses 
– WordNet, DANTE 
• Domain specific ontologies, taxonomies and vocabularies 
– Bioinformatics: Gene Ontology (GO), Entrez 
– Drugs: the Drug Ontology 
– E-commerce: GoodRelations 
10/12/2014 Petar Ristoski, Heiko Paulheim 12
RELATED APPROACHES 
10/12/2014 Petar Ristoski, Heiko Paulheim 13
Standard Feature Selection 
• Wrapper methods 
– Computationally expensive 
• Filter methods 
– Several techniques for scoring the relevance of the features 
• Information Gain 
• χ2 
• Information Gain Ratio 
• Gini Index 
– Often similar results 
10/12/2014 Petar Ristoski, Heiko Paulheim 14
Optimal Feature Selection 
10/12/2014 Petar Ristoski, Heiko Paulheim 15
Standard Feature Selection: Information Gain 
10/12/2014 Petar Ristoski, Heiko Paulheim 16
TSEL Feature Selection 
• Tree-based feature selection (Jeong et al.) 
– Select most representative and most effective feature from each branch 
of the hierarchy 
• 푙푖푓푡 = 
푃(푓|퐶) 
푃(퐶) 
10/12/2014 Petar Ristoski, Heiko Paulheim 17
Bottom-Up Hill-Climbing Feature Selection 
• Bottom-up hill climbing search algorithm to find an optimal subset of 
concepts for document representation (Wang et al.) 
푓 = 1 + 
α − 푛 
α 
∗ β ∗ 
푖∈퐷 
퐷푐푖 , 퐷푐푖⊆ 퐷퐾푁푁푖 푎푛푑 β > 0 
10/12/2014 Petar Ristoski, Heiko Paulheim 18
Greedy Top-Down Feature Selection 
• Greedy based top-down search strategy for feature selection (Lu et al.) 
– Select the most effective nodes from different levels of the hierarchy 
10/12/2014 Petar Ristoski, Heiko Paulheim 19
PROPOSED APPROACH 
10/12/2014 Petar Ristoski, Heiko Paulheim 20
Hierarchical Feature Selection Approach 
(SHSEL) 
• Exploit the hierarchical structure of the feature space 
• Hierarchical relation : vi < vj → (vi = 1 → vj = 1) 
• Relevance similarity: 
– Relevance (Blum et al.) : A feature vi is relevant to a target class C if 
there exists a pair of examples A and B in the instance space such that 
A and B differ only in their assignment to vi and C(A) ≠ C(B) 
• Two features vi and vj have similar relevance if: 
1 − 푅 푣푖 − 푅 푣푗 ≥ 푡, 푡 → [0,1] 
• Goal: Identify features with similar relevance, and select the most 
valuable abstract features, without losing predictive power 
10/12/2014 Petar Ristoski, Heiko Paulheim 21
Hierarchical Feature Selection Approach 
(SHSEL) 
• Initial Selection 
– Identify and filter out ranges of nodes with similar relevance in each 
branch of the hierarchy 
• Pruning 
– Select only the most relevant features from the previously reduced set 
10/12/2014 Petar Ristoski, Heiko Paulheim 22
Initial SHSEL Feature Selection 
1. Identify range of nodes with similar relevance in each branch: 
– Information 푠 푣, 푣gain: = 1 푠− (푣푖 0.45 , 푣푗 ) = − 1 0.5 − = 퐼퐺 0.95 
푣푖 − 퐼퐺(푣푗 ) 
푖 푗 – Correlation: t=0.9 
푠(푣푖 , 푣푗) = 퐶표푟푟푒푙푎푡푖표푛(푣푖 , 푣푗) 
s>t 
2. If the similarity is greater than a user specified threshold, remove 
the more specific feature, based on the hierarchical relation 
10/12/2014 Petar Ristoski, Heiko Paulheim 23
Post SHSEL Feature Selection 
• Select the features with the highest relevance on each path 
– user specified threshold 
– select features with relevance above path average relevance 
퐼퐺(푣푖)=0.2 
AVG(Sp)=0.25 
10/12/2014 Petar Ristoski, Heiko Paulheim 24
EVALUATION 
10/12/2014 Petar Ristoski, Heiko Paulheim 25
Evaluation 
• We use 5 real-world datasets and 6 synthetically generated datasets 
• Classification methods: 
– Naïve Bayes 
– k-Nearest Neighbors (k=3) 
– Support Vector Machine (polynomial kernel function) 
 No parameter optimization 
10/12/2014 Petar Ristoski, Heiko Paulheim 26
Evaluation: Real World Datasets 
Name Features #Instances Class Labels #Features 
Sports Tweets T DBpedia Direct Types 1,179 positive(523); negative(656) 4,082 
Sports Tweets C DBpedia Categories 1,179 positive(523); negative(656) 10,883 
Cities DBpedia Direct Types 212 high(67); medium(106); low(39) 727 
NY Daily Headings DBpedia Direct Types 1,016 positive(580); negative(436) 5,145 
StumbleUpon DMOZ Categories 3,020 positive(1,370); negative(1,650) 3,976 
• Hierarchical features are generated from DBpedia (structured version of 
Wikipedia) 
– The text is annotated with concepts using DBpedia Spotlight 
• The feature generation is independent of the class labels, and it is unbiased 
towards any of the feature selection approaches 
10/12/2014 Petar Ristoski, Heiko Paulheim 27
Evaluation: Synthetic Datasets 
Name #Instances Class Labels #Features 
S-D2-B2 1,000 positive(500); negative(500) 1,201 
S-D2-B5 1,000 positive(500); negative(500) 1,021 
S-D2-B10 1,000 positive(500); negative(500) 961 
S-D4-B2 1,000 positive(500); negative(500) 2,101 
S-D4-B4 1,000 positive(500); negative(500) 1,741 
S-D4-B10 1,000 positive(500); negative(500) 1,621 
• Generate the middle layer using polynomial function 
• Generate the hierarchy upwards and downwards following the 
hierarchical feature implication and transitivity rule 
• The depth and branching factor are controlled with parameters D 
and B 
10/12/2014 Petar Ristoski, Heiko Paulheim 28
Evaluation: Synthetic Datasets 
• Depth = 1 & Branching = 2 
1 
1 1 1 0 
1 0 1 1 0 1 0 0 
0 
0 1 0 0 1 0 0 
10/12/2014 Petar Ristoski, Heiko Paulheim 29
Evaluation: Synthetic Datasets 
Name #Instances Class Labels #Features 
S-D2-B2 1,000 positive(500); negative(500) 1,201 
S-D2-B5 1,000 positive(500); negative(500) 1,021 
S-D2-B10 1,000 positive(500); negative(500) 961 
S-D4-B2 1,000 positive(500); negative(500) 2,101 
S-D4-B4 1,000 positive(500); negative(500) 1,741 
S-D4-B10 1,000 positive(500); negative(500) 1,621 
• Generate the middle layer using polynomial function 
• Generate the hierarchy upwards and downwards following the 
hierarchical feature implication and transitivity rule 
• The depth and branching factor are controlled with parameters D 
and B 
10/12/2014 Petar Ristoski, Heiko Paulheim 30
Evaluation: Approach 
• Testing all approaches using two classification methods 
– Naïve Bayes, KNN and SVM 
• Metrics for performance evaluation 
– Accuracy: Acc V′ = 
퐶표푟푟푒푐푡푙푦 퐶푙푎푠푠푓푖푒푑 퐼푛푠푡푎푛푐푒푠 (푉′) 
푇표푡푎푙 푁푢푚푏푒푟 표푓 퐼푛푠푡푎푛푐푒푠 
– Feature Space Compression: 푐 푉′ = 1 − 
|푉′| 
|푉| 
– Harmonic Mean: 퐻 = 2 ∗ 
퐴푐푐 푉′ ∗푐 푉′ 
퐴푐푐 푉′ +푐 푉′ 
• Results calculated using stratified 10-fold cross validation 
– Feature selection is performed inside each fold 
• Parameter optimization for each feature selection strategy 
10/12/2014 Petar Ristoski, Heiko Paulheim 31
• Classification accuracy when using different relevance similarity threshold 
on the cities dataset 
100.00% 
90.00% 
80.00% 
70.00% 
60.00% 
50.00% 
40.00% 
30.00% 
20.00% 
10.00% 
0.00% 
Relevance Similarity Threshold 
Accuracy 
Compression 
H. Mean 
Evaluation: SHSEL IG 
10/12/2014 Petar Ristoski, Heiko Paulheim 32
Evaluation: Classification Accuracy (NB) 
10/12/2014 Petar Ristoski, Heiko Paulheim 33 
100.00% 
90.00% 
80.00% 
70.00% 
60.00% 
50.00% 
40.00% 
30.00% 
20.00% 
10.00% 
0.00% 
Sports Tweets T Sports Tweets C StumbleUpon Cities NY Daily Headings 
original 
initialSHSEL IG 
initialSHSEL C 
pruneSHSEL IG 
pruneSHSEL C 
SIG 
SC 
TSEL Lift 
TSEL IG 
HillClimbing 
GreedyTopDown 
100.00% 
90.00% 
80.00% 
70.00% 
60.00% 
50.00% 
40.00% 
30.00% 
20.00% 
10.00% 
0.00% 
S_D2_B2 S_D2_B5 S_D2_B10 S_D4_B2 S_D4_B5 S_D4_B10 
original 
initialSHSEL IG 
initialSHSEL C 
pruneSHSEL IG 
pruneSHSEL C 
SIG 
SC 
TSEL Lift 
TSEL IG 
HillClimbing 
GreedyTopDown
Evaluation: Feature Space Compression (NB) 
10/12/2014 Petar Ristoski, Heiko Paulheim 36 
100.00% 
90.00% 
80.00% 
70.00% 
60.00% 
50.00% 
40.00% 
30.00% 
20.00% 
10.00% 
0.00% 
Sports Tweets T Sports Tweets C StumbleUpon Cities NY Daily Headings 
initialSHSEL IG 
initialSHSEL C 
pruneSHSEL IG 
pruneSHSEL C 
SIG 
SC 
TSEL Lift 
TSEL IG 
HillClimbing 
GreedyTopDown 
100.00% 
90.00% 
80.00% 
70.00% 
60.00% 
50.00% 
40.00% 
30.00% 
20.00% 
10.00% 
0.00% 
S_D2_B2 S_D2_B5 S_D2_B10 S_D4_B2 S_D4_B5 S_D4_B10 
initialSHSEL IG 
initialSHSEL C 
pruneSHSEL IG 
pruneSHSEL C 
SIG 
SC 
TSEL Lift 
TSEL IG 
HillClimbing 
GreedyTopDown
Evaluation: Harmonic Mean (NB) 
10/12/2014 Petar Ristoski, Heiko Paulheim 39 
100.00% 
90.00% 
80.00% 
70.00% 
60.00% 
50.00% 
40.00% 
30.00% 
20.00% 
10.00% 
0.00% 
Sports Tweets T Sports Tweets C StumbleUpon Cities NY Daily Headings 
initialSHSEL IG 
initialSHSEL C 
pruneSHSEL IG 
pruneSHSEL C 
SIG 
SC 
TSEL Lift 
TSEL IG 
HillClimbing 
GreedyTopDown 
100.00% 
90.00% 
80.00% 
70.00% 
60.00% 
50.00% 
40.00% 
30.00% 
20.00% 
10.00% 
0.00% 
S_D2_B2 S_D2_B5 S_D2_B10 S_D4_B2 S_D4_B5 S_D4_B10 
initialSHSEL IG 
initialSHSEL C 
pruneSHSEL IG 
pruneSHSEL C 
SIG 
SC 
TSEL Lift 
TSEL IG 
HillClimbing 
GreedyTopDown
Conclusion & Outlook 
• Contribution 
– An approach that exploits hierarchies for feature selection in 
combination with standard metrics 
– The evaluation shows that the approach outperforms standard feature 
selection techniques, and other approaches using hierarchies 
• Future Work 
– Conduct further experiments 
• E.g. text mining, bioinformatics 
– Feature Selection in unsupervised learning 
• E.g. clustering, outlier detection 
• Laplacian Score 
10/12/2014 Petar Ristoski, Heiko Paulheim 43
44 
Feature Selection in 
Hierarchical Feature Spaces 
10/12/2014 Petar Ristoski, Heiko Paulheim

More Related Content

What's hot (19)

PDF
Link Discovery Tutorial Part I: Efficiency
Holistic Benchmarking of Big Linked Data
 
PDF
Tutorial "Linked Data Query Processing" Part 2 "Theoretical Foundations" (WWW...
Olaf Hartig
 
PDF
PhD thesis defense: Large-scale multilingual knowledge extraction, publishin...
Dimitris Kontokostas
 
PDF
Federation and Navigation in SPARQL 1.1
net2-project
 
PPT
Gathering Alternative Surface Forms for DBpedia Entities
Heiko Paulheim
 
PDF
Debunking some “RDF vs. Property Graph” Alternative Facts
Neo4j
 
PDF
Two graph data models : RDF and Property Graphs
andyseaborne
 
PPTX
Semantic Variation Graphs the case for RDF & SPARQL
Jerven Bolleman
 
PDF
Jesús Barrasa
Connected Data World
 
PPTX
Querying the Web of Data
Rinke Hoekstra
 
PDF
inteSearch: An Intelligent Linked Data Information Access Framework
National Inistitute of Informatics (NII), Tokyo, Japann
 
PDF
1.3 introduction to R language, importing dataset in r, data exploration in r
Simple Research
 
PPTX
A year on the Semantic Web @ W3C
Ivan Herman
 
PDF
20110728 datalift-rpi-troy
François Scharffe
 
PDF
Datalift a-catalyser-for-the-web-of-data-fosdem-05-02-2011
François Scharffe
 
PDF
RDF4U: RDF Graph Visualization by Interpreting Linked Data as Knowledge
National Institute of Informatics
 
PDF
Semantic Web Technology
Rathachai Chawuthai
 
PDF
Approximating Numeric Role Fillers via Predictive Clustering Trees for Know...
Giuseppe Rizzo
 
PDF
Li Pei-Temporal RL-VLDB2011
Pei Li
 
Link Discovery Tutorial Part I: Efficiency
Holistic Benchmarking of Big Linked Data
 
Tutorial "Linked Data Query Processing" Part 2 "Theoretical Foundations" (WWW...
Olaf Hartig
 
PhD thesis defense: Large-scale multilingual knowledge extraction, publishin...
Dimitris Kontokostas
 
Federation and Navigation in SPARQL 1.1
net2-project
 
Gathering Alternative Surface Forms for DBpedia Entities
Heiko Paulheim
 
Debunking some “RDF vs. Property Graph” Alternative Facts
Neo4j
 
Two graph data models : RDF and Property Graphs
andyseaborne
 
Semantic Variation Graphs the case for RDF & SPARQL
Jerven Bolleman
 
Jesús Barrasa
Connected Data World
 
Querying the Web of Data
Rinke Hoekstra
 
inteSearch: An Intelligent Linked Data Information Access Framework
National Inistitute of Informatics (NII), Tokyo, Japann
 
1.3 introduction to R language, importing dataset in r, data exploration in r
Simple Research
 
A year on the Semantic Web @ W3C
Ivan Herman
 
20110728 datalift-rpi-troy
François Scharffe
 
Datalift a-catalyser-for-the-web-of-data-fosdem-05-02-2011
François Scharffe
 
RDF4U: RDF Graph Visualization by Interpreting Linked Data as Knowledge
National Institute of Informatics
 
Semantic Web Technology
Rathachai Chawuthai
 
Approximating Numeric Role Fillers via Predictive Clustering Trees for Know...
Giuseppe Rizzo
 
Li Pei-Temporal RL-VLDB2011
Pei Li
 

Viewers also liked (6)

PPTX
Graph-based KNN Algorithm for Spam SMS Detection
SOYEON KIM
 
PPTX
Wrapper feature selection method
Amir Razmjou
 
PPT
Submodularity slides
dragonthu
 
PPTX
Nearest Neighbor Algorithm Zaffar Ahmed
Zaffar Ahmed Shaikh
 
PPTX
Analysis of Feature Selection Algorithms (Branch & Bound and Beam search)
Parinda Rajapaksha
 
PPT
12. Indexing and Hashing in DBMS
koolkampus
 
Graph-based KNN Algorithm for Spam SMS Detection
SOYEON KIM
 
Wrapper feature selection method
Amir Razmjou
 
Submodularity slides
dragonthu
 
Nearest Neighbor Algorithm Zaffar Ahmed
Zaffar Ahmed Shaikh
 
Analysis of Feature Selection Algorithms (Branch & Bound and Beam search)
Parinda Rajapaksha
 
12. Indexing and Hashing in DBMS
koolkampus
 
Ad

Similar to DS2014: Feature selection in hierarchical feature spaces (20)

PPTX
A Comparison of Propositionalization Strategies for Creating Features from Li...
Petar Ristoski
 
PPTX
Towards Automatic Classification of LOD Datasets
Blerina Spahiu
 
PDF
Iaetsd an efficient and large data base using subset selection algorithm
Iaetsd Iaetsd
 
PPT
Nbvtalkonfeatureselection
Nagasuri Bala Venkateswarlu
 
PDF
Feature selection for Big Data: advances and challenges by Verónica Bolón-Can...
Big Data Spain
 
PDF
IRJET- Survey of Feature Selection based on Ant Colony
IRJET Journal
 
PDF
The International Journal of Engineering and Science (The IJES)
theijes
 
PDF
Z suzanne van_den_bosch
Hoopeer Hoopeer
 
PPTX
User Interests Identification From Twitter using Hierarchical Knowledge Base
Artificial Intelligence Institute at UofSC
 
PDF
Computational Methods Of Feature Selection Huan Liu Hiroshi Motoda
rasabigley
 
PPTX
User Interests Identification From Twitter using Hierarchical Knowledge Base
Pavan Kapanipathi
 
PDF
Leveraging high level and low-level features for multimedia event detection.2...
Lu Jiang
 
PDF
Computational Methods of Feature Selection 1st Edition Huan Liu (Editor)
doweyhostel
 
PDF
Feature Selection.pdf
adarshbarnwal5
 
PDF
Feature Selection Algorithm for Supervised and Semisupervised Clustering
Editor IJCATR
 
PDF
M43016571
IJERA Editor
 
PDF
Computational Methods of Feature Selection 1st Edition Huan Liu (Editor)
lemanamiddag
 
DOCX
A fast clustering based feature subset selection algorithm for high-dimension...
IEEEFINALYEARPROJECTS
 
DOCX
JAVA 2013 IEEE DATAMINING PROJECT A fast clustering based feature subset sele...
IEEEGLOBALSOFTTECHNOLOGIES
 
A Comparison of Propositionalization Strategies for Creating Features from Li...
Petar Ristoski
 
Towards Automatic Classification of LOD Datasets
Blerina Spahiu
 
Iaetsd an efficient and large data base using subset selection algorithm
Iaetsd Iaetsd
 
Nbvtalkonfeatureselection
Nagasuri Bala Venkateswarlu
 
Feature selection for Big Data: advances and challenges by Verónica Bolón-Can...
Big Data Spain
 
IRJET- Survey of Feature Selection based on Ant Colony
IRJET Journal
 
The International Journal of Engineering and Science (The IJES)
theijes
 
Z suzanne van_den_bosch
Hoopeer Hoopeer
 
User Interests Identification From Twitter using Hierarchical Knowledge Base
Artificial Intelligence Institute at UofSC
 
Computational Methods Of Feature Selection Huan Liu Hiroshi Motoda
rasabigley
 
User Interests Identification From Twitter using Hierarchical Knowledge Base
Pavan Kapanipathi
 
Leveraging high level and low-level features for multimedia event detection.2...
Lu Jiang
 
Computational Methods of Feature Selection 1st Edition Huan Liu (Editor)
doweyhostel
 
Feature Selection.pdf
adarshbarnwal5
 
Feature Selection Algorithm for Supervised and Semisupervised Clustering
Editor IJCATR
 
M43016571
IJERA Editor
 
Computational Methods of Feature Selection 1st Edition Huan Liu (Editor)
lemanamiddag
 
A fast clustering based feature subset selection algorithm for high-dimension...
IEEEFINALYEARPROJECTS
 
JAVA 2013 IEEE DATAMINING PROJECT A fast clustering based feature subset sele...
IEEEGLOBALSOFTTECHNOLOGIES
 
Ad

Recently uploaded (20)

PDF
High-speedBouldersandtheDebrisFieldinDARTEjecta
Sérgio Sacani
 
PDF
Adding Geochemistry To Understand Recharge Areas - Kinney County, Texas - Jim...
Texas Alliance of Groundwater Districts
 
PPTX
Basal_ganglia_Structure_Function_Importance
muralinath2
 
PPTX
Q1 - W1 - D2 - Models of matter for science.pptx
RyanCudal3
 
PPTX
Bacillus thuringiensis.crops & golden rice
priyadharshini87125
 
PDF
Unit-3 ppt.pdf organic chemistry unit 3 heterocyclic
visionshukla007
 
DOCX
Paper - Suprasegmental Features (Makalah Presentasi)
Sahmiral Amri Rajagukguk
 
PDF
Plant growth promoting bacterial non symbiotic
psuvethapalani
 
DOCX
Critical Book Review (CBR) - "Hate Speech: Linguistic Perspectives"
Sahmiral Amri Rajagukguk
 
DOCX
Paper - Taboo Language (Makalah Presentasi)
Sahmiral Amri Rajagukguk
 
PDF
soil and environmental microbiology.pdf
Divyaprabha67
 
PDF
Service innovation with AI: Transformation of value proposition and market se...
Selcen Ozturkcan
 
PPTX
770043401-q1-Ppt-pe-and-Health-7-week-1-lesson-1.pptx
AizaRazonado
 
PDF
Annual report 2024 - Inria - English version.pdf
Inria
 
PPTX
Cerebellum_ Parts_Structure_Function.pptx
muralinath2
 
PPTX
Diagnostic Features of Common Oral Ulcerative Lesions.pptx
Dr Palak borade
 
PDF
The ALMA-CRISTAL survey: Gas, dust, and stars in star-forming galaxies when t...
Sérgio Sacani
 
PPTX
LESSON 2 PSYCHOSOCIAL DEVELOPMENT.pptx L
JeanCarolColico1
 
PPTX
Class12_Physics_Chapter2 electric potential and capacitance.pptx
mgmahati1234
 
PDF
Unit-5 ppt.pdf unit 5 organic chemistry 3
visionshukla007
 
High-speedBouldersandtheDebrisFieldinDARTEjecta
Sérgio Sacani
 
Adding Geochemistry To Understand Recharge Areas - Kinney County, Texas - Jim...
Texas Alliance of Groundwater Districts
 
Basal_ganglia_Structure_Function_Importance
muralinath2
 
Q1 - W1 - D2 - Models of matter for science.pptx
RyanCudal3
 
Bacillus thuringiensis.crops & golden rice
priyadharshini87125
 
Unit-3 ppt.pdf organic chemistry unit 3 heterocyclic
visionshukla007
 
Paper - Suprasegmental Features (Makalah Presentasi)
Sahmiral Amri Rajagukguk
 
Plant growth promoting bacterial non symbiotic
psuvethapalani
 
Critical Book Review (CBR) - "Hate Speech: Linguistic Perspectives"
Sahmiral Amri Rajagukguk
 
Paper - Taboo Language (Makalah Presentasi)
Sahmiral Amri Rajagukguk
 
soil and environmental microbiology.pdf
Divyaprabha67
 
Service innovation with AI: Transformation of value proposition and market se...
Selcen Ozturkcan
 
770043401-q1-Ppt-pe-and-Health-7-week-1-lesson-1.pptx
AizaRazonado
 
Annual report 2024 - Inria - English version.pdf
Inria
 
Cerebellum_ Parts_Structure_Function.pptx
muralinath2
 
Diagnostic Features of Common Oral Ulcerative Lesions.pptx
Dr Palak borade
 
The ALMA-CRISTAL survey: Gas, dust, and stars in star-forming galaxies when t...
Sérgio Sacani
 
LESSON 2 PSYCHOSOCIAL DEVELOPMENT.pptx L
JeanCarolColico1
 
Class12_Physics_Chapter2 electric potential and capacitance.pptx
mgmahati1234
 
Unit-5 ppt.pdf unit 5 organic chemistry 3
visionshukla007
 

DS2014: Feature selection in hierarchical feature spaces

  • 1. 1 Feature Selection in Hierarchical Feature Spaces 10/12/2014 Petar Ristoski, Heiko Paulheim
  • 2. Motivation: Linked Open Data as Background Knowledge • Linked Open Data is a method for publishing interlinked datasets using machine interpretable semantics • Started 2007 • A collection of ~1,000 datasets – Various domains, e.g. general knowledge, government data, … – Using semantic web standards (HTTP, RDF, SPARQL) • Free of charge • Machine processable • Sophisticated tool stacks 10/12/2014 Petar Ristoski, Heiko Paulheim 2
  • 3. Motivation: Linked Open Data as Background Knowledge 10/12/2014 Petar Ristoski, Heiko Paulheim 3
  • 4. Example: the Auto MPG Dataset • A well-known UCI dataset – Goal: predict fuel consumption of cars • Hypothesis: background knowledge → more accurate predictions • Used background knowledge: – Entity types and categories from DBpedia (=Wikipedia) • Results: M5Rules down to almost half the prediction error – i.e. on average, we are wrong by 1.6 instead of 2.9 MPG Attribute set Linear Regression M5Rules RMSE RE RMSE RE original 3.359 0.118 2.859 0.088 original + direct types 3.334 0.117 2.835 0.091 original + categories 4.474 0.144 2.926 0.090 original + direct types + categories 2.551 0.088 1.574 0.042 10/12/2014 Petar Ristoski, Heiko Paulheim 4
  • 5. Drawbacks • The generated feature sets are rather large – e.g. for dataset of 300 instances, it may generate up to 5,000 features from one source • Increase complexity and runtime • Overfitting for too specific features 10/12/2014 Petar Ristoski, Heiko Paulheim 5
  • 6. Linked Open Data is Backed by Ontologies LOD Graph Excerpt Ontology Excerpt 10/12/2014 Petar Ristoski, Heiko Paulheim 6
  • 7. HIERARCHICAL FEATURE SPACE 10/12/2014 Petar Ristoski, Heiko Paulheim 7
  • 8. Problem Statement • Each instance is an n-dimensional binary feature vector (v1,v2,…,vn), where vi ∈ {0,1} for all 1≤ vi ≤n • Feature space: V={v1,v2,…, vn} • Hierarchic relation between two features vi and vj can be denoted as vi < vj, where vi is more specific than vj • For all hierarchical features, the following implication holds: vi < vj → (vi = 1 → vj = 1) • Transitivity between hierarchical features exists: vi < vj ˄ vj < vk → vi < vk • The problem of feature selection can be defined as finding a projection of V to V’, where V’ ⊆ V and p(V’) ≥ p(V), where p is a performance function: 푝: 푃 푉 → [0,1] 10/12/2014 Petar Ristoski, Heiko Paulheim 8
  • 9. Hierarchical Feature Space: Example Josh Donaldson is the best 3rd baseman in the American League. LeBron James NOT ranked #1 after newly released list of Top NBA players “Two things are infinite: the universe and human stupidity; and I'm not sure about the universe.”―Albert Einstein Nineteen-year-old figure skater Yuzuru Hanyu, who won a gold medal in the Sochi Olympics, is among the 684 peo... http://bit.ly/1kb6W5y In his weekly address, President Barack Obama discusses expanding opportunity for hard-working Americans: http://ofa.bo/ccH Barack Obama cracks jokes at Vladimir Putin's expense http://dlvr.it/5Z7JCR I spotted the Lance Armstrong case in 2006 when everyone thought he was God, and now this case catches my attention. 10/12/2014 Petar Ristoski, Heiko Paulheim 9
  • 10. Hierarchical Feature Space: Example dbpedia-owl: Basketball_Player dbpedia-owl: Baseball_Player dbpedia-owl: Athlete dbpedia:LeBron_James dbpedia:Josh_Donaldson Josh Donaldson is the best 3rd baseman in the American League. LeBron James NOT ranked #1 after newly released list of Top NBA players 10/12/2014 Petar Ristoski, Heiko Paulheim 10
  • 11. Hierarchical Feature Space: Example 10/12/2014 Petar Ristoski, Heiko Paulheim 11
  • 12. Hierarchical Feature Space • Linked Open Data – DBpedia, YAGO, Biperpedia, Google Knowledge Graph • Lexical Databses – WordNet, DANTE • Domain specific ontologies, taxonomies and vocabularies – Bioinformatics: Gene Ontology (GO), Entrez – Drugs: the Drug Ontology – E-commerce: GoodRelations 10/12/2014 Petar Ristoski, Heiko Paulheim 12
  • 13. RELATED APPROACHES 10/12/2014 Petar Ristoski, Heiko Paulheim 13
  • 14. Standard Feature Selection • Wrapper methods – Computationally expensive • Filter methods – Several techniques for scoring the relevance of the features • Information Gain • χ2 • Information Gain Ratio • Gini Index – Often similar results 10/12/2014 Petar Ristoski, Heiko Paulheim 14
  • 15. Optimal Feature Selection 10/12/2014 Petar Ristoski, Heiko Paulheim 15
  • 16. Standard Feature Selection: Information Gain 10/12/2014 Petar Ristoski, Heiko Paulheim 16
  • 17. TSEL Feature Selection • Tree-based feature selection (Jeong et al.) – Select most representative and most effective feature from each branch of the hierarchy • 푙푖푓푡 = 푃(푓|퐶) 푃(퐶) 10/12/2014 Petar Ristoski, Heiko Paulheim 17
  • 18. Bottom-Up Hill-Climbing Feature Selection • Bottom-up hill climbing search algorithm to find an optimal subset of concepts for document representation (Wang et al.) 푓 = 1 + α − 푛 α ∗ β ∗ 푖∈퐷 퐷푐푖 , 퐷푐푖⊆ 퐷퐾푁푁푖 푎푛푑 β > 0 10/12/2014 Petar Ristoski, Heiko Paulheim 18
  • 19. Greedy Top-Down Feature Selection • Greedy based top-down search strategy for feature selection (Lu et al.) – Select the most effective nodes from different levels of the hierarchy 10/12/2014 Petar Ristoski, Heiko Paulheim 19
  • 20. PROPOSED APPROACH 10/12/2014 Petar Ristoski, Heiko Paulheim 20
  • 21. Hierarchical Feature Selection Approach (SHSEL) • Exploit the hierarchical structure of the feature space • Hierarchical relation : vi < vj → (vi = 1 → vj = 1) • Relevance similarity: – Relevance (Blum et al.) : A feature vi is relevant to a target class C if there exists a pair of examples A and B in the instance space such that A and B differ only in their assignment to vi and C(A) ≠ C(B) • Two features vi and vj have similar relevance if: 1 − 푅 푣푖 − 푅 푣푗 ≥ 푡, 푡 → [0,1] • Goal: Identify features with similar relevance, and select the most valuable abstract features, without losing predictive power 10/12/2014 Petar Ristoski, Heiko Paulheim 21
  • 22. Hierarchical Feature Selection Approach (SHSEL) • Initial Selection – Identify and filter out ranges of nodes with similar relevance in each branch of the hierarchy • Pruning – Select only the most relevant features from the previously reduced set 10/12/2014 Petar Ristoski, Heiko Paulheim 22
  • 23. Initial SHSEL Feature Selection 1. Identify range of nodes with similar relevance in each branch: – Information 푠 푣, 푣gain: = 1 푠− (푣푖 0.45 , 푣푗 ) = − 1 0.5 − = 퐼퐺 0.95 푣푖 − 퐼퐺(푣푗 ) 푖 푗 – Correlation: t=0.9 푠(푣푖 , 푣푗) = 퐶표푟푟푒푙푎푡푖표푛(푣푖 , 푣푗) s>t 2. If the similarity is greater than a user specified threshold, remove the more specific feature, based on the hierarchical relation 10/12/2014 Petar Ristoski, Heiko Paulheim 23
  • 24. Post SHSEL Feature Selection • Select the features with the highest relevance on each path – user specified threshold – select features with relevance above path average relevance 퐼퐺(푣푖)=0.2 AVG(Sp)=0.25 10/12/2014 Petar Ristoski, Heiko Paulheim 24
  • 25. EVALUATION 10/12/2014 Petar Ristoski, Heiko Paulheim 25
  • 26. Evaluation • We use 5 real-world datasets and 6 synthetically generated datasets • Classification methods: – Naïve Bayes – k-Nearest Neighbors (k=3) – Support Vector Machine (polynomial kernel function)  No parameter optimization 10/12/2014 Petar Ristoski, Heiko Paulheim 26
  • 27. Evaluation: Real World Datasets Name Features #Instances Class Labels #Features Sports Tweets T DBpedia Direct Types 1,179 positive(523); negative(656) 4,082 Sports Tweets C DBpedia Categories 1,179 positive(523); negative(656) 10,883 Cities DBpedia Direct Types 212 high(67); medium(106); low(39) 727 NY Daily Headings DBpedia Direct Types 1,016 positive(580); negative(436) 5,145 StumbleUpon DMOZ Categories 3,020 positive(1,370); negative(1,650) 3,976 • Hierarchical features are generated from DBpedia (structured version of Wikipedia) – The text is annotated with concepts using DBpedia Spotlight • The feature generation is independent of the class labels, and it is unbiased towards any of the feature selection approaches 10/12/2014 Petar Ristoski, Heiko Paulheim 27
  • 28. Evaluation: Synthetic Datasets Name #Instances Class Labels #Features S-D2-B2 1,000 positive(500); negative(500) 1,201 S-D2-B5 1,000 positive(500); negative(500) 1,021 S-D2-B10 1,000 positive(500); negative(500) 961 S-D4-B2 1,000 positive(500); negative(500) 2,101 S-D4-B4 1,000 positive(500); negative(500) 1,741 S-D4-B10 1,000 positive(500); negative(500) 1,621 • Generate the middle layer using polynomial function • Generate the hierarchy upwards and downwards following the hierarchical feature implication and transitivity rule • The depth and branching factor are controlled with parameters D and B 10/12/2014 Petar Ristoski, Heiko Paulheim 28
  • 29. Evaluation: Synthetic Datasets • Depth = 1 & Branching = 2 1 1 1 1 0 1 0 1 1 0 1 0 0 0 0 1 0 0 1 0 0 10/12/2014 Petar Ristoski, Heiko Paulheim 29
  • 30. Evaluation: Synthetic Datasets Name #Instances Class Labels #Features S-D2-B2 1,000 positive(500); negative(500) 1,201 S-D2-B5 1,000 positive(500); negative(500) 1,021 S-D2-B10 1,000 positive(500); negative(500) 961 S-D4-B2 1,000 positive(500); negative(500) 2,101 S-D4-B4 1,000 positive(500); negative(500) 1,741 S-D4-B10 1,000 positive(500); negative(500) 1,621 • Generate the middle layer using polynomial function • Generate the hierarchy upwards and downwards following the hierarchical feature implication and transitivity rule • The depth and branching factor are controlled with parameters D and B 10/12/2014 Petar Ristoski, Heiko Paulheim 30
  • 31. Evaluation: Approach • Testing all approaches using two classification methods – Naïve Bayes, KNN and SVM • Metrics for performance evaluation – Accuracy: Acc V′ = 퐶표푟푟푒푐푡푙푦 퐶푙푎푠푠푓푖푒푑 퐼푛푠푡푎푛푐푒푠 (푉′) 푇표푡푎푙 푁푢푚푏푒푟 표푓 퐼푛푠푡푎푛푐푒푠 – Feature Space Compression: 푐 푉′ = 1 − |푉′| |푉| – Harmonic Mean: 퐻 = 2 ∗ 퐴푐푐 푉′ ∗푐 푉′ 퐴푐푐 푉′ +푐 푉′ • Results calculated using stratified 10-fold cross validation – Feature selection is performed inside each fold • Parameter optimization for each feature selection strategy 10/12/2014 Petar Ristoski, Heiko Paulheim 31
  • 32. • Classification accuracy when using different relevance similarity threshold on the cities dataset 100.00% 90.00% 80.00% 70.00% 60.00% 50.00% 40.00% 30.00% 20.00% 10.00% 0.00% Relevance Similarity Threshold Accuracy Compression H. Mean Evaluation: SHSEL IG 10/12/2014 Petar Ristoski, Heiko Paulheim 32
  • 33. Evaluation: Classification Accuracy (NB) 10/12/2014 Petar Ristoski, Heiko Paulheim 33 100.00% 90.00% 80.00% 70.00% 60.00% 50.00% 40.00% 30.00% 20.00% 10.00% 0.00% Sports Tweets T Sports Tweets C StumbleUpon Cities NY Daily Headings original initialSHSEL IG initialSHSEL C pruneSHSEL IG pruneSHSEL C SIG SC TSEL Lift TSEL IG HillClimbing GreedyTopDown 100.00% 90.00% 80.00% 70.00% 60.00% 50.00% 40.00% 30.00% 20.00% 10.00% 0.00% S_D2_B2 S_D2_B5 S_D2_B10 S_D4_B2 S_D4_B5 S_D4_B10 original initialSHSEL IG initialSHSEL C pruneSHSEL IG pruneSHSEL C SIG SC TSEL Lift TSEL IG HillClimbing GreedyTopDown
  • 34. Evaluation: Feature Space Compression (NB) 10/12/2014 Petar Ristoski, Heiko Paulheim 36 100.00% 90.00% 80.00% 70.00% 60.00% 50.00% 40.00% 30.00% 20.00% 10.00% 0.00% Sports Tweets T Sports Tweets C StumbleUpon Cities NY Daily Headings initialSHSEL IG initialSHSEL C pruneSHSEL IG pruneSHSEL C SIG SC TSEL Lift TSEL IG HillClimbing GreedyTopDown 100.00% 90.00% 80.00% 70.00% 60.00% 50.00% 40.00% 30.00% 20.00% 10.00% 0.00% S_D2_B2 S_D2_B5 S_D2_B10 S_D4_B2 S_D4_B5 S_D4_B10 initialSHSEL IG initialSHSEL C pruneSHSEL IG pruneSHSEL C SIG SC TSEL Lift TSEL IG HillClimbing GreedyTopDown
  • 35. Evaluation: Harmonic Mean (NB) 10/12/2014 Petar Ristoski, Heiko Paulheim 39 100.00% 90.00% 80.00% 70.00% 60.00% 50.00% 40.00% 30.00% 20.00% 10.00% 0.00% Sports Tweets T Sports Tweets C StumbleUpon Cities NY Daily Headings initialSHSEL IG initialSHSEL C pruneSHSEL IG pruneSHSEL C SIG SC TSEL Lift TSEL IG HillClimbing GreedyTopDown 100.00% 90.00% 80.00% 70.00% 60.00% 50.00% 40.00% 30.00% 20.00% 10.00% 0.00% S_D2_B2 S_D2_B5 S_D2_B10 S_D4_B2 S_D4_B5 S_D4_B10 initialSHSEL IG initialSHSEL C pruneSHSEL IG pruneSHSEL C SIG SC TSEL Lift TSEL IG HillClimbing GreedyTopDown
  • 36. Conclusion & Outlook • Contribution – An approach that exploits hierarchies for feature selection in combination with standard metrics – The evaluation shows that the approach outperforms standard feature selection techniques, and other approaches using hierarchies • Future Work – Conduct further experiments • E.g. text mining, bioinformatics – Feature Selection in unsupervised learning • E.g. clustering, outlier detection • Laplacian Score 10/12/2014 Petar Ristoski, Heiko Paulheim 43
  • 37. 44 Feature Selection in Hierarchical Feature Spaces 10/12/2014 Petar Ristoski, Heiko Paulheim