SlideShare a Scribd company logo
SPARSITY NORMALIZATION:
STABILIZING THE EXPECTED
OUTPUTS OF DEEP NETWORKS
2019. 06. 07.
JoonyoungYi
joonyoung.yi@kaist.ac.kr
2
• Many benchmark datasets differ in the sparsity between the data
instances.









• Variable sparsity problem: the expected value of the output layer
depends on 

the sparsity of the input data instance which makes the training difficult.
• Varying outputs for data instances with similar characteristics under
different sparsity.

VARIABLE SPARSITY PROBLEM
3
• Divide each input data instance by l0:
• So that outputs are not dependent on sparsity (can be applied to CNN
similarly).













• Sparsity Normalization solves various sparsity problem 

(theoretically, experimentally).
• Sparsity in a hidden layer is more stable after applying Sparsity Normalization.
SPARSITY NORMALIZATION
4
• Collaborative filtering datasets: Achieved states-of-the-arts
performance on Movielens 100K & 1M by simply applying Sparsity
Normalization to non-states-of-the-arts model.
• Electronic health records (EHR) dataset: Better AUC & orthogonal to
Dropout.









• Vision datasets: Better accuracy with less capacity & orthogonal to BN.









• 6 UCI datasets: better performance even compared to other missing
handling techniques.
EXPERIMENTAL RESULTS

More Related Content

More from Joonyoung Yi (9)

PDF
Mixture-Rank Matrix Approximation for Collaborative Filtering
Joonyoung Yi
 
PDF
Low-rank Matrix Approximation with Stability
Joonyoung Yi
 
PDF
Introduction to MAML (Model Agnostic Meta Learning) with Discussions
Joonyoung Yi
 
PDF
A Neural Autoregressive Approach to Collaborative Filtering (CF-NADE) Slide
Joonyoung Yi
 
PDF
Introduction to XGBoost
Joonyoung Yi
 
PDF
Why biased matrix factorization works well?
Joonyoung Yi
 
PDF
Dynamically Expandable Network (DEN)
Joonyoung Yi
 
PDF
Introduction to Low-rank Matrix Completion
Joonyoung Yi
 
PDF
Exact Matrix Completion via Convex Optimization Slide (PPT)
Joonyoung Yi
 
Mixture-Rank Matrix Approximation for Collaborative Filtering
Joonyoung Yi
 
Low-rank Matrix Approximation with Stability
Joonyoung Yi
 
Introduction to MAML (Model Agnostic Meta Learning) with Discussions
Joonyoung Yi
 
A Neural Autoregressive Approach to Collaborative Filtering (CF-NADE) Slide
Joonyoung Yi
 
Introduction to XGBoost
Joonyoung Yi
 
Why biased matrix factorization works well?
Joonyoung Yi
 
Dynamically Expandable Network (DEN)
Joonyoung Yi
 
Introduction to Low-rank Matrix Completion
Joonyoung Yi
 
Exact Matrix Completion via Convex Optimization Slide (PPT)
Joonyoung Yi
 

Recently uploaded (20)

PDF
Blockchain Transactions Explained For Everyone
CIFDAQ
 
PDF
The Builder’s Playbook - 2025 State of AI Report.pdf
jeroen339954
 
PPTX
WooCommerce Workshop: Bring Your Laptop
Laura Hartwig
 
PDF
Timothy Rottach - Ramp up on AI Use Cases, from Vector Search to AI Agents wi...
AWS Chicago
 
PDF
Reverse Engineering of Security Products: Developing an Advanced Microsoft De...
nwbxhhcyjv
 
PPT
Interview paper part 3, It is based on Interview Prep
SoumyadeepGhosh39
 
PPTX
Webinar: Introduction to LF Energy EVerest
DanBrown980551
 
PDF
Complete JavaScript Notes: From Basics to Advanced Concepts.pdf
haydendavispro
 
PDF
July Patch Tuesday
Ivanti
 
PDF
"Beyond English: Navigating the Challenges of Building a Ukrainian-language R...
Fwdays
 
PDF
SFWelly Summer 25 Release Highlights July 2025
Anna Loughnan Colquhoun
 
PDF
CIFDAQ Token Spotlight for 9th July 2025
CIFDAQ
 
PDF
Fl Studio 24.2.2 Build 4597 Crack for Windows Free Download 2025
faizk77g
 
PPTX
MSP360 Backup Scheduling and Retention Best Practices.pptx
MSP360
 
PDF
Learn Computer Forensics, Second Edition
AnuraShantha7
 
PDF
LLMs.txt: Easily Control How AI Crawls Your Site
Keploy
 
PPTX
"Autonomy of LLM Agents: Current State and Future Prospects", Oles` Petriv
Fwdays
 
PPTX
OpenID AuthZEN - Analyst Briefing July 2025
David Brossard
 
PDF
Presentation - Vibe Coding The Future of Tech
yanuarsinggih1
 
PDF
HCIP-Data Center Facility Deployment V2.0 Training Material (Without Remarks ...
mcastillo49
 
Blockchain Transactions Explained For Everyone
CIFDAQ
 
The Builder’s Playbook - 2025 State of AI Report.pdf
jeroen339954
 
WooCommerce Workshop: Bring Your Laptop
Laura Hartwig
 
Timothy Rottach - Ramp up on AI Use Cases, from Vector Search to AI Agents wi...
AWS Chicago
 
Reverse Engineering of Security Products: Developing an Advanced Microsoft De...
nwbxhhcyjv
 
Interview paper part 3, It is based on Interview Prep
SoumyadeepGhosh39
 
Webinar: Introduction to LF Energy EVerest
DanBrown980551
 
Complete JavaScript Notes: From Basics to Advanced Concepts.pdf
haydendavispro
 
July Patch Tuesday
Ivanti
 
"Beyond English: Navigating the Challenges of Building a Ukrainian-language R...
Fwdays
 
SFWelly Summer 25 Release Highlights July 2025
Anna Loughnan Colquhoun
 
CIFDAQ Token Spotlight for 9th July 2025
CIFDAQ
 
Fl Studio 24.2.2 Build 4597 Crack for Windows Free Download 2025
faizk77g
 
MSP360 Backup Scheduling and Retention Best Practices.pptx
MSP360
 
Learn Computer Forensics, Second Edition
AnuraShantha7
 
LLMs.txt: Easily Control How AI Crawls Your Site
Keploy
 
"Autonomy of LLM Agents: Current State and Future Prospects", Oles` Petriv
Fwdays
 
OpenID AuthZEN - Analyst Briefing July 2025
David Brossard
 
Presentation - Vibe Coding The Future of Tech
yanuarsinggih1
 
HCIP-Data Center Facility Deployment V2.0 Training Material (Without Remarks ...
mcastillo49
 
Ad

Sparsity Normalization: Stabilizing the Expected Outputs of Deep Networks

  • 1. SPARSITY NORMALIZATION: STABILIZING THE EXPECTED OUTPUTS OF DEEP NETWORKS 2019. 06. 07. JoonyoungYi [email protected]
  • 2. 2 • Many benchmark datasets differ in the sparsity between the data instances.
 
 
 
 
 • Variable sparsity problem: the expected value of the output layer depends on 
 the sparsity of the input data instance which makes the training difficult. • Varying outputs for data instances with similar characteristics under different sparsity.
 VARIABLE SPARSITY PROBLEM
  • 3. 3 • Divide each input data instance by l0: • So that outputs are not dependent on sparsity (can be applied to CNN similarly).
 
 
 
 
 
 
 • Sparsity Normalization solves various sparsity problem 
 (theoretically, experimentally). • Sparsity in a hidden layer is more stable after applying Sparsity Normalization. SPARSITY NORMALIZATION
  • 4. 4 • Collaborative filtering datasets: Achieved states-of-the-arts performance on Movielens 100K & 1M by simply applying Sparsity Normalization to non-states-of-the-arts model. • Electronic health records (EHR) dataset: Better AUC & orthogonal to Dropout.
 
 
 
 
 • Vision datasets: Better accuracy with less capacity & orthogonal to BN.
 
 
 
 
 • 6 UCI datasets: better performance even compared to other missing handling techniques. EXPERIMENTAL RESULTS