SlideShare a Scribd company logo
Learning to learn unlearned feature
for segmentation
Sungyeob Han
Communication and Machine Learning Lab.
Seoul National University
Introduction
• How to transfer with few samples?
Primary cancer Brain metastasis
Outline
1. Training segmentation network
2. Meta-learning
3. Active learning
4. Active meta-tune
5. Applications
Learning to learn unlearned feature for segmentation
Fully convolutional networks
• take input of arbitrary size and produce correspondingly-
sized output
• a feed-forward propagation predicts the labels
• end-to-end, pixel-to-pixel
Long, Jonathan, Evan Shelhamer, and Trevor Darrell. "Fully convolutional networks for semantic segmentation."
Pyramid Scene Parsing Network
pyramid parsing module : harvest different sub-region representations
concatenation : upsampling and concatenation layers
Zhao, Hengshuang, et al. "Pyramid scene parsing network."
Zhao, Hengshuang, et al. "Pyramid scene parsing network."
Details in training segmentation network
• Fast feed-forward time (FCN-based)
• Given the pre-trained encoding parameters (VGGNet), fine-tuning
in stages takes 36 hours on a single GPU.
• Ambiguity : object structure, sparse label
• constrained categories
average loss gives blurry gradient for each
category information
Outline
1. Training segmentation network
2. Meta-learning
3. Active learning
4. Active meta-tune
5. Applications
Learning to learn
• A key aspect of intelligence : versatility
• the capability of doing many different things.
• Meta-learning
• As known as ”learning to learn”
• learn how to learn new tasks faster by reusing previous experience
Few-shot Learning
• In 2015, Brenden et al. show how to learn new concepts from one
or a few instances of that concept.
• learn to learn from
a few examples
• Omniglot
• 1623 character classes
• each with 20 examples
Lake, Brenden M., Ruslan Salakhutdinov, and Joshua B. Tenenbaum. "Human-level concept learning through probabilistic program
induction."
Meta-learner
Ravi, Sachin, and Hugo Larochelle. "Optimization as a model for few-shot learning.“
Meta-learning set-up for few-shot image classification
• 1-shot, 5-class classification task
• one example from each of 5 classes
Model agnostic meta-learning
• Motivation : ambiguity on new task
• From a dynamical systems standpoint
• Maximize the sensitivity of the loss functions of new tasks
Model Agnostic Meta-Learning
• A model 𝑓
• Maps observation 𝐱 to outputs 𝐚
• Each task 𝒯 = ℒ 𝐱1, 𝐚1, ⋯ , 𝐱 𝐻, 𝐚 𝐻 , 𝑞 𝐱1 , 𝑞 𝐱 𝑡+1|𝐱 𝑡, 𝐚 𝑡 , 𝐻
• Loss, Initial distribution, a transition dist., episode length
• ℒ 𝐱1, 𝐚1, ⋯ , 𝐱 𝐻, 𝐚 𝐻
• Task-specific feedback
• A misclassification loss
Algorithm : MAML
• K-shot learning
• Learn a new Task 𝓣𝒊 sampled from 𝒑 𝓣
from only K samples sampled from 𝒒𝒊
• Feedback ℒ 𝑇 𝑖
generated by 𝒯𝑖
• Train with K samples
• Test on new samples from 𝒯𝑖
• improved by considering how the test error
on new data from 𝑞𝑖 changes with respect to
the parameters.
• the test error on sampled tasks 𝒯𝑖 serves as
the training error of the meta-learning
process.
Algorithm : MAML
• A model with parameters 𝑓𝜃 , adapting to an
new task 𝒯𝑖 ∶ 𝜃 → 𝜃𝑖′
𝜃𝑖
′
= 𝜃 − 𝛼𝛻𝜃ℒ 𝒯𝑖
(𝑓𝜃)
• Meta-objective : the model parameters are
trained by optimizing for performance of 𝑓 𝜃′
min
𝜃
෍
𝒯𝑖~𝑝(𝒯)
ℒ 𝒯𝑖
𝑓 𝜃𝑖
′ = ෍
𝒯𝑖~𝑝(𝒯)
ℒ 𝒯𝑖
𝑓𝜃−𝛼𝛻 𝜃ℒ 𝑇 𝑖
(𝑓 𝜃)
• The meta-optimization across tasks
𝜃 ← 𝜃 − 𝛽𝛻𝜃 ෍
𝒯𝑖~𝑝(𝒯)
ℒ 𝒯𝑖
𝑓 𝜃𝑖
′
Experimental Evaluation : Regression
Experimental Evaluation : Supervised classification
Experimental Evaluation : RL
Desinging of meta-learning
Levine, Sergey, and Chelsea Finn, “Meta-learning frontiers: universal, uncertain, and unsupervised.”
Meta-learning with ambiguity
Levine, Sergey, and Chelsea Finn, “Meta-learning frontiers: universal, uncertain, and unsupervised.”
Meta-learning with ambiguity
Levine, Sergey, and Chelsea Finn, “Meta-learning frontiers: universal, uncertain, and unsupervised.”
Outline
1. Training segmentation network
2. Meta-learning
3. Active learning
4. Active meta-tune
5. Applications
Active learning
• pool-based active learning
• queries are selected from
a large pool of unlabeled
instances
• uncertainty sampling
query strategy
• select the instance in the
pool about which the
model is least certain
how to label
Settles, Burr “Active learning literature survey.”
pool-based active learning
• A toy example : logistic regression (400 instances)
• (b) : 30 labeled instances randomly drawn : 70% accuracy
• (c) : 30 actively queried instances using uncertainty sampling : 90%
Uncertainty sampling
• uncertainty measures
Mussmann, Stephen, and Percy Liang. "On the Relationship between Data Efficiency and Error for Uncertainty Sampling.”
How can me measure the uncertainty of networks?
• a small dataset for a new task can simply be too ambiguous to
acquire a single model
Kendall, Alex, and Yarin Gal. "What uncertainties do we need in bayesian deep learning for computer vision?.“
Outline
1. Training segmentation network
2. Meta-learning
3. Active learning
4. Active meta-tune
5. Applications
Catastrophic forgetting
Generate training tasks good task
bad task
𝑻 𝒈𝒐𝒐𝒅
𝑻 𝒃𝒂𝒅
Han, Sungyeob, et al. “Learning to Learn Unlearned Feature for Brain Tumor Segmentation.”
Active meta-tune
Han, Sungyeob, et al. “Learning to Learn Unlearned Feature for Brain Tumor Segmentation.”
Active meta-tune
Han, Sungyeob, et al. “Learning to Learn Unlearned Feature for Brain Tumor Segmentation.”
Active meta-tune
𝜶 𝑻
𝜷 𝑻
Han, Sungyeob, et al. “Learning to Learn Unlearned Feature for Brain Tumor Segmentation.”
Outline
1. Training segmentation network
2. Meta-learning
3. Active learning
4. Active meta-tune
5. Applications
Brain tumor segmentation
FLAIR Ground Truth Proposed ModelBaseline
• 3D FCN on High grade glioma (BRATS)
Mean Dice Score (std)
Whole Active Core
Baseline 0.72
(0.18)
0.54
(0.25)
0.44
(0.24)
Proposed 0.77
(0.13)
0.57
(0.23)
0.51
(0.25)
Convolution Deconvolution
Types of brain tumor – size and location
• Brain tumor treatment options depend on the type of brain tumor
you have, as well as its size and location.
• Types
• Acoustic neuroma, Astrocytoma, Brain metastases
• Choroid plexus carcinoma, Craniopharyngioma
• Embryonal tumors, Ependymoma, Glioblastoma
• Glioma, Medulloblastoma, Meningioma
• Oligodendroglioma, Pediatric brain tumors
• Pineoblastoma, Pituitary tumors
early stage of brain metastasis
Motivation on brain metastasis segmentation
• The target feature is different
• Too many small tumors : automation needed!
x40 slides
for 4 seq.
Active meta-tune on brain metastasis segmentation
Convolution Deconvolution
Active learning
based samplingMAML
task
arrang
ement
Han, Sungyeob, et al. “Learning to Learn Unlearned Feature for Brain Tumor Segmentation.”
Experimental Results – pretrained
Experimental Results – 1 step update
Experimental Results – 5 step updates
Experimental Results
Results on training dataset
Method Dice Score (mean, std)
Baseline 0.66 (on BRaTS dataset)
Naive 0.33 ± 0.3413
Passive 0.41 ± 0.2752
Active 0.45 ± 0.2317
DSC result for enhancing tumor
Results on validation dataset
Conclusion
• Transfer learning method from high grade glioma to brain
metastasis
• Learn unlearned features without forgetting the original task in
brain tumor segmentation
• Show the generalization effect in target domain segmentation
(brain metastasis)
Questions?
Segmentation
Long, Jonathan, Evan Shelhamer, and Trevor Darrell. "Fully convolutional networks for semantic segmentation."
Zhao, Hengshuang, et al. "Pyramid scene parsing network.“
Kendall, Alex, and Yarin Gal. "What uncertainties do we need in bayesian deep learning for computer vision?.“
Meta-learning
Finn, Chelsea, Pieter Abbeel, and Sergey Levine. "Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks.“
Finn, Chelsea, Kelvin Xu, and Sergey Levine. "Probabilistic Model-Agnostic Meta-Learning.“
Ravi, Sachin, and Hugo Larochelle. "Optimization as a model for few-shot learning.“
Lake, Brenden M., Ruslan Salakhutdinov, and Joshua B. Tenenbaum. "Human-level concept learning through probabilistic program induction."
Active learning
Settles, Burr “Active learning literature survey.”
Mussmann, Stephen, and Percy S. Liang. "Uncertainty Sampling is Preconditioned Stochastic Gradient Descent on Zero-One Loss.“
Mussmann, Stephen, and Percy Liang. "On the Relationship between Data Efficiency and Error for Uncertainty Sampling.”
Applications
Menze, Bjoern H., et al. "The multimodal brain tumor image segmentation benchmark (BRATS).“
Han, Sungyeob, et al. “Learning to Learn Unlearned Feature for Brain Tumor Segmentation.”

More Related Content

What's hot (20)

PDF
Learning loss for active learning
NAVER Engineering
 
PDF
[PR12] understanding deep learning requires rethinking generalization
JaeJun Yoo
 
PPTX
Introduction to Interpretable Machine Learning
Nguyen Giang
 
PDF
Matching Network
SuwhanBaek
 
PPTX
Transfer Learning and Fine Tuning for Cross Domain Image Classification with ...
Sujit Pal
 
PDF
Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks
Yoonho Lee
 
PPTX
Convolutional Neural Network for Alzheimer’s disease diagnosis with Neuroim...
Seonho Park
 
PPTX
IROS 2017 Slides
Sungjoon Choi
 
PPTX
Utilizing additional information in factorization methods (research overview,...
Balázs Hidasi
 
PPTX
Deep Learning Models for Question Answering
Sujit Pal
 
PPTX
One shot learning
Vuong Ho Ngoc
 
PDF
[PR12] Spectral Normalization for Generative Adversarial Networks
JaeJun Yoo
 
PDF
Continual Learning with Deep Architectures - Tutorial ICML 2021
Vincenzo Lomonaco
 
PDF
LevDNN
Sungjoon Choi
 
PDF
Task Adaptive Neural Network Search with Meta-Contrastive Learning
MLAI2
 
PPTX
InfoGAIL
Sungjoon Choi
 
PDF
Deep learning and image analytics using Python by Dr Sanparit
BAINIDA
 
PDF
Domain Invariant Representation Learning with Domain Density Transformations
HyunKyu Jeon
 
PPTX
Neural Networks for Pattern Recognition
Vipra Singh
 
PPTX
Artificial Intelligence, Machine Learning and Deep Learning
Sujit Pal
 
Learning loss for active learning
NAVER Engineering
 
[PR12] understanding deep learning requires rethinking generalization
JaeJun Yoo
 
Introduction to Interpretable Machine Learning
Nguyen Giang
 
Matching Network
SuwhanBaek
 
Transfer Learning and Fine Tuning for Cross Domain Image Classification with ...
Sujit Pal
 
Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks
Yoonho Lee
 
Convolutional Neural Network for Alzheimer’s disease diagnosis with Neuroim...
Seonho Park
 
IROS 2017 Slides
Sungjoon Choi
 
Utilizing additional information in factorization methods (research overview,...
Balázs Hidasi
 
Deep Learning Models for Question Answering
Sujit Pal
 
One shot learning
Vuong Ho Ngoc
 
[PR12] Spectral Normalization for Generative Adversarial Networks
JaeJun Yoo
 
Continual Learning with Deep Architectures - Tutorial ICML 2021
Vincenzo Lomonaco
 
Task Adaptive Neural Network Search with Meta-Contrastive Learning
MLAI2
 
InfoGAIL
Sungjoon Choi
 
Deep learning and image analytics using Python by Dr Sanparit
BAINIDA
 
Domain Invariant Representation Learning with Domain Density Transformations
HyunKyu Jeon
 
Neural Networks for Pattern Recognition
Vipra Singh
 
Artificial Intelligence, Machine Learning and Deep Learning
Sujit Pal
 

Similar to Learning to learn unlearned feature for segmentation (20)

PDF
Recent advances of AI for medical imaging : Engineering perspectives
Namkug Kim
 
PPTX
Intro to machine learning
Akshay Kanchan
 
PDF
DEF CON 24 - Clarence Chio - machine duping 101
Felipe Prado
 
PDF
Machine Duping 101: Pwning Deep Learning Systems
Clarence Chio
 
PDF
ppt.pdf
MohanRaj924804
 
PDF
An Efficient PSO Based Ensemble Classification Model on High Dimensional Data...
ijsc
 
PDF
AN EFFICIENT PSO BASED ENSEMBLE CLASSIFICATION MODEL ON HIGH DIMENSIONAL DATA...
ijsc
 
PDF
PhD Defense
Taehoon Lee
 
PDF
Deep Learning: concepts and use cases (October 2018)
Julien SIMON
 
PDF
Cost-effective Interactive Attention Learning with Neural Attention Process
MLAI2
 
PPTX
Deep vs diverse architectures for classification problems
Colleen Farrelly
 
PPT
Statistical learning intro
Pei-shen (James) Wu
 
PPT
deepnet-lourentzou.ppt
yang947066
 
PPT
Deep learning is a subset of machine learning and AI
leradiophysicien1
 
PPT
Overview of Deep Learning and its advantage
aqib296675
 
PPT
Introduction to Deep Learning presentation
johanericka2
 
PDF
The Tower of Knowledge A Generic System Architecture
Distinguished Lecturer Series - Leon The Mathematician
 
PPTX
Application of Neural Network Paradigms to a Study of Nursing Burnout
FelixLadstaetter
 
PDF
Learning to Balance: Bayesian Meta-Learning for Imbalanced and Out-of-distrib...
MLAI2
 
PPTX
SC1.pptx
VaishaliBagewadikar
 
Recent advances of AI for medical imaging : Engineering perspectives
Namkug Kim
 
Intro to machine learning
Akshay Kanchan
 
DEF CON 24 - Clarence Chio - machine duping 101
Felipe Prado
 
Machine Duping 101: Pwning Deep Learning Systems
Clarence Chio
 
An Efficient PSO Based Ensemble Classification Model on High Dimensional Data...
ijsc
 
AN EFFICIENT PSO BASED ENSEMBLE CLASSIFICATION MODEL ON HIGH DIMENSIONAL DATA...
ijsc
 
PhD Defense
Taehoon Lee
 
Deep Learning: concepts and use cases (October 2018)
Julien SIMON
 
Cost-effective Interactive Attention Learning with Neural Attention Process
MLAI2
 
Deep vs diverse architectures for classification problems
Colleen Farrelly
 
Statistical learning intro
Pei-shen (James) Wu
 
deepnet-lourentzou.ppt
yang947066
 
Deep learning is a subset of machine learning and AI
leradiophysicien1
 
Overview of Deep Learning and its advantage
aqib296675
 
Introduction to Deep Learning presentation
johanericka2
 
The Tower of Knowledge A Generic System Architecture
Distinguished Lecturer Series - Leon The Mathematician
 
Application of Neural Network Paradigms to a Study of Nursing Burnout
FelixLadstaetter
 
Learning to Balance: Bayesian Meta-Learning for Imbalanced and Out-of-distrib...
MLAI2
 
Ad

More from NAVER Engineering (20)

PDF
React vac pattern
NAVER Engineering
 
PDF
디자인 시스템에 직방 ZUIX
NAVER Engineering
 
PDF
진화하는 디자인 시스템(걸음마 편)
NAVER Engineering
 
PDF
서비스 운영을 위한 디자인시스템 프로젝트
NAVER Engineering
 
PDF
BPL(Banksalad Product Language) 무야호
NAVER Engineering
 
PDF
이번 생에 디자인 시스템은 처음이라
NAVER Engineering
 
PDF
날고 있는 여러 비행기 넘나 들며 정비하기
NAVER Engineering
 
PDF
쏘카프레임 구축 배경과 과정
NAVER Engineering
 
PDF
플랫폼 디자이너 없이 디자인 시스템을 구축하는 프로덕트 디자이너의 우당탕탕 고통 연대기
NAVER Engineering
 
PDF
200820 NAVER TECH CONCERT 15_Code Review is Horse(코드리뷰는 말이야)(feat.Latte)
NAVER Engineering
 
PDF
200819 NAVER TECH CONCERT 03_화려한 코루틴이 내 앱을 감싸네! 코루틴으로 작성해보는 깔끔한 비동기 코드
NAVER Engineering
 
PDF
200819 NAVER TECH CONCERT 10_맥북에서도 아이맥프로에서 빌드하는 것처럼 빌드 속도 빠르게 하기
NAVER Engineering
 
PDF
200819 NAVER TECH CONCERT 08_성능을 고민하는 슬기로운 개발자 생활
NAVER Engineering
 
PDF
200819 NAVER TECH CONCERT 05_모르면 손해보는 Android 디버깅/분석 꿀팁 대방출
NAVER Engineering
 
PDF
200819 NAVER TECH CONCERT 09_Case.xcodeproj - 좋은 동료로 거듭나기 위한 노하우
NAVER Engineering
 
PDF
200820 NAVER TECH CONCERT 14_야 너두 할 수 있어. 비전공자, COBOL 개발자를 거쳐 네이버에서 FE 개발하게 된...
NAVER Engineering
 
PDF
200820 NAVER TECH CONCERT 13_네이버에서 오픈 소스 개발을 통해 성장하는 방법
NAVER Engineering
 
PDF
200820 NAVER TECH CONCERT 12_상반기 네이버 인턴을 돌아보며
NAVER Engineering
 
PDF
200820 NAVER TECH CONCERT 11_빠르게 성장하는 슈퍼루키로 거듭나기
NAVER Engineering
 
PDF
200819 NAVER TECH CONCERT 07_신입 iOS 개발자 개발업무 적응기
NAVER Engineering
 
React vac pattern
NAVER Engineering
 
디자인 시스템에 직방 ZUIX
NAVER Engineering
 
진화하는 디자인 시스템(걸음마 편)
NAVER Engineering
 
서비스 운영을 위한 디자인시스템 프로젝트
NAVER Engineering
 
BPL(Banksalad Product Language) 무야호
NAVER Engineering
 
이번 생에 디자인 시스템은 처음이라
NAVER Engineering
 
날고 있는 여러 비행기 넘나 들며 정비하기
NAVER Engineering
 
쏘카프레임 구축 배경과 과정
NAVER Engineering
 
플랫폼 디자이너 없이 디자인 시스템을 구축하는 프로덕트 디자이너의 우당탕탕 고통 연대기
NAVER Engineering
 
200820 NAVER TECH CONCERT 15_Code Review is Horse(코드리뷰는 말이야)(feat.Latte)
NAVER Engineering
 
200819 NAVER TECH CONCERT 03_화려한 코루틴이 내 앱을 감싸네! 코루틴으로 작성해보는 깔끔한 비동기 코드
NAVER Engineering
 
200819 NAVER TECH CONCERT 10_맥북에서도 아이맥프로에서 빌드하는 것처럼 빌드 속도 빠르게 하기
NAVER Engineering
 
200819 NAVER TECH CONCERT 08_성능을 고민하는 슬기로운 개발자 생활
NAVER Engineering
 
200819 NAVER TECH CONCERT 05_모르면 손해보는 Android 디버깅/분석 꿀팁 대방출
NAVER Engineering
 
200819 NAVER TECH CONCERT 09_Case.xcodeproj - 좋은 동료로 거듭나기 위한 노하우
NAVER Engineering
 
200820 NAVER TECH CONCERT 14_야 너두 할 수 있어. 비전공자, COBOL 개발자를 거쳐 네이버에서 FE 개발하게 된...
NAVER Engineering
 
200820 NAVER TECH CONCERT 13_네이버에서 오픈 소스 개발을 통해 성장하는 방법
NAVER Engineering
 
200820 NAVER TECH CONCERT 12_상반기 네이버 인턴을 돌아보며
NAVER Engineering
 
200820 NAVER TECH CONCERT 11_빠르게 성장하는 슈퍼루키로 거듭나기
NAVER Engineering
 
200819 NAVER TECH CONCERT 07_신입 iOS 개발자 개발업무 적응기
NAVER Engineering
 
Ad

Recently uploaded (20)

PDF
Bitcoin for Millennials podcast with Bram, Power Laws of Bitcoin
Stephen Perrenod
 
PPTX
WooCommerce Workshop: Bring Your Laptop
Laura Hartwig
 
PPTX
OpenID AuthZEN - Analyst Briefing July 2025
David Brossard
 
PDF
Log-Based Anomaly Detection: Enhancing System Reliability with Machine Learning
Mohammed BEKKOUCHE
 
PDF
HCIP-Data Center Facility Deployment V2.0 Training Material (Without Remarks ...
mcastillo49
 
PDF
Jak MŚP w Europie Środkowo-Wschodniej odnajdują się w świecie AI
dominikamizerska1
 
PPT
Interview paper part 3, It is based on Interview Prep
SoumyadeepGhosh39
 
PDF
DevBcn - Building 10x Organizations Using Modern Productivity Metrics
Justin Reock
 
PDF
The Builder’s Playbook - 2025 State of AI Report.pdf
jeroen339954
 
PDF
Smart Trailers 2025 Update with History and Overview
Paul Menig
 
PPTX
✨Unleashing Collaboration: Salesforce Channels & Community Power in Patna!✨
SanjeetMishra29
 
PDF
[Newgen] NewgenONE Marvin Brochure 1.pdf
darshakparmar
 
PDF
Reverse Engineering of Security Products: Developing an Advanced Microsoft De...
nwbxhhcyjv
 
PDF
Using FME to Develop Self-Service CAD Applications for a Major UK Police Force
Safe Software
 
PDF
Agentic AI lifecycle for Enterprise Hyper-Automation
Debmalya Biswas
 
PDF
Building Real-Time Digital Twins with IBM Maximo & ArcGIS Indoors
Safe Software
 
PDF
Newgen Beyond Frankenstein_Build vs Buy_Digital_version.pdf
darshakparmar
 
PPTX
Building Search Using OpenSearch: Limitations and Workarounds
Sease
 
PDF
CIFDAQ Market Insights for July 7th 2025
CIFDAQ
 
PDF
Python basic programing language for automation
DanialHabibi2
 
Bitcoin for Millennials podcast with Bram, Power Laws of Bitcoin
Stephen Perrenod
 
WooCommerce Workshop: Bring Your Laptop
Laura Hartwig
 
OpenID AuthZEN - Analyst Briefing July 2025
David Brossard
 
Log-Based Anomaly Detection: Enhancing System Reliability with Machine Learning
Mohammed BEKKOUCHE
 
HCIP-Data Center Facility Deployment V2.0 Training Material (Without Remarks ...
mcastillo49
 
Jak MŚP w Europie Środkowo-Wschodniej odnajdują się w świecie AI
dominikamizerska1
 
Interview paper part 3, It is based on Interview Prep
SoumyadeepGhosh39
 
DevBcn - Building 10x Organizations Using Modern Productivity Metrics
Justin Reock
 
The Builder’s Playbook - 2025 State of AI Report.pdf
jeroen339954
 
Smart Trailers 2025 Update with History and Overview
Paul Menig
 
✨Unleashing Collaboration: Salesforce Channels & Community Power in Patna!✨
SanjeetMishra29
 
[Newgen] NewgenONE Marvin Brochure 1.pdf
darshakparmar
 
Reverse Engineering of Security Products: Developing an Advanced Microsoft De...
nwbxhhcyjv
 
Using FME to Develop Self-Service CAD Applications for a Major UK Police Force
Safe Software
 
Agentic AI lifecycle for Enterprise Hyper-Automation
Debmalya Biswas
 
Building Real-Time Digital Twins with IBM Maximo & ArcGIS Indoors
Safe Software
 
Newgen Beyond Frankenstein_Build vs Buy_Digital_version.pdf
darshakparmar
 
Building Search Using OpenSearch: Limitations and Workarounds
Sease
 
CIFDAQ Market Insights for July 7th 2025
CIFDAQ
 
Python basic programing language for automation
DanialHabibi2
 

Learning to learn unlearned feature for segmentation

  • 1. Learning to learn unlearned feature for segmentation Sungyeob Han Communication and Machine Learning Lab. Seoul National University
  • 2. Introduction • How to transfer with few samples? Primary cancer Brain metastasis
  • 3. Outline 1. Training segmentation network 2. Meta-learning 3. Active learning 4. Active meta-tune 5. Applications
  • 5. Fully convolutional networks • take input of arbitrary size and produce correspondingly- sized output • a feed-forward propagation predicts the labels • end-to-end, pixel-to-pixel Long, Jonathan, Evan Shelhamer, and Trevor Darrell. "Fully convolutional networks for semantic segmentation."
  • 6. Pyramid Scene Parsing Network pyramid parsing module : harvest different sub-region representations concatenation : upsampling and concatenation layers Zhao, Hengshuang, et al. "Pyramid scene parsing network."
  • 7. Zhao, Hengshuang, et al. "Pyramid scene parsing network."
  • 8. Details in training segmentation network • Fast feed-forward time (FCN-based) • Given the pre-trained encoding parameters (VGGNet), fine-tuning in stages takes 36 hours on a single GPU. • Ambiguity : object structure, sparse label • constrained categories average loss gives blurry gradient for each category information
  • 9. Outline 1. Training segmentation network 2. Meta-learning 3. Active learning 4. Active meta-tune 5. Applications
  • 10. Learning to learn • A key aspect of intelligence : versatility • the capability of doing many different things. • Meta-learning • As known as ”learning to learn” • learn how to learn new tasks faster by reusing previous experience
  • 11. Few-shot Learning • In 2015, Brenden et al. show how to learn new concepts from one or a few instances of that concept. • learn to learn from a few examples • Omniglot • 1623 character classes • each with 20 examples Lake, Brenden M., Ruslan Salakhutdinov, and Joshua B. Tenenbaum. "Human-level concept learning through probabilistic program induction."
  • 12. Meta-learner Ravi, Sachin, and Hugo Larochelle. "Optimization as a model for few-shot learning.“
  • 13. Meta-learning set-up for few-shot image classification • 1-shot, 5-class classification task • one example from each of 5 classes
  • 14. Model agnostic meta-learning • Motivation : ambiguity on new task • From a dynamical systems standpoint • Maximize the sensitivity of the loss functions of new tasks
  • 15. Model Agnostic Meta-Learning • A model 𝑓 • Maps observation 𝐱 to outputs 𝐚 • Each task 𝒯 = ℒ 𝐱1, 𝐚1, ⋯ , 𝐱 𝐻, 𝐚 𝐻 , 𝑞 𝐱1 , 𝑞 𝐱 𝑡+1|𝐱 𝑡, 𝐚 𝑡 , 𝐻 • Loss, Initial distribution, a transition dist., episode length • ℒ 𝐱1, 𝐚1, ⋯ , 𝐱 𝐻, 𝐚 𝐻 • Task-specific feedback • A misclassification loss
  • 16. Algorithm : MAML • K-shot learning • Learn a new Task 𝓣𝒊 sampled from 𝒑 𝓣 from only K samples sampled from 𝒒𝒊 • Feedback ℒ 𝑇 𝑖 generated by 𝒯𝑖 • Train with K samples • Test on new samples from 𝒯𝑖 • improved by considering how the test error on new data from 𝑞𝑖 changes with respect to the parameters. • the test error on sampled tasks 𝒯𝑖 serves as the training error of the meta-learning process.
  • 17. Algorithm : MAML • A model with parameters 𝑓𝜃 , adapting to an new task 𝒯𝑖 ∶ 𝜃 → 𝜃𝑖′ 𝜃𝑖 ′ = 𝜃 − 𝛼𝛻𝜃ℒ 𝒯𝑖 (𝑓𝜃) • Meta-objective : the model parameters are trained by optimizing for performance of 𝑓 𝜃′ min 𝜃 ෍ 𝒯𝑖~𝑝(𝒯) ℒ 𝒯𝑖 𝑓 𝜃𝑖 ′ = ෍ 𝒯𝑖~𝑝(𝒯) ℒ 𝒯𝑖 𝑓𝜃−𝛼𝛻 𝜃ℒ 𝑇 𝑖 (𝑓 𝜃) • The meta-optimization across tasks 𝜃 ← 𝜃 − 𝛽𝛻𝜃 ෍ 𝒯𝑖~𝑝(𝒯) ℒ 𝒯𝑖 𝑓 𝜃𝑖 ′
  • 19. Experimental Evaluation : Supervised classification
  • 21. Desinging of meta-learning Levine, Sergey, and Chelsea Finn, “Meta-learning frontiers: universal, uncertain, and unsupervised.”
  • 22. Meta-learning with ambiguity Levine, Sergey, and Chelsea Finn, “Meta-learning frontiers: universal, uncertain, and unsupervised.”
  • 23. Meta-learning with ambiguity Levine, Sergey, and Chelsea Finn, “Meta-learning frontiers: universal, uncertain, and unsupervised.”
  • 24. Outline 1. Training segmentation network 2. Meta-learning 3. Active learning 4. Active meta-tune 5. Applications
  • 25. Active learning • pool-based active learning • queries are selected from a large pool of unlabeled instances • uncertainty sampling query strategy • select the instance in the pool about which the model is least certain how to label Settles, Burr “Active learning literature survey.”
  • 26. pool-based active learning • A toy example : logistic regression (400 instances) • (b) : 30 labeled instances randomly drawn : 70% accuracy • (c) : 30 actively queried instances using uncertainty sampling : 90%
  • 27. Uncertainty sampling • uncertainty measures Mussmann, Stephen, and Percy Liang. "On the Relationship between Data Efficiency and Error for Uncertainty Sampling.”
  • 28. How can me measure the uncertainty of networks? • a small dataset for a new task can simply be too ambiguous to acquire a single model Kendall, Alex, and Yarin Gal. "What uncertainties do we need in bayesian deep learning for computer vision?.“
  • 29. Outline 1. Training segmentation network 2. Meta-learning 3. Active learning 4. Active meta-tune 5. Applications
  • 31. Generate training tasks good task bad task 𝑻 𝒈𝒐𝒐𝒅 𝑻 𝒃𝒂𝒅 Han, Sungyeob, et al. “Learning to Learn Unlearned Feature for Brain Tumor Segmentation.”
  • 32. Active meta-tune Han, Sungyeob, et al. “Learning to Learn Unlearned Feature for Brain Tumor Segmentation.”
  • 33. Active meta-tune Han, Sungyeob, et al. “Learning to Learn Unlearned Feature for Brain Tumor Segmentation.”
  • 34. Active meta-tune 𝜶 𝑻 𝜷 𝑻 Han, Sungyeob, et al. “Learning to Learn Unlearned Feature for Brain Tumor Segmentation.”
  • 35. Outline 1. Training segmentation network 2. Meta-learning 3. Active learning 4. Active meta-tune 5. Applications
  • 36. Brain tumor segmentation FLAIR Ground Truth Proposed ModelBaseline • 3D FCN on High grade glioma (BRATS) Mean Dice Score (std) Whole Active Core Baseline 0.72 (0.18) 0.54 (0.25) 0.44 (0.24) Proposed 0.77 (0.13) 0.57 (0.23) 0.51 (0.25) Convolution Deconvolution
  • 37. Types of brain tumor – size and location • Brain tumor treatment options depend on the type of brain tumor you have, as well as its size and location. • Types • Acoustic neuroma, Astrocytoma, Brain metastases • Choroid plexus carcinoma, Craniopharyngioma • Embryonal tumors, Ependymoma, Glioblastoma • Glioma, Medulloblastoma, Meningioma • Oligodendroglioma, Pediatric brain tumors • Pineoblastoma, Pituitary tumors early stage of brain metastasis
  • 38. Motivation on brain metastasis segmentation • The target feature is different • Too many small tumors : automation needed! x40 slides for 4 seq.
  • 39. Active meta-tune on brain metastasis segmentation Convolution Deconvolution Active learning based samplingMAML task arrang ement Han, Sungyeob, et al. “Learning to Learn Unlearned Feature for Brain Tumor Segmentation.”
  • 41. Experimental Results – 1 step update
  • 42. Experimental Results – 5 step updates
  • 43. Experimental Results Results on training dataset Method Dice Score (mean, std) Baseline 0.66 (on BRaTS dataset) Naive 0.33 ± 0.3413 Passive 0.41 ± 0.2752 Active 0.45 ± 0.2317 DSC result for enhancing tumor Results on validation dataset
  • 44. Conclusion • Transfer learning method from high grade glioma to brain metastasis • Learn unlearned features without forgetting the original task in brain tumor segmentation • Show the generalization effect in target domain segmentation (brain metastasis)
  • 45. Questions? Segmentation Long, Jonathan, Evan Shelhamer, and Trevor Darrell. "Fully convolutional networks for semantic segmentation." Zhao, Hengshuang, et al. "Pyramid scene parsing network.“ Kendall, Alex, and Yarin Gal. "What uncertainties do we need in bayesian deep learning for computer vision?.“ Meta-learning Finn, Chelsea, Pieter Abbeel, and Sergey Levine. "Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks.“ Finn, Chelsea, Kelvin Xu, and Sergey Levine. "Probabilistic Model-Agnostic Meta-Learning.“ Ravi, Sachin, and Hugo Larochelle. "Optimization as a model for few-shot learning.“ Lake, Brenden M., Ruslan Salakhutdinov, and Joshua B. Tenenbaum. "Human-level concept learning through probabilistic program induction." Active learning Settles, Burr “Active learning literature survey.” Mussmann, Stephen, and Percy S. Liang. "Uncertainty Sampling is Preconditioned Stochastic Gradient Descent on Zero-One Loss.“ Mussmann, Stephen, and Percy Liang. "On the Relationship between Data Efficiency and Error for Uncertainty Sampling.” Applications Menze, Bjoern H., et al. "The multimodal brain tumor image segmentation benchmark (BRATS).“ Han, Sungyeob, et al. “Learning to Learn Unlearned Feature for Brain Tumor Segmentation.”