SlideShare a Scribd company logo
YongGang Hu, Chao Xue, IBM
Hyper-Parameter Selection
and Adaptive Model Tuning
for Deep Neural Networks
#AssignedHashtagGoesHere
From	Hyper	parameter/Network	Search	to	On-line	Tuning
Image
recognition
DeepLearningApplications
Object
detection
Translation
Others
Minimize ( , )
Outline
• Hyper-parameter selections & Neural network search
– Bayesian optimization
– Hyperband
– Reinforcement learning
• High performance AutoML
– Transfer AutoML - Using historical hyper-parameter configurations for different
tasks
– Joint optimization with AutoML and finetune
• Training visualization and Interactive tuning
• Real world experience and Implementation
3
From	Hyper	parameter/Network	Search	to	On-line	Tuning
Monitor
Advice
Hyper-parameter & Network
Search
criterion and
model
parameters
optimizing
procedure
parameters
Real-time monitor for running application
Real-time adviser for running application
overflow
underfitting
overfitting
divergence
convergence
checkthetraining
process
learning curve
weight/gradient/a
ctivation
histogram and
norm
worst cases of
training samples
example: learning
rate, batch size,
momentum, learning
rate scheme
example: number of
hidden units, number
of layers, filter kernel
size, max-pool size,
weight decay
Image
recognition
DeepLearningApplications
Object
detection
Translation
Others
Optimize
Standard AutoML--from Random search, Bayesian Optimization to
Reinforcement Learning
5
Adaptive
Random
Search
(Hyperband)
Reinforceme
nt Learning
Bayesian
Optimization
Neural Network auto
selected
Hyperparameters auto
selected
Qs1
sN
u1 uM
Results of AdaRandom and Bayesian Optimization for Object detection
6
• The raw data (no data augment)
0
0.2
0.4
0.6
0.8
avg Best case Worst case
mAP
default
AdaRandom(HyperBand)
Bayesian Optimization
Machine IBM P8
GPU Tesla K80
GPU MEM 10G
CPU ppc64le
CPU cores 160
MEM 55G
Frequency 3.5 GHz
OS Ubuntu
14.04
caffe version Py-faster-
rcnn
Dataset User-defined
Model vgg16
§ The average results for the three (HPP)
hyperparameters combinations at 4000
iterations are (0.49, 0.52, 0.55), that is,
using AdaRandom and Bayesian
optimization recommended HPP, you can
gain 6% and 12% in the average results
comparing to default setting. AdaRandom
method has more variance among
different tries(train-test dataset split). The
Bayesian models are more stable while
having better average performance.
§ The below pictures show accuracy during
the 0-4000 iterations, with different tries,
under the two HPP configurations. We
can see that: 1) It can be early stopped at
about 1300 iterations. 2) the performance
with different tries differ significantly, it
caused by in some tries, training dataset
has the invariance property according to
test dataset, but some doesn’t have. It
need to augment data to gain the stable
performance. 3) different HPP
combinations(models) may have different
sensitivity to the different tries.
Neural Network:
C: Convolution, P: pooling, FC: Full
connection.
Default manual: (9 layers)
C(3,32)+C(3,32)+P(3,3)+C(3,64)+C(3,64)+P
(3,3)+FC(64)+FC(32)+FC(6)
AdaRandom generated: (6 layers)
C(3,128)+C(4,64)+C(3,32)+P(3,3)+FC(16)+
FC(6)
Reinforce generated: (5 layers)
C(3,32)+P(2,2)+P(3,2)+C(5,32)+FC(6)
Evaluations with Acoustic Applications
• The best results for the three networks are (0.703,0.673, 0.695) (the smaller the better), that is, using AdaRandom and Reinforce recommended
models, you can gain 4.3% and 1.1% in the best results comparisons. The average result of the three networks is (0.817,0.776, 0.763), that is, the
DL Insight recommended modes can increase about 5.0% and 6.6% in the average case performance. And from the standard deviation view, the
recommended models are clearly more stable.
• The CDF (cumulative distribution function) curve is more intuitive to illustrate the comparison of the three models(the more left the better). For
example, using reinforce recommended model, ER has more than 60% probability (frequency) less than 0.75, while the default only has the 30%.
7
• We implement the AdaRandom (adaptive random search scheme) and Reinforce (reinforcement learning search
scheme) methods to generate deep learning neural network automatically.
• We are trying the new methods in different areas. Here is the example for acoustic. Default is the best scheme by manual
tuning.
Better accuracy More stableLower complexity
8
Traditional fine-tune
Standalone AutoML
Dataset
Virtual Dataset
Model
ParameterWeights and bias
Neural network Hyper-parameter
Dataset group
Dog Car Unknown
Collaborative AutoML
Transfer AutoML Architecture
Upload
datasets
White-box
analysis
Black-box
analysis
Benchmark
models
selection
Clients
Clients
Knowledge base
Model
selection
Joint Optimization
with transfer
learning
RL
HyperBand
Bayesian
ParallelizationB
Virtual datasets group
AutoML process
Path 1
Path1
Server
Server
Spark
Challenges for AutoML with Transfer Learning
• Training small user dataset leads to convergence problem à Transfer learning is needed
• When considering transfer learning, the pretrained model need to be chosen, usually in the
computer vision, we choose image-net as the base dataset to get the initial weights as the
pretrained model, but it can’t fit many specific user datasets.
• To solve this transfer learning’s problem, we can let the user to classify his dataset into some
predefined categories, and in each category, the pretrained model was trained separately. It
can improve the performance of transfer learning but involve user’s intervention with their
datasets.
• Using AutoML with transfer learning can improve transfer learning’s performance without user’s
intervention. But considering the transfer learning’s properties, there are two challenges for
AutoML:
– Since reusing the initial weights, transfer learning limits the searching space of AutoML, how to use
AutoML based on the pretrained model is a question.
– We can’t use AutoML to build one model for every user dataset, it is too expensive. How to reuse the
model for transfer learning is a question.
10
Joint optimization: AutoML with the fine-tune
Search space:
• lr_policy: LR_POLICY
stepsize: STEPSIZE
gamma: GAMMA
momentum: MOMENTUM
solver_mode: GPU
max_iter: MAX_ITER
test_iter: TEST_ITER
test_interval: TEST_INTERVAL
base_lr: BASE_LR
weight_decay: WEIGHT_DECAY
solver_type: SGD
layer {
param {
lr_mult: LR_MULT_C_W_0
decay_mult:
DECAY_MULT_C_W_0
}
param {
lr_mult: LR_MULT_C_B_0
decay_mult:
DECAY_MULT_C_B_0
}
convolution_param {
num_output: NUM_OUTPUT_0
pad: PAD_0
kernel_size: KERNEL_SIZE_C_0
group: GROUP_0
weight_filler {
type: TYPE_C_W_0
std: STD_C_W_0
}
bias_filler {
type: TYPE_C_B_0
std: STD_C_B_0
}
}
}
layer {
name: "conv0_relu"
type: TYPE_C_AF_0
bottom: "conv0"
top: "conv0"
}
layer {
name: "pool1"
type: "Pooling"
bottom: "conv0"
top: "pool1"
pooling_param {
pool: AVE
kernel_size:
KERNEL_SIZE_P_0
stride: STRIDE_P_0
}
}
layer {
name: "last_fc"
type: "InnerProduct"
bottom: "pool1"
top: "last_fc"
param {
lr_mult: LR_MULT_FC_W_0
decay_mult:
DECAY_MULT_FC_W_0
}
param {
lr_mult: LR_MULT_FC_B_0
decay_mult:
DECAY_MULT_FC_B_0
}
inner_product_param {
num_output: OUTPUT_NUMS
weight_filler {
type: TYPE_FC_W_0
std: STD_FC_W_0
}
bias_filler {
type: TYPE_FC_B_0
std: STD_FC_B_0
}
}
}
c p c p fc
1) Neural network at the last stage:
Example:
2) Below hyper-parameter
Results and Analysis
0
0.2
0.4
0.6
0.8
1
1.2
Food(30 epoch) TV show(10
epoch)
Car(50 epoch) Sence(100
epoch)
Action(20
epoch)
Accuracy
GoogleCAM+Finetune(Imagenet)
GoogleCAM+Finetune(imagenet+ same category dataset)
AutoML+Finetune(Imagenet)
Advantages of AutoML
with finetune:
(+) Best accuracy
(+) Most stable
(+) Don’t need separate
pretrained models by
predefining dataset
categories. No user’s
interventions.
14
Expert optimization advice
for hyper parameter
selection and tuning
Traffic light alerting
for required parameter
optimization with early
stop advice and more
CPU, GPU, memory
utilization info, comms
overhead, +++
Interactive Tuning of Hyper Parameters – DL Insight
15
CIFAR-10
Traditional method: underfitting
Our method: still going down
Auto-detection method for training process
16
Traditional method: underfitting Our method: good game
MNIST
Auto-detection method for training process
17
Interactive Tuning – Example - 1st run of Caffe
CIFAR-10
We will run three times of caffe on spark, to show DL Insight’s capabilities for DL
Interactive Tuning – Example - 1st run of Caffe
18
The DL Insight suggests user to stop the
training early because overflow detected, and
give the user advisor to add clip gradient to
0.1
Interactive Tuning – Example - 2nd run of Caffe
19
On the second run of Caffe, we accept the
suggestion which DL Insight gave, the add
clip gradient to 0.1, what happened then?
Interactive Tuning – Example - 2nd run of Caffe
20
DL Insight suggests user to stop the training
Interactive Tuning – Example -3rd run of Caffe
21
On the third run of Caffe, we accept the
suggestion which DL Insight gave, and
more layers such as RELU, what
happened then?
Interactive Tuning – Example - 3rd run of Caffe
22
No more suggestion. You can do other
things or have a rest while training the
dataset. When you feel worried about it,
just have a look at DL InsightJ
2
base_lr weight_decay momentum early stop
batch
size/rpn
batch size
split ratio max_iters display
applause01 0.001 0.0005 0.9
yes
divergence
128/256 0.7(94/39) 2000 10 appl = 0.3409 0.3409
iter=370
mAP=0.221060
runs
Hyperparameters in suggestions
AP mAP
mAP at early
stop iteration
Parameters
Interactive Tuning - Example: “Applause” datasets
Good judgement
2
base_lr
weight_de
cay
momentu
m
early stop
batch
size/rpn
batch size
split ratio max_iters display
applause-aug01-01 0.001 0.0005 0.9
yes
divergence
128/256 0.7(98/42) 2000 10 AP for appl = 0.4848 0.4848
iter=350
mAP=0.313152
runs
Hyperparameters in suggestions Parameters
AP mAP
mAP at early
stop iteration
Good judgement
Learning Curve Trend Judgement – ‘Applause’ (Continue)
2
Good judgement
base_lr
weight_de
cay
momentu
m
early stop
batch
size/rpn
batch size
split ratio max_iters display
applause-aug03-01 0.001 0.0005 0.9 128/256 0.7(164/69) 2000 10 AP for appl = 0.7847 0.7847
iter=1510
mAP=0.784068
runs
Hyperparameters in suggestions Parameters
AP mAP
mAP at early
stop iteration
Learning Curve Trend Judgement – ‘Applause’ (Continue)
Hyper-Parameter Search Implementation
26
Search hyper-parameters
space :
– Learning rate
– Decay rate
– Batch size
– Optimizer:
• GradientDecedent,
• Adadelta,
• …
– Momentum (for some optimizers)
– LSTM hidden unit size
Random, Bayesian, TPE
Based Search Types
IBM CONFIDENTIAL
Hyper-Parameter Search Implementation
Think 2018 / 5613A.pdf / March 22, 2018 / © 2018 IBM Corporation 27
Spark search jobs are generated dynamically and executed in parallel
Random
TPE
Tree-based Parzen Estimator Bayesian
Multitenant Spark Cluster
IBM Spectrum Conductor
Hyper-Parameter Search Implementation
ModelTuningMgrl
startModelAutoTuning()
stopModelAutoTuning()
deleteModelAutoTuning()
TuningTask
(Thread)
run()
TuningJobCtr
runTuningJobs()
SparkTuningJobCtrI
TuningFrameworkMgr frameworkMgr
HPTAlgorithm
TuneInputParam
inputP
TuningJobCtr jobctl
search()
BayesianAlg
RandomAlg
TPEAlg
PythonAlg
UserPlugInAl
g
TuningFrameworkMgr
initialize()
getLossValueFromLog()
prepareJob()
CoSFrameWorkMgrl
TfFrameworkMgrl
PytorchFrameworkMg
rl
Enterprise Class Deep Learning Solution
IBM Spectrum Conductor Deep Learning Impact, IBM PowerAI, IBM Storage
29
Monitoring&Reporting
Workload Management / Scheduling
Resource Management & Orchestration
Native Services Management
Services & Support
Deep
Learning
Impact
Tensorflow
Caffe
Red Hat Linux
x86…
IBM Power Systems
IBM Storage
IBM PowerAI
IBM CONFIDENTIAL
Reference
• [1] David Schaffer, Darrell Whitley and Larry J Eshelman, Combinations of genetic algorithms and neural networks: A survey of the
state of the art. International Workshop on Combinations of Genetic Algorithms and Neural Networks, 1992.
• [2] J.Snoek, H.Larochelle and R.P.Adams, Practical Bayesian optimization of machine learning algorithms. In Advances in Neural
Information Processing Systems(NIPS), 2012. 

• [3] Bergstra, James and Yoshua Bengio, Random search for hyper-parameter optimization. Journal of Machine Learning Research,
2012. 

• [4] Lisha Li, Kevin Jamieson and Giulia DeSalvo, HYPERBAND: BANDIT- BASED CONFIGURATION EVALUATION FOR
HYPERPARAMETER OPTIMIZATION. ICLR, 2017. 

• [5] James Bergstra, etc. Algorithms for Hyper-Parameter Optimization. Proceedings of the IEEE, 2012. 

• [6] Bowen Baker, Otkrist Gupta, Nikhil Naik and RameshRaskar, DESIGNING NEURAL NETWORK ARCHITECTURES USING
REINFORCEMENT LEARNING. ICLR, 2017.
Notices and disclaimers
Think 2018 / 5613A.pdf / March 22, 2018 / © 2018 IBM Corporation
•© 2018 International Business Machines Corporation. No part of this
document may be reproduced or transmitted in any form without
written permission from IBM.
•U.S. Government Users Restricted Rights — use, duplication or
disclosure restricted by GSA ADP Schedule Contract with IBM.
•Information in these presentations (including information relating to
products that have not yet been announced by IBM) has been
reviewed for accuracy as of the date of initial publication and could
include unintentional technical or typographical errors. IBM shall have
no responsibility to update this information. This document is
distributed “as is” without any warranty, either express or
implied. In no event, shall IBM be liable for any damage arising
from the use of this information, including but not limited to, loss
of data, business interruption, loss of profit or loss of
opportunity. IBM products and services are warranted per the terms
and conditions of the agreements under which they are provided.
•IBM products are manufactured from new parts or new and used
parts.
In some cases, a product may not be new and may have been
previously installed. Regardless, our warranty terms apply.”
•Any statements regarding IBM's future direction, intent or
product plans are subject to change or withdrawal without
notice.
•Performance data contained herein was generally obtained in a
controlled, isolated environments. Customer examples are presented
as illustrations of how those
•customers have used IBM products and the results they may have
achieved. Actual performance, cost, savings or other results in other
operating environments may vary.
•References in this document to IBM products, programs, or services
does not imply that IBM intends to make such products, programs or
services available in all countries in which IBM operates or does
business.
•Workshops, sessions and associated materials may have been
prepared by independent session speakers, and do not necessarily
reflect the views of IBM. All materials and discussions are provided for
informational purposes only, and are neither intended to, nor shall
constitute legal or other guidance or advice to any individual
participant or their specific situation.
•It is the customer’s responsibility to insure its own compliance
with legal requirements and to obtain advice of competent legal
counsel as to the identification and interpretation of any relevant laws
and regulatory requirements that may affect the customer’s business
and any actions the customer may need to take to comply with such
laws. IBM does not provide legal advice or represent or warrant that
its services or products will ensure that the customer follows any law.
31
Notices and disclaimers
continued
Think 2018 / 5613A.pdf / March 22, 2018 / © 2018 IBM Corporation
•Information concerning non-IBM products was obtained from the
suppliers of those products, their published announcements or other
publicly available sources. IBM has not tested those products about
this publication and cannot confirm the accuracy of performance,
compatibility or any other claims related to non-IBM
products. Questions on the capabilities of non-IBM products should be
addressed to the suppliers of those products. IBM does not warrant
the quality of any third-party products, or the ability of any such third-
party products to interoperate with IBM’s products. IBM expressly
disclaims all warranties, expressed or implied, including but not
limited to, the implied warranties of merchantability and fitness
for a purpose.
•The provision of the information contained herein is not intended to,
and does not, grant any right or license under any IBM patents,
copyrights, trademarks or other intellectual property right.
•IBM, the IBM logo, ibm.com and [names of other referenced IBM
products and services used in the presentation] are trademarks of
International Business Machines Corporation, registered in many
jurisdictions worldwide. Other product and service names might
be trademarks of IBM or other companies. A current list of IBM
trademarks is available on the Web at "Copyright and trademark
information" at: www.ibm.com/legal/copytrade.shtml.
32
Date, Time, Location & Duration Session title and Speaker
Tue, June 5 | 11 AM
2010-2012, 30 mins
Productionizing Spark ML Pipelines with the Portable Format for Analytics
Nick Pentreath (IBM)
Tue, June 5 | 2 PM
2018, 30 mins
Making PySpark Amazing—From Faster UDFs to Dependency Management and Graphing!
Holden Karau (Google) Bryan Cutler (IBM)
Tue, June 5 | 2 PM
Nook by 2001, 30 mins
Making Data and AI Accessible for All
Armand Ruiz Gabernet (IBM)
Tue, June 5 | 2:40 PM
2002-2004, 30 mins
Cognitive Database: An Apache Spark-Based AI-Enabled Relational Database System
Rajesh Bordawekar (IBM T.J. Watson Research Center)
Tue, June 5 | 3:20 PM
3016-3022, 30 mins
Dynamic Priorities for Apache Spark Application’s Resource Allocations
Michael Feiman (IBM Spectrum Computing) Shinnosuke Okada (IBM Canada Ltd.)
Tue, June 5 | 3:20 PM
2001-2005, 30 mins
Model Parallelism in Spark ML Cross-Validation
Nick Pentreath (IBM) Bryan Cutler (IBM)
Tue, June 5 | 3:20 PM
2007, 30 mins
Serverless Machine Learning on Modern Hardware Using Apache Spark
Patrick Stuedi (IBM)
Tue, June 5 | 5:40 PM
2002-2004, 30 mins
Create a Loyal Customer Base by Knowing Their Personality Using AI-Based Personality Recommendation Engine;
Sourav Mazumder (IBM Analytics) Aradhna Tiwari (University of South Florida)
Tue, June 5 | 5:40 PM
2007, 30 mins
Transparent GPU Exploitation on Apache Spark
Dr. Kazuaki Ishizaki (IBM) Madhusudanan Kandasamy (IBM)
Tue, June 5 | 5:40 PM
2009-2011, 30 mins
Apache Spark Based Hyper-Parameter Selection and Adaptive Model Tuning for Deep Neural Networks
Yonggang Hu (IBM) Chao Xue (IBM)
IBM Sessions at Spark+AI Summit 2018 (Tuesday, June 5)
IBM Cloud / Watson and Cloud Platform / © 2018 IBM Corporation 33
Date, Time, Location & Duration Session title and Speaker
Wed, June 6 | 12:50 PM Birds of a Feather: Apache Arrow in Spark and More
Bryan Cutler (IBM) Li Jin (Two Sigma Investments, LP)
Wed, June 6 | 2 PM
2002-2004, 30 mins
Deep Learning for Recommender Systems
Nick Pentreath (IBM) )
Wed, June 6 | 3:20 PM
2018, 30 mins
Bringing an AI Ecosystem to the Domain Expert and Enterprise AI Developer
Frederick Reiss (IBM) Vijay Bommireddipalli (IBM Center for Open-Source Data & AI Technologies)
IBM Sessions at Spark+AI Summit 2018 (Wednesday, June 6)
IBM Cloud / Watson and Cloud Platform / © 2018 IBM Corporation 34
Meet us at IBM booth in the Expo area.

More Related Content

What's hot (20)

PDF
Chris Fregly, Research Scientist, PipelineIO at MLconf ATL 2016
MLconf
 
PDF
Oleksii Moskalenko "Continuous Delivery of ML Pipelines to Production"
Fwdays
 
PPTX
Tom Peters, Software Engineer, Ufora at MLconf ATL 2016
MLconf
 
PDF
Building A Machine Learning Platform At Quora (1)
Nikhil Garg
 
PDF
Erin LeDell, Machine Learning Scientist, H2O.ai at MLconf ATL 2016
MLconf
 
PDF
Nikhil Garg, Engineering Manager, Quora at MLconf SF 2016
MLconf
 
PDF
Balancing Automation and Explanation in Machine Learning
Databricks
 
PPTX
Josh Patterson, Advisor, Skymind – Deep learning for Industry at MLconf ATL 2016
MLconf
 
PPTX
Managing and Versioning Machine Learning Models in Python
Simon Frid
 
PDF
Challenges on Distributed Machine Learning
jie cao
 
PDF
MLeap: Productionize Data Science Workflows Using Spark
Jen Aman
 
PDF
Machine learning pipeline with spark ml
datamantra
 
PPTX
Braden Hancock "Programmatically creating and managing training data with Sno...
Fwdays
 
PDF
Ryan Curtin, Principal Research Scientist, Symantec at MLconf ATL 2016
MLconf
 
PDF
Kaz Sato, Evangelist, Google at MLconf ATL 2016
MLconf
 
PDF
MLflow: Infrastructure for a Complete Machine Learning Life Cycle with Mani ...
Databricks
 
PDF
Continuous Delivery of Deep Transformer-Based NLP Models Using MLflow and AWS...
Databricks
 
PPTX
MALT: Distributed Data-Parallelism for Existing ML Applications (Distributed ...
asimkadav
 
PDF
MLLeap, or How to Productionize Data Science Workflows Using Spark by Mikha...
Spark Summit
 
PDF
Concept Drift: Monitoring Model Quality In Streaming ML Applications
Lightbend
 
Chris Fregly, Research Scientist, PipelineIO at MLconf ATL 2016
MLconf
 
Oleksii Moskalenko "Continuous Delivery of ML Pipelines to Production"
Fwdays
 
Tom Peters, Software Engineer, Ufora at MLconf ATL 2016
MLconf
 
Building A Machine Learning Platform At Quora (1)
Nikhil Garg
 
Erin LeDell, Machine Learning Scientist, H2O.ai at MLconf ATL 2016
MLconf
 
Nikhil Garg, Engineering Manager, Quora at MLconf SF 2016
MLconf
 
Balancing Automation and Explanation in Machine Learning
Databricks
 
Josh Patterson, Advisor, Skymind – Deep learning for Industry at MLconf ATL 2016
MLconf
 
Managing and Versioning Machine Learning Models in Python
Simon Frid
 
Challenges on Distributed Machine Learning
jie cao
 
MLeap: Productionize Data Science Workflows Using Spark
Jen Aman
 
Machine learning pipeline with spark ml
datamantra
 
Braden Hancock "Programmatically creating and managing training data with Sno...
Fwdays
 
Ryan Curtin, Principal Research Scientist, Symantec at MLconf ATL 2016
MLconf
 
Kaz Sato, Evangelist, Google at MLconf ATL 2016
MLconf
 
MLflow: Infrastructure for a Complete Machine Learning Life Cycle with Mani ...
Databricks
 
Continuous Delivery of Deep Transformer-Based NLP Models Using MLflow and AWS...
Databricks
 
MALT: Distributed Data-Parallelism for Existing ML Applications (Distributed ...
asimkadav
 
MLLeap, or How to Productionize Data Science Workflows Using Spark by Mikha...
Spark Summit
 
Concept Drift: Monitoring Model Quality In Streaming ML Applications
Lightbend
 

Similar to Apache Spark Based Hyper-Parameter Selection and Adaptive Model Tuning for Deep Neural Networks with Yonggang Hu and Chao Xue (20)

PDF
Apache Spark Based Hyper-Parameter Selection and Adaptive Model Tuning for De...
Databricks
 
PDF
Deep learning
Rouyun Pan
 
PDF
Python for Image Understanding: Deep Learning with Convolutional Neural Nets
Roelof Pieters
 
PDF
Using SigOpt to Tune Deep Learning Models with Nervana Cloud
SigOpt
 
PPTX
deeplearningpresentation-180625071236.pptx
JeetDesai14
 
PDF
C3 w1
Ajay Taneja
 
PDF
Using Bayesian Optimization to Tune Machine Learning Models
Scott Clark
 
PDF
Using Bayesian Optimization to Tune Machine Learning Models
SigOpt
 
PPTX
Deep learning presentation
Tunde Ajose-Ismail
 
PDF
Automated Hyperparameter Tuning, Scaling and Tracking
Databricks
 
PPTX
B4UConference_machine learning_deeplearning
Hoa Le
 
PPTX
Session_2_Introduction_to_Deep_Learning.pptx
aljeboorymuhammed
 
PPT
deepnet-lourentzou.ppt
yang947066
 
PPT
Overview of Deep Learning and its advantage
aqib296675
 
PPT
Introduction to Deep Learning presentation
johanericka2
 
PPT
Deep learning is a subset of machine learning and AI
leradiophysicien1
 
PPTX
Deep learning summary
ankit_ppt
 
PPTX
AI Class Topic 6: Easy Way to Learn Deep Learning AI Technologies
Value Amplify Consulting
 
PDF
ML in Android
Jose Antonio Corbacho
 
PDF
Taming Your Deep Learning Workflow by Determined AI
desmondchanatdet
 
Apache Spark Based Hyper-Parameter Selection and Adaptive Model Tuning for De...
Databricks
 
Deep learning
Rouyun Pan
 
Python for Image Understanding: Deep Learning with Convolutional Neural Nets
Roelof Pieters
 
Using SigOpt to Tune Deep Learning Models with Nervana Cloud
SigOpt
 
deeplearningpresentation-180625071236.pptx
JeetDesai14
 
Using Bayesian Optimization to Tune Machine Learning Models
Scott Clark
 
Using Bayesian Optimization to Tune Machine Learning Models
SigOpt
 
Deep learning presentation
Tunde Ajose-Ismail
 
Automated Hyperparameter Tuning, Scaling and Tracking
Databricks
 
B4UConference_machine learning_deeplearning
Hoa Le
 
Session_2_Introduction_to_Deep_Learning.pptx
aljeboorymuhammed
 
deepnet-lourentzou.ppt
yang947066
 
Overview of Deep Learning and its advantage
aqib296675
 
Introduction to Deep Learning presentation
johanericka2
 
Deep learning is a subset of machine learning and AI
leradiophysicien1
 
Deep learning summary
ankit_ppt
 
AI Class Topic 6: Easy Way to Learn Deep Learning AI Technologies
Value Amplify Consulting
 
ML in Android
Jose Antonio Corbacho
 
Taming Your Deep Learning Workflow by Determined AI
desmondchanatdet
 
Ad

More from Databricks (20)

PPTX
DW Migration Webinar-March 2022.pptx
Databricks
 
PPTX
Data Lakehouse Symposium | Day 1 | Part 1
Databricks
 
PPT
Data Lakehouse Symposium | Day 1 | Part 2
Databricks
 
PPTX
Data Lakehouse Symposium | Day 2
Databricks
 
PPTX
Data Lakehouse Symposium | Day 4
Databricks
 
PDF
5 Critical Steps to Clean Your Data Swamp When Migrating Off of Hadoop
Databricks
 
PDF
Democratizing Data Quality Through a Centralized Platform
Databricks
 
PDF
Learn to Use Databricks for Data Science
Databricks
 
PDF
Why APM Is Not the Same As ML Monitoring
Databricks
 
PDF
The Function, the Context, and the Data—Enabling ML Ops at Stitch Fix
Databricks
 
PDF
Stage Level Scheduling Improving Big Data and AI Integration
Databricks
 
PDF
Simplify Data Conversion from Spark to TensorFlow and PyTorch
Databricks
 
PDF
Scaling your Data Pipelines with Apache Spark on Kubernetes
Databricks
 
PDF
Scaling and Unifying SciKit Learn and Apache Spark Pipelines
Databricks
 
PDF
Sawtooth Windows for Feature Aggregations
Databricks
 
PDF
Redis + Apache Spark = Swiss Army Knife Meets Kitchen Sink
Databricks
 
PDF
Re-imagine Data Monitoring with whylogs and Spark
Databricks
 
PDF
Raven: End-to-end Optimization of ML Prediction Queries
Databricks
 
PDF
Processing Large Datasets for ADAS Applications using Apache Spark
Databricks
 
PDF
Massive Data Processing in Adobe Using Delta Lake
Databricks
 
DW Migration Webinar-March 2022.pptx
Databricks
 
Data Lakehouse Symposium | Day 1 | Part 1
Databricks
 
Data Lakehouse Symposium | Day 1 | Part 2
Databricks
 
Data Lakehouse Symposium | Day 2
Databricks
 
Data Lakehouse Symposium | Day 4
Databricks
 
5 Critical Steps to Clean Your Data Swamp When Migrating Off of Hadoop
Databricks
 
Democratizing Data Quality Through a Centralized Platform
Databricks
 
Learn to Use Databricks for Data Science
Databricks
 
Why APM Is Not the Same As ML Monitoring
Databricks
 
The Function, the Context, and the Data—Enabling ML Ops at Stitch Fix
Databricks
 
Stage Level Scheduling Improving Big Data and AI Integration
Databricks
 
Simplify Data Conversion from Spark to TensorFlow and PyTorch
Databricks
 
Scaling your Data Pipelines with Apache Spark on Kubernetes
Databricks
 
Scaling and Unifying SciKit Learn and Apache Spark Pipelines
Databricks
 
Sawtooth Windows for Feature Aggregations
Databricks
 
Redis + Apache Spark = Swiss Army Knife Meets Kitchen Sink
Databricks
 
Re-imagine Data Monitoring with whylogs and Spark
Databricks
 
Raven: End-to-end Optimization of ML Prediction Queries
Databricks
 
Processing Large Datasets for ADAS Applications using Apache Spark
Databricks
 
Massive Data Processing in Adobe Using Delta Lake
Databricks
 
Ad

Recently uploaded (20)

PPTX
ER_Model_Relationship_in_DBMS_Presentation.pptx
dharaadhvaryu1992
 
PDF
List of all the AI prompt cheat codes.pdf
Avijit Kumar Roy
 
PDF
apidays Helsinki & North 2025 - Monetizing AI APIs: The New API Economy, Alla...
apidays
 
PPTX
Climate Action.pptx action plan for climate
justfortalabat
 
PPTX
recruitment Presentation.pptxhdhshhshshhehh
devraj40467
 
PDF
apidays Helsinki & North 2025 - How (not) to run a Graphql Stewardship Group,...
apidays
 
PPT
Data base management system Transactions.ppt
gandhamcharan2006
 
PPTX
ER_Model_with_Diagrams_Presentation.pptx
dharaadhvaryu1992
 
PPTX
apidays Helsinki & North 2025 - API access control strategies beyond JWT bear...
apidays
 
PDF
AUDITABILITY & COMPLIANCE OF AI SYSTEMS IN HEALTHCARE
GAHI Youssef
 
PDF
Avatar for apidays apidays PRO June 07, 2025 0 5 apidays Helsinki & North 2...
apidays
 
PDF
OPPOTUS - Malaysias on Malaysia 1Q2025.pdf
Oppotus
 
PPTX
Numbers of a nation: how we estimate population statistics | Accessible slides
Office for National Statistics
 
PDF
Context Engineering for AI Agents, approaches, memories.pdf
Tamanna
 
PDF
Copia de Strategic Roadmap Infographics by Slidesgo.pptx (1).pdf
ssuserd4c6911
 
PDF
Data Chunking Strategies for RAG in 2025.pdf
Tamanna
 
PDF
WEF_Future_of_Global_Fintech_Second_Edition_2025.pdf
AproximacionAlFuturo
 
PDF
MusicVideoProjectRubric Animation production music video.pdf
ALBERTIANCASUGA
 
PDF
JavaScript - Good or Bad? Tips for Google Tag Manager
📊 Markus Baersch
 
PPTX
apidays Helsinki & North 2025 - Agentic AI: A Friend or Foe?, Merja Kajava (A...
apidays
 
ER_Model_Relationship_in_DBMS_Presentation.pptx
dharaadhvaryu1992
 
List of all the AI prompt cheat codes.pdf
Avijit Kumar Roy
 
apidays Helsinki & North 2025 - Monetizing AI APIs: The New API Economy, Alla...
apidays
 
Climate Action.pptx action plan for climate
justfortalabat
 
recruitment Presentation.pptxhdhshhshshhehh
devraj40467
 
apidays Helsinki & North 2025 - How (not) to run a Graphql Stewardship Group,...
apidays
 
Data base management system Transactions.ppt
gandhamcharan2006
 
ER_Model_with_Diagrams_Presentation.pptx
dharaadhvaryu1992
 
apidays Helsinki & North 2025 - API access control strategies beyond JWT bear...
apidays
 
AUDITABILITY & COMPLIANCE OF AI SYSTEMS IN HEALTHCARE
GAHI Youssef
 
Avatar for apidays apidays PRO June 07, 2025 0 5 apidays Helsinki & North 2...
apidays
 
OPPOTUS - Malaysias on Malaysia 1Q2025.pdf
Oppotus
 
Numbers of a nation: how we estimate population statistics | Accessible slides
Office for National Statistics
 
Context Engineering for AI Agents, approaches, memories.pdf
Tamanna
 
Copia de Strategic Roadmap Infographics by Slidesgo.pptx (1).pdf
ssuserd4c6911
 
Data Chunking Strategies for RAG in 2025.pdf
Tamanna
 
WEF_Future_of_Global_Fintech_Second_Edition_2025.pdf
AproximacionAlFuturo
 
MusicVideoProjectRubric Animation production music video.pdf
ALBERTIANCASUGA
 
JavaScript - Good or Bad? Tips for Google Tag Manager
📊 Markus Baersch
 
apidays Helsinki & North 2025 - Agentic AI: A Friend or Foe?, Merja Kajava (A...
apidays
 

Apache Spark Based Hyper-Parameter Selection and Adaptive Model Tuning for Deep Neural Networks with Yonggang Hu and Chao Xue

  • 1. YongGang Hu, Chao Xue, IBM Hyper-Parameter Selection and Adaptive Model Tuning for Deep Neural Networks #AssignedHashtagGoesHere
  • 3. Outline • Hyper-parameter selections & Neural network search – Bayesian optimization – Hyperband – Reinforcement learning • High performance AutoML – Transfer AutoML - Using historical hyper-parameter configurations for different tasks – Joint optimization with AutoML and finetune • Training visualization and Interactive tuning • Real world experience and Implementation 3
  • 4. From Hyper parameter/Network Search to On-line Tuning Monitor Advice Hyper-parameter & Network Search criterion and model parameters optimizing procedure parameters Real-time monitor for running application Real-time adviser for running application overflow underfitting overfitting divergence convergence checkthetraining process learning curve weight/gradient/a ctivation histogram and norm worst cases of training samples example: learning rate, batch size, momentum, learning rate scheme example: number of hidden units, number of layers, filter kernel size, max-pool size, weight decay Image recognition DeepLearningApplications Object detection Translation Others Optimize
  • 5. Standard AutoML--from Random search, Bayesian Optimization to Reinforcement Learning 5 Adaptive Random Search (Hyperband) Reinforceme nt Learning Bayesian Optimization Neural Network auto selected Hyperparameters auto selected Qs1 sN u1 uM
  • 6. Results of AdaRandom and Bayesian Optimization for Object detection 6 • The raw data (no data augment) 0 0.2 0.4 0.6 0.8 avg Best case Worst case mAP default AdaRandom(HyperBand) Bayesian Optimization Machine IBM P8 GPU Tesla K80 GPU MEM 10G CPU ppc64le CPU cores 160 MEM 55G Frequency 3.5 GHz OS Ubuntu 14.04 caffe version Py-faster- rcnn Dataset User-defined Model vgg16 § The average results for the three (HPP) hyperparameters combinations at 4000 iterations are (0.49, 0.52, 0.55), that is, using AdaRandom and Bayesian optimization recommended HPP, you can gain 6% and 12% in the average results comparing to default setting. AdaRandom method has more variance among different tries(train-test dataset split). The Bayesian models are more stable while having better average performance. § The below pictures show accuracy during the 0-4000 iterations, with different tries, under the two HPP configurations. We can see that: 1) It can be early stopped at about 1300 iterations. 2) the performance with different tries differ significantly, it caused by in some tries, training dataset has the invariance property according to test dataset, but some doesn’t have. It need to augment data to gain the stable performance. 3) different HPP combinations(models) may have different sensitivity to the different tries.
  • 7. Neural Network: C: Convolution, P: pooling, FC: Full connection. Default manual: (9 layers) C(3,32)+C(3,32)+P(3,3)+C(3,64)+C(3,64)+P (3,3)+FC(64)+FC(32)+FC(6) AdaRandom generated: (6 layers) C(3,128)+C(4,64)+C(3,32)+P(3,3)+FC(16)+ FC(6) Reinforce generated: (5 layers) C(3,32)+P(2,2)+P(3,2)+C(5,32)+FC(6) Evaluations with Acoustic Applications • The best results for the three networks are (0.703,0.673, 0.695) (the smaller the better), that is, using AdaRandom and Reinforce recommended models, you can gain 4.3% and 1.1% in the best results comparisons. The average result of the three networks is (0.817,0.776, 0.763), that is, the DL Insight recommended modes can increase about 5.0% and 6.6% in the average case performance. And from the standard deviation view, the recommended models are clearly more stable. • The CDF (cumulative distribution function) curve is more intuitive to illustrate the comparison of the three models(the more left the better). For example, using reinforce recommended model, ER has more than 60% probability (frequency) less than 0.75, while the default only has the 30%. 7 • We implement the AdaRandom (adaptive random search scheme) and Reinforce (reinforcement learning search scheme) methods to generate deep learning neural network automatically. • We are trying the new methods in different areas. Here is the example for acoustic. Default is the best scheme by manual tuning. Better accuracy More stableLower complexity
  • 8. 8 Traditional fine-tune Standalone AutoML Dataset Virtual Dataset Model ParameterWeights and bias Neural network Hyper-parameter Dataset group Dog Car Unknown Collaborative AutoML Transfer AutoML Architecture
  • 9. Upload datasets White-box analysis Black-box analysis Benchmark models selection Clients Clients Knowledge base Model selection Joint Optimization with transfer learning RL HyperBand Bayesian ParallelizationB Virtual datasets group AutoML process Path 1 Path1 Server Server Spark
  • 10. Challenges for AutoML with Transfer Learning • Training small user dataset leads to convergence problem à Transfer learning is needed • When considering transfer learning, the pretrained model need to be chosen, usually in the computer vision, we choose image-net as the base dataset to get the initial weights as the pretrained model, but it can’t fit many specific user datasets. • To solve this transfer learning’s problem, we can let the user to classify his dataset into some predefined categories, and in each category, the pretrained model was trained separately. It can improve the performance of transfer learning but involve user’s intervention with their datasets. • Using AutoML with transfer learning can improve transfer learning’s performance without user’s intervention. But considering the transfer learning’s properties, there are two challenges for AutoML: – Since reusing the initial weights, transfer learning limits the searching space of AutoML, how to use AutoML based on the pretrained model is a question. – We can’t use AutoML to build one model for every user dataset, it is too expensive. How to reuse the model for transfer learning is a question. 10
  • 11. Joint optimization: AutoML with the fine-tune
  • 12. Search space: • lr_policy: LR_POLICY stepsize: STEPSIZE gamma: GAMMA momentum: MOMENTUM solver_mode: GPU max_iter: MAX_ITER test_iter: TEST_ITER test_interval: TEST_INTERVAL base_lr: BASE_LR weight_decay: WEIGHT_DECAY solver_type: SGD layer { param { lr_mult: LR_MULT_C_W_0 decay_mult: DECAY_MULT_C_W_0 } param { lr_mult: LR_MULT_C_B_0 decay_mult: DECAY_MULT_C_B_0 } convolution_param { num_output: NUM_OUTPUT_0 pad: PAD_0 kernel_size: KERNEL_SIZE_C_0 group: GROUP_0 weight_filler { type: TYPE_C_W_0 std: STD_C_W_0 } bias_filler { type: TYPE_C_B_0 std: STD_C_B_0 } } } layer { name: "conv0_relu" type: TYPE_C_AF_0 bottom: "conv0" top: "conv0" } layer { name: "pool1" type: "Pooling" bottom: "conv0" top: "pool1" pooling_param { pool: AVE kernel_size: KERNEL_SIZE_P_0 stride: STRIDE_P_0 } } layer { name: "last_fc" type: "InnerProduct" bottom: "pool1" top: "last_fc" param { lr_mult: LR_MULT_FC_W_0 decay_mult: DECAY_MULT_FC_W_0 } param { lr_mult: LR_MULT_FC_B_0 decay_mult: DECAY_MULT_FC_B_0 } inner_product_param { num_output: OUTPUT_NUMS weight_filler { type: TYPE_FC_W_0 std: STD_FC_W_0 } bias_filler { type: TYPE_FC_B_0 std: STD_FC_B_0 } } } c p c p fc 1) Neural network at the last stage: Example: 2) Below hyper-parameter
  • 13. Results and Analysis 0 0.2 0.4 0.6 0.8 1 1.2 Food(30 epoch) TV show(10 epoch) Car(50 epoch) Sence(100 epoch) Action(20 epoch) Accuracy GoogleCAM+Finetune(Imagenet) GoogleCAM+Finetune(imagenet+ same category dataset) AutoML+Finetune(Imagenet) Advantages of AutoML with finetune: (+) Best accuracy (+) Most stable (+) Don’t need separate pretrained models by predefining dataset categories. No user’s interventions.
  • 14. 14 Expert optimization advice for hyper parameter selection and tuning Traffic light alerting for required parameter optimization with early stop advice and more CPU, GPU, memory utilization info, comms overhead, +++ Interactive Tuning of Hyper Parameters – DL Insight
  • 15. 15 CIFAR-10 Traditional method: underfitting Our method: still going down Auto-detection method for training process
  • 16. 16 Traditional method: underfitting Our method: good game MNIST Auto-detection method for training process
  • 17. 17 Interactive Tuning – Example - 1st run of Caffe CIFAR-10 We will run three times of caffe on spark, to show DL Insight’s capabilities for DL
  • 18. Interactive Tuning – Example - 1st run of Caffe 18 The DL Insight suggests user to stop the training early because overflow detected, and give the user advisor to add clip gradient to 0.1
  • 19. Interactive Tuning – Example - 2nd run of Caffe 19 On the second run of Caffe, we accept the suggestion which DL Insight gave, the add clip gradient to 0.1, what happened then?
  • 20. Interactive Tuning – Example - 2nd run of Caffe 20 DL Insight suggests user to stop the training
  • 21. Interactive Tuning – Example -3rd run of Caffe 21 On the third run of Caffe, we accept the suggestion which DL Insight gave, and more layers such as RELU, what happened then?
  • 22. Interactive Tuning – Example - 3rd run of Caffe 22 No more suggestion. You can do other things or have a rest while training the dataset. When you feel worried about it, just have a look at DL InsightJ
  • 23. 2 base_lr weight_decay momentum early stop batch size/rpn batch size split ratio max_iters display applause01 0.001 0.0005 0.9 yes divergence 128/256 0.7(94/39) 2000 10 appl = 0.3409 0.3409 iter=370 mAP=0.221060 runs Hyperparameters in suggestions AP mAP mAP at early stop iteration Parameters Interactive Tuning - Example: “Applause” datasets Good judgement
  • 24. 2 base_lr weight_de cay momentu m early stop batch size/rpn batch size split ratio max_iters display applause-aug01-01 0.001 0.0005 0.9 yes divergence 128/256 0.7(98/42) 2000 10 AP for appl = 0.4848 0.4848 iter=350 mAP=0.313152 runs Hyperparameters in suggestions Parameters AP mAP mAP at early stop iteration Good judgement Learning Curve Trend Judgement – ‘Applause’ (Continue)
  • 25. 2 Good judgement base_lr weight_de cay momentu m early stop batch size/rpn batch size split ratio max_iters display applause-aug03-01 0.001 0.0005 0.9 128/256 0.7(164/69) 2000 10 AP for appl = 0.7847 0.7847 iter=1510 mAP=0.784068 runs Hyperparameters in suggestions Parameters AP mAP mAP at early stop iteration Learning Curve Trend Judgement – ‘Applause’ (Continue)
  • 26. Hyper-Parameter Search Implementation 26 Search hyper-parameters space : – Learning rate – Decay rate – Batch size – Optimizer: • GradientDecedent, • Adadelta, • … – Momentum (for some optimizers) – LSTM hidden unit size Random, Bayesian, TPE Based Search Types IBM CONFIDENTIAL
  • 27. Hyper-Parameter Search Implementation Think 2018 / 5613A.pdf / March 22, 2018 / © 2018 IBM Corporation 27 Spark search jobs are generated dynamically and executed in parallel Random TPE Tree-based Parzen Estimator Bayesian Multitenant Spark Cluster IBM Spectrum Conductor
  • 28. Hyper-Parameter Search Implementation ModelTuningMgrl startModelAutoTuning() stopModelAutoTuning() deleteModelAutoTuning() TuningTask (Thread) run() TuningJobCtr runTuningJobs() SparkTuningJobCtrI TuningFrameworkMgr frameworkMgr HPTAlgorithm TuneInputParam inputP TuningJobCtr jobctl search() BayesianAlg RandomAlg TPEAlg PythonAlg UserPlugInAl g TuningFrameworkMgr initialize() getLossValueFromLog() prepareJob() CoSFrameWorkMgrl TfFrameworkMgrl PytorchFrameworkMg rl
  • 29. Enterprise Class Deep Learning Solution IBM Spectrum Conductor Deep Learning Impact, IBM PowerAI, IBM Storage 29 Monitoring&Reporting Workload Management / Scheduling Resource Management & Orchestration Native Services Management Services & Support Deep Learning Impact Tensorflow Caffe Red Hat Linux x86… IBM Power Systems IBM Storage IBM PowerAI IBM CONFIDENTIAL
  • 30. Reference • [1] David Schaffer, Darrell Whitley and Larry J Eshelman, Combinations of genetic algorithms and neural networks: A survey of the state of the art. International Workshop on Combinations of Genetic Algorithms and Neural Networks, 1992. • [2] J.Snoek, H.Larochelle and R.P.Adams, Practical Bayesian optimization of machine learning algorithms. In Advances in Neural Information Processing Systems(NIPS), 2012. 
 • [3] Bergstra, James and Yoshua Bengio, Random search for hyper-parameter optimization. Journal of Machine Learning Research, 2012. 
 • [4] Lisha Li, Kevin Jamieson and Giulia DeSalvo, HYPERBAND: BANDIT- BASED CONFIGURATION EVALUATION FOR HYPERPARAMETER OPTIMIZATION. ICLR, 2017. 
 • [5] James Bergstra, etc. Algorithms for Hyper-Parameter Optimization. Proceedings of the IEEE, 2012. 
 • [6] Bowen Baker, Otkrist Gupta, Nikhil Naik and RameshRaskar, DESIGNING NEURAL NETWORK ARCHITECTURES USING REINFORCEMENT LEARNING. ICLR, 2017.
  • 31. Notices and disclaimers Think 2018 / 5613A.pdf / March 22, 2018 / © 2018 IBM Corporation •© 2018 International Business Machines Corporation. No part of this document may be reproduced or transmitted in any form without written permission from IBM. •U.S. Government Users Restricted Rights — use, duplication or disclosure restricted by GSA ADP Schedule Contract with IBM. •Information in these presentations (including information relating to products that have not yet been announced by IBM) has been reviewed for accuracy as of the date of initial publication and could include unintentional technical or typographical errors. IBM shall have no responsibility to update this information. This document is distributed “as is” without any warranty, either express or implied. In no event, shall IBM be liable for any damage arising from the use of this information, including but not limited to, loss of data, business interruption, loss of profit or loss of opportunity. IBM products and services are warranted per the terms and conditions of the agreements under which they are provided. •IBM products are manufactured from new parts or new and used parts. In some cases, a product may not be new and may have been previously installed. Regardless, our warranty terms apply.” •Any statements regarding IBM's future direction, intent or product plans are subject to change or withdrawal without notice. •Performance data contained herein was generally obtained in a controlled, isolated environments. Customer examples are presented as illustrations of how those •customers have used IBM products and the results they may have achieved. Actual performance, cost, savings or other results in other operating environments may vary. •References in this document to IBM products, programs, or services does not imply that IBM intends to make such products, programs or services available in all countries in which IBM operates or does business. •Workshops, sessions and associated materials may have been prepared by independent session speakers, and do not necessarily reflect the views of IBM. All materials and discussions are provided for informational purposes only, and are neither intended to, nor shall constitute legal or other guidance or advice to any individual participant or their specific situation. •It is the customer’s responsibility to insure its own compliance with legal requirements and to obtain advice of competent legal counsel as to the identification and interpretation of any relevant laws and regulatory requirements that may affect the customer’s business and any actions the customer may need to take to comply with such laws. IBM does not provide legal advice or represent or warrant that its services or products will ensure that the customer follows any law. 31
  • 32. Notices and disclaimers continued Think 2018 / 5613A.pdf / March 22, 2018 / © 2018 IBM Corporation •Information concerning non-IBM products was obtained from the suppliers of those products, their published announcements or other publicly available sources. IBM has not tested those products about this publication and cannot confirm the accuracy of performance, compatibility or any other claims related to non-IBM products. Questions on the capabilities of non-IBM products should be addressed to the suppliers of those products. IBM does not warrant the quality of any third-party products, or the ability of any such third- party products to interoperate with IBM’s products. IBM expressly disclaims all warranties, expressed or implied, including but not limited to, the implied warranties of merchantability and fitness for a purpose. •The provision of the information contained herein is not intended to, and does not, grant any right or license under any IBM patents, copyrights, trademarks or other intellectual property right. •IBM, the IBM logo, ibm.com and [names of other referenced IBM products and services used in the presentation] are trademarks of International Business Machines Corporation, registered in many jurisdictions worldwide. Other product and service names might be trademarks of IBM or other companies. A current list of IBM trademarks is available on the Web at "Copyright and trademark information" at: www.ibm.com/legal/copytrade.shtml. 32
  • 33. Date, Time, Location & Duration Session title and Speaker Tue, June 5 | 11 AM 2010-2012, 30 mins Productionizing Spark ML Pipelines with the Portable Format for Analytics Nick Pentreath (IBM) Tue, June 5 | 2 PM 2018, 30 mins Making PySpark Amazing—From Faster UDFs to Dependency Management and Graphing! Holden Karau (Google) Bryan Cutler (IBM) Tue, June 5 | 2 PM Nook by 2001, 30 mins Making Data and AI Accessible for All Armand Ruiz Gabernet (IBM) Tue, June 5 | 2:40 PM 2002-2004, 30 mins Cognitive Database: An Apache Spark-Based AI-Enabled Relational Database System Rajesh Bordawekar (IBM T.J. Watson Research Center) Tue, June 5 | 3:20 PM 3016-3022, 30 mins Dynamic Priorities for Apache Spark Application’s Resource Allocations Michael Feiman (IBM Spectrum Computing) Shinnosuke Okada (IBM Canada Ltd.) Tue, June 5 | 3:20 PM 2001-2005, 30 mins Model Parallelism in Spark ML Cross-Validation Nick Pentreath (IBM) Bryan Cutler (IBM) Tue, June 5 | 3:20 PM 2007, 30 mins Serverless Machine Learning on Modern Hardware Using Apache Spark Patrick Stuedi (IBM) Tue, June 5 | 5:40 PM 2002-2004, 30 mins Create a Loyal Customer Base by Knowing Their Personality Using AI-Based Personality Recommendation Engine; Sourav Mazumder (IBM Analytics) Aradhna Tiwari (University of South Florida) Tue, June 5 | 5:40 PM 2007, 30 mins Transparent GPU Exploitation on Apache Spark Dr. Kazuaki Ishizaki (IBM) Madhusudanan Kandasamy (IBM) Tue, June 5 | 5:40 PM 2009-2011, 30 mins Apache Spark Based Hyper-Parameter Selection and Adaptive Model Tuning for Deep Neural Networks Yonggang Hu (IBM) Chao Xue (IBM) IBM Sessions at Spark+AI Summit 2018 (Tuesday, June 5) IBM Cloud / Watson and Cloud Platform / © 2018 IBM Corporation 33
  • 34. Date, Time, Location & Duration Session title and Speaker Wed, June 6 | 12:50 PM Birds of a Feather: Apache Arrow in Spark and More Bryan Cutler (IBM) Li Jin (Two Sigma Investments, LP) Wed, June 6 | 2 PM 2002-2004, 30 mins Deep Learning for Recommender Systems Nick Pentreath (IBM) ) Wed, June 6 | 3:20 PM 2018, 30 mins Bringing an AI Ecosystem to the Domain Expert and Enterprise AI Developer Frederick Reiss (IBM) Vijay Bommireddipalli (IBM Center for Open-Source Data & AI Technologies) IBM Sessions at Spark+AI Summit 2018 (Wednesday, June 6) IBM Cloud / Watson and Cloud Platform / © 2018 IBM Corporation 34 Meet us at IBM booth in the Expo area.