SlideShare a Scribd company logo
Best Practices for Hyperparameter Tuning
with
Joseph Bradley
April 24, 2019
Spark + AI Summit
About me
Joseph Bradley
• Software engineer at Databricks
• Apache Spark committer & PMC member
TEAM
About Databricks
Started Spark project (now Apache Spark) at UC Berkeley in 2009
PRODUCT
Unified Analytics Platform
MISSION
Making Big Data Simple
Try for free today.
databricks.com
Hyperparameters
• Express high-level concepts, such as statistical assumptions
• Are fixed before training or are hard to learn from data
• Affect objective, test time performance, computational cost
E.g.:
• Linear Regression: regularization, # iterations of optimization
• Neural Network: learning rate, # hidden layers
Tuning hyperparameters
E.g.: Fitting a
polynomial
Common goals:
• More flexible modeling process
• Reduced generalization error
• Faster training
• Plug & play ML
Challenges in tuning
Curse of dimensionality
Non-convex optimization
Computational cost
Unintuitive hyperparameters
Tuning in the Data Science workflow
Data
Tuning in the Data Science workflow
Training Data Test Data
ML Model
Tuning in the Data Science workflow
Training
Data
Validation
Data
Test Data
Final
ML Model
ML Model 1
ML Model 2
ML Model 3
Tuning in the Data Science workflow
ML Model
Featurization
Model family
selection
Hyperparameter
tuning
“AutoML” includes hyperparameter tuning.
This talk
Popular methods for hyperparameter tuning
• Overview of methods
• Comparing methods
• Open-source tools
Tuning in practice with MLflow
• Instrument tuning
• Analyze results
• Productionize models
Beyond this talk
This talk
Popular methods for hyperparameter tuning
• Overview of methods
• Comparing methods
• Open-source tools
Tuning in practice with MLflow
• Instrument tuning
• Analyze results
• Productionize models
Beyond this talk
Overview of tuning methods
• Manual search
• Grid search
• Random search
• Population-based algorithms
• Bayesian algorithms
Manual search
Select hyperparameter settings to try based on human intuition.
2 hyperparameters:
• [0, ..., 5]
• {A, B, ..., F}
Expert knowledge tells us to try:
(2,C), (2,D), (2,E), (3,C), (3,D), (3,E)
A B C D E F
0
1
2
3
4
5
Grid Search
Try points on a grid defined by ranges and step sizes
X-axis: {A,...,F}
Y-axis: 0-5, step = 1
A B C D E F
0
1
2
3
4
5
A B C D E F
0
1
2
3
4
5
Random Search
Sample from distributions over ranges
X-axis: Uniform({A,...,F})
Y-axis: Uniform([0,5])
Start with random search, then iterate:
• Use the previous “generation” to
inform the next generation
• E.g., sample from best performers &
then perturb them
Population Based Algorithms
A B C D E F
0
1
2
3
4
5
Start with random search, then iterate:
• Use the previous “generation” to
inform the next generation
• E.g., sample from best performers &
then perturb them
Population Based Algorithms
A B C D E F
0
1
2
3
4
5
Start with random search, then iterate:
• Use the previous “generation” to
inform the next generation
• E.g., sample from best performers &
then perturb them
Population Based Algorithms
A B C D E F
0
1
2
3
4
5
Model the loss function:
Hyperparameters à loss
Iteratively search space, trading off
between exploration and exploitation
A B C D E F
0
1
2
3
4
5
Bayesian Optimization
Bayesian OptimizationPerformance
Parameter Space
Bayesian OptimizationPerformance
Parameter Space
Bayesian OptimizationPerformance
Parameter Space
Bayesian OptimizationPerformance
Parameter Space
Bayesian OptimizationPerformance
Parameter Space
Bayesian OptimizationPerformance
Parameter Space
Bayesian OptimizationPerformance
Parameter Space
Bayesian OptimizationPerformance
Parameter Space
Bayesian OptimizationPerformance
Parameter Space
Bayesian OptimizationPerformance
Parameter Space
Bayesian OptimizationPerformance
Parameter Space
Bayesian OptimizationPerformance
Parameter Space
Bayesian OptimizationPerformance
Parameter Space
Bayesian OptimizationPerformance
Parameter Space
Bayesian OptimizationPerformance
Parameter Space
Comparing tuning methods
Iterative /
adaptive?
# evaluations
for P params
Model of
param space
Grid search No O(c^P) none
Random search No O(k) none
Population-based Yes O(k) implicit
Bayesian Yes O(k) explicit
Open-source tools for tuning
Grid
search
Random
search
Population
-based
Bayesian PyPi
downloads
last month
Github
stars
License
scikit-learn Yes Yes --- --- BSD
MLlib Yes --- --- Apache 2.0
scikit-
optimize
Yes 49,189 1,278 BSD
Hyperopt Yes Yes 98,282 3,286 BSD
DEAP Yes 26,700 2,789 LGPL v3
TPOT Yes 9,057 5,609 LGPL v3
GPyOpt Yes 4,959 451 BSD
As of mid-April 2019
This talk
Popular methods for hyperparameter tuning
• Overview of methods
• Comparing methods
• Open-source tools
Tuning in practice with MLflow
• Instrument tuning
• Analyze results
• Productionize models
Beyond this talk
This talk
Popular methods for hyperparameter tuning
• Overview of methods
• Comparing methods
• Open-source tools
Tuning in practice with MLflow
• Instrument tuning
• Analyze results
• Productionize models
Beyond this talk
Tracking
• Experiments
• Runs
• Parameters
• Metrics
• Tags & artifacts
Projects
• Directory or git
repository
• Entry points
• Environments
Models
• Storage format
• Flavors
• Deployment
tools
Organizing with
Training Data Validation Data Test Data
Final ML ModelML Model 1
ML Model 2
ML Model 3
Experiment
Main run
Child runs
Instrumenting tuning with
What to track in a run for a model
• Hyperparameters: all vs. ones being tuned
• Metric(s): training & validation, loss & objective, multiple objectives
• Tags: provenance, simple metadata
• Artifacts: serialized model, large metadata
Tip: Tune full pipeline, not 1 model.
Analyzing how tuning performs
Questions to answer
• Am I tuning the right hyperparameters?
• Am I exploring the right parts of the search space?
• Do I need to do another round of tuning?
Examining results
• Simple case: visualize param vs metric
• Challenges: multiple params and metrics, iterative experimentation
Moving models to production
Repeatable experiments via MLflow Projects
• Code checkpoints
• Environments
Model serialization via MLflow Models
• Flavors: TensorFlow, Keras, Spark, MLeap, ...
Deployment to prediction services
• Azure ML, AWS Sagemaker, Spark UDF
Auto-tracking MLlib with
Training Data Validation Data Test Data
Final ML ModelML Model 1
ML Model 2
ML Model 3
Experiment
Main run
Child runs
In Databricks
• CrossValidator &
TrainValidationSplit
• 1 run per setting of
hyperparameters
• Avg metrics for CV folds(demo)
This talk
Popular methods for hyperparameter tuning
• Overview of methods
• Comparing methods
• Open-source tools
Tuning in practice with MLflow
• Instrument tuning
• Analyze results
• Productionize models
Beyond this talk
This talk
Popular methods for hyperparameter tuning
• Overview of methods
• Comparing methods
• Open-source tools
Tuning in practice with MLflow
• Instrument tuning
• Analyze results
• Productionize models
Beyond this talk
Advanced topics
Efficient tuning
• Parallelizing hyperparameter
search
• Early stopping
• Transfer learning
Fancy tuning
• Multi-metric optimization
• Conditional/awkward
parameter spaces
Check out Maneesh Bhide’s talk:
"Advanced Hyperparameter
Optimization for Deep Learning"
to hear about early stopping,
multi-metric, & conditionals
Thursday @ 3:30pm, Room 3014
Advanced topics
Efficient tuning: Parallelizing
hyperparameter search
Challenge in analyzing results:
multiple parameters or
multiple metrics
Hyperopt + Apache Spark
+ MLflow integration
• Hyperopt: general tuning
library for ML in Python
• Spark integration:
parallelize model tuning in
batches
• MLflow integration: track
runs, analogous to MLlib +
MLflow integration
(demo)
Getting started
MLflow: http://mlflow.org
MLlib tuning
• Databricks auto-tracking with MLflow in private preview now, public
preview mid-May
Hyperopt
• Distributed tuning via Apache Spark: working to open-source the code
• Databricks auto-tracking with MLflow in public preview mid-May
Thank You!
Questions?
AMA @ DevLounge Theater
Thursday @ 10:30-11am
Thanks to Maneesh Bhide for
material for this talk!

More Related Content

PDF
Advanced Hyperparameter Optimization for Deep Learning with MLflow
PDF
Building a Feature Store around Dataframes and Apache Spark
PDF
Automated Hyperparameter Tuning, Scaling and Tracking
PDF
Apache Kylin: Speed Up Cubing with Apache Spark with Luke Han and Shaofeng Shi
PDF
MLflow: A Platform for Production Machine Learning
PPTX
The Evolution of Apache Kylin
PPTX
Apache Spark MLlib
PDF
KNIME Software Overview
Advanced Hyperparameter Optimization for Deep Learning with MLflow
Building a Feature Store around Dataframes and Apache Spark
Automated Hyperparameter Tuning, Scaling and Tracking
Apache Kylin: Speed Up Cubing with Apache Spark with Luke Han and Shaofeng Shi
MLflow: A Platform for Production Machine Learning
The Evolution of Apache Kylin
Apache Spark MLlib
KNIME Software Overview

What's hot (20)

PDF
Advanced MLflow: Multi-Step Workflows, Hyperparameter Tuning and Integrating ...
PDF
Netflix SIRT - Culture and Tech -Trainman
PDF
Jupyter, A Platform for Data Science at Scale
PDF
Collaborative filtering
PDF
Apache Superset at Airbnb
PDF
Machine Learning using Apache Spark MLlib
PDF
Intro to Neo4j and Graph Databases
PPT
Aggregate fact tables
PDF
Scipy 2011 Time Series Analysis in Python
PPTX
Basics of AB testing in online products
PPTX
Data cleansing
PDF
Approximate nearest neighbor methods and vector models – NYC ML meetup
PPTX
Anomaly Detection Technique
PDF
Hive Bucketing in Apache Spark with Tejas Patil
PPTX
Apache Beam: A unified model for batch and stream processing data
PDF
Stage Level Scheduling Improving Big Data and AI Integration
PDF
Feature Engineering
PPTX
Recommendation Systems Basics
PDF
Hyperspace for Delta Lake
PDF
MLflow Model Serving
Advanced MLflow: Multi-Step Workflows, Hyperparameter Tuning and Integrating ...
Netflix SIRT - Culture and Tech -Trainman
Jupyter, A Platform for Data Science at Scale
Collaborative filtering
Apache Superset at Airbnb
Machine Learning using Apache Spark MLlib
Intro to Neo4j and Graph Databases
Aggregate fact tables
Scipy 2011 Time Series Analysis in Python
Basics of AB testing in online products
Data cleansing
Approximate nearest neighbor methods and vector models – NYC ML meetup
Anomaly Detection Technique
Hive Bucketing in Apache Spark with Tejas Patil
Apache Beam: A unified model for batch and stream processing data
Stage Level Scheduling Improving Big Data and AI Integration
Feature Engineering
Recommendation Systems Basics
Hyperspace for Delta Lake
MLflow Model Serving
Ad

Similar to Best Practices for Hyperparameter Tuning with MLflow (20)

PDF
Evolving The Optimal Relevancy Scoring Model at Dice.com: Presented by Simon ...
PDF
Tuning ML Models: Scaling, Workflows, and Architecture
PPTX
Building High Available and Scalable Machine Learning Applications
PPTX
Deep-Dive into Deep Learning Pipelines with Sue Ann Hong and Tim Hunter
PDF
Spark Summit EU 2015: Combining the Strengths of MLlib, scikit-learn, and R
PDF
Guiding through a typical Machine Learning Pipeline
PDF
Strata San Jose 2016: Scalable Ensemble Learning with H2O
PPTX
Evolving the Optimal Relevancy Ranking Model at Dice.com
PPTX
Combining Machine Learning frameworks with Apache Spark
PPTX
Machine learning
PPTX
Apache Spark Model Deployment
PDF
The Power of Auto ML and How Does it Work
PDF
Scalable Automatic Machine Learning in H2O
PDF
From Labelling Open data images to building a private recommender system
PDF
201906 04 Overview of Automated ML June 2019
PDF
AlphaPy: A Data Science Pipeline in Python
PDF
AlphaPy
PDF
Taking the Pain out of Data Science - RecSys Machine Learning Framework Over ...
PPTX
Net campus2015 antimomusone
PPTX
PREDICT THE FUTURE , MACHINE LEARNING & BIG DATA
Evolving The Optimal Relevancy Scoring Model at Dice.com: Presented by Simon ...
Tuning ML Models: Scaling, Workflows, and Architecture
Building High Available and Scalable Machine Learning Applications
Deep-Dive into Deep Learning Pipelines with Sue Ann Hong and Tim Hunter
Spark Summit EU 2015: Combining the Strengths of MLlib, scikit-learn, and R
Guiding through a typical Machine Learning Pipeline
Strata San Jose 2016: Scalable Ensemble Learning with H2O
Evolving the Optimal Relevancy Ranking Model at Dice.com
Combining Machine Learning frameworks with Apache Spark
Machine learning
Apache Spark Model Deployment
The Power of Auto ML and How Does it Work
Scalable Automatic Machine Learning in H2O
From Labelling Open data images to building a private recommender system
201906 04 Overview of Automated ML June 2019
AlphaPy: A Data Science Pipeline in Python
AlphaPy
Taking the Pain out of Data Science - RecSys Machine Learning Framework Over ...
Net campus2015 antimomusone
PREDICT THE FUTURE , MACHINE LEARNING & BIG DATA
Ad

More from Databricks (20)

PPTX
DW Migration Webinar-March 2022.pptx
PPTX
Data Lakehouse Symposium | Day 1 | Part 1
PPT
Data Lakehouse Symposium | Day 1 | Part 2
PPTX
Data Lakehouse Symposium | Day 2
PPTX
Data Lakehouse Symposium | Day 4
PDF
5 Critical Steps to Clean Your Data Swamp When Migrating Off of Hadoop
PDF
Democratizing Data Quality Through a Centralized Platform
PDF
Learn to Use Databricks for Data Science
PDF
Why APM Is Not the Same As ML Monitoring
PDF
The Function, the Context, and the Data—Enabling ML Ops at Stitch Fix
PDF
Simplify Data Conversion from Spark to TensorFlow and PyTorch
PDF
Scaling your Data Pipelines with Apache Spark on Kubernetes
PDF
Scaling and Unifying SciKit Learn and Apache Spark Pipelines
PDF
Sawtooth Windows for Feature Aggregations
PDF
Redis + Apache Spark = Swiss Army Knife Meets Kitchen Sink
PDF
Re-imagine Data Monitoring with whylogs and Spark
PDF
Raven: End-to-end Optimization of ML Prediction Queries
PDF
Processing Large Datasets for ADAS Applications using Apache Spark
PDF
Massive Data Processing in Adobe Using Delta Lake
PDF
Machine Learning CI/CD for Email Attack Detection
DW Migration Webinar-March 2022.pptx
Data Lakehouse Symposium | Day 1 | Part 1
Data Lakehouse Symposium | Day 1 | Part 2
Data Lakehouse Symposium | Day 2
Data Lakehouse Symposium | Day 4
5 Critical Steps to Clean Your Data Swamp When Migrating Off of Hadoop
Democratizing Data Quality Through a Centralized Platform
Learn to Use Databricks for Data Science
Why APM Is Not the Same As ML Monitoring
The Function, the Context, and the Data—Enabling ML Ops at Stitch Fix
Simplify Data Conversion from Spark to TensorFlow and PyTorch
Scaling your Data Pipelines with Apache Spark on Kubernetes
Scaling and Unifying SciKit Learn and Apache Spark Pipelines
Sawtooth Windows for Feature Aggregations
Redis + Apache Spark = Swiss Army Knife Meets Kitchen Sink
Re-imagine Data Monitoring with whylogs and Spark
Raven: End-to-end Optimization of ML Prediction Queries
Processing Large Datasets for ADAS Applications using Apache Spark
Massive Data Processing in Adobe Using Delta Lake
Machine Learning CI/CD for Email Attack Detection

Recently uploaded (20)

PPTX
AI Strategy room jwfjksfksfjsjsjsjsjfsjfsj
PPTX
Introduction to Knowledge Engineering Part 1
PPTX
Database Infoormation System (DBIS).pptx
PPTX
IB Computer Science - Internal Assessment.pptx
PPTX
IBA_Chapter_11_Slides_Final_Accessible.pptx
PPTX
oil_refinery_comprehensive_20250804084928 (1).pptx
PDF
Clinical guidelines as a resource for EBP(1).pdf
PPTX
SAP 2 completion done . PRESENTATION.pptx
PPT
Miokarditis (Inflamasi pada Otot Jantung)
PDF
168300704-gasification-ppt.pdfhghhhsjsjhsuxush
PPTX
Business Ppt On Nestle.pptx huunnnhhgfvu
PPT
Quality review (1)_presentation of this 21
PPTX
STUDY DESIGN details- Lt Col Maksud (21).pptx
PDF
Business Analytics and business intelligence.pdf
PDF
[EN] Industrial Machine Downtime Prediction
PDF
Galatica Smart Energy Infrastructure Startup Pitch Deck
PDF
annual-report-2024-2025 original latest.
PDF
Recruitment and Placement PPT.pdfbjfibjdfbjfobj
PDF
Fluorescence-microscope_Botany_detailed content
PDF
Mega Projects Data Mega Projects Data
AI Strategy room jwfjksfksfjsjsjsjsjfsjfsj
Introduction to Knowledge Engineering Part 1
Database Infoormation System (DBIS).pptx
IB Computer Science - Internal Assessment.pptx
IBA_Chapter_11_Slides_Final_Accessible.pptx
oil_refinery_comprehensive_20250804084928 (1).pptx
Clinical guidelines as a resource for EBP(1).pdf
SAP 2 completion done . PRESENTATION.pptx
Miokarditis (Inflamasi pada Otot Jantung)
168300704-gasification-ppt.pdfhghhhsjsjhsuxush
Business Ppt On Nestle.pptx huunnnhhgfvu
Quality review (1)_presentation of this 21
STUDY DESIGN details- Lt Col Maksud (21).pptx
Business Analytics and business intelligence.pdf
[EN] Industrial Machine Downtime Prediction
Galatica Smart Energy Infrastructure Startup Pitch Deck
annual-report-2024-2025 original latest.
Recruitment and Placement PPT.pdfbjfibjdfbjfobj
Fluorescence-microscope_Botany_detailed content
Mega Projects Data Mega Projects Data

Best Practices for Hyperparameter Tuning with MLflow

  • 1. Best Practices for Hyperparameter Tuning with Joseph Bradley April 24, 2019 Spark + AI Summit
  • 2. About me Joseph Bradley • Software engineer at Databricks • Apache Spark committer & PMC member
  • 3. TEAM About Databricks Started Spark project (now Apache Spark) at UC Berkeley in 2009 PRODUCT Unified Analytics Platform MISSION Making Big Data Simple Try for free today. databricks.com
  • 4. Hyperparameters • Express high-level concepts, such as statistical assumptions • Are fixed before training or are hard to learn from data • Affect objective, test time performance, computational cost E.g.: • Linear Regression: regularization, # iterations of optimization • Neural Network: learning rate, # hidden layers
  • 5. Tuning hyperparameters E.g.: Fitting a polynomial Common goals: • More flexible modeling process • Reduced generalization error • Faster training • Plug & play ML
  • 6. Challenges in tuning Curse of dimensionality Non-convex optimization Computational cost Unintuitive hyperparameters
  • 7. Tuning in the Data Science workflow Data
  • 8. Tuning in the Data Science workflow Training Data Test Data ML Model
  • 9. Tuning in the Data Science workflow Training Data Validation Data Test Data Final ML Model ML Model 1 ML Model 2 ML Model 3
  • 10. Tuning in the Data Science workflow ML Model Featurization Model family selection Hyperparameter tuning “AutoML” includes hyperparameter tuning.
  • 11. This talk Popular methods for hyperparameter tuning • Overview of methods • Comparing methods • Open-source tools Tuning in practice with MLflow • Instrument tuning • Analyze results • Productionize models Beyond this talk
  • 12. This talk Popular methods for hyperparameter tuning • Overview of methods • Comparing methods • Open-source tools Tuning in practice with MLflow • Instrument tuning • Analyze results • Productionize models Beyond this talk
  • 13. Overview of tuning methods • Manual search • Grid search • Random search • Population-based algorithms • Bayesian algorithms
  • 14. Manual search Select hyperparameter settings to try based on human intuition. 2 hyperparameters: • [0, ..., 5] • {A, B, ..., F} Expert knowledge tells us to try: (2,C), (2,D), (2,E), (3,C), (3,D), (3,E) A B C D E F 0 1 2 3 4 5
  • 15. Grid Search Try points on a grid defined by ranges and step sizes X-axis: {A,...,F} Y-axis: 0-5, step = 1 A B C D E F 0 1 2 3 4 5
  • 16. A B C D E F 0 1 2 3 4 5 Random Search Sample from distributions over ranges X-axis: Uniform({A,...,F}) Y-axis: Uniform([0,5])
  • 17. Start with random search, then iterate: • Use the previous “generation” to inform the next generation • E.g., sample from best performers & then perturb them Population Based Algorithms A B C D E F 0 1 2 3 4 5
  • 18. Start with random search, then iterate: • Use the previous “generation” to inform the next generation • E.g., sample from best performers & then perturb them Population Based Algorithms A B C D E F 0 1 2 3 4 5
  • 19. Start with random search, then iterate: • Use the previous “generation” to inform the next generation • E.g., sample from best performers & then perturb them Population Based Algorithms A B C D E F 0 1 2 3 4 5
  • 20. Model the loss function: Hyperparameters à loss Iteratively search space, trading off between exploration and exploitation A B C D E F 0 1 2 3 4 5 Bayesian Optimization
  • 36. Comparing tuning methods Iterative / adaptive? # evaluations for P params Model of param space Grid search No O(c^P) none Random search No O(k) none Population-based Yes O(k) implicit Bayesian Yes O(k) explicit
  • 37. Open-source tools for tuning Grid search Random search Population -based Bayesian PyPi downloads last month Github stars License scikit-learn Yes Yes --- --- BSD MLlib Yes --- --- Apache 2.0 scikit- optimize Yes 49,189 1,278 BSD Hyperopt Yes Yes 98,282 3,286 BSD DEAP Yes 26,700 2,789 LGPL v3 TPOT Yes 9,057 5,609 LGPL v3 GPyOpt Yes 4,959 451 BSD As of mid-April 2019
  • 38. This talk Popular methods for hyperparameter tuning • Overview of methods • Comparing methods • Open-source tools Tuning in practice with MLflow • Instrument tuning • Analyze results • Productionize models Beyond this talk
  • 39. This talk Popular methods for hyperparameter tuning • Overview of methods • Comparing methods • Open-source tools Tuning in practice with MLflow • Instrument tuning • Analyze results • Productionize models Beyond this talk
  • 40. Tracking • Experiments • Runs • Parameters • Metrics • Tags & artifacts Projects • Directory or git repository • Entry points • Environments Models • Storage format • Flavors • Deployment tools
  • 41. Organizing with Training Data Validation Data Test Data Final ML ModelML Model 1 ML Model 2 ML Model 3 Experiment Main run Child runs
  • 42. Instrumenting tuning with What to track in a run for a model • Hyperparameters: all vs. ones being tuned • Metric(s): training & validation, loss & objective, multiple objectives • Tags: provenance, simple metadata • Artifacts: serialized model, large metadata Tip: Tune full pipeline, not 1 model.
  • 43. Analyzing how tuning performs Questions to answer • Am I tuning the right hyperparameters? • Am I exploring the right parts of the search space? • Do I need to do another round of tuning? Examining results • Simple case: visualize param vs metric • Challenges: multiple params and metrics, iterative experimentation
  • 44. Moving models to production Repeatable experiments via MLflow Projects • Code checkpoints • Environments Model serialization via MLflow Models • Flavors: TensorFlow, Keras, Spark, MLeap, ... Deployment to prediction services • Azure ML, AWS Sagemaker, Spark UDF
  • 45. Auto-tracking MLlib with Training Data Validation Data Test Data Final ML ModelML Model 1 ML Model 2 ML Model 3 Experiment Main run Child runs In Databricks • CrossValidator & TrainValidationSplit • 1 run per setting of hyperparameters • Avg metrics for CV folds(demo)
  • 46. This talk Popular methods for hyperparameter tuning • Overview of methods • Comparing methods • Open-source tools Tuning in practice with MLflow • Instrument tuning • Analyze results • Productionize models Beyond this talk
  • 47. This talk Popular methods for hyperparameter tuning • Overview of methods • Comparing methods • Open-source tools Tuning in practice with MLflow • Instrument tuning • Analyze results • Productionize models Beyond this talk
  • 48. Advanced topics Efficient tuning • Parallelizing hyperparameter search • Early stopping • Transfer learning Fancy tuning • Multi-metric optimization • Conditional/awkward parameter spaces Check out Maneesh Bhide’s talk: "Advanced Hyperparameter Optimization for Deep Learning" to hear about early stopping, multi-metric, & conditionals Thursday @ 3:30pm, Room 3014
  • 49. Advanced topics Efficient tuning: Parallelizing hyperparameter search Challenge in analyzing results: multiple parameters or multiple metrics Hyperopt + Apache Spark + MLflow integration • Hyperopt: general tuning library for ML in Python • Spark integration: parallelize model tuning in batches • MLflow integration: track runs, analogous to MLlib + MLflow integration (demo)
  • 50. Getting started MLflow: http://mlflow.org MLlib tuning • Databricks auto-tracking with MLflow in private preview now, public preview mid-May Hyperopt • Distributed tuning via Apache Spark: working to open-source the code • Databricks auto-tracking with MLflow in public preview mid-May
  • 51. Thank You! Questions? AMA @ DevLounge Theater Thursday @ 10:30-11am Thanks to Maneesh Bhide for material for this talk!