SlideShare a Scribd company logo
AutoML-Zero: Evolving Machine Learning
Algorithms From Scratch
Esteban Real et al (Google Brain / Research). In <ICML 2020>
발표자 : 윤지상
Graduate School of Information. Yonsei Univ.
Machine Learning & Computational Finance Lab.
1. Introduction
2. AutoML-Zero
3. Methods
4. Results
4.1. Finding Simple Neural Nets in a Difficult Space
4.2. Searching with Minimal Human Input
4.3. Discovering Algorithm Adaptations
5. Conclusion
INDEX
1 Introduction
1. Introduction
Machine create the ReLU as activation
function to improve its prediction
performance
1. Introduction
What is the best architecture to tackle this problem?
What is the best strategy to improve this model well?
Drop Out
Batch Normalization
Convolutional Neural Network
Transformer
GPT
YOLO
Res-Net
How Deep
Adam
Learning rate Decay
LSTM
PEGASUS
Cycle-GAN
BERT
NAS
Cut-Mix
Auto-Encoder
Seq2Seq
Anomaly Detection
Recommend System
Capsule-Net
Time series Forecasting Weight Initialization
1. Introduction
Eventually to make ‘Strong AI’,
machine should be able to find the best model
for each problems by itself.
1. Introduction
‘AutoML’ aims to automate progress in modeling for ML
by spending machine compute time instead of human research time
<NAS>
<NAS Net>
1. Introduction
<Mfas>
<PNAS>
<ENAS>
1. Introduction
Many Architecture search methods constrain their search space
by human such as methods like employing hand-designed layers as
building blocks and respecting the rules of backpropagation
Drawback :
1. Reduce the innovation potential of AutoML
2. Bias the search results depending on search space
3. Create a new burden on researchers to make search space
1. Introduction
Many Architecture search methods constrain their search space
by human such as methods like employing hand-designed layers as
building blocks and respecting the rules of backpropagation
Drawback :
1. Reduce the innovation potential of AutoML
2. Bias the search results depending on search space
3. Create a new burden on researchers to make search space
AutoML with little restriction on form and with simple mathematical operations as
building blocks is essential AutoML-Zero
2Auto-ML Zero
2. Auto-ML Zero
Find the strongest algorithm (the best model) in the JUNGLE
over the generations
using the evolutionary algorithm
2. Auto-ML Zero
Operation set
2. Auto-ML Zero
def Setup():
…
def Predict():
…
def Learn():
…
To exploit supervised learning method
population(= 𝑃)
1 2 3
4 5
2. Auto-ML Zero
def Setup():
…
def Predict():
…
def Learn():
…
Sample algorithms
X
1 2 3
4 5
Remove the oldest algorithm
𝑇 sample algorithms
2. Auto-ML Zero
def Setup():
…
def Predict():
…
def Learn():
…
Sample algorithms
Best
Select the best algorithm
Best
X
1 2 3
4 5
2. Auto-ML Zero
Operation set
def Setup():
…
def Predict():
…
def Learn():
…
Sample algorithms
Best
Select the best algorithm
Best
Mutate
New
Parent
child
X
1 2 3
4 5
2. Auto-ML Zero
Operation set
def Setup():
…
def Predict():
…
def Learn():
…
New operation
…
New
(6)
Sample algorithms
Best
Select the best algorithm
Best
Mutate
New
Parent
child
Put mutation to populations
2 3 5
4
2. Auto-ML Zero
Operation set
New
Remove and Sample
algorithms
Best
Select the best algorithm
Best
Mutate
New
Parent
child
Last
After N
iterations
Put mutation to population
Last candidates
Evaluate
2. Auto-ML Zero
Operation set
New
Remove and Sample
algorithms
Best
Select the best algorithm
Best
Mutate
New
Parent
child
Last
After N
iterations
Put mutation to population
2. Auto-ML Zero
 Contribution
1. AutoML-Zero, the proposal to automatically search for ML algorithms
from scratch with minimal human design
2. A novel framework with open-sourced code and a search space that
combines only basic mathematical operations
3. Detailed results to show potential through the discovery of nuanced
ML algorithms using evolutionary search.
3 Methods
3. Methods
 Components of Search space
1. 64 mathematical operations of high-school
level
2. Address integers at which scalar / vector /
matrix variable is located(e.g. m1, v0).
Range of these integers are restricted.
3. Real-valued constants of some operation as
input(e.g. gaussian(-0.033, 0.01))
3. Methods
Given set of ML Tasks 𝒯(e.g. binary classification of animals),
• subset 𝒯
𝑠𝑒𝑎𝑟𝑐ℎ ⊂ 𝒯(e.g. binary classification of {(dogs, cats), (lions, tigers)})
is used to find the candidate algorithms
• subset 𝒯
𝑠𝑒𝑙𝑒𝑐𝑡 ⊂ 𝒯(e.g. binary classification of {(wolves, foxes), (cows, pigs)})
is used to select the best algorithms
𝒯
𝑠𝑒𝑎𝑟𝑐ℎ is only used
during N iterations
Last
candidates
Evaluate and select the
best model using 𝒯
𝑠𝑒𝑙𝑒𝑐𝑡
( )
( )
3. Methods
Each task in 𝒯
𝑠𝑒𝑎𝑟𝑐ℎ train set / validation set exist
Using train set,
weights in each
algorithm are
updated
Using validation
set, evaluate
each algorithm
3. Methods
Each task in 𝒯
𝑠𝑒𝑎𝑟𝑐ℎ train set / validation set exist
Using train set,
weights in each
algorithm are
updated
Using validation
set, evaluate
each algorithm
mean_loss_1, mean_loss_2, … , mean_loss_len(𝒯
𝑠𝑒𝑎𝑟𝑐ℎ)
Sample 𝐷 mean_loss and select median
median_1, median_2, … , median_len(sample algo-)
3. Methods
1. Choose an instruction function randomly
from {Setup(), Predict(), Learn()}
2. Choose an action type to mutate randomly
from {Type(i), Type(ii), Type(iii)}
3. Methods
Add or remove
Randomize all
Modify an
argument
1. Choose an instruction function randomly
from {Setup(), Predict(), Learn()}
2. Choose an action type to mutate randomly
from {Type(i), Type(ii), Type(iii)}
<Mutation example>
3. Methods
In order to achieve speedup during model search,
1. Use functional equivalence checking (FEC)
(shortly, reusing same accuracy algorithm to reduce computation time)
2. Add Hurdles which is used as threshold of early-stopping
3. Exploit Parallel search with 𝑊 worker processes
4 Results
4. Results
4.1. Finding Simple Neural Nets in a Difficult Space
4.2. Searching with Minimal Human Input
4.3. Discovering Algorithm Adaptations
4. Results
4.1. Finding Simple Neural Nets in a Difficult Space
Experiments are conducted with AutoML-Zero of using
1) Random Search
2) Evolutionary Algorithm
4. Results
4.1. Finding Simple Neural Nets in a Difficult Space
type Linear Affine
formula 𝑢 ∙ 𝑥𝑖 𝑢 ∙ 𝑥𝑖 + 𝑎
1. Linear regression
Experiments are conducted with AutoML-Zero of using
1) Random Search
2) Evolutionary Algorithm
Experiment : Generate simple regression tasks and Restrict search
space with respect to tasks then Search the best model at each task
2. Nonlinear regression
type Nonlinear
formula 𝑢 ∙ 𝑅𝑒𝐿𝑈(𝑀𝑥𝑖)
4. Results
4.1. Finding Simple Neural Nets in a Difficult Space
The number of iterations for
which Random Search method
can get acceptable algorithm first
• With Evolutionary Algorithm, we can find acceptable algorithms (e.g. those with lower
mean RMS error than a hand-designed reference) more than when we use Random Search
1. Linear regression
4. Results
4.1. Finding Simple Neural Nets in a Difficult Space
• We can no longer find solutions with RS since search space is too large to find solution
• But, with the Evolutionary Algorithm, we can find the exact solution 𝑢 ∙ 𝑅𝑒𝐿𝑈(𝑀𝑥𝑖) when
𝐷 = 1
• As the numbers of 𝐷 and 𝒯
𝑠𝑒𝑎𝑟𝑐ℎ are increasing, evolution ‘invents’ forward pass and back-
propagation due to necessity of generalization
2. Nonlinear regression
( )
( )
( )
4. Results
4.2. Searching with Minimal Human Input
Experiments are conducted with binary classification tasks of
1) MNIST (total 45 tasks)
2) CIFAR-10 (total 45 tasks)
- Binary classification tasks are made by combining different labels such as (1,9) or
(4,6). So (10 x 9) / 2 = 45 tasks we can obtain
Experiment : Binary classification tasks of CIFAR-10
 To search more generalized architecture, we train population with
CIFAR-10 as well as MNIST data
 But performance of architecture will be evaluated only using CIFAR-10
4. Results
4.2. Searching with Minimal Human Input
4. Results
4.2. Searching with Minimal Human Input
• The last candidate algorithms in 13 out of 20 experiments perform better than hand-
designed 2-layer fully connected NN
• Result of various tasks is following:
model
task
AutoML-Zero
Logistic
Regression
2-layer FCNN
CIFAR-10 84.06±0.10% 77.65±0.22% 82.22±0.17%
SVHN 88.12% 59.58% 85.14%
ImageNet
(down sampled)
80.78% 76.44% 78.44%
Fashion MNIST 98.6% 97.9% 98.21%
4. Results
4.3. Discovering Algorithm Adaptations
Experiment : Binary classification of CIFAR-10 with only 80 train data for 100 epochs
1. Few training examples
Experiment : Binary classification of CIFAR-10 with 800 train data for only 10 epochs
2. Fast training
With the architecture, called SGD model, we already obtain, conduct 3 experiments
Experiment : 10 classification of CIFAR-10 with 45k train data
3. Multiple classes
4. Results
4.3. Discovering Algorithm Adaptations
1. Few training examples
2. Fast training
With the architecture, called SGD model, we already obtain, conduct 3 experiments
4. Results
4.3. Discovering Algorithm Adaptations
With the architecture, called SGD model, we already obtain, conduct 3 experiments
3. Multiple classes
5 Conclusion
5. Conclusion
 Search method can be improved by more sophisticated evolutionary
approaches such as Bayesian optimization, Reinforcement learning.
 Search space can be enhanced allowing for AutoML model to choose
more complicated operations.
 Interpreting evolved algorithms is required effort due to the complexity
of their raw code.
 AutoML can go further

More Related Content

PPTX
Radial basis function network ppt bySheetal,Samreen and Dhanashri
PPTX
Spiking neural network: an introduction I
PDF
Computing DFT using Matrix method
PPTX
TensorFlow Tutorial | Deep Learning With TensorFlow | TensorFlow Tutorial For...
PPT
Artificial neural network
PDF
matlab code of shifting and folding of two sequences
PDF
Image Restoration for 3D Computer Vision
Radial basis function network ppt bySheetal,Samreen and Dhanashri
Spiking neural network: an introduction I
Computing DFT using Matrix method
TensorFlow Tutorial | Deep Learning With TensorFlow | TensorFlow Tutorial For...
Artificial neural network
matlab code of shifting and folding of two sequences
Image Restoration for 3D Computer Vision

What's hot (20)

PDF
Notes from Coursera Deep Learning courses by Andrew Ng
PPTX
CNN Machine learning DeepLearning
PPTX
Particle Swarm Optimization
PPTX
Sunflower optimization algorithm
PPT
lecture12-clustering.ppt
PDF
GAN - Theory and Applications
PDF
1D Convolutional Neural Networks for Time Series Modeling - Nathan Janos, Jef...
PPTX
PPT - Enhancing the Locality and Breaking the Memory Bottleneck of Transforme...
PPTX
cnn ppt.pptx
PPTX
Introduction to CNN
PPTX
Convolutional Neural Network - CNN | How CNN Works | Deep Learning Course | S...
PDF
Introduction to Neural Networks
PDF
MATLAB Programming
PPTX
K nearest neighbor
PPT
PPTX
Graph Neural Network - Introduction
PDF
Neural Networks: Radial Bases Functions (RBF)
PPTX
Discrete Fourier Transform
PPTX
Attention in Deep Learning
Notes from Coursera Deep Learning courses by Andrew Ng
CNN Machine learning DeepLearning
Particle Swarm Optimization
Sunflower optimization algorithm
lecture12-clustering.ppt
GAN - Theory and Applications
1D Convolutional Neural Networks for Time Series Modeling - Nathan Janos, Jef...
PPT - Enhancing the Locality and Breaking the Memory Bottleneck of Transforme...
cnn ppt.pptx
Introduction to CNN
Convolutional Neural Network - CNN | How CNN Works | Deep Learning Course | S...
Introduction to Neural Networks
MATLAB Programming
K nearest neighbor
Graph Neural Network - Introduction
Neural Networks: Radial Bases Functions (RBF)
Discrete Fourier Transform
Attention in Deep Learning
Ad

Similar to PPT - AutoML-Zero: Evolving Machine Learning Algorithms From Scratch (20)

PDF
PR-232: AutoML-Zero:Evolving Machine Learning Algorithms From Scratch
PPTX
Everything you need to know about AutoML
PPTX
Automated Machine Learning (Auto ML)
PDF
AutoML lectures (ACDL 2019)
PPTX
Master defence 2020 -Volodymyr Lut-Neural Architecture Search: a Probabilisti...
PDF
Automatic Machine Learning, AutoML
PDF
Intro to machine learning
PDF
Cse 7th-sem-machine-learning-laboratory-csml1819
PPTX
Introduction to Machine Learningg
PDF
01_introduction.pdfbnmelllleitrthnjjjkkk
PDF
The Power of Auto ML and How Does it Work
PPTX
Machine learning to solve bioinformatics problems
PDF
PDF
01_introduction to machine learning algorithms and basics .pdf
PDF
A General Overview of Machine Learning
PDF
Overview of machine learning
PPTX
Machine learning
PDF
AutoML for Data Science Productivity and Toward Better Digital Decisions
PPTX
Launching into machine learning
PDF
GDG DEvFest Hellas 2020 - Automated ML - Panagiotis Papaemmanouil
PR-232: AutoML-Zero:Evolving Machine Learning Algorithms From Scratch
Everything you need to know about AutoML
Automated Machine Learning (Auto ML)
AutoML lectures (ACDL 2019)
Master defence 2020 -Volodymyr Lut-Neural Architecture Search: a Probabilisti...
Automatic Machine Learning, AutoML
Intro to machine learning
Cse 7th-sem-machine-learning-laboratory-csml1819
Introduction to Machine Learningg
01_introduction.pdfbnmelllleitrthnjjjkkk
The Power of Auto ML and How Does it Work
Machine learning to solve bioinformatics problems
01_introduction to machine learning algorithms and basics .pdf
A General Overview of Machine Learning
Overview of machine learning
Machine learning
AutoML for Data Science Productivity and Toward Better Digital Decisions
Launching into machine learning
GDG DEvFest Hellas 2020 - Automated ML - Panagiotis Papaemmanouil
Ad

Recently uploaded (20)

PPTX
advance b rammar.pptxfdgdfgdfsgdfgsdgfdfgdfgsdfgdfgdfg
PPTX
mbdjdhjjodule 5-1 rhfhhfjtjjhafbrhfnfbbfnb
PDF
“Getting Started with Data Analytics Using R – Concepts, Tools & Case Studies”
PPTX
MODULE 8 - DISASTER risk PREPAREDNESS.pptx
PPT
Reliability_Chapter_ presentation 1221.5784
PPTX
Acceptance and paychological effects of mandatory extra coach I classes.pptx
PPT
ISS -ESG Data flows What is ESG and HowHow
PPTX
climate analysis of Dhaka ,Banglades.pptx
PPTX
STUDY DESIGN details- Lt Col Maksud (21).pptx
PPTX
The THESIS FINAL-DEFENSE-PRESENTATION.pptx
PPTX
Introduction to Knowledge Engineering Part 1
PDF
Lecture1 pattern recognition............
PPTX
oil_refinery_comprehensive_20250804084928 (1).pptx
PPTX
Microsoft-Fabric-Unifying-Analytics-for-the-Modern-Enterprise Solution.pptx
PPTX
1_Introduction to advance data techniques.pptx
PDF
.pdf is not working space design for the following data for the following dat...
PDF
22.Patil - Early prediction of Alzheimer’s disease using convolutional neural...
PPTX
IB Computer Science - Internal Assessment.pptx
PDF
annual-report-2024-2025 original latest.
PPTX
Database Infoormation System (DBIS).pptx
advance b rammar.pptxfdgdfgdfsgdfgsdgfdfgdfgsdfgdfgdfg
mbdjdhjjodule 5-1 rhfhhfjtjjhafbrhfnfbbfnb
“Getting Started with Data Analytics Using R – Concepts, Tools & Case Studies”
MODULE 8 - DISASTER risk PREPAREDNESS.pptx
Reliability_Chapter_ presentation 1221.5784
Acceptance and paychological effects of mandatory extra coach I classes.pptx
ISS -ESG Data flows What is ESG and HowHow
climate analysis of Dhaka ,Banglades.pptx
STUDY DESIGN details- Lt Col Maksud (21).pptx
The THESIS FINAL-DEFENSE-PRESENTATION.pptx
Introduction to Knowledge Engineering Part 1
Lecture1 pattern recognition............
oil_refinery_comprehensive_20250804084928 (1).pptx
Microsoft-Fabric-Unifying-Analytics-for-the-Modern-Enterprise Solution.pptx
1_Introduction to advance data techniques.pptx
.pdf is not working space design for the following data for the following dat...
22.Patil - Early prediction of Alzheimer’s disease using convolutional neural...
IB Computer Science - Internal Assessment.pptx
annual-report-2024-2025 original latest.
Database Infoormation System (DBIS).pptx

PPT - AutoML-Zero: Evolving Machine Learning Algorithms From Scratch

  • 1. AutoML-Zero: Evolving Machine Learning Algorithms From Scratch Esteban Real et al (Google Brain / Research). In <ICML 2020> 발표자 : 윤지상 Graduate School of Information. Yonsei Univ. Machine Learning & Computational Finance Lab.
  • 2. 1. Introduction 2. AutoML-Zero 3. Methods 4. Results 4.1. Finding Simple Neural Nets in a Difficult Space 4.2. Searching with Minimal Human Input 4.3. Discovering Algorithm Adaptations 5. Conclusion INDEX
  • 4. 1. Introduction Machine create the ReLU as activation function to improve its prediction performance
  • 5. 1. Introduction What is the best architecture to tackle this problem? What is the best strategy to improve this model well? Drop Out Batch Normalization Convolutional Neural Network Transformer GPT YOLO Res-Net How Deep Adam Learning rate Decay LSTM PEGASUS Cycle-GAN BERT NAS Cut-Mix Auto-Encoder Seq2Seq Anomaly Detection Recommend System Capsule-Net Time series Forecasting Weight Initialization
  • 6. 1. Introduction Eventually to make ‘Strong AI’, machine should be able to find the best model for each problems by itself.
  • 7. 1. Introduction ‘AutoML’ aims to automate progress in modeling for ML by spending machine compute time instead of human research time <NAS> <NAS Net>
  • 9. 1. Introduction Many Architecture search methods constrain their search space by human such as methods like employing hand-designed layers as building blocks and respecting the rules of backpropagation Drawback : 1. Reduce the innovation potential of AutoML 2. Bias the search results depending on search space 3. Create a new burden on researchers to make search space
  • 10. 1. Introduction Many Architecture search methods constrain their search space by human such as methods like employing hand-designed layers as building blocks and respecting the rules of backpropagation Drawback : 1. Reduce the innovation potential of AutoML 2. Bias the search results depending on search space 3. Create a new burden on researchers to make search space AutoML with little restriction on form and with simple mathematical operations as building blocks is essential AutoML-Zero
  • 12. 2. Auto-ML Zero Find the strongest algorithm (the best model) in the JUNGLE over the generations using the evolutionary algorithm
  • 14. 2. Auto-ML Zero def Setup(): … def Predict(): … def Learn(): … To exploit supervised learning method population(= 𝑃) 1 2 3 4 5
  • 15. 2. Auto-ML Zero def Setup(): … def Predict(): … def Learn(): … Sample algorithms X 1 2 3 4 5 Remove the oldest algorithm 𝑇 sample algorithms
  • 16. 2. Auto-ML Zero def Setup(): … def Predict(): … def Learn(): … Sample algorithms Best Select the best algorithm Best X 1 2 3 4 5
  • 17. 2. Auto-ML Zero Operation set def Setup(): … def Predict(): … def Learn(): … Sample algorithms Best Select the best algorithm Best Mutate New Parent child X 1 2 3 4 5
  • 18. 2. Auto-ML Zero Operation set def Setup(): … def Predict(): … def Learn(): … New operation … New (6) Sample algorithms Best Select the best algorithm Best Mutate New Parent child Put mutation to populations 2 3 5 4
  • 19. 2. Auto-ML Zero Operation set New Remove and Sample algorithms Best Select the best algorithm Best Mutate New Parent child Last After N iterations Put mutation to population Last candidates Evaluate
  • 20. 2. Auto-ML Zero Operation set New Remove and Sample algorithms Best Select the best algorithm Best Mutate New Parent child Last After N iterations Put mutation to population
  • 21. 2. Auto-ML Zero  Contribution 1. AutoML-Zero, the proposal to automatically search for ML algorithms from scratch with minimal human design 2. A novel framework with open-sourced code and a search space that combines only basic mathematical operations 3. Detailed results to show potential through the discovery of nuanced ML algorithms using evolutionary search.
  • 23. 3. Methods  Components of Search space 1. 64 mathematical operations of high-school level 2. Address integers at which scalar / vector / matrix variable is located(e.g. m1, v0). Range of these integers are restricted. 3. Real-valued constants of some operation as input(e.g. gaussian(-0.033, 0.01))
  • 24. 3. Methods Given set of ML Tasks 𝒯(e.g. binary classification of animals), • subset 𝒯 𝑠𝑒𝑎𝑟𝑐ℎ ⊂ 𝒯(e.g. binary classification of {(dogs, cats), (lions, tigers)}) is used to find the candidate algorithms • subset 𝒯 𝑠𝑒𝑙𝑒𝑐𝑡 ⊂ 𝒯(e.g. binary classification of {(wolves, foxes), (cows, pigs)}) is used to select the best algorithms 𝒯 𝑠𝑒𝑎𝑟𝑐ℎ is only used during N iterations Last candidates Evaluate and select the best model using 𝒯 𝑠𝑒𝑙𝑒𝑐𝑡 ( ) ( )
  • 25. 3. Methods Each task in 𝒯 𝑠𝑒𝑎𝑟𝑐ℎ train set / validation set exist Using train set, weights in each algorithm are updated Using validation set, evaluate each algorithm
  • 26. 3. Methods Each task in 𝒯 𝑠𝑒𝑎𝑟𝑐ℎ train set / validation set exist Using train set, weights in each algorithm are updated Using validation set, evaluate each algorithm mean_loss_1, mean_loss_2, … , mean_loss_len(𝒯 𝑠𝑒𝑎𝑟𝑐ℎ) Sample 𝐷 mean_loss and select median median_1, median_2, … , median_len(sample algo-)
  • 27. 3. Methods 1. Choose an instruction function randomly from {Setup(), Predict(), Learn()} 2. Choose an action type to mutate randomly from {Type(i), Type(ii), Type(iii)}
  • 28. 3. Methods Add or remove Randomize all Modify an argument 1. Choose an instruction function randomly from {Setup(), Predict(), Learn()} 2. Choose an action type to mutate randomly from {Type(i), Type(ii), Type(iii)} <Mutation example>
  • 29. 3. Methods In order to achieve speedup during model search, 1. Use functional equivalence checking (FEC) (shortly, reusing same accuracy algorithm to reduce computation time) 2. Add Hurdles which is used as threshold of early-stopping 3. Exploit Parallel search with 𝑊 worker processes
  • 31. 4. Results 4.1. Finding Simple Neural Nets in a Difficult Space 4.2. Searching with Minimal Human Input 4.3. Discovering Algorithm Adaptations
  • 32. 4. Results 4.1. Finding Simple Neural Nets in a Difficult Space Experiments are conducted with AutoML-Zero of using 1) Random Search 2) Evolutionary Algorithm
  • 33. 4. Results 4.1. Finding Simple Neural Nets in a Difficult Space type Linear Affine formula 𝑢 ∙ 𝑥𝑖 𝑢 ∙ 𝑥𝑖 + 𝑎 1. Linear regression Experiments are conducted with AutoML-Zero of using 1) Random Search 2) Evolutionary Algorithm Experiment : Generate simple regression tasks and Restrict search space with respect to tasks then Search the best model at each task 2. Nonlinear regression type Nonlinear formula 𝑢 ∙ 𝑅𝑒𝐿𝑈(𝑀𝑥𝑖)
  • 34. 4. Results 4.1. Finding Simple Neural Nets in a Difficult Space The number of iterations for which Random Search method can get acceptable algorithm first • With Evolutionary Algorithm, we can find acceptable algorithms (e.g. those with lower mean RMS error than a hand-designed reference) more than when we use Random Search 1. Linear regression
  • 35. 4. Results 4.1. Finding Simple Neural Nets in a Difficult Space • We can no longer find solutions with RS since search space is too large to find solution • But, with the Evolutionary Algorithm, we can find the exact solution 𝑢 ∙ 𝑅𝑒𝐿𝑈(𝑀𝑥𝑖) when 𝐷 = 1 • As the numbers of 𝐷 and 𝒯 𝑠𝑒𝑎𝑟𝑐ℎ are increasing, evolution ‘invents’ forward pass and back- propagation due to necessity of generalization 2. Nonlinear regression ( ) ( ) ( )
  • 36. 4. Results 4.2. Searching with Minimal Human Input Experiments are conducted with binary classification tasks of 1) MNIST (total 45 tasks) 2) CIFAR-10 (total 45 tasks) - Binary classification tasks are made by combining different labels such as (1,9) or (4,6). So (10 x 9) / 2 = 45 tasks we can obtain Experiment : Binary classification tasks of CIFAR-10  To search more generalized architecture, we train population with CIFAR-10 as well as MNIST data  But performance of architecture will be evaluated only using CIFAR-10
  • 37. 4. Results 4.2. Searching with Minimal Human Input
  • 38. 4. Results 4.2. Searching with Minimal Human Input • The last candidate algorithms in 13 out of 20 experiments perform better than hand- designed 2-layer fully connected NN • Result of various tasks is following: model task AutoML-Zero Logistic Regression 2-layer FCNN CIFAR-10 84.06±0.10% 77.65±0.22% 82.22±0.17% SVHN 88.12% 59.58% 85.14% ImageNet (down sampled) 80.78% 76.44% 78.44% Fashion MNIST 98.6% 97.9% 98.21%
  • 39. 4. Results 4.3. Discovering Algorithm Adaptations Experiment : Binary classification of CIFAR-10 with only 80 train data for 100 epochs 1. Few training examples Experiment : Binary classification of CIFAR-10 with 800 train data for only 10 epochs 2. Fast training With the architecture, called SGD model, we already obtain, conduct 3 experiments Experiment : 10 classification of CIFAR-10 with 45k train data 3. Multiple classes
  • 40. 4. Results 4.3. Discovering Algorithm Adaptations 1. Few training examples 2. Fast training With the architecture, called SGD model, we already obtain, conduct 3 experiments
  • 41. 4. Results 4.3. Discovering Algorithm Adaptations With the architecture, called SGD model, we already obtain, conduct 3 experiments 3. Multiple classes
  • 43. 5. Conclusion  Search method can be improved by more sophisticated evolutionary approaches such as Bayesian optimization, Reinforcement learning.  Search space can be enhanced allowing for AutoML model to choose more complicated operations.  Interpreting evolved algorithms is required effort due to the complexity of their raw code.  AutoML can go further

Editor's Notes

  • #11: Feasibility 를 보여줌. 결과는 간단하게, parameter는 질문으로.