SlideShare a Scribd company logo
Keras with Tensorflow
back-end in R and Python
Longhow Lam
Agenda
• Introduction to neural networks &Deep learning
• Keras some examples
• Train from scratch
• Use pretrained models
• Fine tune
Introduction to neural networks
Neural network NEURAL NETWORK FOR BEGINNERS, JUST LINEAR REGRESSION
f Y = f(X,w) = w1 + w2X2 + w3X3 + w4X41
X2
X3
X4
w4
w3
w1
w2 Neural network compute node
f is the so-called activation function. This could be
the logit function, but other choices are possible.
There are no hidden layers.
There are four weights w’s that have to
be determined
Neural networks ONE HIDDEN LAYER, MATHEMATICAL FORMULATION
Age
Income
Region
Gender
X1
X2
X3
X4
Z1
Z2
Z3
f
X inputs Hidden layer z outputs
α1
β1
neural net prediction f = 𝑔 𝑇𝑌
𝑇𝑌 = 𝛽0𝑌 + 𝛽 𝑌
𝑇
𝑍
𝑍 𝑚 = 𝜎 𝛼0𝑚 + 𝛼 𝑚
𝑇 𝑋
The function σ is defined as:
𝜎(𝑥) =
1
1+𝑒−𝑥
𝝈 is also called the activation function,
In case of regression the function g is the Identify function I
In case of a binary classifier, g is the softmax 𝑔 𝑇𝑌 =
𝑒 𝑇 𝑌
𝑒 𝑇 𝑁+𝑒 𝑇 𝑌
The model weights w = (α , β) have to be estimated from the data
m = 1, ... ,M
number of nodes / neurons
in the hidden layer
Neural networks
Back propagation algorithm
is just gradient descent in numerical optimization terms
Randomly choose small values for all wi’ s. For each data point (observation) i :
• Calculate the neural net prediction fi
• Calculate the error, for example for regression squared error (yi – fi)2
• Calculate the sum of all errors: E = Σ (yi – fi)2
Adjust weights w according to:
A run through all observations is called an epoch
Stop if error E is small enough.
Training the weights
𝑤𝑖
𝑛𝑒𝑤
= 𝑤𝑖 + ∆𝑤𝑖
∆𝑤𝑖 = −𝛼
𝜕𝐸
𝜕𝑤𝑖
Deep learning
Deep learning LOOSELY DEFINED:
NEURAL NET WORK WITH MORE THAN 2 HIDDEN LAYERS
Don’t use deep learning for ‘simple’ business
analytics problems… it is really an overkill!
Keep it simple if you have ‘classical’ churn or
response models: logistics regression, trees,
or forests.
In this example all layers are fully connected (or also called dense layers)
Convolutional networks
For computer vision special structures are used.Usually not all layers fully connected.
We have so-called Convolutional layers and pooling layers.
Convolutional layer A, takes only from a
local window inputs from previous layer
Pooling layer ‘max’, takes max value of a bunch of inputs
But pictures are arrays…. No problem
These blocks of numbers are called “tensors” in linear algebra terms.
Calculations on these tensors can be done very fast in parallel on GPU’s
Training images
VGG19 deep learning networks structure
The model achieves 92.7% top-5 test
accuracy in ImageNet , which is a dataset of
over 14 million images belonging to 1000
classes. 143.mln weights!
Target output:
1000 classes
KERAS on Tensorflow
Keras
• Keras is a high-level neural networks API, written in Python and
capable of running on top of either TensorFlow or Theano.
• It was developed with a focus on enabling fast experimentation.
• Being able to go from idea to result with the least possible delay is
key to doing good research.
• Specifying models in keras is at a higher level than tensorflow, but you
still have lot’s of options
• There is now also an R interface (of course created by Rstudio… )
Simpel set-up “Architecture”
Tensorflow installed on a (linux) machine
Ideally with lots of GPU’s 
pip install keras
You’re good to go in
Python
(Jupyter notebooks)
install_github("rstudio/keras")
You’re good to go in
R / RStudio
Training from scratch: MNIST example
MNIST data:
70.000 handwritten digits with a
label (“0”, “1”,…,”9”)
Each image has a resolution of
28*28 pixels, so a 28 by 28 matrix
First a simple neural network in R
Treat image as a vector. It has length 784 (28by28), the number of
pixels. One hidden layer (fully connected)
Pixel 3
Pixel 2
Pixel 1
Pixel 783
Pixel 784
neuron 1
neuron 256
Label 0
Label 9
First a simple neural network in R
N of neurons time for 50 epochs Test accuracy
5 39 s 0.8988
15 51 s 0.9486
25 44 s 0.9626
50 51 s 0.9741
100 73 s 0.9751
256 125 s 0.9796
512 213 s 0.9825
1024 314 s 0.9830
2 dense (fully connected) layers
2 layer sec Test acc
64 *64 58 0.9705
128*128 88 0.9768
256*256 143 0.9797
512*512 349 0.9819
1024*1024 900 0.9835
Pixel 3
Pixel 2
Pixel 1
Pixel 783
Pixel 784
Label 0
Label 9
A more complex model in Python
Images are treated as matrices / arrays
• Convolutional layers
• Pooling layer
• Dropouts
• Dense last layer
Test loss: 0.028
Test Accuracy: 0.9912
Run time 25 minutes
Now compare with GPU
Some extra steps:
1. Spin up: Microsoft NC6 machine: 1 X Tesla K80 GPU ($1.084/hr)
2. Install CUDA toolkit / install cuDNN
3. pip install tensorflow-gpu
Run same model as in previous slide: Now it takes 2.9 minutes
Example predictions some are wrong….
1 vs 9 7 vs 2 8 vs 8 0 vs 6
Tensorboard
TensorBoard is a visualization tool included with TensorFlow
It enables you to visualize dynamic graphs of your Keras training and
test metrics, as well as activation histograms for the different layers in
your model.
model %>% fit(
x_train, y_train,
batch_size = batch_size,
epochs = epochs,
verbose = 2,
callbacks = callback_tensorboard(
log_dir = "logs/run_1",
write_images = TRUE
),
validation_split = 0.2
)
Now open a shell and start tensorboard, providing the log directory
Pre trained neural networks
Using pre-trained models
Image classifiers have been trained on big GPU machines
for weeks with millions of pictures on very large networks
Not many people do that from scratch. Instead, one can
use pre-trained networks and start from there.
predict image class using pretrained models
RTL NIEUWS Images trough resnet and vgg16
Link to trellisJS app
Images from Videos
Use ffmpeg: open source tool for video analyses Example call for Dutch series Family Kruys trailer
ffmpeg –i
"FAMILIE_KRUYS_TRAILER.mp4"
-s 600x400 –ss 00:00:05.000
-t 1200 -r 2.0
"FamKruys%03d.jpg"
And now tag them with vgg16
See my video analyser shiny app on github
Extract features using pre-trained models
Remove top layers for feature extraction
We have a 7*7*512 ‘feature’ tensor = 25.088 values
Only a few lines of R code
RTL NIEUWS Image similarity
1024 RTL Nieuws Sample pictures. Compute for each image the 25.088 feature values.
Calculate for each image the top 10 closest images, based on cosine similarity.
Little Shiny APP
Examples RTL Nieuws image similarities
Examples RTL Nieuws image similarities
Examples RTL Nieuws image similarities
Same can be done for Videoland ‘boxarts’
See little shiny app
The Brad Pitt similarity index
Take five Brad Pitt pictures
Run them trough the pre-trained
vgg16 and extract feature vectors.
This is a 5 by 25088 matrix
The brad Pit Index
Take other images, run them through the VGG16
Calculate the distances with the five Brad Pitt pictures and average:
0.771195 0.802654 0.714752 0.792587 0.8291976 0.8096944 0.665990 0.9737212
0.6273 0.5908 0.8231 0.7711 0.8839 0.8975 0.6934 0.9659
Focusing on only the face!!
Transfer learning / Fine tune
pre trained models
Transfer learning or
finetune pre-trained models
Train new image classifiers on limited training cases
• Get a pretrained model, say VGG16
• Remove existing top layers
• Add your own (fully) connected layer(s)
• Fix all the parameters except for your layers
• Use your (limited) samples as train cases to train the
weights of your layers.
Python code example
base_model = VGG16(weights='imagenet', include_top=False)
x = base_model.output
x = GlobalAveragePooling2D()(x)
# let's add a fully-connected layer
x = Dense(256, activation='relu')(x)
# and a logistic layer -- 2 classes dogs and cats
predictions = Dense(2, activation='softmax')(x)
# this is the model we will train
model = Model(inputs=base_model.input, outputs=predictions)
# first: train only the top layers (which were randomly initialized)
# i.e. freeze all convolutional layers
for layer in base_model.layers:
layer.trainable = False
# compile the model (should be done *after* setting layers to non-trainable)
model.compile(optimizer='rmsprop', loss='categorical_crossentropy', metrics =['accuracy'])
Python code example
1000 cats and dogs example

More Related Content

PPTX
Transformers AI PPT.pptx
PDF
Deep Learning for NLP (without Magic) - Richard Socher and Christopher Manning
PDF
PhD Defense
PPTX
neural network
PPTX
1.Introduction to deep learning
PDF
[Paper review] contrastive language image pre-training, open ai, 2020
PDF
TensorFlow and Keras: An Overview
PDF
Machine Learning
Transformers AI PPT.pptx
Deep Learning for NLP (without Magic) - Richard Socher and Christopher Manning
PhD Defense
neural network
1.Introduction to deep learning
[Paper review] contrastive language image pre-training, open ai, 2020
TensorFlow and Keras: An Overview
Machine Learning

What's hot (20)

PDF
Neural Networks: Rosenblatt's Perceptron
PPTX
AI Algorithms
PDF
Best Python Libraries For Data Science & Machine Learning | Edureka
PDF
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
PPTX
Python Scipy Numpy
PDF
Recurrent Neural Networks
PPT
Introduction to Natural Language Processing
PDF
Tableauから始める機械学習ーやってみようPython連携_2019-05-23
PDF
Recurrent Neural Networks (RNN) | RNN LSTM | Deep Learning Tutorial | Tensorf...
PPTX
Recurrent neural network
PDF
Hands on Explainable Recommender Systems with Knowledge Graphs @ RecSys22
PPTX
TensorFlow Tutorial | Deep Learning With TensorFlow | TensorFlow Tutorial For...
PPTX
What is TensorFlow? | Introduction to TensorFlow | TensorFlow Tutorial For Be...
PDF
P, NP, NP-Complete, and NP-Hard
PDF
kaggle Freesound Audio Tagging 2019 4th place solution
PPTX
Classification Algorithm.
PPT
artificial neural network
PDF
1 seaborn introduction
PPTX
Semi-Supervised Learning
PPTX
ONNX and MLflow
Neural Networks: Rosenblatt's Perceptron
AI Algorithms
Best Python Libraries For Data Science & Machine Learning | Edureka
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
Python Scipy Numpy
Recurrent Neural Networks
Introduction to Natural Language Processing
Tableauから始める機械学習ーやってみようPython連携_2019-05-23
Recurrent Neural Networks (RNN) | RNN LSTM | Deep Learning Tutorial | Tensorf...
Recurrent neural network
Hands on Explainable Recommender Systems with Knowledge Graphs @ RecSys22
TensorFlow Tutorial | Deep Learning With TensorFlow | TensorFlow Tutorial For...
What is TensorFlow? | Introduction to TensorFlow | TensorFlow Tutorial For Be...
P, NP, NP-Complete, and NP-Hard
kaggle Freesound Audio Tagging 2019 4th place solution
Classification Algorithm.
artificial neural network
1 seaborn introduction
Semi-Supervised Learning
ONNX and MLflow
Ad

Similar to Keras on tensorflow in R & Python (20)

PDF
Separating Hype from Reality in Deep Learning with Sameer Farooqui
PPTX
Learn to Build an App to Find Similar Images using Deep Learning- Piotr Teterwak
PPTX
Android and Deep Learning
PPTX
Machine Learning, Deep Learning and Data Analysis Introduction
PPTX
Diving into Deep Learning (Silicon Valley Code Camp 2017)
PPTX
Introduction to Deep Learning and Tensorflow
PDF
Deep Dive on Deep Learning (June 2018)
PDF
A Platform for Accelerating Machine Learning Applications
PDF
Neural network image recognition
PPT
backpropagation in neural networks
PDF
BMVA summer school MATLAB programming tutorial
PPTX
Final training course
PDF
Introduction to Applied Machine Learning
PPTX
Deep Learning: R with Keras and TensorFlow
PDF
Transfer Learning (20230516)
 
PDF
NVIDIA深度學習教育機構 (DLI): Object detection with jetson
PDF
Overview of Chainer and Its Features
PDF
Handwritten Digit Recognition using Convolutional Neural Networks
PPTX
Java and Deep Learning (Introduction)
PPTX
Deep learning
Separating Hype from Reality in Deep Learning with Sameer Farooqui
Learn to Build an App to Find Similar Images using Deep Learning- Piotr Teterwak
Android and Deep Learning
Machine Learning, Deep Learning and Data Analysis Introduction
Diving into Deep Learning (Silicon Valley Code Camp 2017)
Introduction to Deep Learning and Tensorflow
Deep Dive on Deep Learning (June 2018)
A Platform for Accelerating Machine Learning Applications
Neural network image recognition
backpropagation in neural networks
BMVA summer school MATLAB programming tutorial
Final training course
Introduction to Applied Machine Learning
Deep Learning: R with Keras and TensorFlow
Transfer Learning (20230516)
 
NVIDIA深度學習教育機構 (DLI): Object detection with jetson
Overview of Chainer and Its Features
Handwritten Digit Recognition using Convolutional Neural Networks
Java and Deep Learning (Introduction)
Deep learning
Ad

More from Longhow Lam (14)

PPTX
Xomia_20220602.pptx
PDF
A Unifying theory for blockchain and AI
PDF
Data science inspiratie_sessie
PDF
Jaap Huisprijzen, GTST, The Bold, IKEA en Iens
PDF
text2vec SatRDay Amsterdam
PDF
Dataiku meetup 12 july 2018 Amsterdam
PPTX
Data science in action
PPTX
MasterSearch_Meetup_AdvancedAnalytics
PDF
Latent transwarp neural networks
PDF
MathPaperPublished
PPTX
Heliview 29sep2015 slideshare
PDF
Parameter estimation in a non stationary markov model
PDF
The analysis of doubly censored survival data
PPTX
Machine learning overview (with SAS software)
Xomia_20220602.pptx
A Unifying theory for blockchain and AI
Data science inspiratie_sessie
Jaap Huisprijzen, GTST, The Bold, IKEA en Iens
text2vec SatRDay Amsterdam
Dataiku meetup 12 july 2018 Amsterdam
Data science in action
MasterSearch_Meetup_AdvancedAnalytics
Latent transwarp neural networks
MathPaperPublished
Heliview 29sep2015 slideshare
Parameter estimation in a non stationary markov model
The analysis of doubly censored survival data
Machine learning overview (with SAS software)

Recently uploaded (20)

PPTX
DISORDERS OF THE LIVER, GALLBLADDER AND PANCREASE (1).pptx
PPTX
Major-Components-ofNKJNNKNKNKNKronment.pptx
PPTX
Global journeys: estimating international migration
PPTX
Introduction to Firewall Analytics - Interfirewall and Transfirewall.pptx
PPT
Chapter 3 METAL JOINING.pptnnnnnnnnnnnnn
PPT
Reliability_Chapter_ presentation 1221.5784
PPTX
Moving the Public Sector (Government) to a Digital Adoption
PPTX
The THESIS FINAL-DEFENSE-PRESENTATION.pptx
PPTX
Introduction to Knowledge Engineering Part 1
PPTX
Supervised vs unsupervised machine learning algorithms
PDF
Lecture1 pattern recognition............
PDF
Clinical guidelines as a resource for EBP(1).pdf
PPTX
advance b rammar.pptxfdgdfgdfsgdfgsdgfdfgdfgsdfgdfgdfg
PPTX
Business Ppt On Nestle.pptx huunnnhhgfvu
PPTX
oil_refinery_comprehensive_20250804084928 (1).pptx
PPTX
Business Acumen Training GuidePresentation.pptx
PPTX
Data_Analytics_and_PowerBI_Presentation.pptx
PPTX
MODULE 8 - DISASTER risk PREPAREDNESS.pptx
PDF
“Getting Started with Data Analytics Using R – Concepts, Tools & Case Studies”
DISORDERS OF THE LIVER, GALLBLADDER AND PANCREASE (1).pptx
Major-Components-ofNKJNNKNKNKNKronment.pptx
Global journeys: estimating international migration
Introduction to Firewall Analytics - Interfirewall and Transfirewall.pptx
Chapter 3 METAL JOINING.pptnnnnnnnnnnnnn
Reliability_Chapter_ presentation 1221.5784
Moving the Public Sector (Government) to a Digital Adoption
The THESIS FINAL-DEFENSE-PRESENTATION.pptx
Introduction to Knowledge Engineering Part 1
Supervised vs unsupervised machine learning algorithms
Lecture1 pattern recognition............
Clinical guidelines as a resource for EBP(1).pdf
advance b rammar.pptxfdgdfgdfsgdfgsdgfdfgdfgsdfgdfgdfg
Business Ppt On Nestle.pptx huunnnhhgfvu
oil_refinery_comprehensive_20250804084928 (1).pptx
Business Acumen Training GuidePresentation.pptx
Data_Analytics_and_PowerBI_Presentation.pptx
MODULE 8 - DISASTER risk PREPAREDNESS.pptx
“Getting Started with Data Analytics Using R – Concepts, Tools & Case Studies”

Keras on tensorflow in R & Python

  • 1. Keras with Tensorflow back-end in R and Python Longhow Lam
  • 2. Agenda • Introduction to neural networks &Deep learning • Keras some examples • Train from scratch • Use pretrained models • Fine tune
  • 4. Neural network NEURAL NETWORK FOR BEGINNERS, JUST LINEAR REGRESSION f Y = f(X,w) = w1 + w2X2 + w3X3 + w4X41 X2 X3 X4 w4 w3 w1 w2 Neural network compute node f is the so-called activation function. This could be the logit function, but other choices are possible. There are no hidden layers. There are four weights w’s that have to be determined
  • 5. Neural networks ONE HIDDEN LAYER, MATHEMATICAL FORMULATION Age Income Region Gender X1 X2 X3 X4 Z1 Z2 Z3 f X inputs Hidden layer z outputs α1 β1 neural net prediction f = 𝑔 𝑇𝑌 𝑇𝑌 = 𝛽0𝑌 + 𝛽 𝑌 𝑇 𝑍 𝑍 𝑚 = 𝜎 𝛼0𝑚 + 𝛼 𝑚 𝑇 𝑋 The function σ is defined as: 𝜎(𝑥) = 1 1+𝑒−𝑥 𝝈 is also called the activation function, In case of regression the function g is the Identify function I In case of a binary classifier, g is the softmax 𝑔 𝑇𝑌 = 𝑒 𝑇 𝑌 𝑒 𝑇 𝑁+𝑒 𝑇 𝑌 The model weights w = (α , β) have to be estimated from the data m = 1, ... ,M number of nodes / neurons in the hidden layer
  • 6. Neural networks Back propagation algorithm is just gradient descent in numerical optimization terms Randomly choose small values for all wi’ s. For each data point (observation) i : • Calculate the neural net prediction fi • Calculate the error, for example for regression squared error (yi – fi)2 • Calculate the sum of all errors: E = Σ (yi – fi)2 Adjust weights w according to: A run through all observations is called an epoch Stop if error E is small enough. Training the weights 𝑤𝑖 𝑛𝑒𝑤 = 𝑤𝑖 + ∆𝑤𝑖 ∆𝑤𝑖 = −𝛼 𝜕𝐸 𝜕𝑤𝑖
  • 8. Deep learning LOOSELY DEFINED: NEURAL NET WORK WITH MORE THAN 2 HIDDEN LAYERS Don’t use deep learning for ‘simple’ business analytics problems… it is really an overkill! Keep it simple if you have ‘classical’ churn or response models: logistics regression, trees, or forests. In this example all layers are fully connected (or also called dense layers)
  • 9. Convolutional networks For computer vision special structures are used.Usually not all layers fully connected. We have so-called Convolutional layers and pooling layers. Convolutional layer A, takes only from a local window inputs from previous layer Pooling layer ‘max’, takes max value of a bunch of inputs
  • 10. But pictures are arrays…. No problem These blocks of numbers are called “tensors” in linear algebra terms. Calculations on these tensors can be done very fast in parallel on GPU’s
  • 11. Training images VGG19 deep learning networks structure The model achieves 92.7% top-5 test accuracy in ImageNet , which is a dataset of over 14 million images belonging to 1000 classes. 143.mln weights! Target output: 1000 classes
  • 13. Keras • Keras is a high-level neural networks API, written in Python and capable of running on top of either TensorFlow or Theano. • It was developed with a focus on enabling fast experimentation. • Being able to go from idea to result with the least possible delay is key to doing good research. • Specifying models in keras is at a higher level than tensorflow, but you still have lot’s of options • There is now also an R interface (of course created by Rstudio… )
  • 14. Simpel set-up “Architecture” Tensorflow installed on a (linux) machine Ideally with lots of GPU’s  pip install keras You’re good to go in Python (Jupyter notebooks) install_github("rstudio/keras") You’re good to go in R / RStudio
  • 15. Training from scratch: MNIST example MNIST data: 70.000 handwritten digits with a label (“0”, “1”,…,”9”) Each image has a resolution of 28*28 pixels, so a 28 by 28 matrix
  • 16. First a simple neural network in R Treat image as a vector. It has length 784 (28by28), the number of pixels. One hidden layer (fully connected) Pixel 3 Pixel 2 Pixel 1 Pixel 783 Pixel 784 neuron 1 neuron 256 Label 0 Label 9
  • 17. First a simple neural network in R N of neurons time for 50 epochs Test accuracy 5 39 s 0.8988 15 51 s 0.9486 25 44 s 0.9626 50 51 s 0.9741 100 73 s 0.9751 256 125 s 0.9796 512 213 s 0.9825 1024 314 s 0.9830
  • 18. 2 dense (fully connected) layers 2 layer sec Test acc 64 *64 58 0.9705 128*128 88 0.9768 256*256 143 0.9797 512*512 349 0.9819 1024*1024 900 0.9835 Pixel 3 Pixel 2 Pixel 1 Pixel 783 Pixel 784 Label 0 Label 9
  • 19. A more complex model in Python Images are treated as matrices / arrays • Convolutional layers • Pooling layer • Dropouts • Dense last layer
  • 20. Test loss: 0.028 Test Accuracy: 0.9912 Run time 25 minutes
  • 21. Now compare with GPU Some extra steps: 1. Spin up: Microsoft NC6 machine: 1 X Tesla K80 GPU ($1.084/hr) 2. Install CUDA toolkit / install cuDNN 3. pip install tensorflow-gpu Run same model as in previous slide: Now it takes 2.9 minutes
  • 22. Example predictions some are wrong…. 1 vs 9 7 vs 2 8 vs 8 0 vs 6
  • 23. Tensorboard TensorBoard is a visualization tool included with TensorFlow It enables you to visualize dynamic graphs of your Keras training and test metrics, as well as activation histograms for the different layers in your model. model %>% fit( x_train, y_train, batch_size = batch_size, epochs = epochs, verbose = 2, callbacks = callback_tensorboard( log_dir = "logs/run_1", write_images = TRUE ), validation_split = 0.2 )
  • 24. Now open a shell and start tensorboard, providing the log directory
  • 25. Pre trained neural networks
  • 26. Using pre-trained models Image classifiers have been trained on big GPU machines for weeks with millions of pictures on very large networks Not many people do that from scratch. Instead, one can use pre-trained networks and start from there.
  • 27. predict image class using pretrained models
  • 28. RTL NIEUWS Images trough resnet and vgg16 Link to trellisJS app
  • 29. Images from Videos Use ffmpeg: open source tool for video analyses Example call for Dutch series Family Kruys trailer ffmpeg –i "FAMILIE_KRUYS_TRAILER.mp4" -s 600x400 –ss 00:00:05.000 -t 1200 -r 2.0 "FamKruys%03d.jpg"
  • 30. And now tag them with vgg16
  • 31. See my video analyser shiny app on github
  • 32. Extract features using pre-trained models Remove top layers for feature extraction We have a 7*7*512 ‘feature’ tensor = 25.088 values
  • 33. Only a few lines of R code
  • 34. RTL NIEUWS Image similarity 1024 RTL Nieuws Sample pictures. Compute for each image the 25.088 feature values. Calculate for each image the top 10 closest images, based on cosine similarity. Little Shiny APP
  • 35. Examples RTL Nieuws image similarities
  • 36. Examples RTL Nieuws image similarities
  • 37. Examples RTL Nieuws image similarities
  • 38. Same can be done for Videoland ‘boxarts’ See little shiny app
  • 39. The Brad Pitt similarity index
  • 40. Take five Brad Pitt pictures Run them trough the pre-trained vgg16 and extract feature vectors. This is a 5 by 25088 matrix The brad Pit Index Take other images, run them through the VGG16 Calculate the distances with the five Brad Pitt pictures and average: 0.771195 0.802654 0.714752 0.792587 0.8291976 0.8096944 0.665990 0.9737212
  • 41. 0.6273 0.5908 0.8231 0.7711 0.8839 0.8975 0.6934 0.9659 Focusing on only the face!!
  • 42. Transfer learning / Fine tune pre trained models
  • 43. Transfer learning or finetune pre-trained models Train new image classifiers on limited training cases • Get a pretrained model, say VGG16 • Remove existing top layers • Add your own (fully) connected layer(s) • Fix all the parameters except for your layers • Use your (limited) samples as train cases to train the weights of your layers.
  • 44. Python code example base_model = VGG16(weights='imagenet', include_top=False) x = base_model.output x = GlobalAveragePooling2D()(x) # let's add a fully-connected layer x = Dense(256, activation='relu')(x) # and a logistic layer -- 2 classes dogs and cats predictions = Dense(2, activation='softmax')(x) # this is the model we will train model = Model(inputs=base_model.input, outputs=predictions) # first: train only the top layers (which were randomly initialized) # i.e. freeze all convolutional layers for layer in base_model.layers: layer.trainable = False # compile the model (should be done *after* setting layers to non-trainable) model.compile(optimizer='rmsprop', loss='categorical_crossentropy', metrics =['accuracy'])
  • 46. 1000 cats and dogs example

Editor's Notes

  • #6: binaire clasificatie  2 output nodes: Y en N 4 input variabelen  4 input nodes: X = (X1,..,X4) 1 hidden layer met 3 hidden nodes: Z = (Z1, Z2, Z3)