SlideShare a Scribd company logo
1
Deep Learning
Vlad Ovidiu Mihalca
PhD student – AI & Vision
Mechatronics @ University of Oradea
16th of January, 2018
2
Series of talks
●
Part I: theoretical base concepts
●
Part II: frameworks and DL industry use
●
Part III: usage in robotics and Computer Vision, state of
the art in research
3
2016 = Year of Deep Learning (1 / 7)
4
2016 = Year of Deep Learning (2 / 7)
5
2016 = Year of Deep Learning (3 / 7)
6
2016 = Year of Deep Learning (4 / 7)
7
2016 = Year of Deep Learning (5 / 7)
8
2016 = Year of Deep Learning (6 / 7)
9
2016 = Year of Deep Learning (7 / 7)
10
About Deep Learning. Why now?
●
D.L. = topic in Artificial Intelligence, specifically belonging to Machine
Learning
●
Methods used are based on learning models
●
These techniques are in contrast to algorithms with manually-crafted
features
●
Neural nets knowledge existed for decades. Why now the emergence
of D.L.?
–
Availability of an extensive dataset
–
GPU progress and better hardware, cheaper products
–
Improved techniques
11
Inspiration for D.L. models
●
The human brain
–
Comes with an initial model => approximates reality
–
While growing up => receives sensory inputs => approximates
the experience
●
External factors (= our parents)
–
Confirm or correct our approximations => the model is
reinforced / is adjusted
●
Similar ideas to the above inspired Machine Learning =>
self-adjusting models that learn from example
12
The linear perceptron
●
A simple learning model:
–
A mathematical function h(x, θ)
–
x = model input vector
–
θ = internal parameters vector
●
Example: an exam result guess model => above/below
average exam result – based on sleep hours and study
hours
h(x ,θ)=
{
−1,x
T
⋅
[θ 1
θ 2]+θ 0<0
1,xT
⋅
[θ 1
θ 2
]+θ 0≥0
x=[x1, x2]T
θ=[θ 1,θ 2,θ 0]T
x1
– sleep hours
x2
– study hours
13
Perceptron limitations
●
Geometrical interpretation:
separates points on a plane using
a straight line
●
Clearly delimited sets of points 
can decide which set the input
belongs to
●
Points are inseparable using
straight line => the perceptron can’t
learn the classification rule
14
Artificial neural nets
●
Biologically inspired structure, similar to the brain =>
connected artificial neurons
●
Connection = intensity & information flow direction
●
Neuron outputs = inputs for neurons in the following
layers
15
Artificial neuron
●
Model that contains:
–
Weighted inputs
–
An output
–
Activation function
S=∑
i=1
n
xi wi
o=F(S)
16
Types of artificial neurons
●
Activation function dictates the neuron type
●
Linear function => neural net reduced to perceptron
●
Activation function creates nonlinearity
●
3 common types of neurons:
–
Sigmoid
–
Tanh (hyperbolic tangent)
–
ReLU (Rectified Linear Unit)
17
Sigmoid
F(x)=
1
1+e
− x
18
Tanh
F(x)=tanh(x)=
ex
−e− x
e
x
+e
−x
19
ReLU
F(x)=max(0,x)
20
The learning process
●
Adjusting edge weights in an iterative way, to output
desired answers
●
Several techniques and algorithms
●
The backpropagation algorithm: ripple the difference
between result and objective backwards on previous
layers
21
Backpropagation - outline
1) Use a vector as input => get result as output
2) Compare output with desired vector
3) The difference is propagated backwards in the net
4) Adjust weights according to an error-minimizing
algorithm
22
The error function
●
Assuming t(i) the right answer for the i-th sample and
y(i) the neural net output => the following error function:
Error minimization = an optimization task
=> various approaches to solving it:
●
Gradient descent (common technique)
●
Genetic algorithms
●
Swarm intelligence algorithms (PSO, ACO, GSO)
E=
1
2
∑i
(t
(i)
−y
(i)
)
2
23
Gradient descent
●
Assume 2 weights: w1 and w2 => XY plane made of
[w1, w2] pairs
●
Z axis: error value at [w1, w2] coordinates =>
somewhere on the graph surface
●
Need to descend on the slope towards a minimum point
●
Steepest curve = perpendicular line to level ellipse =>
descend on the gradient of the error function
Δ wk=−ϵ
∂ E
∂wk
=...=∑
i
ϵ xk
(i)
(t
(i)
− y
(i)
)
ϵ = learning rate xk(i) = k-th input of i-th sample
i = i-th sample t(i),y(i) = desired outcome / actual output
24
Backpropagation & gradient descent
●
Weight adjustment în hidden layers:
–
Calculate error change depending on hidden outputs
–
Calculate error change depending on individual weights
●
For this we can use a dynamic programming approach
–
Store previously calculated values in a table and reuse them
as necessary
25
Convolutional neural nets
●
Classical neural net density gets very large for images
●
Simplify graph by calculating a subgraph
●
Multiple filters are repeatedly applied to parts of the image
=> smaller array
●
The operation is known as convolution
●
It is a linear operation => nonlinearity is later added through
ReLU or sigmoids
●
The neural net trains on these feature maps
●
More details about these in a future session...
26
Thank you for attending!
27
Bibliography
●
MIT 6.S191 course: Intro to Deep Learning
●
Nikhil Buduma – Fundamentals of Deep Learning
●
Ecaterina Vladu – Inteligenţa artificială
●
Wikipedia. Deep learning

More Related Content

PDF
The Perceptron (DLAI D1L2 2017 UPC Deep Learning for Artificial Intelligence)
PDF
Loss functions (DLAI D4L2 2017 UPC Deep Learning for Artificial Intelligence)
PDF
Session II - Estimation methods and accuracy Li-Chun Zhang Discussion: Sess...
PDF
Loss Functions for Deep Learning - Javier Ruiz Hidalgo - UPC Barcelona 2018
PDF
Weakly supervised semantic segmentation of 3D point cloud
PDF
The Perceptron (D1L1 Insight@DCU Machine Learning Workshop 2017)
PDF
Embeddings the geometry of relational algebra
PPTX
A brief introduction of Artificial neural network by example
The Perceptron (DLAI D1L2 2017 UPC Deep Learning for Artificial Intelligence)
Loss functions (DLAI D4L2 2017 UPC Deep Learning for Artificial Intelligence)
Session II - Estimation methods and accuracy Li-Chun Zhang Discussion: Sess...
Loss Functions for Deep Learning - Javier Ruiz Hidalgo - UPC Barcelona 2018
Weakly supervised semantic segmentation of 3D point cloud
The Perceptron (D1L1 Insight@DCU Machine Learning Workshop 2017)
Embeddings the geometry of relational algebra
A brief introduction of Artificial neural network by example

What's hot (20)

PPTX
Multilayer Perceptron (DLAI D1L2 2017 UPC Deep Learning for Artificial Intell...
PPT
PPTX
Summary of survey papers on deep learning method to 3D data
PDF
PPTX
Introduction to ml and dl
PDF
Variational Autoencoders VAE - Santiago Pascual - UPC Barcelona 2018
PDF
Multilayer Perceptron - Elisa Sayrol - UPC Barcelona 2018
PPTX
Digital Electronics -Boolean Algebra
PPTX
DIGITAL ELECTRONICS- Minimization Technique Karnaugh Map
PDF
Introductions to Neural Networks,Basic concepts
PDF
PDF
The Perceptron (D1L2 Deep Learning for Speech and Language)
PDF
Convolutional Neural Networks - Veronica Vilaplana - UPC Barcelona 2018
PDF
Auto-encoding variational bayes
PPTX
DIGITAL ELECTRONICS- Logic Gates
DOC
PDF
Generative adversarial networks
PDF
CSC446: Pattern Recognition (LN6)
PPTX
Pyramid Vector Quantization
PPTX
Machine Learning
Multilayer Perceptron (DLAI D1L2 2017 UPC Deep Learning for Artificial Intell...
Summary of survey papers on deep learning method to 3D data
Introduction to ml and dl
Variational Autoencoders VAE - Santiago Pascual - UPC Barcelona 2018
Multilayer Perceptron - Elisa Sayrol - UPC Barcelona 2018
Digital Electronics -Boolean Algebra
DIGITAL ELECTRONICS- Minimization Technique Karnaugh Map
Introductions to Neural Networks,Basic concepts
The Perceptron (D1L2 Deep Learning for Speech and Language)
Convolutional Neural Networks - Veronica Vilaplana - UPC Barcelona 2018
Auto-encoding variational bayes
DIGITAL ELECTRONICS- Logic Gates
Generative adversarial networks
CSC446: Pattern Recognition (LN6)
Pyramid Vector Quantization
Machine Learning
Ad

Similar to Deep learning @ University of Oradea - part I (16 Jan. 2018) (20)

PPTX
Deep Learning in Recommender Systems - RecSys Summer School 2017
PPTX
Deep learning: the future of recommendations
PPT
deep learning UNIT-1 Introduction Part-1.ppt
PDF
Hardware Acceleration for Machine Learning
PDF
Neural network book. Interesting and precise
PDF
Review_of_Deep_Learning_Algorithms_and_Architectures.pdf
PDF
MLIP - Chapter 3 - Introduction to deep learning
PPTX
Deep learning introduction
PDF
Mastering Advanced Deep Learning Techniques | IABAC
PPTX
Auto encoders in Deep Learning
PDF
Hands on machine learning with scikit-learn and tensor flow by ahmed yousry
PDF
Deep learning: Cutting through the Myths and Hype
PDF
Deep learning
PDF
MLIP - Chapter 2 - Preliminaries to deep learning
PPTX
Introduction to Deep learning and H2O for beginner's
PPTX
Deep Learning Sample Class (Jon Lederman)
PPTX
Neural Networks and Deep Learning Basics
PPTX
Deep Learning Jump Start
PDF
Deep Learning & Tensor flow: An Intro
PPTX
Deep learning from a novice perspective
Deep Learning in Recommender Systems - RecSys Summer School 2017
Deep learning: the future of recommendations
deep learning UNIT-1 Introduction Part-1.ppt
Hardware Acceleration for Machine Learning
Neural network book. Interesting and precise
Review_of_Deep_Learning_Algorithms_and_Architectures.pdf
MLIP - Chapter 3 - Introduction to deep learning
Deep learning introduction
Mastering Advanced Deep Learning Techniques | IABAC
Auto encoders in Deep Learning
Hands on machine learning with scikit-learn and tensor flow by ahmed yousry
Deep learning: Cutting through the Myths and Hype
Deep learning
MLIP - Chapter 2 - Preliminaries to deep learning
Introduction to Deep learning and H2O for beginner's
Deep Learning Sample Class (Jon Lederman)
Neural Networks and Deep Learning Basics
Deep Learning Jump Start
Deep Learning & Tensor flow: An Intro
Deep learning from a novice perspective
Ad

Recently uploaded (20)

PPTX
ap-psych-ch-1-introduction-to-psychology-presentation.pptx
PDF
Worlds Next Door: A Candidate Giant Planet Imaged in the Habitable Zone of ↵ ...
PPTX
endocrine - management of adrenal incidentaloma.pptx
PDF
Unit 5 Preparations, Reactions, Properties and Isomersim of Organic Compounds...
PPTX
Microbes in human welfare class 12 .pptx
PPTX
Substance Disorders- part different drugs change body
PPTX
PMR- PPT.pptx for students and doctors tt
PPTX
GREEN FIELDS SCHOOL PPT ON HOLIDAY HOMEWORK
PPTX
perinatal infections 2-171220190027.pptx
PDF
Cosmic Outliers: Low-spin Halos Explain the Abundance, Compactness, and Redsh...
PPTX
Welcome-grrewfefweg-students-of-2024.pptx
PPT
1. INTRODUCTION TO EPIDEMIOLOGY.pptx for community medicine
PPT
Mutation in dna of bacteria and repairss
PPTX
Understanding the Circulatory System……..
PDF
Assessment of environmental effects of quarrying in Kitengela subcountyof Kaj...
PPT
Heredity-grade-9 Heredity-grade-9. Heredity-grade-9.
PPTX
Lesson-1-Introduction-to-the-Study-of-Chemistry.pptx
PPTX
INTRODUCTION TO PAEDIATRICS AND PAEDIATRIC HISTORY TAKING-1.pptx
PPT
Animal tissues, epithelial, muscle, connective, nervous tissue
PPT
veterinary parasitology ````````````.ppt
ap-psych-ch-1-introduction-to-psychology-presentation.pptx
Worlds Next Door: A Candidate Giant Planet Imaged in the Habitable Zone of ↵ ...
endocrine - management of adrenal incidentaloma.pptx
Unit 5 Preparations, Reactions, Properties and Isomersim of Organic Compounds...
Microbes in human welfare class 12 .pptx
Substance Disorders- part different drugs change body
PMR- PPT.pptx for students and doctors tt
GREEN FIELDS SCHOOL PPT ON HOLIDAY HOMEWORK
perinatal infections 2-171220190027.pptx
Cosmic Outliers: Low-spin Halos Explain the Abundance, Compactness, and Redsh...
Welcome-grrewfefweg-students-of-2024.pptx
1. INTRODUCTION TO EPIDEMIOLOGY.pptx for community medicine
Mutation in dna of bacteria and repairss
Understanding the Circulatory System……..
Assessment of environmental effects of quarrying in Kitengela subcountyof Kaj...
Heredity-grade-9 Heredity-grade-9. Heredity-grade-9.
Lesson-1-Introduction-to-the-Study-of-Chemistry.pptx
INTRODUCTION TO PAEDIATRICS AND PAEDIATRIC HISTORY TAKING-1.pptx
Animal tissues, epithelial, muscle, connective, nervous tissue
veterinary parasitology ````````````.ppt

Deep learning @ University of Oradea - part I (16 Jan. 2018)

  • 1. 1 Deep Learning Vlad Ovidiu Mihalca PhD student – AI & Vision Mechatronics @ University of Oradea 16th of January, 2018
  • 2. 2 Series of talks ● Part I: theoretical base concepts ● Part II: frameworks and DL industry use ● Part III: usage in robotics and Computer Vision, state of the art in research
  • 3. 3 2016 = Year of Deep Learning (1 / 7)
  • 4. 4 2016 = Year of Deep Learning (2 / 7)
  • 5. 5 2016 = Year of Deep Learning (3 / 7)
  • 6. 6 2016 = Year of Deep Learning (4 / 7)
  • 7. 7 2016 = Year of Deep Learning (5 / 7)
  • 8. 8 2016 = Year of Deep Learning (6 / 7)
  • 9. 9 2016 = Year of Deep Learning (7 / 7)
  • 10. 10 About Deep Learning. Why now? ● D.L. = topic in Artificial Intelligence, specifically belonging to Machine Learning ● Methods used are based on learning models ● These techniques are in contrast to algorithms with manually-crafted features ● Neural nets knowledge existed for decades. Why now the emergence of D.L.? – Availability of an extensive dataset – GPU progress and better hardware, cheaper products – Improved techniques
  • 11. 11 Inspiration for D.L. models ● The human brain – Comes with an initial model => approximates reality – While growing up => receives sensory inputs => approximates the experience ● External factors (= our parents) – Confirm or correct our approximations => the model is reinforced / is adjusted ● Similar ideas to the above inspired Machine Learning => self-adjusting models that learn from example
  • 12. 12 The linear perceptron ● A simple learning model: – A mathematical function h(x, θ) – x = model input vector – θ = internal parameters vector ● Example: an exam result guess model => above/below average exam result – based on sleep hours and study hours h(x ,θ)= { −1,x T ⋅ [θ 1 θ 2]+θ 0<0 1,xT ⋅ [θ 1 θ 2 ]+θ 0≥0 x=[x1, x2]T θ=[θ 1,θ 2,θ 0]T x1 – sleep hours x2 – study hours
  • 13. 13 Perceptron limitations ● Geometrical interpretation: separates points on a plane using a straight line ● Clearly delimited sets of points  can decide which set the input belongs to ● Points are inseparable using straight line => the perceptron can’t learn the classification rule
  • 14. 14 Artificial neural nets ● Biologically inspired structure, similar to the brain => connected artificial neurons ● Connection = intensity & information flow direction ● Neuron outputs = inputs for neurons in the following layers
  • 15. 15 Artificial neuron ● Model that contains: – Weighted inputs – An output – Activation function S=∑ i=1 n xi wi o=F(S)
  • 16. 16 Types of artificial neurons ● Activation function dictates the neuron type ● Linear function => neural net reduced to perceptron ● Activation function creates nonlinearity ● 3 common types of neurons: – Sigmoid – Tanh (hyperbolic tangent) – ReLU (Rectified Linear Unit)
  • 20. 20 The learning process ● Adjusting edge weights in an iterative way, to output desired answers ● Several techniques and algorithms ● The backpropagation algorithm: ripple the difference between result and objective backwards on previous layers
  • 21. 21 Backpropagation - outline 1) Use a vector as input => get result as output 2) Compare output with desired vector 3) The difference is propagated backwards in the net 4) Adjust weights according to an error-minimizing algorithm
  • 22. 22 The error function ● Assuming t(i) the right answer for the i-th sample and y(i) the neural net output => the following error function: Error minimization = an optimization task => various approaches to solving it: ● Gradient descent (common technique) ● Genetic algorithms ● Swarm intelligence algorithms (PSO, ACO, GSO) E= 1 2 ∑i (t (i) −y (i) ) 2
  • 23. 23 Gradient descent ● Assume 2 weights: w1 and w2 => XY plane made of [w1, w2] pairs ● Z axis: error value at [w1, w2] coordinates => somewhere on the graph surface ● Need to descend on the slope towards a minimum point ● Steepest curve = perpendicular line to level ellipse => descend on the gradient of the error function Δ wk=−ϵ ∂ E ∂wk =...=∑ i ϵ xk (i) (t (i) − y (i) ) ϵ = learning rate xk(i) = k-th input of i-th sample i = i-th sample t(i),y(i) = desired outcome / actual output
  • 24. 24 Backpropagation & gradient descent ● Weight adjustment în hidden layers: – Calculate error change depending on hidden outputs – Calculate error change depending on individual weights ● For this we can use a dynamic programming approach – Store previously calculated values in a table and reuse them as necessary
  • 25. 25 Convolutional neural nets ● Classical neural net density gets very large for images ● Simplify graph by calculating a subgraph ● Multiple filters are repeatedly applied to parts of the image => smaller array ● The operation is known as convolution ● It is a linear operation => nonlinearity is later added through ReLU or sigmoids ● The neural net trains on these feature maps ● More details about these in a future session...
  • 26. 26 Thank you for attending!
  • 27. 27 Bibliography ● MIT 6.S191 course: Intro to Deep Learning ● Nikhil Buduma – Fundamentals of Deep Learning ● Ecaterina Vladu – Inteligenţa artificială ● Wikipedia. Deep learning