SlideShare a Scribd company logo
Deep Learning
Jin Sakuma
Neural Network
● Neural network is what in the deep learning
● Designed to simulate human brain
● Consist of several layers and data will be passed
one by one
Input Layer
Hidden Layer
Hidden Layer
Output Layer
Data Input
Data Output
Perceptron
● In one perceptron, output value is computed by
● Where xi is the input, wi is weight and f is activation function
y
x1
x2
x3
f z
One Layer
● For input vector x, output of one layer is computed by
y = Wx+b
z = f(y)
Where W is weight of the layer, b is bias, f is activation function.
● W and b are parameters because we will modify
them to optimize the model
Activation Function
● Without activation function, multiple layers would be meaningless
● Good activation function is non-linear, differentiable, monotonically increasing.
● Logistic function
● Hyperbolic Tangent
● Retified linear function (ReLU)
For Various Problems
● The activation function of the output layer depends on the type of problem
● For regression:
○ Activation function: Identity function
○ Length of output vector: Any
● For binary classification:
○ Activation function: Logistic function
○ Length of output vector: 1
● For multi-class classification:
Example Task
● Task setting:
○ Given picture of hand-written numbers 0-9, we want to tell which number
it is
○ Training dataset consist of a lot of pictures and all picture is labeled with
correct answer.
● Analyze task:
○ Problem type: Multi-class classification
○ Activation function of output layer: Softmax function
Learn: Minimize Error
● Consider error function E(xi, di; W, b) which represents how off the model is
from the true value for the ith picture. Here xi is the vector representation of
picture.
● In our example we use
Here, dik is 1 only if xi is actually picture of number k and otherwise 0.
● We want to modify W and b to minimize error function.
Learn: Gradient Descent
We use Gradient Descent to modify parameters W and b.
● Consider yourself in mountain and willing to reach the top of mountain, but you
lost map and you can’t look distance because of smog. How do you reach the
top?
➔Move to direction that bring you to the highest.
● The vector that indicate the direction to move is called gradient
● Since we want to minimize (instead of maximize), we update parameters by
subtracting gradient.
Learn: Backpropagation
● Neural network with multiple layers are too complex to
simply compute gradients.
● This problem was one bottleneck in early stage of
development of deep learning.
● Backpropagation compute gradient from output layer
to input layer (A lot of chain rules).
Input Layer
Hidden Layer
Hidden Layer
Output Layer
Data Input
Data Output
Why Deep Learning?
● Deep learning has tons of parameters (things that we can change to optimize
the model)
➔Better accuracy
➔Hard to optimize
➔Need a lot of data
● Flexible model
➔Can be used to different types of problems
➔Easy to modify models for various situations
Experiment
● Run python script for the example task
● Input vector is given as vector of length 784
● Configurations are
○ 2 hidden layers
○ 1000 perceptrons for each layers
○ ReLU function for activation function
○ Softmax function for activation function for the output layer
○ Batch size: 100
Result
● For this experiment, I used example code of chainer
● Execution time: about 45min
● Final Validation Loss: 0.107
● Final Validation Accuracy: 0.98
Use case of Deep Learning
● Convolutional Neural Network: Deep learning specific to picture data
○ Object identification
○ Face recognition
● Recurrent Neural Network: Deep learning for sequential data
○ Speech recognition
○ For text
● DQN: Combination of deep learning and Q learning
○ Alpha Go uses DQN and won top level Go player

More Related Content

PDF
Loss functions (DLAI D4L2 2017 UPC Deep Learning for Artificial Intelligence)
PDF
Deep Learning for Computer Vision: Unsupervised Learning (UPC 2016)
PDF
Deep Learning for Computer Vision: Data Augmentation (UPC 2016)
PDF
The Perceptron (DLAI D1L2 2017 UPC Deep Learning for Artificial Intelligence)
PDF
Loss Functions for Deep Learning - Javier Ruiz Hidalgo - UPC Barcelona 2018
PDF
A Comparison of Loss Function on Deep Embedding
PDF
The Perceptron (D1L1 Insight@DCU Machine Learning Workshop 2017)
PDF
Deep Learning for Computer Vision: Memory usage and computational considerati...
Loss functions (DLAI D4L2 2017 UPC Deep Learning for Artificial Intelligence)
Deep Learning for Computer Vision: Unsupervised Learning (UPC 2016)
Deep Learning for Computer Vision: Data Augmentation (UPC 2016)
The Perceptron (DLAI D1L2 2017 UPC Deep Learning for Artificial Intelligence)
Loss Functions for Deep Learning - Javier Ruiz Hidalgo - UPC Barcelona 2018
A Comparison of Loss Function on Deep Embedding
The Perceptron (D1L1 Insight@DCU Machine Learning Workshop 2017)
Deep Learning for Computer Vision: Memory usage and computational considerati...

What's hot (20)

PPTX
Accelerated Logistic Regression on GPU(s)
PPTX
Multilayer Perceptron (DLAI D1L2 2017 UPC Deep Learning for Artificial Intell...
PDF
Transfer Learning and Domain Adaptation (DLAI D5L2 2017 UPC Deep Learning for...
PDF
Optimization for Deep Networks (D2L1 2017 UPC Deep Learning for Computer Vision)
PDF
Set Transfomer: A Framework for Attention-based Permutaion-Invariant Neural N...
PDF
Deep Learning for Computer Vision: Deep Networks (UPC 2016)
PDF
Optimizing Deep Networks (D1L6 Insight@DCU Machine Learning Workshop 2017)
PPTX
Semantic Segmentation on Satellite Imagery
PDF
Multilayer Perceptron - Elisa Sayrol - UPC Barcelona 2018
PDF
Quantization and Training of Neural Networks for Efficient Integer-Arithmetic...
PDF
Transfer Learning and Domain Adaptation - Ramon Morros - UPC Barcelona 2018
PDF
Convolutional Neural Networks (D1L3 2017 UPC Deep Learning for Computer Vision)
PDF
Centernet
PDF
Deep Learning for Computer Vision: Attention Models (UPC 2016)
PDF
Training Deep Networks with Backprop (D1L4 Insight@DCU Machine Learning Works...
PDF
Deep Neural Networks (D1L2 Insight@DCU Machine Learning Workshop 2017)
PDF
Backpropagation - Elisa Sayrol - UPC Barcelona 2018
PDF
Skip RNN: Learning to Skip State Updates in Recurrent Neural Networks
PDF
Deep Learning for Computer Vision: Backward Propagation (UPC 2016)
PDF
Generative Models and Adversarial Training (D2L3 Insight@DCU Machine Learning...
Accelerated Logistic Regression on GPU(s)
Multilayer Perceptron (DLAI D1L2 2017 UPC Deep Learning for Artificial Intell...
Transfer Learning and Domain Adaptation (DLAI D5L2 2017 UPC Deep Learning for...
Optimization for Deep Networks (D2L1 2017 UPC Deep Learning for Computer Vision)
Set Transfomer: A Framework for Attention-based Permutaion-Invariant Neural N...
Deep Learning for Computer Vision: Deep Networks (UPC 2016)
Optimizing Deep Networks (D1L6 Insight@DCU Machine Learning Workshop 2017)
Semantic Segmentation on Satellite Imagery
Multilayer Perceptron - Elisa Sayrol - UPC Barcelona 2018
Quantization and Training of Neural Networks for Efficient Integer-Arithmetic...
Transfer Learning and Domain Adaptation - Ramon Morros - UPC Barcelona 2018
Convolutional Neural Networks (D1L3 2017 UPC Deep Learning for Computer Vision)
Centernet
Deep Learning for Computer Vision: Attention Models (UPC 2016)
Training Deep Networks with Backprop (D1L4 Insight@DCU Machine Learning Works...
Deep Neural Networks (D1L2 Insight@DCU Machine Learning Workshop 2017)
Backpropagation - Elisa Sayrol - UPC Barcelona 2018
Skip RNN: Learning to Skip State Updates in Recurrent Neural Networks
Deep Learning for Computer Vision: Backward Propagation (UPC 2016)
Generative Models and Adversarial Training (D2L3 Insight@DCU Machine Learning...
Ad

Viewers also liked (14)

PPT
No 2º Período foram desenvolvidos trabalhos subordinados aos temas "Moda" e "...
PPTX
Portafolio fotocopias yenny
PPTX
PDF
South korea tablet listening
PDF
Prwr625 bizplan menu
PPT
8 aula 6 república velha
PPT
Mono lds2012
PDF
GOVT2305SyllabusHCC
PPTX
шинагулова даметай+одежда для дома+создать магазин одежды для дома
PDF
20151107 - The Influences on Motivation in Online Educational Environments
PPTX
Europa e Europeus - Celtas
PPT
Stock and Share Market Technical Analysis
PDF
The back of the napkin Presentation
No 2º Período foram desenvolvidos trabalhos subordinados aos temas "Moda" e "...
Portafolio fotocopias yenny
South korea tablet listening
Prwr625 bizplan menu
8 aula 6 república velha
Mono lds2012
GOVT2305SyllabusHCC
шинагулова даметай+одежда для дома+создать магазин одежды для дома
20151107 - The Influences on Motivation in Online Educational Environments
Europa e Europeus - Celtas
Stock and Share Market Technical Analysis
The back of the napkin Presentation
Ad

Similar to Deep learning (20)

PPTX
Deep Learning Module 2A Training MLP.pptx
PPTX
Deep Learning Tutorial
ODP
Machine Learning With Neural Networks
PDF
Neural networks
PPTX
Fully Homomorphic Encryption (1).pptx
PPTX
AUTO ENCODERS (Deep Learning fundamentals)
PDF
Introduction to Machine Learning with Spark
PPTX
Introduction to Neural Networks and Perceptron Learning Algorithm.pptx
PPTX
Practical ML
PPTX
Jyduydufyuyf8yfiyfiyfifiyfiyviyviyfiugiuy8f7dd64d4yrsxyfhgdhfjhvjhv
PPTX
Lecture02_Updated_Shallow Neural Networks.pptx
PDF
物件偵測與辨識技術
PPTX
Artificial Neural Networks presentations
PPTX
Neural network basic and introduction of Deep learning
PDF
Intro to TensorFlow and PyTorch Workshop at Tubular Labs
PDF
Introduction to Applied Machine Learning
PPTX
cnn.pptx Convolutional neural network used for image classication
PPT
deep learning UNIT-1 Introduction Part-1.ppt
PPTX
2Wisjshsbebe pehele isienew Dorene isksnwnw
PDF
Fundamental of deep learning
Deep Learning Module 2A Training MLP.pptx
Deep Learning Tutorial
Machine Learning With Neural Networks
Neural networks
Fully Homomorphic Encryption (1).pptx
AUTO ENCODERS (Deep Learning fundamentals)
Introduction to Machine Learning with Spark
Introduction to Neural Networks and Perceptron Learning Algorithm.pptx
Practical ML
Jyduydufyuyf8yfiyfiyfifiyfiyviyviyfiugiuy8f7dd64d4yrsxyfhgdhfjhvjhv
Lecture02_Updated_Shallow Neural Networks.pptx
物件偵測與辨識技術
Artificial Neural Networks presentations
Neural network basic and introduction of Deep learning
Intro to TensorFlow and PyTorch Workshop at Tubular Labs
Introduction to Applied Machine Learning
cnn.pptx Convolutional neural network used for image classication
deep learning UNIT-1 Introduction Part-1.ppt
2Wisjshsbebe pehele isienew Dorene isksnwnw
Fundamental of deep learning

Recently uploaded (20)

PDF
ENT215_Completing-a-large-scale-migration-and-modernization-with-AWS.pdf
PDF
DASA ADMISSION 2024_FirstRound_FirstRank_LastRank.pdf
PPTX
SOPHOS-XG Firewall Administrator PPT.pptx
PDF
7 ChatGPT Prompts to Help You Define Your Ideal Customer Profile.pdf
PDF
Encapsulation_ Review paper, used for researhc scholars
PDF
WOOl fibre morphology and structure.pdf for textiles
PPTX
Programs and apps: productivity, graphics, security and other tools
PDF
Mushroom cultivation and it's methods.pdf
PDF
Assigned Numbers - 2025 - Bluetooth® Document
PDF
MIND Revenue Release Quarter 2 2025 Press Release
PDF
Unlocking AI with Model Context Protocol (MCP)
PDF
Web App vs Mobile App What Should You Build First.pdf
PDF
Profit Center Accounting in SAP S/4HANA, S4F28 Col11
PPTX
TLE Review Electricity (Electricity).pptx
PPTX
OMC Textile Division Presentation 2021.pptx
PDF
project resource management chapter-09.pdf
PPTX
Group 1 Presentation -Planning and Decision Making .pptx
PDF
August Patch Tuesday
PDF
DP Operators-handbook-extract for the Mautical Institute
PDF
Encapsulation theory and applications.pdf
ENT215_Completing-a-large-scale-migration-and-modernization-with-AWS.pdf
DASA ADMISSION 2024_FirstRound_FirstRank_LastRank.pdf
SOPHOS-XG Firewall Administrator PPT.pptx
7 ChatGPT Prompts to Help You Define Your Ideal Customer Profile.pdf
Encapsulation_ Review paper, used for researhc scholars
WOOl fibre morphology and structure.pdf for textiles
Programs and apps: productivity, graphics, security and other tools
Mushroom cultivation and it's methods.pdf
Assigned Numbers - 2025 - Bluetooth® Document
MIND Revenue Release Quarter 2 2025 Press Release
Unlocking AI with Model Context Protocol (MCP)
Web App vs Mobile App What Should You Build First.pdf
Profit Center Accounting in SAP S/4HANA, S4F28 Col11
TLE Review Electricity (Electricity).pptx
OMC Textile Division Presentation 2021.pptx
project resource management chapter-09.pdf
Group 1 Presentation -Planning and Decision Making .pptx
August Patch Tuesday
DP Operators-handbook-extract for the Mautical Institute
Encapsulation theory and applications.pdf

Deep learning

  • 2. Neural Network ● Neural network is what in the deep learning ● Designed to simulate human brain ● Consist of several layers and data will be passed one by one Input Layer Hidden Layer Hidden Layer Output Layer Data Input Data Output
  • 3. Perceptron ● In one perceptron, output value is computed by ● Where xi is the input, wi is weight and f is activation function y x1 x2 x3 f z
  • 4. One Layer ● For input vector x, output of one layer is computed by y = Wx+b z = f(y) Where W is weight of the layer, b is bias, f is activation function. ● W and b are parameters because we will modify them to optimize the model
  • 5. Activation Function ● Without activation function, multiple layers would be meaningless ● Good activation function is non-linear, differentiable, monotonically increasing. ● Logistic function ● Hyperbolic Tangent ● Retified linear function (ReLU)
  • 6. For Various Problems ● The activation function of the output layer depends on the type of problem ● For regression: ○ Activation function: Identity function ○ Length of output vector: Any ● For binary classification: ○ Activation function: Logistic function ○ Length of output vector: 1 ● For multi-class classification:
  • 7. Example Task ● Task setting: ○ Given picture of hand-written numbers 0-9, we want to tell which number it is ○ Training dataset consist of a lot of pictures and all picture is labeled with correct answer. ● Analyze task: ○ Problem type: Multi-class classification ○ Activation function of output layer: Softmax function
  • 8. Learn: Minimize Error ● Consider error function E(xi, di; W, b) which represents how off the model is from the true value for the ith picture. Here xi is the vector representation of picture. ● In our example we use Here, dik is 1 only if xi is actually picture of number k and otherwise 0. ● We want to modify W and b to minimize error function.
  • 9. Learn: Gradient Descent We use Gradient Descent to modify parameters W and b. ● Consider yourself in mountain and willing to reach the top of mountain, but you lost map and you can’t look distance because of smog. How do you reach the top? ➔Move to direction that bring you to the highest. ● The vector that indicate the direction to move is called gradient ● Since we want to minimize (instead of maximize), we update parameters by subtracting gradient.
  • 10. Learn: Backpropagation ● Neural network with multiple layers are too complex to simply compute gradients. ● This problem was one bottleneck in early stage of development of deep learning. ● Backpropagation compute gradient from output layer to input layer (A lot of chain rules). Input Layer Hidden Layer Hidden Layer Output Layer Data Input Data Output
  • 11. Why Deep Learning? ● Deep learning has tons of parameters (things that we can change to optimize the model) ➔Better accuracy ➔Hard to optimize ➔Need a lot of data ● Flexible model ➔Can be used to different types of problems ➔Easy to modify models for various situations
  • 12. Experiment ● Run python script for the example task ● Input vector is given as vector of length 784 ● Configurations are ○ 2 hidden layers ○ 1000 perceptrons for each layers ○ ReLU function for activation function ○ Softmax function for activation function for the output layer ○ Batch size: 100
  • 13. Result ● For this experiment, I used example code of chainer ● Execution time: about 45min ● Final Validation Loss: 0.107 ● Final Validation Accuracy: 0.98
  • 14. Use case of Deep Learning ● Convolutional Neural Network: Deep learning specific to picture data ○ Object identification ○ Face recognition ● Recurrent Neural Network: Deep learning for sequential data ○ Speech recognition ○ For text ● DQN: Combination of deep learning and Q learning ○ Alpha Go uses DQN and won top level Go player

Editor's Notes

  • #2: Deep learningの全体的な話をする
  • #3: 目の例を使って、人間の脳をもとに考えられたことを説明 一つ一つの層が何をしているかをざっくりと説明して次のスライドにつなげる
  • #4: 一つの神経細胞内で何が起こるかの説明。activation functionに関しては、次に説明するからってことにする
  • #5: 一つの層の線形性について
  • #6: Activation Functionの条件についてちゃんと話す
  • #7: ここまででdeep learningの全体像については一通り終わり。次に行く前にまとめをする