SlideShare a Scribd company logo
{
Deep Learning
Machine Perception and Its Applications
Adam Gibson
deeplearning4j.org // blix.io // zipfian academy
 Deep Learning is a subset of Machine Learning
 Machine Learning is a subset of Artificial
Intelligence
 AI is nothing more than a collection
of algorithms that repeatedly optimize
themselves.
 Deep learning is pattern recognition, a way for
machines to classify what they perceive.
DL is a subset of AI
 Deep learning algorithms are called neural
nets. They are mathematical models.
 They mirror the neurons of the human brain.
 In the brain, sets of neurons learn to recognize
certain patterns or phenomena, like faces,
birdcalls or grammatical sequences.
 These models have names like:
 Restricted Boltzmann Machine
 Deep-Belief Net
 Convolutional Net
 Stacked Denoising Autoencoder
 Recursive Neural Tensor Network
Deep learning’s algorithms
 Deep learning understands numbers, so
anything that can be converted to numbers is
fair game:
 Digital media. Anything you can see or here.
DL can analyze sights, sounds and text.
 Sensor output. DL can work with data about
temperature, pressure, motion and chemical
composition.
 Time-series data. DL handles prices and their
movement over time; e.g. the stock market, real
estate, weather and economic indicators.
What DL can handle
 Recommendation engines: DL can identify
patterns of human behavior and predict what
you will want to buy.
 Anomaly detection: DL can identify signals that
indicate bad outcomes. It can point out fraud in
e-commerce; tumors in X-rays; and loan
applicants likely to default.
 Signal processing: Deep learning can tell you
what to expect, whether its customer lifetime
value, how much inventory to stock, or
whether the market on the verge of a flash
crash. It has predictive capacity.
What can you do with it?
 Faces can be represented by a collection of
images.
 Those images have persistent patterns of pixels.
 Those pixel patterns are known as features; i.e.
highly granular facial features.
 Deep-learning nets learn to identify features in
data, and use them to classify faces as faces and
to label them by name; e.g. John or Sarah.
 Nets train themselves by reconstructing faces
from features again and again, and measuring
their work against a benchmark.
Facial recognition
Facial reconstructions…
 Deep learning networks learn from the data
you feed them.
 Initial data is known as the training set, and
you know what it’s made of.
 The net learns the faces of the training set by
trying to reconstruct them, again and again.
 Reconstruction is a process of finding which
facial features are indicative of larger forms.
 When a net can rebuild the training set, it is
ready to work with unsupervised data.
How did it do that?
 Nets measure the difference between what they
produce and a benchmark you set.
 They try to minimize that difference.
 They do that by altering their own parameters
– the way they treat the data – and testing how
that affects their own results.
 This test is known as a “loss function.”
No really, how did it do that?
Learning looks like this.
 Facebook uses facial recognition to make itself
stickier, and to know more about us.
 Government agencies use facial recognition to
secure national borders.
 Video game makers use facial recognition to
construct more realistic worlds.
 Stores use it to identify customers and track
behavior.
What are faces for?
 Sentiment analysis is a form of Natural-
Language Processing.
 With it, software classifies the affective content
of sentences, their emotional tone, bias and
intensity.
 Are they positive or negative about the subject
in question?
 This can be very useful in ranking movies,
books, media and just about anything humans
consume.
 Including politicians.
Sentiment Analysis & Text
 By reading sentiment, you read many things.
 Corporations can measure customer
satisfaction.
 Governments can monitor popular unrest.
 Event organizers can track audience
engagement.
 Employers can measure job applicant fit.
 Celebrities can gauge fame and track scandal.
Who cares what they say?
 Recurrent neural net
 Restricted Boltzmann machine (RBM)
 Deep-belief network: A stack of RBMs
 Deep Autoencoder: 2 DBNs
 Denoising Autoencoder (yay, noise!)
 Convolutional net (ConvNet)
 Recursive neural tensor network (RNTN)
A Neural Nets Taxonomy
 Two layers of neuron-like nodes.
 The first layer is the visible, or input, layer
 The second is the hidden layer, which identifies
features in the input
 This simple network is symmetrically
connected.
 “Restricted” means there are no visible-visible
or hidden-hidden connections; i.e. all
connections happen *between* layers.
Restricted Boltzmann
Machine (RBMs)
 A deep-belief net is a stack of RBMs.
 Each RBM’s hidden layer becomes the next
RBM’s visible/input layer.
 In this manner, a DBN learns more and more
complex features
 A machine vision example: 1) Pixels are input;
2) H1 learns an edge or line; 3) H2 learns a
corner or set of lines; 4) H3 learns two groups
of lines forming an object, maybe a face.
 The final layer of a DBN classifies feature
groups. It groups them in buckets: e.g. sunset,
elephant, flower.
Deep-belief net (DBN)
 A deep autoencoder consists of two DBNs.
 The first DBN *encodes* the data into a vector
of 10-30 numbers. This is pre-training.
 The second DBN decodes the data into its
original state.
 Backprop happens solely on the second DBN
 This is the fine-tuning stage and it’s carried out
with reconstruction entropy.
 Deep autoencoders will reduce any document
or image to a highly compact vector.
 Those vectors are useful in search, QA and
information retrieval.
Deep Autoencoder
 Autoencoders are useful for dimensionality
reduction.
 The risk they run is learning the identity
function of the input.
 Dropout is one way to address that risk.
 Noise is another.
 Noise is the stochastic, or random, corruption
of the input.
 The machine then learns features despite the
noise. It “denoises” the input.
 A stacked denoising encoder is exactly what
you’d think.
 Good for unsupervised pre-training, which
initializes the weights.
Denoising Autoencoder
 ConvNets are a type of RBM.
 The difference is they’re asymmetric.
 In an RBM, each node in the visible layer
connects to each node in the hidden layer.
 In a ConvNet, each node connects to the node
straight ahead of it, and to the two others
immediately to the right and left of it.
 This means that ConvNets learn data like
images in patches.
 Each piece learned is then woven together in
the whole.
Convolutional Net
 Recursive nets are top-down, hierarchical nets
rather than feed-forward like DBNs.
 RNTNs handle sequence-based classification,
windows of several events, entire scenes rather
than images.
 The features themselves are vectors.
 A tensor is a multi-dimensional matrix, or
multiple matrices of the same size.
Recursive Neural Tensor Net
RNTNs & Scene Composition
RNTNs & Sentence Parsing

More Related Content

PDF
San Francisco Hacker News - Machine Learning for Hackers
PPTX
Sf data mining_meetup
PDF
Top 10 deep learning algorithms you should know in
PDF
Donner - Deep Learning - Overview and practical aspects
PPTX
Deep Learning Tutorial
PDF
Deep Learning
PDF
Handwritten Recognition using Deep Learning with R
PDF
Deep learning
San Francisco Hacker News - Machine Learning for Hackers
Sf data mining_meetup
Top 10 deep learning algorithms you should know in
Donner - Deep Learning - Overview and practical aspects
Deep Learning Tutorial
Deep Learning
Handwritten Recognition using Deep Learning with R
Deep learning

What's hot (20)

PDF
From Conventional Machine Learning to Deep Learning and Beyond.pptx
PDF
Emotion detection using cnn.pptx
PPTX
Neural networks and deep learning
PDF
Deep learning - what is it and why now?
PPTX
Cat and dog classification
PDF
Neural Networks and Deep Learning
PPTX
Deep learning tutorial 9/2019
PDF
Deep Learning: Application & Opportunity
PDF
Deep learning - Conceptual understanding and applications
PPTX
Machine model to classify dogs and cat
PPTX
Deep Learning - A Literature survey
PPTX
Deep learning at nmc devin jones
PPTX
Deep Learning Projects - Anomaly Detection Using Deep Learning
PPTX
Handwritten bangla-digit-recognition-using-deep-learning
PPTX
Introduction to deep learning
PPTX
Introduction to Deep Learning
PDF
Deep Learning and the state of AI / 2016
PDF
Deep learning frameworks v0.40
PDF
Intro to Deep Learning for Computer Vision
PPTX
Deep learning presentation
From Conventional Machine Learning to Deep Learning and Beyond.pptx
Emotion detection using cnn.pptx
Neural networks and deep learning
Deep learning - what is it and why now?
Cat and dog classification
Neural Networks and Deep Learning
Deep learning tutorial 9/2019
Deep Learning: Application & Opportunity
Deep learning - Conceptual understanding and applications
Machine model to classify dogs and cat
Deep Learning - A Literature survey
Deep learning at nmc devin jones
Deep Learning Projects - Anomaly Detection Using Deep Learning
Handwritten bangla-digit-recognition-using-deep-learning
Introduction to deep learning
Introduction to Deep Learning
Deep Learning and the state of AI / 2016
Deep learning frameworks v0.40
Intro to Deep Learning for Computer Vision
Deep learning presentation
Ad

Viewers also liked (20)

PDF
Intro to DeepLearning4J on ApacheSpark SDS DL Workshop 16
PDF
Deep learning on a mixed cluster with deeplearning4j and spark
PDF
Skymind & Deeplearning4j: Deep Learning for the Enterprise
PDF
DeepLearning4J and Spark: Successes and Challenges - François Garillot
PDF
DeepLearning4J and Spark: Successes and Challenges - François Garillot
PDF
DL4J at Workday Meetup
ODP
Deep Learning meetup
PPTX
Deep Learning and Recurrent Neural Networks in the Enterprise
PDF
Introduction to Deeplearning4j
PDF
Advanced Spark and Tensorflow Meetup - London - Nov 15, 2016 - Deploy Spark M...
PDF
Neural Networks, Spark MLlib, Deep Learning
PDF
Advanced Spark and TensorFlow Meetup May 26, 2016
PPTX
Suneel Marthi - Deep Learning with Apache Flink and DL4J
PDF
Which Is Deeper - Comparison Of Deep Learning Frameworks On Spark
PPTX
TensorFrames: Google Tensorflow on Apache Spark
PPTX
Wordpress Shortcode
PPTX
Semantic Twitter Analyzing Tweets For Real Time Event Notification
PPT
semantic social network analysis
PDF
IEEE 2016-2017 SOFTWARE TITLE
PDF
Ijcai ip-2015 cyberbullying-final
Intro to DeepLearning4J on ApacheSpark SDS DL Workshop 16
Deep learning on a mixed cluster with deeplearning4j and spark
Skymind & Deeplearning4j: Deep Learning for the Enterprise
DeepLearning4J and Spark: Successes and Challenges - François Garillot
DeepLearning4J and Spark: Successes and Challenges - François Garillot
DL4J at Workday Meetup
Deep Learning meetup
Deep Learning and Recurrent Neural Networks in the Enterprise
Introduction to Deeplearning4j
Advanced Spark and Tensorflow Meetup - London - Nov 15, 2016 - Deploy Spark M...
Neural Networks, Spark MLlib, Deep Learning
Advanced Spark and TensorFlow Meetup May 26, 2016
Suneel Marthi - Deep Learning with Apache Flink and DL4J
Which Is Deeper - Comparison Of Deep Learning Frameworks On Spark
TensorFrames: Google Tensorflow on Apache Spark
Wordpress Shortcode
Semantic Twitter Analyzing Tweets For Real Time Event Notification
semantic social network analysis
IEEE 2016-2017 SOFTWARE TITLE
Ijcai ip-2015 cyberbullying-final
Ad

Similar to Ersatz meetup - DeepLearning4j Demo (20)

PPTX
Deep learning intro and examples and types
DOCX
Deep learning vxcvbfsdfaegsr gsgfgsdg sd gdgd gdgd gse
PDF
Deep Learning - The Past, Present and Future of Artificial Intelligence
PDF
Unit 4 Deep Generative Models Unit 4 Deep Generative Model
PPTX
UNIT 3- Introduction to DL(Deep Learning) in AI
PPTX
Artificial intelligence Presentation.pptx
PPTX
Deep learning
PDF
Deep learning - A Visual Introduction
PPTX
Deep learning Techniques JNTU R20 UNIT 2
PDF
A Survey of Deep Learning Algorithms for Malware Detection
PDF
Deep learning seminar report
PPTX
Unlocking Human Emotions: Facial Emotion Recognition with Deep Learning
PPTX
Decoding Human Emotions with Deep Learning: Facial Expression Recognition Exp...
PPTX
Weed Detection and Identification using Deep learning Techniques
PDF
The Magic Behind AI
PPTX
Handwritten digit and symbol recognition using CNN.pptx
PDF
A SURVEY ON DEEP LEARNING METHOD USED FOR CHARACTER RECOGNITION
PPTX
PDF
Neural network
PPTX
MachinaFiesta: A Vision into Machine Learning 🚀
Deep learning intro and examples and types
Deep learning vxcvbfsdfaegsr gsgfgsdg sd gdgd gdgd gse
Deep Learning - The Past, Present and Future of Artificial Intelligence
Unit 4 Deep Generative Models Unit 4 Deep Generative Model
UNIT 3- Introduction to DL(Deep Learning) in AI
Artificial intelligence Presentation.pptx
Deep learning
Deep learning - A Visual Introduction
Deep learning Techniques JNTU R20 UNIT 2
A Survey of Deep Learning Algorithms for Malware Detection
Deep learning seminar report
Unlocking Human Emotions: Facial Emotion Recognition with Deep Learning
Decoding Human Emotions with Deep Learning: Facial Expression Recognition Exp...
Weed Detection and Identification using Deep learning Techniques
The Magic Behind AI
Handwritten digit and symbol recognition using CNN.pptx
A SURVEY ON DEEP LEARNING METHOD USED FOR CHARACTER RECOGNITION
Neural network
MachinaFiesta: A Vision into Machine Learning 🚀

More from Adam Gibson (20)

PDF
End to end MLworkflows
PDF
World Artificial Intelligence Conference Shanghai 2018
PPTX
Deploying signature verification with deep learning
ODP
Self driving computers active learning workflows with human interpretable ve...
PDF
Anomaly Detection and Automatic Labeling with Deep Learning
PDF
Strata Beijing 2017: Jumpy, a python interface for nd4j
PPTX
Boolan machine learning summit
PDF
Advanced deeplearning4j features
PDF
Deep Learning with GPUs in Production - AI By the Bay
PDF
Big Data Analytics Tokyo
PDF
Wrangleconf Big Data Malaysia 2016
PDF
Distributed deep rl on spark strata singapore
PDF
Deep learning in production with the best
PPTX
Dl4j in the wild
PDF
SKIL - Dl4j in the wild meetup
PDF
Strata Beijing - Deep Learning in Production on Spark
PPTX
Anomaly detection in deep learning (Updated) English
PPTX
Skymind - Udacity China presentation
PDF
Anomaly Detection in Deep Learning (Updated)
PPTX
Hadoop summit 2016
End to end MLworkflows
World Artificial Intelligence Conference Shanghai 2018
Deploying signature verification with deep learning
Self driving computers active learning workflows with human interpretable ve...
Anomaly Detection and Automatic Labeling with Deep Learning
Strata Beijing 2017: Jumpy, a python interface for nd4j
Boolan machine learning summit
Advanced deeplearning4j features
Deep Learning with GPUs in Production - AI By the Bay
Big Data Analytics Tokyo
Wrangleconf Big Data Malaysia 2016
Distributed deep rl on spark strata singapore
Deep learning in production with the best
Dl4j in the wild
SKIL - Dl4j in the wild meetup
Strata Beijing - Deep Learning in Production on Spark
Anomaly detection in deep learning (Updated) English
Skymind - Udacity China presentation
Anomaly Detection in Deep Learning (Updated)
Hadoop summit 2016

Recently uploaded (20)

PDF
BMEC211 - INTRODUCTION TO MECHATRONICS-1.pdf
PDF
Operating System & Kernel Study Guide-1 - converted.pdf
PPTX
Lecture Notes Electrical Wiring System Components
PPTX
additive manufacturing of ss316l using mig welding
PDF
SM_6th-Sem__Cse_Internet-of-Things.pdf IOT
PDF
PRIZ Academy - 9 Windows Thinking Where to Invest Today to Win Tomorrow.pdf
PPT
Mechanical Engineering MATERIALS Selection
PPTX
UNIT 4 Total Quality Management .pptx
PDF
Model Code of Practice - Construction Work - 21102022 .pdf
PDF
Evaluating the Democratization of the Turkish Armed Forces from a Normative P...
PPTX
Welding lecture in detail for understanding
PDF
July 2025 - Top 10 Read Articles in International Journal of Software Enginee...
PPTX
FINAL REVIEW FOR COPD DIANOSIS FOR PULMONARY DISEASE.pptx
PDF
Digital Logic Computer Design lecture notes
PPTX
UNIT-1 - COAL BASED THERMAL POWER PLANTS
PPTX
Foundation to blockchain - A guide to Blockchain Tech
PPTX
Construction Project Organization Group 2.pptx
PPTX
CARTOGRAPHY AND GEOINFORMATION VISUALIZATION chapter1 NPTE (2).pptx
PPTX
Engineering Ethics, Safety and Environment [Autosaved] (1).pptx
PDF
Mohammad Mahdi Farshadian CV - Prospective PhD Student 2026
BMEC211 - INTRODUCTION TO MECHATRONICS-1.pdf
Operating System & Kernel Study Guide-1 - converted.pdf
Lecture Notes Electrical Wiring System Components
additive manufacturing of ss316l using mig welding
SM_6th-Sem__Cse_Internet-of-Things.pdf IOT
PRIZ Academy - 9 Windows Thinking Where to Invest Today to Win Tomorrow.pdf
Mechanical Engineering MATERIALS Selection
UNIT 4 Total Quality Management .pptx
Model Code of Practice - Construction Work - 21102022 .pdf
Evaluating the Democratization of the Turkish Armed Forces from a Normative P...
Welding lecture in detail for understanding
July 2025 - Top 10 Read Articles in International Journal of Software Enginee...
FINAL REVIEW FOR COPD DIANOSIS FOR PULMONARY DISEASE.pptx
Digital Logic Computer Design lecture notes
UNIT-1 - COAL BASED THERMAL POWER PLANTS
Foundation to blockchain - A guide to Blockchain Tech
Construction Project Organization Group 2.pptx
CARTOGRAPHY AND GEOINFORMATION VISUALIZATION chapter1 NPTE (2).pptx
Engineering Ethics, Safety and Environment [Autosaved] (1).pptx
Mohammad Mahdi Farshadian CV - Prospective PhD Student 2026

Ersatz meetup - DeepLearning4j Demo

  • 1. { Deep Learning Machine Perception and Its Applications Adam Gibson deeplearning4j.org // blix.io // zipfian academy
  • 2.  Deep Learning is a subset of Machine Learning  Machine Learning is a subset of Artificial Intelligence  AI is nothing more than a collection of algorithms that repeatedly optimize themselves.  Deep learning is pattern recognition, a way for machines to classify what they perceive. DL is a subset of AI
  • 3.  Deep learning algorithms are called neural nets. They are mathematical models.  They mirror the neurons of the human brain.  In the brain, sets of neurons learn to recognize certain patterns or phenomena, like faces, birdcalls or grammatical sequences.  These models have names like:  Restricted Boltzmann Machine  Deep-Belief Net  Convolutional Net  Stacked Denoising Autoencoder  Recursive Neural Tensor Network Deep learning’s algorithms
  • 4.  Deep learning understands numbers, so anything that can be converted to numbers is fair game:  Digital media. Anything you can see or here. DL can analyze sights, sounds and text.  Sensor output. DL can work with data about temperature, pressure, motion and chemical composition.  Time-series data. DL handles prices and their movement over time; e.g. the stock market, real estate, weather and economic indicators. What DL can handle
  • 5.  Recommendation engines: DL can identify patterns of human behavior and predict what you will want to buy.  Anomaly detection: DL can identify signals that indicate bad outcomes. It can point out fraud in e-commerce; tumors in X-rays; and loan applicants likely to default.  Signal processing: Deep learning can tell you what to expect, whether its customer lifetime value, how much inventory to stock, or whether the market on the verge of a flash crash. It has predictive capacity. What can you do with it?
  • 6.  Faces can be represented by a collection of images.  Those images have persistent patterns of pixels.  Those pixel patterns are known as features; i.e. highly granular facial features.  Deep-learning nets learn to identify features in data, and use them to classify faces as faces and to label them by name; e.g. John or Sarah.  Nets train themselves by reconstructing faces from features again and again, and measuring their work against a benchmark. Facial recognition
  • 8.  Deep learning networks learn from the data you feed them.  Initial data is known as the training set, and you know what it’s made of.  The net learns the faces of the training set by trying to reconstruct them, again and again.  Reconstruction is a process of finding which facial features are indicative of larger forms.  When a net can rebuild the training set, it is ready to work with unsupervised data. How did it do that?
  • 9.  Nets measure the difference between what they produce and a benchmark you set.  They try to minimize that difference.  They do that by altering their own parameters – the way they treat the data – and testing how that affects their own results.  This test is known as a “loss function.” No really, how did it do that?
  • 11.  Facebook uses facial recognition to make itself stickier, and to know more about us.  Government agencies use facial recognition to secure national borders.  Video game makers use facial recognition to construct more realistic worlds.  Stores use it to identify customers and track behavior. What are faces for?
  • 12.  Sentiment analysis is a form of Natural- Language Processing.  With it, software classifies the affective content of sentences, their emotional tone, bias and intensity.  Are they positive or negative about the subject in question?  This can be very useful in ranking movies, books, media and just about anything humans consume.  Including politicians. Sentiment Analysis & Text
  • 13.  By reading sentiment, you read many things.  Corporations can measure customer satisfaction.  Governments can monitor popular unrest.  Event organizers can track audience engagement.  Employers can measure job applicant fit.  Celebrities can gauge fame and track scandal. Who cares what they say?
  • 14.  Recurrent neural net  Restricted Boltzmann machine (RBM)  Deep-belief network: A stack of RBMs  Deep Autoencoder: 2 DBNs  Denoising Autoencoder (yay, noise!)  Convolutional net (ConvNet)  Recursive neural tensor network (RNTN) A Neural Nets Taxonomy
  • 15.  Two layers of neuron-like nodes.  The first layer is the visible, or input, layer  The second is the hidden layer, which identifies features in the input  This simple network is symmetrically connected.  “Restricted” means there are no visible-visible or hidden-hidden connections; i.e. all connections happen *between* layers. Restricted Boltzmann Machine (RBMs)
  • 16.  A deep-belief net is a stack of RBMs.  Each RBM’s hidden layer becomes the next RBM’s visible/input layer.  In this manner, a DBN learns more and more complex features  A machine vision example: 1) Pixels are input; 2) H1 learns an edge or line; 3) H2 learns a corner or set of lines; 4) H3 learns two groups of lines forming an object, maybe a face.  The final layer of a DBN classifies feature groups. It groups them in buckets: e.g. sunset, elephant, flower. Deep-belief net (DBN)
  • 17.  A deep autoencoder consists of two DBNs.  The first DBN *encodes* the data into a vector of 10-30 numbers. This is pre-training.  The second DBN decodes the data into its original state.  Backprop happens solely on the second DBN  This is the fine-tuning stage and it’s carried out with reconstruction entropy.  Deep autoencoders will reduce any document or image to a highly compact vector.  Those vectors are useful in search, QA and information retrieval. Deep Autoencoder
  • 18.  Autoencoders are useful for dimensionality reduction.  The risk they run is learning the identity function of the input.  Dropout is one way to address that risk.  Noise is another.  Noise is the stochastic, or random, corruption of the input.  The machine then learns features despite the noise. It “denoises” the input.  A stacked denoising encoder is exactly what you’d think.  Good for unsupervised pre-training, which initializes the weights. Denoising Autoencoder
  • 19.  ConvNets are a type of RBM.  The difference is they’re asymmetric.  In an RBM, each node in the visible layer connects to each node in the hidden layer.  In a ConvNet, each node connects to the node straight ahead of it, and to the two others immediately to the right and left of it.  This means that ConvNets learn data like images in patches.  Each piece learned is then woven together in the whole. Convolutional Net
  • 20.  Recursive nets are top-down, hierarchical nets rather than feed-forward like DBNs.  RNTNs handle sequence-based classification, windows of several events, entire scenes rather than images.  The features themselves are vectors.  A tensor is a multi-dimensional matrix, or multiple matrices of the same size. Recursive Neural Tensor Net
  • 21. RNTNs & Scene Composition
  • 22. RNTNs & Sentence Parsing