SlideShare a Scribd company logo
Deep Learning Internals
SEMINAR
contents
☛ Description
☛ Key Skills
☛ Prerequisites
☛ Instructional Method
☛ course contents
Mobile: +91 7719882295/ 9730463630
Email: sales@anikatechnologies.com
Website:www.anikatechnologies.com
Deep Learning Internals
SEMINAR
course contents
☛ Introduction to Deep Learning Day1
☛ Deep Neural Network
☛ Convolutional Neural Network
☛ Deep Models for Text and Sequence
☛ Crash Course in GPU
☛ NLP with Deep Learning
☛ Software Tools
☛ Deep Feedforward Networks Day2
☛ Regularization for Deep Learning
☛ Optimization for Training Deep Models
☛ Sequence Modeling: Recurrent and Recursive
Nets
☛ Practical Methodology
☛ Applications
☛ Linear Factor Models
☛ Autoencoders Day3
☛ Representation Learning
☛ Structured Probabilistic Models for Deep
Learning
☛ Monte Carlo Methods
☛ Confronting the Partition Function
☛ Approximate Inference
☛ Deep Generative Models
☛ Reinforcement Learning
Mobile: +91 7719882295/ 9730463630
Email: sales@anikatechnologies.com
Website:www.anikatechnologies.com
Description:
■ You’ve probably heard that Deep Learning is making news across the world as one
of the most promising techniques in machine learning, especially for analyzing image
data. With every industry dedicating resources to unlock the deep learning potential,
to be competitive, you will want to use these models in tasks such as image tagging,
object recognition, speech recognition, and text analysis. In this training session you
will build deep learning models using neural networks, explore what they are, what
they do, and how. To remove the barrier introduced by designing, training, and
tuning networks, and to be able to achieve high performance with less labeled data,
you will also build deep learning classifiers tailored to your specific task using pre-
trained models, which we call deep features. Also, you’ll develop a clear
understanding of the motivation for deep learning, and design intelligent systems that
learn from complex and/or large-scale datasets.
Key Skills:
■
■
■
■
■
■
Combine different types of layers and activation functions to obtain
better performance
Describe how these models can be applied in computer vision, text analytics
and speech recognition
Describe how a neural network model is represented and how it encodes non-
linear features
Use pretrained models, such as deep features, for new classification
tasks You will learn how to Prototype ideas and then productionize
Explore a dataset of products, reviews and images
Prerequisites:
■
■
This is an advanced level session and it assumes that you have good familiarity
with Machine learning.
Machine Learning Internals
Instructional Method:
■ This is an instructor led course provides lecture topics and the practical application
of Deep Learning and the underlying technologies. It pictorially presents most
concepts and there is a detailed case study that strings together the technologies,
patterns and design.
Mobile: +91 7719882295/ 9730463630
Email: sales@anikatechnologies.com
Website:www.anikatechnologies.com
Deep Learning Internals
■
■
■
Introduction to Deep Learning
• Parameter Hyperspace
• Minimizing Cost Entropy
• Normalized Inputs And Initial Weights
• Transition Into Practical Aspects Of Learning
• Measuring Performance
• Stochastic Gradient Descent
• Training your Logistic Classifier
• Transition: Overfitting -> Dataset Size
• Momentum And Learning Rate Decay
• Solving Problems
• Supervised Classification
• Lather Rinse Repeat
• Optimizing A Logistic Classifier
• Cross Entropy
• What is Deep Learning
Deep Neural Network
• "2-layer" neural network
• Dropout
• Network Of ReLUs
• Intro to Deep Neural Network
• No Neurons
• Backprop
• Regularization Intro
• Linear Models Are Limited
• The Chain Rule
• Dropout Pt-2
• Regularization
• Training A Deep Learning Network
Convolutional Neural Network
Mobile: +91 7719882295/ 9730463630
Email: sales@anikatechnologies.com
Website:www.anikatechnologies.com
• Statistical Invariance
• Intro To CNNs
• Inception Module
• Convolutional Networks
• Explore The Design Space
• x Convolutions
• Convolutions Continued
■ Deep Models for Text and Sequence
• Play Legos
• WordVec
• Beam Search
• Train A Text Embedding Model
• Embeddings
• WordVec Details
• TSNE
• LSTM
• RNNs
• Semantic Ambiguity
• Captioning And Translation
• Memory Cel
• Regularization
• Vanishing / Exploding Gradients
• Unsupervised Learning
• Analogies
• Sequences Of Varying Length
• LSTM Cell
• Backprop Through Time
■ Crash Course in GPU
• Introduction to CUDA and OpenCL
• Fundamentals of GPU Algorithms(Applications of Sort and Scan)
• Dynamic Parallelism
• Optimizing GPU Programs
Mobile: +91 7719882295/ 9730463630
Email: sales@anikatechnologies.com
Website:www.anikatechnologies.com
• Parallel Computation Patterns
• The GPU Hardware and Parallel Communication Patterns
• The GPU programming Model
• Parallel Optimization Patterns
• Fundamentals of GPU Algorithms(Reduce,Scan,Histograms)
• Deep Learning use of GPU
■ NLP with Deep Learning
• Word representations
• Compositional Vector Grammars: Parsing
• Matrix-Vector RNNs: Relation classification
• Unsupervised word vector learning
• Learning word-level classifiers: POS and NER
• Backpropagation Training
• Recursive Neural Tensor Networks: Sentiment Analysis
• Recursive Neural Networks for Parsing
• Recursive Autoencoders: Paraphrase Detection
• Optimization and Backpropagation Through Structure
• Sharing statistical strength
• Assorted Speech and NLP applications
■ Software Tools
• Tensorflow
• Matlab
• Octave
■ Deep Feedforward Networks
• Hidden Units
• Architecture Design
• Back-Propagation and Other Differentiation Algorithms
• Gradient-Based Learning
• Learning XOR
■ Regularization for Deep Learning
• Bagging and Other Ensemble Methods
• Dataset Augmentation
Mobile: +91 7719882295/ 9730463630
Email: sales@anikatechnologies.com
Website:www.anikatechnologies.com
• Tangent Distance, Tangent Prop, and Manifold Tangent Classifier
• Parameter Tying and Parameter Sharing
• Semi-Supervised Learning
• Early Stopping
• Sparse Representations
• Multi-Task Learning
• Regularization and Under-Constrained Problems
• Dropout
• Norm Penalties as Constrained Optimization
• Noise Robustness
• Parameter Norm Penalties
• Adversarial Training
■ Optimization for Training Deep Models
• Random or Unsupervised Features
• Convolutional Networks
• Structured Outputs
• Efficient Convolution Algorithms
• Challenges in Neural Network Optimizatio
• Variants of the Basic Convolution Function
• The Convolution Operation
• Pooling
• Parameter Initialization Strategies
• Motivation
• How Learning Differs from Pure Optimization
• Basic Algorithms
• Optimization Strategies and Meta-Algorithms
• The Neuroscientific Basis for Convolutional Networks
• Convolution and Pooling as an Infinitely Strong Prior
• Data Types
• Approximate Second-Order Methods
• Algorithms with Adaptive Learning Rates
■ Sequence Modeling: Recurrent and Recursive Nets
Mobile: +91 7719882295/ 9730463630
Email: sales@anikatechnologies.com
Website:www.anikatechnologies.com
• Unfolding Computational Graphs
• The Challenge of Long-Term Dependencies
• Bidirectional RNNs
• Echo State Networks
• Deep Recurrent Networks
• Recursive Neural Networks
• Leaky Units and Other Strategies for Multiple Time Scales
• Explicit Memory
• The Long Short-Term Memory and Other Gated RNNs
• Encoder-Decoder Sequence-to-Sequence Architectures
• Recurrent Neural Networks
• Optimization for Long-Term Dependencies
■ Practical Methodology
• Selecting Hyperparameters
• Debugging Strategies
• Example : Facial Recognition
• Performance Metrics
• Determining Whether to Gather More Data
• Default Baseline Models
■ Applications
• Other Applications
• Computer Vision
• Natural Language Processing
• Large Scale Deep Learning
• Speech Recognition
■ Linear Factor Models
• Probabilistic PCA and Factor Analysis
• Independent Component Analysis (ICA)
• Manifold Interpretation of PCA
• Slow Feature Analysis
• Sparse Coding
Mobile: +91 7719882295/ 9730463630
Email: sales@anikatechnologies.com
Website:www.anikatechnologies.com
■ Autoencoders
• Representational Power, Layer Size and Depth
• Contractive Autoencoders
• Stochastic Encoders and Decoders
• Predictive Sparse Decomposition
• Undercomplete Autoencoders
• Regularized Autoencoders
• Learning Manifolds with Autoencoders
• Applications of Autoencoders
• Denoising Autoencoders
■ Representation Learning
• Distributed Representation
• Transfer Learning and Domain Adaptation
• Greedy Layer-Wise Unsupervised Pretraining
• Providing Clues to Discover Underlying Causes
• Semi-Supervised Disentangling of Causal Factors
• Exponential Gains from Depth
■ Structured Probabilistic Models for Deep Learning
• The Deep Learning Approach to Structured Probabilistic Models
• Advantages of Structured Modeling
• Inference and Approximate Inference
• Using Graphs to Describe Model Structure
• Sampling from Graphical Models
• Learning about Dependencies
• The Challenge of Unstructured Modeling
■ Monte Carlo Methods
• The Challenge of Mixing between Separated Modes
• Gibbs Sampling
• Sampling and Monte Carlo Methods
• Markov Chain Monte Carlo Methods
• Importance Sampling
Mobile: +91 7719882295/ 9730463630
Email: sales@anikatechnologies.com
Website:www.anikatechnologies.com
■ Confronting the Partition Function
• The Log-Likelihood Gradient
• Noise-Contrastive Estimation
• Stochastic Maximum Likelihood and Contrastive Divergence
• Score Matching and Ratio Matching
• Denoising Score Matching
• Pseudolikelihood
• Estimating the Partition Function
■ Approximate Inference
• Variational Inference and Learning
• MAP Inference and Sparse Coding
• Inference as Optimization
• Learned Approximate Inference
• Expectation Maximization
■ Deep Generative Models
• Deep Boltzmann Machines
• Back-Propagation through Random Operations
• Restricted Boltzmann Machines
• Generative Stochastic Networks
• Boltzmann Machines for Structured or Sequential Outputs
• Boltzmann Machines
• Other Boltzmann Machines
• Other Generation Schemes
• Directed Generative Nets
• Boltzmann Machines for Real-Valued Data
• Evaluating Generative Models
• Drawing Samples from Autoencoders
• Deep Belief Networks
• Convolutional Boltzmann Machines
■ Reinforcement Learning
• Integrating Learning and Planning
• Policy Gradient Methods
Mobile: +91 7719882295/ 9730463630
Email: sales@anikatechnologies.com
Website:www.anikatechnologies.com
• Model-Free Control
• Exploration and Exploitation
• Markov Decision Processes
• Case Study: RL in Classic Games
• Introduction to Reinforcement Learning
• Planning by Dynamic Programming
• Model-Free Prediction
• Value Function Approximation
Mobile: +91 7719882295/ 9730463630
Email: sales@anikatechnologies.com
Website:www.anikatechnologies.com

More Related Content

PPTX
Natural language processing techniques transition from machine learning to de...
PPTX
Word embeddings, RNN, GRU and LSTM
PDF
Mastering Advanced Deep Learning Techniques | IABAC
PDF
Mastering Advanced Deep Learning Techniques
DOCX
Title_ Deep Learning Explained_ What You Should Be Aware of in Data Science a...
PPTX
Deep_Learning_Introduction for newbe.pptx
PPTX
Deep learning from a novice perspective
PPTX
Deep learning introduction
Natural language processing techniques transition from machine learning to de...
Word embeddings, RNN, GRU and LSTM
Mastering Advanced Deep Learning Techniques | IABAC
Mastering Advanced Deep Learning Techniques
Title_ Deep Learning Explained_ What You Should Be Aware of in Data Science a...
Deep_Learning_Introduction for newbe.pptx
Deep learning from a novice perspective
Deep learning introduction

Similar to Deep learning internals (20)

PPT
Introduction_to_DEEP_LEARNING.ppt
PPT
Introduction_to_DEEP_LEARNING ppt 101ppt
PPT
Introduction_to_DEEP_LEARNING.sfsdafsadfsadfsdafsdppt
PDF
Deep learning: Cutting through the Myths and Hype
PPTX
Introduction-to-Deep-Learning about new technologies
PPTX
Introduction to deep learning
PPTX
Introduction to deep learning
PPTX
Deep_Learning_Algorithms_Presentation.pptx
PDF
Deep learning
PDF
What is Deep Learning? A Comprehensive Guide
PDF
Separating Hype from Reality in Deep Learning with Sameer Farooqui
PDF
Deep learning @ University of Oradea - part I (16 Jan. 2018)
PPTX
Deep learning short introduction
PDF
MLIP - Chapter 3 - Introduction to deep learning
PDF
A Platform for Accelerating Machine Learning Applications
PPT
Introduction_to_DEEP_LEARNING.ppt machine learning that uses data, loads ...
PDF
AI and Deep Learning
PPTX
How to architect Deep Learning
PPTX
Deep-Learning-Basics-Introduction-RAJA M
PPTX
Deep_Learning_Demo_Class_Detailed.pptx sn
Introduction_to_DEEP_LEARNING.ppt
Introduction_to_DEEP_LEARNING ppt 101ppt
Introduction_to_DEEP_LEARNING.sfsdafsadfsadfsdafsdppt
Deep learning: Cutting through the Myths and Hype
Introduction-to-Deep-Learning about new technologies
Introduction to deep learning
Introduction to deep learning
Deep_Learning_Algorithms_Presentation.pptx
Deep learning
What is Deep Learning? A Comprehensive Guide
Separating Hype from Reality in Deep Learning with Sameer Farooqui
Deep learning @ University of Oradea - part I (16 Jan. 2018)
Deep learning short introduction
MLIP - Chapter 3 - Introduction to deep learning
A Platform for Accelerating Machine Learning Applications
Introduction_to_DEEP_LEARNING.ppt machine learning that uses data, loads ...
AI and Deep Learning
How to architect Deep Learning
Deep-Learning-Basics-Introduction-RAJA M
Deep_Learning_Demo_Class_Detailed.pptx sn
Ad

More from Anand Narayanan (12)

PDF
Scrum Foundation Training by Anika Technologies
PDF
Agile Essentials Training by Anika Technologies
PDF
Smart Staffing using Regression Analysis Model
PDF
Sentiment analysis using nlp
PDF
Deep learning with_computer_vision
PDF
Spark Internals Training | Apache Spark | Spark | Anika Technologies
PDF
Advanced Elastic Search | Elastic Search | Kibana | Logstash
PDF
JVM and Java Performance Tuning | JVM Tuning | Java Performance
PDF
Java Concurrency and Performance | Multi Threading | Concurrency | Java Conc...
PDF
Understanding and Designing Ultra low latency systems | Low Latency | Ultra L...
PDF
Big Data Analytics and Artifical Intelligence
PDF
SynopsisLowLatencySeminar.PDF
Scrum Foundation Training by Anika Technologies
Agile Essentials Training by Anika Technologies
Smart Staffing using Regression Analysis Model
Sentiment analysis using nlp
Deep learning with_computer_vision
Spark Internals Training | Apache Spark | Spark | Anika Technologies
Advanced Elastic Search | Elastic Search | Kibana | Logstash
JVM and Java Performance Tuning | JVM Tuning | Java Performance
Java Concurrency and Performance | Multi Threading | Concurrency | Java Conc...
Understanding and Designing Ultra low latency systems | Low Latency | Ultra L...
Big Data Analytics and Artifical Intelligence
SynopsisLowLatencySeminar.PDF
Ad

Recently uploaded (20)

PPTX
Understanding_Digital_Forensics_Presentation.pptx
PDF
Profit Center Accounting in SAP S/4HANA, S4F28 Col11
PDF
The Rise and Fall of 3GPP – Time for a Sabbatical?
PDF
Build a system with the filesystem maintained by OSTree @ COSCUP 2025
PDF
Agricultural_Statistics_at_a_Glance_2022_0.pdf
PDF
MIND Revenue Release Quarter 2 2025 Press Release
PDF
NewMind AI Weekly Chronicles - August'25 Week I
PDF
Building Integrated photovoltaic BIPV_UPV.pdf
PDF
Empathic Computing: Creating Shared Understanding
PDF
cuic standard and advanced reporting.pdf
PPTX
20250228 LYD VKU AI Blended-Learning.pptx
PDF
TokAI - TikTok AI Agent : The First AI Application That Analyzes 10,000+ Vira...
PDF
Blue Purple Modern Animated Computer Science Presentation.pdf.pdf
PDF
Unlocking AI with Model Context Protocol (MCP)
PPTX
Cloud computing and distributed systems.
PDF
Per capita expenditure prediction using model stacking based on satellite ima...
PDF
Advanced methodologies resolving dimensionality complications for autism neur...
PPTX
MYSQL Presentation for SQL database connectivity
PDF
Review of recent advances in non-invasive hemoglobin estimation
PPTX
ACSFv1EN-58255 AWS Academy Cloud Security Foundations.pptx
Understanding_Digital_Forensics_Presentation.pptx
Profit Center Accounting in SAP S/4HANA, S4F28 Col11
The Rise and Fall of 3GPP – Time for a Sabbatical?
Build a system with the filesystem maintained by OSTree @ COSCUP 2025
Agricultural_Statistics_at_a_Glance_2022_0.pdf
MIND Revenue Release Quarter 2 2025 Press Release
NewMind AI Weekly Chronicles - August'25 Week I
Building Integrated photovoltaic BIPV_UPV.pdf
Empathic Computing: Creating Shared Understanding
cuic standard and advanced reporting.pdf
20250228 LYD VKU AI Blended-Learning.pptx
TokAI - TikTok AI Agent : The First AI Application That Analyzes 10,000+ Vira...
Blue Purple Modern Animated Computer Science Presentation.pdf.pdf
Unlocking AI with Model Context Protocol (MCP)
Cloud computing and distributed systems.
Per capita expenditure prediction using model stacking based on satellite ima...
Advanced methodologies resolving dimensionality complications for autism neur...
MYSQL Presentation for SQL database connectivity
Review of recent advances in non-invasive hemoglobin estimation
ACSFv1EN-58255 AWS Academy Cloud Security Foundations.pptx

Deep learning internals

  • 1. Deep Learning Internals SEMINAR contents ☛ Description ☛ Key Skills ☛ Prerequisites ☛ Instructional Method ☛ course contents Mobile: +91 7719882295/ 9730463630 Email: sales@anikatechnologies.com Website:www.anikatechnologies.com
  • 2. Deep Learning Internals SEMINAR course contents ☛ Introduction to Deep Learning Day1 ☛ Deep Neural Network ☛ Convolutional Neural Network ☛ Deep Models for Text and Sequence ☛ Crash Course in GPU ☛ NLP with Deep Learning ☛ Software Tools ☛ Deep Feedforward Networks Day2 ☛ Regularization for Deep Learning ☛ Optimization for Training Deep Models ☛ Sequence Modeling: Recurrent and Recursive Nets ☛ Practical Methodology ☛ Applications ☛ Linear Factor Models ☛ Autoencoders Day3 ☛ Representation Learning ☛ Structured Probabilistic Models for Deep Learning ☛ Monte Carlo Methods ☛ Confronting the Partition Function ☛ Approximate Inference ☛ Deep Generative Models ☛ Reinforcement Learning Mobile: +91 7719882295/ 9730463630 Email: sales@anikatechnologies.com Website:www.anikatechnologies.com
  • 3. Description: ■ You’ve probably heard that Deep Learning is making news across the world as one of the most promising techniques in machine learning, especially for analyzing image data. With every industry dedicating resources to unlock the deep learning potential, to be competitive, you will want to use these models in tasks such as image tagging, object recognition, speech recognition, and text analysis. In this training session you will build deep learning models using neural networks, explore what they are, what they do, and how. To remove the barrier introduced by designing, training, and tuning networks, and to be able to achieve high performance with less labeled data, you will also build deep learning classifiers tailored to your specific task using pre- trained models, which we call deep features. Also, you’ll develop a clear understanding of the motivation for deep learning, and design intelligent systems that learn from complex and/or large-scale datasets. Key Skills: ■ ■ ■ ■ ■ ■ Combine different types of layers and activation functions to obtain better performance Describe how these models can be applied in computer vision, text analytics and speech recognition Describe how a neural network model is represented and how it encodes non- linear features Use pretrained models, such as deep features, for new classification tasks You will learn how to Prototype ideas and then productionize Explore a dataset of products, reviews and images Prerequisites: ■ ■ This is an advanced level session and it assumes that you have good familiarity with Machine learning. Machine Learning Internals Instructional Method: ■ This is an instructor led course provides lecture topics and the practical application of Deep Learning and the underlying technologies. It pictorially presents most concepts and there is a detailed case study that strings together the technologies, patterns and design. Mobile: +91 7719882295/ 9730463630 Email: sales@anikatechnologies.com Website:www.anikatechnologies.com
  • 4. Deep Learning Internals ■ ■ ■ Introduction to Deep Learning • Parameter Hyperspace • Minimizing Cost Entropy • Normalized Inputs And Initial Weights • Transition Into Practical Aspects Of Learning • Measuring Performance • Stochastic Gradient Descent • Training your Logistic Classifier • Transition: Overfitting -> Dataset Size • Momentum And Learning Rate Decay • Solving Problems • Supervised Classification • Lather Rinse Repeat • Optimizing A Logistic Classifier • Cross Entropy • What is Deep Learning Deep Neural Network • "2-layer" neural network • Dropout • Network Of ReLUs • Intro to Deep Neural Network • No Neurons • Backprop • Regularization Intro • Linear Models Are Limited • The Chain Rule • Dropout Pt-2 • Regularization • Training A Deep Learning Network Convolutional Neural Network Mobile: +91 7719882295/ 9730463630 Email: sales@anikatechnologies.com Website:www.anikatechnologies.com
  • 5. • Statistical Invariance • Intro To CNNs • Inception Module • Convolutional Networks • Explore The Design Space • x Convolutions • Convolutions Continued ■ Deep Models for Text and Sequence • Play Legos • WordVec • Beam Search • Train A Text Embedding Model • Embeddings • WordVec Details • TSNE • LSTM • RNNs • Semantic Ambiguity • Captioning And Translation • Memory Cel • Regularization • Vanishing / Exploding Gradients • Unsupervised Learning • Analogies • Sequences Of Varying Length • LSTM Cell • Backprop Through Time ■ Crash Course in GPU • Introduction to CUDA and OpenCL • Fundamentals of GPU Algorithms(Applications of Sort and Scan) • Dynamic Parallelism • Optimizing GPU Programs Mobile: +91 7719882295/ 9730463630 Email: sales@anikatechnologies.com Website:www.anikatechnologies.com
  • 6. • Parallel Computation Patterns • The GPU Hardware and Parallel Communication Patterns • The GPU programming Model • Parallel Optimization Patterns • Fundamentals of GPU Algorithms(Reduce,Scan,Histograms) • Deep Learning use of GPU ■ NLP with Deep Learning • Word representations • Compositional Vector Grammars: Parsing • Matrix-Vector RNNs: Relation classification • Unsupervised word vector learning • Learning word-level classifiers: POS and NER • Backpropagation Training • Recursive Neural Tensor Networks: Sentiment Analysis • Recursive Neural Networks for Parsing • Recursive Autoencoders: Paraphrase Detection • Optimization and Backpropagation Through Structure • Sharing statistical strength • Assorted Speech and NLP applications ■ Software Tools • Tensorflow • Matlab • Octave ■ Deep Feedforward Networks • Hidden Units • Architecture Design • Back-Propagation and Other Differentiation Algorithms • Gradient-Based Learning • Learning XOR ■ Regularization for Deep Learning • Bagging and Other Ensemble Methods • Dataset Augmentation Mobile: +91 7719882295/ 9730463630 Email: sales@anikatechnologies.com Website:www.anikatechnologies.com
  • 7. • Tangent Distance, Tangent Prop, and Manifold Tangent Classifier • Parameter Tying and Parameter Sharing • Semi-Supervised Learning • Early Stopping • Sparse Representations • Multi-Task Learning • Regularization and Under-Constrained Problems • Dropout • Norm Penalties as Constrained Optimization • Noise Robustness • Parameter Norm Penalties • Adversarial Training ■ Optimization for Training Deep Models • Random or Unsupervised Features • Convolutional Networks • Structured Outputs • Efficient Convolution Algorithms • Challenges in Neural Network Optimizatio • Variants of the Basic Convolution Function • The Convolution Operation • Pooling • Parameter Initialization Strategies • Motivation • How Learning Differs from Pure Optimization • Basic Algorithms • Optimization Strategies and Meta-Algorithms • The Neuroscientific Basis for Convolutional Networks • Convolution and Pooling as an Infinitely Strong Prior • Data Types • Approximate Second-Order Methods • Algorithms with Adaptive Learning Rates ■ Sequence Modeling: Recurrent and Recursive Nets Mobile: +91 7719882295/ 9730463630 Email: sales@anikatechnologies.com Website:www.anikatechnologies.com
  • 8. • Unfolding Computational Graphs • The Challenge of Long-Term Dependencies • Bidirectional RNNs • Echo State Networks • Deep Recurrent Networks • Recursive Neural Networks • Leaky Units and Other Strategies for Multiple Time Scales • Explicit Memory • The Long Short-Term Memory and Other Gated RNNs • Encoder-Decoder Sequence-to-Sequence Architectures • Recurrent Neural Networks • Optimization for Long-Term Dependencies ■ Practical Methodology • Selecting Hyperparameters • Debugging Strategies • Example : Facial Recognition • Performance Metrics • Determining Whether to Gather More Data • Default Baseline Models ■ Applications • Other Applications • Computer Vision • Natural Language Processing • Large Scale Deep Learning • Speech Recognition ■ Linear Factor Models • Probabilistic PCA and Factor Analysis • Independent Component Analysis (ICA) • Manifold Interpretation of PCA • Slow Feature Analysis • Sparse Coding Mobile: +91 7719882295/ 9730463630 Email: sales@anikatechnologies.com Website:www.anikatechnologies.com
  • 9. ■ Autoencoders • Representational Power, Layer Size and Depth • Contractive Autoencoders • Stochastic Encoders and Decoders • Predictive Sparse Decomposition • Undercomplete Autoencoders • Regularized Autoencoders • Learning Manifolds with Autoencoders • Applications of Autoencoders • Denoising Autoencoders ■ Representation Learning • Distributed Representation • Transfer Learning and Domain Adaptation • Greedy Layer-Wise Unsupervised Pretraining • Providing Clues to Discover Underlying Causes • Semi-Supervised Disentangling of Causal Factors • Exponential Gains from Depth ■ Structured Probabilistic Models for Deep Learning • The Deep Learning Approach to Structured Probabilistic Models • Advantages of Structured Modeling • Inference and Approximate Inference • Using Graphs to Describe Model Structure • Sampling from Graphical Models • Learning about Dependencies • The Challenge of Unstructured Modeling ■ Monte Carlo Methods • The Challenge of Mixing between Separated Modes • Gibbs Sampling • Sampling and Monte Carlo Methods • Markov Chain Monte Carlo Methods • Importance Sampling Mobile: +91 7719882295/ 9730463630 Email: sales@anikatechnologies.com Website:www.anikatechnologies.com
  • 10. ■ Confronting the Partition Function • The Log-Likelihood Gradient • Noise-Contrastive Estimation • Stochastic Maximum Likelihood and Contrastive Divergence • Score Matching and Ratio Matching • Denoising Score Matching • Pseudolikelihood • Estimating the Partition Function ■ Approximate Inference • Variational Inference and Learning • MAP Inference and Sparse Coding • Inference as Optimization • Learned Approximate Inference • Expectation Maximization ■ Deep Generative Models • Deep Boltzmann Machines • Back-Propagation through Random Operations • Restricted Boltzmann Machines • Generative Stochastic Networks • Boltzmann Machines for Structured or Sequential Outputs • Boltzmann Machines • Other Boltzmann Machines • Other Generation Schemes • Directed Generative Nets • Boltzmann Machines for Real-Valued Data • Evaluating Generative Models • Drawing Samples from Autoencoders • Deep Belief Networks • Convolutional Boltzmann Machines ■ Reinforcement Learning • Integrating Learning and Planning • Policy Gradient Methods Mobile: +91 7719882295/ 9730463630 Email: sales@anikatechnologies.com Website:www.anikatechnologies.com
  • 11. • Model-Free Control • Exploration and Exploitation • Markov Decision Processes • Case Study: RL in Classic Games • Introduction to Reinforcement Learning • Planning by Dynamic Programming • Model-Free Prediction • Value Function Approximation Mobile: +91 7719882295/ 9730463630 Email: sales@anikatechnologies.com Website:www.anikatechnologies.com