DEEP LEARNING IN LIMITED
RESOURCE ENVIRONMENTS
OGUZ VURUSKANER
OVERVIEW
➢ Limited Resource Environments
➢ Training Improvements
➢ Self-Adversarial Training
➢ Arcihtectural Improvements
➢ Model Quantization
➢ Depthwise Separable Convolutions
➢ References
LIMITED RESOURCE ENVIRONMENTS
➢ In actual, the supply of a resource is always limited at any point of time.
➢ Virtually unlimited resources: On-demand extensions are available.Training
environments mostly have virtually unlimited resources. ( e.g. data centers,
cloud services )
➢ Limited resources: Not extendable. ( e.g. Perseverance (Mars Rover),
embedded devices, mobile phones )
Model Improvements
Training Improvements Architectural Improvements
FIRE DETECTION DATASET
• It is a benchmark dataset for model experiments.
• In the following months, it is going to be released public.
• 4200 training images , 672 validation images
TRAINING IMPROVEMENTS
SELF-ADVERSARIAL TRAINING
• By adding small but intentional worst-case perturbations, perturbed input
results in the model outputting an incorrect answer with high confidence.[1]
• Even though deep learning models have a complex non-linear computational
graph, they can be deceived by simple linear method which is called Fast
Gradient Sign Method.
• In our experiments, we’ve used Fast Gradient Sign Method.
FAST GRADIENT SIGN METHOD
+ =
FIRE DETECTION RESULTS
MODEL CORRECT ALARM FALSE ALARM
ResNet-18 w/ Adversarial 91.1% 2.9%
ResNet-18 91.0% 3.2%
CONCLUSION
• FGSM is a valid data augmentation strategy. It has improved performance with
considerably small training time drawback.
• One advantage of FGSM is its perturbation vector strictly depends on current
state of the trained model. It is a self-evolving data augmentation strategy.
ARCHITECTURAL IMPROVEMENTS
MODEL QUANTIZATION
• Quantization converts a real value to an integer value. Reverse of this process
is called Dequantization.
• In general, quantization converts from 32-bit floating point to 1-byte which is
x4 memory saving!
Typical Quantization Schema[2]
S is called scale, Z is called zero-point. Together, they
define an affine transformation between real values
and integer values.
MODEL QUANTIZATION
Quantization mapping between floating point and signed byte with Scale=0.024
and Zero-point=0
-2 0 3 4
-127 127
-3.048
-83 0 125
QUANTIZATION AWARE TRAINING
• This technique readjusts floating point weights to the nearest quantization level
after every training step in the given quantization interval [a,b].
Quantization
Step
Quantization
Level
Clamp function translates input domain to quantization
interval.
FIRE DETECTION RESULTS
MODEL CORRECT ALARM FALSE ALARM
ResNet-18 QAT 90.3% 2.8%
ResNet-18 91.0% 3.2%
RESOURCE USAGE
CONCLUSION
• In single batch inference, quantized inference outperforms approximately
doubles up in speed. However, in general performance, it seems that there are
inconsistencies on inference.
• When the results are compared with respect to inference, still, standard FP-32
inference has better results. It has higher average inference time but less
deviation.
DEPTHWISE SEPERABLE
CONVOLUTIONS (MOBILENET[3])
Naïve Convolution Depthwise Seperable Convolution
DEPTHWISE SEPARABLE
CONVOLUTIONS (MOBILENET)
• Naïve convolution complexity
• Depthwise Separable Convolution complexity
USE CASE
• Unsupervised anomaly detection on real time streams requires continuous
training of deep learning model.
• To increase inference speed and memory, we’ve proposed using depthwise
separable convolutions.
• A hourglass model is trained with normal video frames and then tested with
anormal video frames.
USE CASE
An example hourglass network architecture
RESULTS
Naïve convolution – 537K parameters
Average InferenceTime: 0.106s
DS convolution – 93.8K parameters
Average InferenceTime: 0.144s
CONCLUSION
• While replacing naïve convolutions with depthwise separable convolutions, 2
extra layers has been added.That’s why inference speed may have reduced
even there are less parameters in DS Convolution.
• Real-time anomaly detection with self-trained models are still active research
field.
FUTURE WORK
• Student-Teacher Models
• Feature-Based Knowledge Distillation
• Response-Based Knowledge Distillation
• Pseudo Labels
• Confident Learning : Dataset Labels Improvement
• Pseudo Labels combined with student-teacher models : Meta Pseudo Labels
REFERENCES
1. Goodfellow, Ian J., Jonathon Shlens, and Christian Szegedy.
"Explaining and harnessing adversarial examples." arXiv preprint
arXiv:1412.6572 (2014).
2. Jacob, Benoit, et al. "Quantization and training of neural networks for
efficient integer-arithmetic-only inference." Proceedings of the IEEE
Conference on Computer Vision and Pattern Recognition. 2018.
3. Howard, Andrew G., et al. "Mobilenets: Efficient convolutional neural
networks for mobile vision applications." arXiv preprint
arXiv:1704.04861 (2017).
Deep Learning in Limited Resource Environments

More Related Content

PPTX
Deep gradient compression
PDF
PR-231: A Simple Framework for Contrastive Learning of Visual Representations
PDF
Efficient de cvpr_2020_paper
PDF
Super resolution in deep learning era - Jaejun Yoo
PDF
PR-144: SqueezeNext: Hardware-Aware Neural Network Design
PDF
In datacenter performance analysis of a tensor processing unit
PDF
PR-270: PP-YOLO: An Effective and Efficient Implementation of Object Detector
PDF
PR243: Designing Network Design Spaces
Deep gradient compression
PR-231: A Simple Framework for Contrastive Learning of Visual Representations
Efficient de cvpr_2020_paper
Super resolution in deep learning era - Jaejun Yoo
PR-144: SqueezeNext: Hardware-Aware Neural Network Design
In datacenter performance analysis of a tensor processing unit
PR-270: PP-YOLO: An Effective and Efficient Implementation of Object Detector
PR243: Designing Network Design Spaces

What's hot (20)

PPTX
Fast AutoAugment
PDF
Unsupervised visual representation learning overview: Toward Self-Supervision
PPTX
Survey on contrastive self supervised l earning
PDF
PR-217: EfficientDet: Scalable and Efficient Object Detection
PDF
Efficient Neural Architecture Search via Parameter Sharing
PDF
PR-197: One ticket to win them all: generalizing lottery ticket initializatio...
PDF
PR-169: EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks
PDF
PR-284: End-to-End Object Detection with Transformers(DETR)
PDF
[PR-325] Pixel-BERT: Aligning Image Pixels with Text by Deep Multi-Modal Tran...
PPTX
EfficientNet
PDF
ResNeSt: Split-Attention Networks
PDF
PR-317: MLP-Mixer: An all-MLP Architecture for Vision
PDF
PR-344: A Battle of Network Structures: An Empirical Study of CNN, Transforme...
PDF
"Quantizing Deep Networks for Efficient Inference at the Edge," a Presentatio...
PDF
Audio tagging system using densely connected convolutional networks (DCASE201...
PDF
PR-155: Exploring Randomly Wired Neural Networks for Image Recognition
PDF
Automated Speech Recognition
PDF
[2020 CVPR Efficient DET paper review]
PDF
CNN Quantization
PDF
DQN Variants: A quick glance
Fast AutoAugment
Unsupervised visual representation learning overview: Toward Self-Supervision
Survey on contrastive self supervised l earning
PR-217: EfficientDet: Scalable and Efficient Object Detection
Efficient Neural Architecture Search via Parameter Sharing
PR-197: One ticket to win them all: generalizing lottery ticket initializatio...
PR-169: EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks
PR-284: End-to-End Object Detection with Transformers(DETR)
[PR-325] Pixel-BERT: Aligning Image Pixels with Text by Deep Multi-Modal Tran...
EfficientNet
ResNeSt: Split-Attention Networks
PR-317: MLP-Mixer: An all-MLP Architecture for Vision
PR-344: A Battle of Network Structures: An Empirical Study of CNN, Transforme...
"Quantizing Deep Networks for Efficient Inference at the Edge," a Presentatio...
Audio tagging system using densely connected convolutional networks (DCASE201...
PR-155: Exploring Randomly Wired Neural Networks for Image Recognition
Automated Speech Recognition
[2020 CVPR Efficient DET paper review]
CNN Quantization
DQN Variants: A quick glance
Ad

Similar to Deep Learning in Limited Resource Environments (20)

PDF
DLD_WeightSharing_Slide
PDF
IEEE CIS Webinar Sustainable futures.pdf
PPTX
Batch normalization presentation
PPT
Unexpected Challenges in Large Scale Machine Learning by Charles Parker
PDF
PR-108: MobileNetV2: Inverted Residuals and Linear Bottlenecks
PPTX
Deep Learning
PDF
Big Data and Small Devices by Katharina Morik
PPTX
Deep learning
PPTX
Deep Learning for Image Analysis
PDF
Hardware Acceleration for Machine Learning
PDF
A Survey of Deep Learning Algorithms for Malware Detection
PPTX
Introduction to Deep Learning
PDF
unit4 Neural Networks and Deep Learning.pdf
PDF
1801.06434
PDF
Client side machine learning
PPT
deepnet-lourentzou.ppt
PPT
Overview of Deep Learning and its advantage
PPT
Deep learning is a subset of machine learning and AI
PPT
Introduction to Deep Learning presentation
PDF
Low compute Convolution Neural Network models
DLD_WeightSharing_Slide
IEEE CIS Webinar Sustainable futures.pdf
Batch normalization presentation
Unexpected Challenges in Large Scale Machine Learning by Charles Parker
PR-108: MobileNetV2: Inverted Residuals and Linear Bottlenecks
Deep Learning
Big Data and Small Devices by Katharina Morik
Deep learning
Deep Learning for Image Analysis
Hardware Acceleration for Machine Learning
A Survey of Deep Learning Algorithms for Malware Detection
Introduction to Deep Learning
unit4 Neural Networks and Deep Learning.pdf
1801.06434
Client side machine learning
deepnet-lourentzou.ppt
Overview of Deep Learning and its advantage
Deep learning is a subset of machine learning and AI
Introduction to Deep Learning presentation
Low compute Convolution Neural Network models
Ad

Recently uploaded (20)

DOCX
search engine optimization ppt fir known well about this
PPTX
observCloud-Native Containerability and monitoring.pptx
PDF
Hybrid model detection and classification of lung cancer
PPTX
Web Crawler for Trend Tracking Gen Z Insights.pptx
PDF
From MVP to Full-Scale Product A Startup’s Software Journey.pdf
PPT
Geologic Time for studying geology for geologist
PPTX
Chapter 5: Probability Theory and Statistics
PDF
Zenith AI: Advanced Artificial Intelligence
PDF
A review of recent deep learning applications in wood surface defect identifi...
PDF
Five Habits of High-Impact Board Members
PDF
STKI Israel Market Study 2025 version august
PPTX
O2C Customer Invoices to Receipt V15A.pptx
PDF
How ambidextrous entrepreneurial leaders react to the artificial intelligence...
PDF
A Late Bloomer's Guide to GenAI: Ethics, Bias, and Effective Prompting - Boha...
PDF
Hybrid horned lizard optimization algorithm-aquila optimizer for DC motor
PDF
Univ-Connecticut-ChatGPT-Presentaion.pdf
PDF
ENT215_Completing-a-large-scale-migration-and-modernization-with-AWS.pdf
PDF
Video forgery: An extensive analysis of inter-and intra-frame manipulation al...
PPTX
The various Industrial Revolutions .pptx
PDF
Enhancing emotion recognition model for a student engagement use case through...
search engine optimization ppt fir known well about this
observCloud-Native Containerability and monitoring.pptx
Hybrid model detection and classification of lung cancer
Web Crawler for Trend Tracking Gen Z Insights.pptx
From MVP to Full-Scale Product A Startup’s Software Journey.pdf
Geologic Time for studying geology for geologist
Chapter 5: Probability Theory and Statistics
Zenith AI: Advanced Artificial Intelligence
A review of recent deep learning applications in wood surface defect identifi...
Five Habits of High-Impact Board Members
STKI Israel Market Study 2025 version august
O2C Customer Invoices to Receipt V15A.pptx
How ambidextrous entrepreneurial leaders react to the artificial intelligence...
A Late Bloomer's Guide to GenAI: Ethics, Bias, and Effective Prompting - Boha...
Hybrid horned lizard optimization algorithm-aquila optimizer for DC motor
Univ-Connecticut-ChatGPT-Presentaion.pdf
ENT215_Completing-a-large-scale-migration-and-modernization-with-AWS.pdf
Video forgery: An extensive analysis of inter-and intra-frame manipulation al...
The various Industrial Revolutions .pptx
Enhancing emotion recognition model for a student engagement use case through...

Deep Learning in Limited Resource Environments

  • 1. DEEP LEARNING IN LIMITED RESOURCE ENVIRONMENTS OGUZ VURUSKANER
  • 2. OVERVIEW ➢ Limited Resource Environments ➢ Training Improvements ➢ Self-Adversarial Training ➢ Arcihtectural Improvements ➢ Model Quantization ➢ Depthwise Separable Convolutions ➢ References
  • 3. LIMITED RESOURCE ENVIRONMENTS ➢ In actual, the supply of a resource is always limited at any point of time. ➢ Virtually unlimited resources: On-demand extensions are available.Training environments mostly have virtually unlimited resources. ( e.g. data centers, cloud services ) ➢ Limited resources: Not extendable. ( e.g. Perseverance (Mars Rover), embedded devices, mobile phones )
  • 4. Model Improvements Training Improvements Architectural Improvements
  • 5. FIRE DETECTION DATASET • It is a benchmark dataset for model experiments. • In the following months, it is going to be released public. • 4200 training images , 672 validation images
  • 7. SELF-ADVERSARIAL TRAINING • By adding small but intentional worst-case perturbations, perturbed input results in the model outputting an incorrect answer with high confidence.[1] • Even though deep learning models have a complex non-linear computational graph, they can be deceived by simple linear method which is called Fast Gradient Sign Method. • In our experiments, we’ve used Fast Gradient Sign Method.
  • 8. FAST GRADIENT SIGN METHOD + =
  • 9. FIRE DETECTION RESULTS MODEL CORRECT ALARM FALSE ALARM ResNet-18 w/ Adversarial 91.1% 2.9% ResNet-18 91.0% 3.2%
  • 10. CONCLUSION • FGSM is a valid data augmentation strategy. It has improved performance with considerably small training time drawback. • One advantage of FGSM is its perturbation vector strictly depends on current state of the trained model. It is a self-evolving data augmentation strategy.
  • 12. MODEL QUANTIZATION • Quantization converts a real value to an integer value. Reverse of this process is called Dequantization. • In general, quantization converts from 32-bit floating point to 1-byte which is x4 memory saving! Typical Quantization Schema[2] S is called scale, Z is called zero-point. Together, they define an affine transformation between real values and integer values.
  • 13. MODEL QUANTIZATION Quantization mapping between floating point and signed byte with Scale=0.024 and Zero-point=0 -2 0 3 4 -127 127 -3.048 -83 0 125
  • 14. QUANTIZATION AWARE TRAINING • This technique readjusts floating point weights to the nearest quantization level after every training step in the given quantization interval [a,b]. Quantization Step Quantization Level Clamp function translates input domain to quantization interval.
  • 15. FIRE DETECTION RESULTS MODEL CORRECT ALARM FALSE ALARM ResNet-18 QAT 90.3% 2.8% ResNet-18 91.0% 3.2%
  • 17. CONCLUSION • In single batch inference, quantized inference outperforms approximately doubles up in speed. However, in general performance, it seems that there are inconsistencies on inference. • When the results are compared with respect to inference, still, standard FP-32 inference has better results. It has higher average inference time but less deviation.
  • 18. DEPTHWISE SEPERABLE CONVOLUTIONS (MOBILENET[3]) Naïve Convolution Depthwise Seperable Convolution
  • 19. DEPTHWISE SEPARABLE CONVOLUTIONS (MOBILENET) • Naïve convolution complexity • Depthwise Separable Convolution complexity
  • 20. USE CASE • Unsupervised anomaly detection on real time streams requires continuous training of deep learning model. • To increase inference speed and memory, we’ve proposed using depthwise separable convolutions. • A hourglass model is trained with normal video frames and then tested with anormal video frames.
  • 21. USE CASE An example hourglass network architecture
  • 22. RESULTS Naïve convolution – 537K parameters Average InferenceTime: 0.106s DS convolution – 93.8K parameters Average InferenceTime: 0.144s
  • 23. CONCLUSION • While replacing naïve convolutions with depthwise separable convolutions, 2 extra layers has been added.That’s why inference speed may have reduced even there are less parameters in DS Convolution. • Real-time anomaly detection with self-trained models are still active research field.
  • 24. FUTURE WORK • Student-Teacher Models • Feature-Based Knowledge Distillation • Response-Based Knowledge Distillation • Pseudo Labels • Confident Learning : Dataset Labels Improvement • Pseudo Labels combined with student-teacher models : Meta Pseudo Labels
  • 25. REFERENCES 1. Goodfellow, Ian J., Jonathon Shlens, and Christian Szegedy. "Explaining and harnessing adversarial examples." arXiv preprint arXiv:1412.6572 (2014). 2. Jacob, Benoit, et al. "Quantization and training of neural networks for efficient integer-arithmetic-only inference." Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2018. 3. Howard, Andrew G., et al. "Mobilenets: Efficient convolutional neural networks for mobile vision applications." arXiv preprint arXiv:1704.04861 (2017).