SlideShare a Scribd company logo
© 2018, Amazon Web Services, Inc. or Its Affiliates. All rights reserved.
Julien Simon, Principal AI/ML Evangelist, Amazon Web Services
@julsimon
Simone Mangiante, Research and Standards Specialist, Vodafone
Simone.Mangiante@Vodafone.com
Machine Learning Inference at the Edge
Agenda
• Deep Learning at the Edge?
• Apache MXNet
• Predicting in the Cloud or at the Edge?
• Case study: Driver Monitoring by Vodafone
• New services: AWS Greengrass ML and AWS DeepLens
• Resources
Deep Learning at the Edge?
Use Cases
Voice/sound
recognition
Collision
avoidance
Image
recognition
Anomaly
detection
More
!
Smart
Agriculture
Predictive
maintenance
Self-driving
cars
Video
surveillance
Robotics
Deep Learning challenges at the Edge
• Resource-constrained devices
• CPU, memory, storage, power consumption.
• Network connectivity
• Availability, cost, bandwidth, latency.
• On-device prediction may be the only option.
• Deployment
• Updating code and models on a fleet of devices
is not easy.
Deep Learning wishlist at the Edge
• Rely on cloud-based services for seamless training and deployment.
• Have the option to use cloud-based prediction.
• Be able to run device-based prediction with good performance.
• Support different technical environments (CPUs, languages).
Apache MXNet
Apache MXNet: Open Source library for Deep Learning
Programmable Portable High Performance
Near linear scaling
across hundreds of
GPUs
Highly efficient
models for
mobile
and IoT
Simple syntax,
multiple
languages
Most Open Best On AWS
Optimized for
Deep Learning on AWS
Accepted into the
Apache Incubator
Apache MXNet for IoT
1. Flexible experimentation in the Cloud.
2. Scalable training in the Cloud.
3. Good prediction performance at the Edge.
4. Prediction in the Cloud or at the Edge.
1 - Flexible experimentation in the Cloud
• API for Python, R, Perl, Matlab, Scala, C++.
• Gluon
• Imperative programming aka ‘define-by-run’.
• Inspect, debug and modify models during training.
• Extensive model zoo
• Pre-trained computer vision models.
• DenseNet, SqueezeNet for resource-constrained devices.
2 - Scalable training in the Cloud
Amazon SageMaker AWS Deep Learning AMI
Amazon EC2 c5 p3
3 - Good prediction performance at the Edge
• MXNet is written in C++.
• Gluon networks can be ‘hybridized’ for additional speed.
• Two libraries boost performance on CPU-only devices
• Fast implementation of math primitives
• Hardware-specific instructions, e.g. Intel AVX or ARM NEON
• Intel Math Kernel Library https://software.intel.com/en-us/mkl
• NNPACK https://github.com/Maratyszcza/NNPACK
• Mixed precision training
• Use float16 instead of float32 for weights and activations
• Almost 2x reduction in model size, no loss of accuracy, faster inference
• https://devblogs.nvidia.com/parallelforall/mixed-precision-training-deep-neural-networks/
4 - Predicting in the Cloud or at the Edge
• Cloud-based: invoke a Lambda function with AWS IoT.
• Cloud-based: invoke a SageMaker endpoint with HTTP.
• Device-based: bring your own code and model.
• Device-based: deploy your code and model with AWS Greengrass.
Invoking a Lambda function with AWS IoT
• Train a model in SageMaker (or bring your own).
• Host it in S3 (or embed it in a Lambda function).
• Write a Lambda function performing prediction.
• Invoke it through AWS IoT.
Best when
Devices can support neither HTTP nor local
inference (e.g. Arduino).
Costs must be kept as low as possible.
Requirements
Network is available and reliable
(MQTT is less demanding than HTTP).
Devices are provisioned in AWS IoT (certificate, keys).
https://aws.amazon.com/blogs/compute/seamlessly-scale-predictions-with-aws-lambda-and-mxnet/
Invoking a SageMaker endpoint with HTTP
• Train a model in SageMaker (or bring your own).
• Deploy it to a prediction endpoint.
• Invoke the HTTP endpoint from your devices.
Best when
Devices are not powerful enough for local inference.
Models can’t be easily deployed to devices.
Additional cloud-based data is required for prediction.
Prediction activity must be centralized.
Requirements
Network is available and reliable.
Devices support HTTP.
Bring your own code and model
• Train a model in SageMaker (or bring your own).
• Bring your own application code.
• Provision devices at manufacturing time
(or use your own update mechanism).
Best when
You don’t want to or can’t rely on cloud services
(no network connectivity?)
Requirements
Devices are powerful enough for local inference.
Models don’t need to be updated, if ever.
DIY!
Deploy your code and model with AWS Greengrass
• Train a model in SageMaker (or bring your own).
• Write a Lambda function performing prediction.
• Add both as resources in your Greengrass group.
• Let Greengrass handle deployment and updates.
Best when
You want the same programming model in the Cloud
and at the Edge.
Code and models need to be updated, even if
network connectivity is infrequent or unreliable.
One device in the group should be able to perform
prediction on behalf on other devices.
Requirements
Devices are powerful enough to run Greengrass
(XXX HW requirements)
Devices are provisioned in AWS IoT (certificate, keys).
Distributing ML at the edge
Dr Simone Mangiante
simone.mangiante@vodafone.com
27 March 2018
Distributed, low latency machine
learning
Increasing amount of in-vehicle
data (40+ TB/h)
Desire for more mobile-driven
services in cars harnessing
advanced network functions
Needs to be developer friendly
ML at the edge: challenges and benefits
Challenges from use cases, benefits from deployment at the telco network edge
Reduce in-vehicle bill-of-material by
offloading computing resources
Optimise service cost with shared resources
Seamlessly upgrade in a cloud-based
environment
Evolve at the pace of technology leveraging
the latest advances
Real-time alerts that enhance safety
What is Mobile/Multi-access Edge Computing?
MEC is a network architectural concept
that enables cloud-computing capabilities
at the edge of the network
MEC offers applications and content
providers cloud-computing capabilities at
the edge of the network
CloudMobile Edge CloudRadio Site
• MEC software (i.e. MEC Platform) runs as a
VNF on a cloud-based edge infrastructure
• 3rd party applications can be deployed on
MEC platforms
• MEC platform can expose network APIs to
applications
4G/5G Services Edge Cloud Infrastructure
VNF
Edge Cloud Services & API
Edge Application
MEC Platform
Virtual
Network
Functions
Edge DeploymentEdge Deployment
The rationale for edge computing goes beyond latency
Latency argument is sometimes ‘overhyped’
Mobile Core network
MEC driver monitoring proof of concept
Saguna vEdge
AWS Greengrass
Driver Monitoring App
Vodafone 4G/LTE
MEC Edge Cloud
Internet
IPsec tunnel IPsec tunnel
1
2
3
4 5
1 • Camera device
(Raspberry Pi)
• Connected via cellular
radio
• Video streamed to Driver
Monitoring App
• Filters traffic based on
traffic rules: e.g. pass-
through to Internet,
mirror, redirect to local
application
• Can ‘chain’
applications
2 • Camera device is in
Greengrass Group
• Runs in a VM
• It’s a ‘MEC application’
• Receives traffic from radio
access as per configured
traffic rules
3 4 • Edge app
• Business logic
• Includes neural
network
5 • Training
MQTT
messages
Notifications/results
Model updates
Local breakout
Driver monitoring application
Convolutional Neural Network
• Two architectures explored:
Inception, MobileNet
• MobileNet chosen for its speed
advantage and ability to run on a
light platform (as for a PoC)
Training data
• Source: Kaggle.com
• Dataset: State Farm Distracted
Driver Detection
Business logic
• Classifies a series of sequential video
frames
• Then makes decision on whether
“distracted driving” detected
• For PoC, sends message to Raspberry Pi
for local web server and displaying via
browser on monitor
Stream video
Receive alerts
In-Vehicle
AWS IoT SDK
MEC Edge Cloud
Receive video
Send alerts
Perform inference
using ML model
AWS Greengrass core
Create and train
the driver
monitoring model
AWS Cloud
AWS SageMaker
Developed by
The future is exciting.
Ready?
ML Inference using AWS Greengrass
Deploy model
and Lambda
Send
insightsPredict and take
actions locally
on device
AWS Cloud
PREVIEW
AVAILABLE
AWS Greengrass ML
PREVIEW
AVAILABLE
AWS DeepLens
A new way to learn
Custom built for Deep Learning
Broad Framework Support
Deploy models from Amazon SageMaker
Integrated with AWS
Fully programmable with AWS Lambda
AWS DeepLens
W o r l d ’ s f i r s t D e e p L e a r n i n g e n a b l e d v i d e o c a m e r a f o r d e v e l o p e r s
AWS DeepLens
Object detection with AWS DeepLens
Resources
Resources
https://mxnet.incubator.apache.org
http://gluon.mxnet.io
https://aws.amazon.com/sagemaker (free tier available)
An overview of Amazon SageMaker: https://www.youtube.com/watch?v=ym7NEYEx9x4
https://github.com/awslabs/amazon-sagemaker-examples
https://aws.amazon.com/greengrass (free tier available)
https://aws.amazon.com/deeplens
https://medium.com/@julsimon
© 2018, Amazon Web Services, Inc. or Its Affiliates. All rights reserved.
Thank you!
Julien Simon, Principal AI/ML Evangelist, Amazon Web Services
@julsimon
Simone Mangiante, Research and Standards Specialist, Vodafone
Simone.Mangiante@Vodafone.com

More Related Content

PPTX
Introduction-to-LLM-Developers-Guide.pptx
PDF
Evaluating LLM Models for Production Systems Methods and Practices -
PPTX
Chatbot and Virtual AI Assistant Implementation in Natural Language Processing
PDF
Use Case Patterns for LLM Applications (1).pdf
PPTX
AI-Plugins-Planners-Persona-SemanticKernel.pptx
PDF
PDF
Implications of GPT-3
PPT
Chat bots and AI
Introduction-to-LLM-Developers-Guide.pptx
Evaluating LLM Models for Production Systems Methods and Practices -
Chatbot and Virtual AI Assistant Implementation in Natural Language Processing
Use Case Patterns for LLM Applications (1).pdf
AI-Plugins-Planners-Persona-SemanticKernel.pptx
Implications of GPT-3
Chat bots and AI

What's hot (20)

DOCX
Advantages and disadvantages of machine learning language
PDF
How to build a generative AI solution From prototyping to production.pdf
PDF
Generative AI Roadmap By ScholarHat PDF.pdf
PPTX
A Beginner's Guide to Large Language Models
PPTX
Generative AI in CSharp with Semantic Kernel.pptx
PPTX
OpenCL Heterogeneous Parallel Computing
PPTX
CHATGPT VS BARD AI
PDF
Whitepaper Robotic Process Automation
PDF
From gaming to the metaverse
PPTX
GenerativeAI and Automation - IEEE ACSOS 2023.pptx
PDF
Generative AI
PPTX
Open ai openpower
PDF
[SP02] Developing autonomous vehicles with AirSim
PDF
Build and Modernize Intelligent Apps​
PDF
Using Large Language Models in 10 Lines of Code
PPT
M05 Metamodel
PPTX
Cassie Kozyrkov. Journey to AI
PDF
AI and ML Series - Introduction to Generative AI and LLMs - Session 1
PDF
ChatGPT for Academic
PDF
ML DL AI DS BD - An Introduction
Advantages and disadvantages of machine learning language
How to build a generative AI solution From prototyping to production.pdf
Generative AI Roadmap By ScholarHat PDF.pdf
A Beginner's Guide to Large Language Models
Generative AI in CSharp with Semantic Kernel.pptx
OpenCL Heterogeneous Parallel Computing
CHATGPT VS BARD AI
Whitepaper Robotic Process Automation
From gaming to the metaverse
GenerativeAI and Automation - IEEE ACSOS 2023.pptx
Generative AI
Open ai openpower
[SP02] Developing autonomous vehicles with AirSim
Build and Modernize Intelligent Apps​
Using Large Language Models in 10 Lines of Code
M05 Metamodel
Cassie Kozyrkov. Journey to AI
AI and ML Series - Introduction to Generative AI and LLMs - Session 1
ChatGPT for Academic
ML DL AI DS BD - An Introduction
Ad

Similar to Machine Learning Inference at the Edge (20)

PPTX
Machine Learning inference at the Edge
PDF
Deep Learning at the Edge
PPTX
AWS re:Invent 2018 - AIM302 - Machine Learning at the Edge
PDF
ML at the Edge: Building Your Production Pipeline with Apache Spark and Tens...
PPTX
ML At the Edge: Building Your Production Pipeline With Apache Spark and Tens...
PPTX
Machine Learning Meets Communication Networks.pptx
PPTX
AWS re:Invent 2018 - Machine Learning recap (December 2018)
PDF
엔터프라이즈를 위한 머신러닝 그리고 AWS (김일호 솔루션즈 아키텍트, AWS) :: AWS Techforum 2018
PDF
Get started with Machine Learning and Computer Vision Using AWS DeepLens (Feb...
PPTX
Deep Learning with TensorFlow and Apache MXNet on Amazon SageMaker (March 2019)
PDF
AIM410R Deep Learning Applications with TensorFlow, featuring Mobileye (Decem...
PPTX
Optimize your machine learning workloads on AWS (March 2019)
PDF
AI & Machine Learning at AWS - An Introduction
PPTX
Aran Khanna, Software Engineer, Amazon Web Services at MLconf ATL 2017
PDF
Deep Learning with Tensorflow and Apache MXNet on AWS (April 2019)
PDF
Artificial Intelligence (Machine Learning) on AWS: How to Start
PDF
Maschinelles Lernen auf AWS für Entwickler, Data Scientists und Experten
PDF
Transversal Delivery Pipeline by Mike Nescot and Nick Grace
PDF
Transversal Delivery Pipeline by Mike Nescot and Nick Grace
PDF
Tiny intelligent computers and sensors - Open Hardware Event 2020
Machine Learning inference at the Edge
Deep Learning at the Edge
AWS re:Invent 2018 - AIM302 - Machine Learning at the Edge
ML at the Edge: Building Your Production Pipeline with Apache Spark and Tens...
ML At the Edge: Building Your Production Pipeline With Apache Spark and Tens...
Machine Learning Meets Communication Networks.pptx
AWS re:Invent 2018 - Machine Learning recap (December 2018)
엔터프라이즈를 위한 머신러닝 그리고 AWS (김일호 솔루션즈 아키텍트, AWS) :: AWS Techforum 2018
Get started with Machine Learning and Computer Vision Using AWS DeepLens (Feb...
Deep Learning with TensorFlow and Apache MXNet on Amazon SageMaker (March 2019)
AIM410R Deep Learning Applications with TensorFlow, featuring Mobileye (Decem...
Optimize your machine learning workloads on AWS (March 2019)
AI & Machine Learning at AWS - An Introduction
Aran Khanna, Software Engineer, Amazon Web Services at MLconf ATL 2017
Deep Learning with Tensorflow and Apache MXNet on AWS (April 2019)
Artificial Intelligence (Machine Learning) on AWS: How to Start
Maschinelles Lernen auf AWS für Entwickler, Data Scientists und Experten
Transversal Delivery Pipeline by Mike Nescot and Nick Grace
Transversal Delivery Pipeline by Mike Nescot and Nick Grace
Tiny intelligent computers and sensors - Open Hardware Event 2020
Ad

More from Julien SIMON (20)

PDF
Implementing high-quality and cost-effiient AI applications with small langua...
PDF
Optimiser vos workloads AI/ML sur Amazon EC2 et AWS Graviton
PDF
Trying to figure out MCP by actually building an app from scratch with open s...
PDF
Arcee AI - building and working with small language models (06/25)
PDF
deep_dive_multihead_latent_attention.pdf
PDF
Deep Dive: Model Distillation with DistillKit
PDF
Deep Dive: Parameter-Efficient Model Adaptation with LoRA and Spectrum
PDF
Building High-Quality Domain-Specific Models with Mergekit
PDF
Tailoring Small Language Models for Enterprise Use Cases
PDF
Tailoring Small Language Models for Enterprise Use Cases
PDF
Julien Simon - Deep Dive: Compiling Deep Learning Models
PDF
Tailoring Small Language Models for Enterprise Use Cases
PDF
Julien Simon - Deep Dive - Optimizing LLM Inference
PDF
Julien Simon - Deep Dive - Accelerating Models with Better Attention Layers
PDF
Julien Simon - Deep Dive - Quantizing LLMs
PDF
Julien Simon - Deep Dive - Model Merging
PDF
An introduction to computer vision with Hugging Face
PDF
Reinventing Deep Learning
 with Hugging Face Transformers
PDF
Building NLP applications with Transformers
PPTX
Building Machine Learning Models Automatically (June 2020)
Implementing high-quality and cost-effiient AI applications with small langua...
Optimiser vos workloads AI/ML sur Amazon EC2 et AWS Graviton
Trying to figure out MCP by actually building an app from scratch with open s...
Arcee AI - building and working with small language models (06/25)
deep_dive_multihead_latent_attention.pdf
Deep Dive: Model Distillation with DistillKit
Deep Dive: Parameter-Efficient Model Adaptation with LoRA and Spectrum
Building High-Quality Domain-Specific Models with Mergekit
Tailoring Small Language Models for Enterprise Use Cases
Tailoring Small Language Models for Enterprise Use Cases
Julien Simon - Deep Dive: Compiling Deep Learning Models
Tailoring Small Language Models for Enterprise Use Cases
Julien Simon - Deep Dive - Optimizing LLM Inference
Julien Simon - Deep Dive - Accelerating Models with Better Attention Layers
Julien Simon - Deep Dive - Quantizing LLMs
Julien Simon - Deep Dive - Model Merging
An introduction to computer vision with Hugging Face
Reinventing Deep Learning
 with Hugging Face Transformers
Building NLP applications with Transformers
Building Machine Learning Models Automatically (June 2020)

Recently uploaded (20)

PDF
Mobile App Security Testing_ A Comprehensive Guide.pdf
PDF
Per capita expenditure prediction using model stacking based on satellite ima...
PDF
Architecting across the Boundaries of two Complex Domains - Healthcare & Tech...
PPTX
sap open course for s4hana steps from ECC to s4
PPTX
Big Data Technologies - Introduction.pptx
PDF
Encapsulation_ Review paper, used for researhc scholars
PDF
Chapter 3 Spatial Domain Image Processing.pdf
PPTX
Spectroscopy.pptx food analysis technology
PDF
TokAI - TikTok AI Agent : The First AI Application That Analyzes 10,000+ Vira...
PDF
MIND Revenue Release Quarter 2 2025 Press Release
PPT
Teaching material agriculture food technology
PDF
Build a system with the filesystem maintained by OSTree @ COSCUP 2025
PPTX
Machine Learning_overview_presentation.pptx
PDF
Review of recent advances in non-invasive hemoglobin estimation
PDF
Agricultural_Statistics_at_a_Glance_2022_0.pdf
DOCX
The AUB Centre for AI in Media Proposal.docx
PDF
Approach and Philosophy of On baking technology
PDF
Unlocking AI with Model Context Protocol (MCP)
PDF
gpt5_lecture_notes_comprehensive_20250812015547.pdf
PPTX
Programs and apps: productivity, graphics, security and other tools
Mobile App Security Testing_ A Comprehensive Guide.pdf
Per capita expenditure prediction using model stacking based on satellite ima...
Architecting across the Boundaries of two Complex Domains - Healthcare & Tech...
sap open course for s4hana steps from ECC to s4
Big Data Technologies - Introduction.pptx
Encapsulation_ Review paper, used for researhc scholars
Chapter 3 Spatial Domain Image Processing.pdf
Spectroscopy.pptx food analysis technology
TokAI - TikTok AI Agent : The First AI Application That Analyzes 10,000+ Vira...
MIND Revenue Release Quarter 2 2025 Press Release
Teaching material agriculture food technology
Build a system with the filesystem maintained by OSTree @ COSCUP 2025
Machine Learning_overview_presentation.pptx
Review of recent advances in non-invasive hemoglobin estimation
Agricultural_Statistics_at_a_Glance_2022_0.pdf
The AUB Centre for AI in Media Proposal.docx
Approach and Philosophy of On baking technology
Unlocking AI with Model Context Protocol (MCP)
gpt5_lecture_notes_comprehensive_20250812015547.pdf
Programs and apps: productivity, graphics, security and other tools

Machine Learning Inference at the Edge

  • 1. © 2018, Amazon Web Services, Inc. or Its Affiliates. All rights reserved. Julien Simon, Principal AI/ML Evangelist, Amazon Web Services @julsimon Simone Mangiante, Research and Standards Specialist, Vodafone Simone.Mangiante@Vodafone.com Machine Learning Inference at the Edge
  • 2. Agenda • Deep Learning at the Edge? • Apache MXNet • Predicting in the Cloud or at the Edge? • Case study: Driver Monitoring by Vodafone • New services: AWS Greengrass ML and AWS DeepLens • Resources
  • 3. Deep Learning at the Edge?
  • 5. Deep Learning challenges at the Edge • Resource-constrained devices • CPU, memory, storage, power consumption. • Network connectivity • Availability, cost, bandwidth, latency. • On-device prediction may be the only option. • Deployment • Updating code and models on a fleet of devices is not easy.
  • 6. Deep Learning wishlist at the Edge • Rely on cloud-based services for seamless training and deployment. • Have the option to use cloud-based prediction. • Be able to run device-based prediction with good performance. • Support different technical environments (CPUs, languages).
  • 8. Apache MXNet: Open Source library for Deep Learning Programmable Portable High Performance Near linear scaling across hundreds of GPUs Highly efficient models for mobile and IoT Simple syntax, multiple languages Most Open Best On AWS Optimized for Deep Learning on AWS Accepted into the Apache Incubator
  • 9. Apache MXNet for IoT 1. Flexible experimentation in the Cloud. 2. Scalable training in the Cloud. 3. Good prediction performance at the Edge. 4. Prediction in the Cloud or at the Edge.
  • 10. 1 - Flexible experimentation in the Cloud • API for Python, R, Perl, Matlab, Scala, C++. • Gluon • Imperative programming aka ‘define-by-run’. • Inspect, debug and modify models during training. • Extensive model zoo • Pre-trained computer vision models. • DenseNet, SqueezeNet for resource-constrained devices.
  • 11. 2 - Scalable training in the Cloud Amazon SageMaker AWS Deep Learning AMI Amazon EC2 c5 p3
  • 12. 3 - Good prediction performance at the Edge • MXNet is written in C++. • Gluon networks can be ‘hybridized’ for additional speed. • Two libraries boost performance on CPU-only devices • Fast implementation of math primitives • Hardware-specific instructions, e.g. Intel AVX or ARM NEON • Intel Math Kernel Library https://software.intel.com/en-us/mkl • NNPACK https://github.com/Maratyszcza/NNPACK • Mixed precision training • Use float16 instead of float32 for weights and activations • Almost 2x reduction in model size, no loss of accuracy, faster inference • https://devblogs.nvidia.com/parallelforall/mixed-precision-training-deep-neural-networks/
  • 13. 4 - Predicting in the Cloud or at the Edge • Cloud-based: invoke a Lambda function with AWS IoT. • Cloud-based: invoke a SageMaker endpoint with HTTP. • Device-based: bring your own code and model. • Device-based: deploy your code and model with AWS Greengrass.
  • 14. Invoking a Lambda function with AWS IoT • Train a model in SageMaker (or bring your own). • Host it in S3 (or embed it in a Lambda function). • Write a Lambda function performing prediction. • Invoke it through AWS IoT. Best when Devices can support neither HTTP nor local inference (e.g. Arduino). Costs must be kept as low as possible. Requirements Network is available and reliable (MQTT is less demanding than HTTP). Devices are provisioned in AWS IoT (certificate, keys). https://aws.amazon.com/blogs/compute/seamlessly-scale-predictions-with-aws-lambda-and-mxnet/
  • 15. Invoking a SageMaker endpoint with HTTP • Train a model in SageMaker (or bring your own). • Deploy it to a prediction endpoint. • Invoke the HTTP endpoint from your devices. Best when Devices are not powerful enough for local inference. Models can’t be easily deployed to devices. Additional cloud-based data is required for prediction. Prediction activity must be centralized. Requirements Network is available and reliable. Devices support HTTP.
  • 16. Bring your own code and model • Train a model in SageMaker (or bring your own). • Bring your own application code. • Provision devices at manufacturing time (or use your own update mechanism). Best when You don’t want to or can’t rely on cloud services (no network connectivity?) Requirements Devices are powerful enough for local inference. Models don’t need to be updated, if ever. DIY!
  • 17. Deploy your code and model with AWS Greengrass • Train a model in SageMaker (or bring your own). • Write a Lambda function performing prediction. • Add both as resources in your Greengrass group. • Let Greengrass handle deployment and updates. Best when You want the same programming model in the Cloud and at the Edge. Code and models need to be updated, even if network connectivity is infrequent or unreliable. One device in the group should be able to perform prediction on behalf on other devices. Requirements Devices are powerful enough to run Greengrass (XXX HW requirements) Devices are provisioned in AWS IoT (certificate, keys).
  • 18. Distributing ML at the edge Dr Simone Mangiante simone.mangiante@vodafone.com 27 March 2018
  • 19. Distributed, low latency machine learning Increasing amount of in-vehicle data (40+ TB/h) Desire for more mobile-driven services in cars harnessing advanced network functions Needs to be developer friendly ML at the edge: challenges and benefits Challenges from use cases, benefits from deployment at the telco network edge Reduce in-vehicle bill-of-material by offloading computing resources Optimise service cost with shared resources Seamlessly upgrade in a cloud-based environment Evolve at the pace of technology leveraging the latest advances Real-time alerts that enhance safety
  • 20. What is Mobile/Multi-access Edge Computing? MEC is a network architectural concept that enables cloud-computing capabilities at the edge of the network MEC offers applications and content providers cloud-computing capabilities at the edge of the network CloudMobile Edge CloudRadio Site • MEC software (i.e. MEC Platform) runs as a VNF on a cloud-based edge infrastructure • 3rd party applications can be deployed on MEC platforms • MEC platform can expose network APIs to applications 4G/5G Services Edge Cloud Infrastructure VNF Edge Cloud Services & API Edge Application MEC Platform Virtual Network Functions Edge DeploymentEdge Deployment
  • 21. The rationale for edge computing goes beyond latency Latency argument is sometimes ‘overhyped’
  • 22. Mobile Core network MEC driver monitoring proof of concept Saguna vEdge AWS Greengrass Driver Monitoring App Vodafone 4G/LTE MEC Edge Cloud Internet IPsec tunnel IPsec tunnel 1 2 3 4 5 1 • Camera device (Raspberry Pi) • Connected via cellular radio • Video streamed to Driver Monitoring App • Filters traffic based on traffic rules: e.g. pass- through to Internet, mirror, redirect to local application • Can ‘chain’ applications 2 • Camera device is in Greengrass Group • Runs in a VM • It’s a ‘MEC application’ • Receives traffic from radio access as per configured traffic rules 3 4 • Edge app • Business logic • Includes neural network 5 • Training MQTT messages Notifications/results Model updates Local breakout
  • 23. Driver monitoring application Convolutional Neural Network • Two architectures explored: Inception, MobileNet • MobileNet chosen for its speed advantage and ability to run on a light platform (as for a PoC) Training data • Source: Kaggle.com • Dataset: State Farm Distracted Driver Detection Business logic • Classifies a series of sequential video frames • Then makes decision on whether “distracted driving” detected • For PoC, sends message to Raspberry Pi for local web server and displaying via browser on monitor Stream video Receive alerts In-Vehicle AWS IoT SDK MEC Edge Cloud Receive video Send alerts Perform inference using ML model AWS Greengrass core Create and train the driver monitoring model AWS Cloud AWS SageMaker Developed by
  • 24. The future is exciting. Ready?
  • 25. ML Inference using AWS Greengrass Deploy model and Lambda Send insightsPredict and take actions locally on device AWS Cloud PREVIEW AVAILABLE
  • 28. A new way to learn Custom built for Deep Learning Broad Framework Support Deploy models from Amazon SageMaker Integrated with AWS Fully programmable with AWS Lambda AWS DeepLens W o r l d ’ s f i r s t D e e p L e a r n i n g e n a b l e d v i d e o c a m e r a f o r d e v e l o p e r s
  • 30. Object detection with AWS DeepLens
  • 32. Resources https://mxnet.incubator.apache.org http://gluon.mxnet.io https://aws.amazon.com/sagemaker (free tier available) An overview of Amazon SageMaker: https://www.youtube.com/watch?v=ym7NEYEx9x4 https://github.com/awslabs/amazon-sagemaker-examples https://aws.amazon.com/greengrass (free tier available) https://aws.amazon.com/deeplens https://medium.com/@julsimon
  • 33. © 2018, Amazon Web Services, Inc. or Its Affiliates. All rights reserved. Thank you! Julien Simon, Principal AI/ML Evangelist, Amazon Web Services @julsimon Simone Mangiante, Research and Standards Specialist, Vodafone Simone.Mangiante@Vodafone.com