SlideShare a Scribd company logo
Sensor-Aware Augmented Reality
Addressing Real-World HMI Challenges
Dr. Liu Ren
Global Head and Chief Scientist, HMI
Bosch Research North America
Palo Alto, CA
Sensor-Aware Augmented Reality
Research and Technology Center North America | Liu Ren | 12/20/2016
© 2016 Robert Bosch LLC and affiliates. All rights reserved.
2
Bosch Overview
Mobility Solutions Industrial Technology
Energy and Building
Technology
Consumer Goods
Bosch is one of the world’s leading international providers of technology and services
• 375,0001 Bosch associates
• More than 4401 subsidiary companies and regional subsidiaries in some 601 countries
• Including its sales and service partners, Bosch is represented in some 1501 countries.
1 As of Dec. 2015
Home Appliance
Personal
Assistant
Home Robots
Garden Tools Smart Home
Internet of Things (IoT)
Thermothenology
Security
Systems
Smart Cities
Assembly Technology
Industry 4.0
Packaging Technology
Industrial
Robots
Car Infotainment
Concept Car
Autonomous
Driving
Automotive Aftermarket
Sensor-Aware Augmented Reality
Research and Technology Center North America | Liu Ren | 12/20/2016
© 2016 Robert Bosch LLC and affiliates. All rights reserved.
3
Industry 4.0Smart HomeRoboticsAftermarket Repair Shops
Car Infotainment
Highly Automated Driving
Human Machine Interaction (HMI) Research in Bosch
Industry 4.0Smart HomeRoboticsAftermarket Repair Shops
Car Infotainment
Highly Automated Driving
Sensor-Aware Augmented Reality
Research and Technology Center North America | Liu Ren | 12/20/2016
© 2016 Robert Bosch LLC and affiliates. All rights reserved.
4
Key Success Factors of HMI Products
Intuitive Interactive Intelligent
Human Machine Interaction
• Global Head and Chief
Scientist, HMI, Bosch
Research
• Ph.D. and M.Sc. in
Computer Science,
Carnegie Mellon
University
• B.Sc. in Computer
Science, Zhejiang
University, P.R. China
Sensor-Aware Augmented Reality
Research and Technology Center North America | Liu Ren | 12/20/2016
© 2016 Robert Bosch LLC and affiliates. All rights reserved.
5
Global HMI Research Team
Renningen, Germany
Shanghai, China
Headquarters
Palo Alto, USA
Liu Ren Short Bio HMI teams in Bosch Research
Sensor-Aware Augmented Reality
Research and Technology Center North America | Liu Ren | 12/20/2016
© 2016 Robert Bosch LLC and affiliates. All rights reserved.
6
Real-World HMI Challenges for (Wearable) AR
Hardware
Form Factor
SoftwareField-of-View
Comfort
Battery Life
Context-Aware
Visualization
Scalable & Easy
Content Generation
Natural Interaction
(Speech, Gesture, etc.)
AI
(Perception,
Understanding, etc.)
Sensor-Aware Augmented Reality
Research and Technology Center North America | Liu Ren | 12/20/2016
© 2016 Robert Bosch LLC and affiliates. All rights reserved.
7
Bosch Product: CAP (Common Augmented Reality Platform)
Software
Context-Aware
Visualization
Scalable & Easy
Content Generation
Natural Interaction
Bosch CAP enables implementation of complete enterprises AR solutions
• Integrates the production of visual and digital content directly into the authoring process.
• Existing CAD, image and video data were used and save the expense of creating new content.
Bosch CAP
Production and
manufacturing
Target/actual comparison
and collision planning
Plant and system
planning
Education and
training Maintenance, service
and repair
Marketing, trade shows
and distribution
Technical doc. and digital
operating instructions
Sensor-Aware Augmented Reality
Research and Technology Center North America | Liu Ren | 12/20/2016
© 2016 Robert Bosch LLC and affiliates. All rights reserved.
8
Our Sensor-Aware Solutions
Software
Context-Aware
Visualization
Scalable & Easy
Content Generation
Natural Interaction
(Speech, Gesture, etc.)
Dynamic occlusion handling
ISMAR 2016[1]
• Enhance realistic depth
perception
Robust visual tracking
ISMAR 2016[2]
• Improve tracking robustness
and accuracy
[1] Chao Du, Yen-Lin Chen, Mao Ye, and Liu Ren, “Edge Snapping-Based Depth Enhancement for Dynamic Occlusion Handling in
Augmented Reality”, IEEE International Symposium on Mixed and Augmented Reality (ISMAR) 2016.
[2] Benzun Wisely Babu, Soohwan Kim, Zhixin Yan, and Liu Ren, “σ-DVO: Sensor Noise Model Meets Dense Visual Odometry”, IEEE
International Symposium on Mixed and Augmented Reality (ISMAR) 2016.
Sensor-Aware Augmented Reality
Research and Technology Center North America | Liu Ren | 12/20/2016
© 2016 Robert Bosch LLC and affiliates. All rights reserved.
9
Dynamic Occlusion Handling
Software
Context-Aware
Visualization
Scalable & Easy
Content Generation
Natural Interaction
(Speech, Gesture, etc.)
Robust visual tracking
ISMAR 2016[2]
• Improve tracking robustness
and accuracy
Dynamic occlusion handling
ISMAR 2016[1]
• Enhance realistic depth
perception
1. Compact setup: single sensor
2. Dynamic occlusion handling
Sensor-Aware Augmented Reality
Research and Technology Center North America | Liu Ren | 12/20/2016
© 2016 Robert Bosch LLC and affiliates. All rights reserved.
10
Dynamic Occlusion Handling: Motivation
One near-range RGBD sensor
Optical see-through head-
mounted display (HMD)
Challenges
• Performance requirements for real-time AR applications
• Limited computational resources (e.g., on tablet)
Goals
Sensor-Aware Augmented Reality
Research and Technology Center North America | Liu Ren | 12/20/2016
© 2016 Robert Bosch LLC and affiliates. All rights reserved.
11
Dynamic Occlusion Handling: Our Sensor-Aware Solution
Target Object Boundary
Boundary from Depth
Align Object Boundary (Edge-Snapping)1
Depth
Color
Enhance Depth Map2
Raw Depth Map Enhanced Depth Map
• Use color images as guidance
• Snap object boundaries in depth data towards the
edges in color images
• Formulated as an optimization problem, efficiently
solved via dynamic programming
Edge Snapping-Based Algorithm
Depth data not reliable at the object boundary
• Structured Light/Stereo: matching is not accurate
at the boundary
• Time of Flight: light signal reaching object
boundary barely bounce back to the sensor
Knowledge on RGBD Sensor
Sensor-Aware Augmented Reality
Research and Technology Center North America | Liu Ren | 12/20/2016
© 2016 Robert Bosch LLC and affiliates. All rights reserved.
12
Dynamic Occlusion Handling: Experimental Results
Sensor-Aware Augmented Reality
Research and Technology Center North America | Liu Ren | 12/20/2016
© 2016 Robert Bosch LLC and affiliates. All rights reserved.
13
Robust Visual Tracking
Software
Context-Aware
Visualization
Scalable & Easy
Content Generation
Natural Interaction
(Speech, Gesture, etc.)
Dynamic occlusion handling
ISMAR 2016[1]
• Enhance realistic depth
perception
Robust visual tracking
ISMAR 2016[2]
• Improve tracking robustness
and accuracy
Sensor-Aware Augmented Reality
Research and Technology Center North America | Liu Ren | 12/20/2016
© 2016 Robert Bosch LLC and affiliates. All rights reserved.
14
Robust Visual Tracking: Motivation
1. Visual tracking is an essential AR
component
• 6 DoF camera pose
• Correctly place virtual objects in real world
2. Markerless Visual Tracking:
• Visual SLAM (simultaneous localization
and mapping)
Background
Challenges
Textureless Blurry image Lighting condition change
Sensor-Aware Augmented Reality
Research and Technology Center North America | Liu Ren | 12/20/2016
© 2016 Robert Bosch LLC and affiliates. All rights reserved.
15
Robust Visual Tracking: Our Sensor-Aware Solution (σ-DVO)
• Working well with textureless environments
• Less sensitive to lighting condition changes
• Noise of depth measurement grows quadratically as depth
increases
• Estimate the relative pose between two given frames based on
residuals (front-end of visual SLAM)
• Utilize all pixels from RGBD images
RGBD Dense Visual OdometryKnowledge on RGBD Sensor
Previous Frame
− =
Current Frame
Relative Camera Pose
Warped Current Frame Residuals
Color Residual Depth Residual
Non-linear Optimization
Previous Frame Current Frame
Relative Camera Pose
x
Color Weights Depth Weights
Weights
Weight decreases as
noise of depth
measurement grows
Sensor-Aware
Weighting
• Incorporate sensor noise model to guide pose optimization
Previous Frame
− =
Current Frame
Warped Current Frame
Non-linear Optimization
Relative Camera Pose
Residuals
Color Residual Depth Residual
Sensor-Aware Augmented Reality
Research and Technology Center North America | Liu Ren | 12/20/2016
© 2016 Robert Bosch LLC and affiliates. All rights reserved.
16
Robust Visual Tracking: Our Sensor-Aware Solution (σ-DVO)
• For better robustness and accuracy, we formulated the optimization problem in a
Bayesian framework
Bayesian Framework
𝑝 𝑝𝑜𝑠𝑒|𝑟𝑒𝑠𝑖𝑑𝑢𝑎𝑙𝑠 ∝ 𝒑 𝒓𝒆𝒔𝒊𝒅𝒖𝒂𝒍𝒔|𝒑𝒐𝒔𝒆 ∙ 𝑝 𝑝𝑜𝑠𝑒
• Assume uniform
distribution of residuals
 All the pixels share the
same weight
Early approaches
• Find an empirical distribution via
experiments
 Weights only depends on residuals
The state-of-the-art approach (DVO[1])
• Explore the source of residuals, especially based on sensor
characteristics
• Develop a sensor noise model to generate distribution of
residuals
 Decrease weights of pixels with either noisy sensor
measurement or high residuals
• Easily incorporate sensor-specific noise model for different
sensors to customize pose optimization for best performance
Our sensor-aware approach (σ-DVO)
Sensor Measurement Noise
One near-range
RGBD sensor
x
Color Weights Depth Weights
Weights
[1] Christian Kerl, Jürgen Sturm, and Daniel Cremers. "Dense visual SLAM for RGB-D cameras." 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems. IEEE, 2013.
Visual SLAM *
Sensor-Aware Augmented Reality
Research and Technology Center North America | Liu Ren | 12/20/2016
© 2016 Robert Bosch LLC and affiliates. All rights reserved.
17
Robust Visual Tracking: Experimental Results
Dataset
RGB-D
SLAM[2]
MRSM
ap[3]
Kintinu
ous[4]
ElasticFu
sion[5]
DVO
SLAM[1]
Our SLAM
approach
r1/desk 0.023 0.043 0.037 0.020 0.021 0.019
fr2/xyz 0.008 0.020 0.029 0.011 0.018 0.018
fr3/office 0.032 0.042 0.030 0.017 0.035 0.015
fr1/360 0.079 0.069 - - 0.083 0.061
Dataset
DVO[1] Our approach (σ-DVO )
ATE RPE ATE RPE
fr1/360 0.415 0.153 0.229 0.110
fr1/desk 0.109 0.048 0.067 0.039
fr1/desk2 0.261 0.074 0.088 0.065
fr1/floor 0.242 0.070 0.226 0.053
fr1/room 0.459 0.092 0.314 0.063
fr1/rpy 0.216 0.065 0.072 0.046
fr1/xyz 0.102 0.05 0.052 0.036
fr2/desk 0.561 0.038 0.184 0.016
fr2/large 4.370 0.240 0.724 0.134
fr2/rpy 0.501 0.039 0.188 0.012
fr2/xyz 0.497 0.030 0.188 0.010
fr3/office 0.485 0.044 0.164 0.014
average 0.684 0.067 0.208 0.050
Absolute Tracking Error [m]
ATE: Absolute Tracking Error [m], RPE: Relative Pose Error [m/s]
Visual Odometry
• Our σ-DVO outperforms DVO significantly in all the datasets. On average, 70%
reduction in ATE and 25% reduction in RPE
• σ-DVO SLAM outperforms the state-of-the-art SLAM algorithms
in most of the RGB-D datasets. On average 25% reduction in ATE
* σ-DVO is extended to σ-DVO SLAM by combining the front end (visual odometry)
with the backend (pose-graph optimization)
[1] Kerl, et al. "Dense visual SLAM for RGB-D cameras." 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems.
[2] Endres, et al. "An evaluation of the RGB-D SLAM system." Robotics and Automation (ICRA), 2012.
[3] Stückler, et al. “Model Learning and Real-Time Tracking using Multi-Resolution Surfel Maps”. AAAI, 2012.
[4] Whelan, et al. "Kintinuous: Spatially extended kinectfusion." Proc. Workshop RGB-D, Adv. Reason. Depth Cameras, 2012.
[5] Whelan, et al. "ElasticFusion: Dense SLAM without a pose graph." Proc. Robotics: Science and Systems, 2015.
Sensor-Aware Augmented Reality
Research and Technology Center North America | Liu Ren | 12/20/2016
© 2016 Robert Bosch LLC and affiliates. All rights reserved.
18
Robust Visual Tracking: Experimental Results
Sensor-Aware Augmented Reality
Research and Technology Center North America | Liu Ren | 12/20/2016
© 2016 Robert Bosch LLC and affiliates. All rights reserved.
19
Deep Learning for Augmented Reality?
Software
Context-Aware
Visualization
Scalable & Easy
Content Generation
Natural Interaction
(Speech, Gesture, etc.)
?
• Require semantic understanding of the
environments & context
• Modern AI technologies, e.g., Deep
Learning, could be effective approaches
Sensor-Aware Augmented Reality
Research and Technology Center North America | Liu Ren | 12/20/2016
© 2016 Robert Bosch LLC and affiliates. All rights reserved.
20
Summary and Outlook
1 The three “I”s (Intuitive, Interactive, Intelligent) are key success factors of
Human Machine Interaction (HMI) solutions.
2 Sensor-aware approaches that leverage sensor knowledge and machine
learning are effective to address real-world HMI challenges.
3 Using the right AI technology to address the right problem. Deep Learning,
could be effective for core AR solutions.
We’re looking for good research scientists and interns!
(Send CV to rbvisualcomp@gmail.com)
Thank You

More Related Content

PDF
Graphcore presenting at Project Juno Machine Intelligence Showcase
PDF
Roland Memisevic at AI Frontiers: Common sense video understanding at TwentyBN
PDF
May 2016 Embedded Vision Summit Introductory Presentation (Day 2)
PPTX
biggest technology trends
PDF
"Computer Vision 2.0: Where We Are and Where We're Going," a Presentation fro...
PDF
AI gold rush, tool vendors and the next big thing
PDF
Nvidia needsfor atonomousvehicles
PDF
Why Microservices Are The New Innovation Enablers For Enterprises
Graphcore presenting at Project Juno Machine Intelligence Showcase
Roland Memisevic at AI Frontiers: Common sense video understanding at TwentyBN
May 2016 Embedded Vision Summit Introductory Presentation (Day 2)
biggest technology trends
"Computer Vision 2.0: Where We Are and Where We're Going," a Presentation fro...
AI gold rush, tool vendors and the next big thing
Nvidia needsfor atonomousvehicles
Why Microservices Are The New Innovation Enablers For Enterprises

What's hot (20)

PDF
Silicom Ventures Talk Aug 2013 - GPUs and Parallel Programming create new opp...
PDF
CES 2012 Jen-Hsun Huang Press Conference
PDF
May 2021 Embedded Vision Summit Opening Remarks (May 28)
PDF
The IBM Software Story
PPTX
H2O.ai - Road Ahead - keynote presentation by Sri Ambati
PDF
AWE Tel Aviv Startup Pitch: Mark Goldfarb with Sixdof Space
PDF
Junli Gu at AI Frontiers: Autonomous Driving Revolution
PDF
Preparing your team for a new XR platform; 7 key take-aways
PDF
Maximizing the Business Value of Connected Devices by Transforming the CIO's ...
PPTX
Artificial Intelligence in Small Embedded System
PDF
Scott Montgomerie (Scope AR): AR’s Influence on the Workforce of Tomorrow: Jo...
DOCX
GT C Tour 2018 Highlights
PDF
"A Practitioner’s Guide to Commercializing Applications of Computer Vision," ...
PDF
Find Your Passion and Make a Difference in Your Career
PPTX
Technology and AI sharing - From 2016 to Y2017 and Beyond
PDF
T Bytes IoT & AR
PPTX
ImageCon CTO keynote
PDF
Qualcomm: How to take advantage of XR over 5G in 2019: Understanding XR Viewers
PDF
Highlights from NVIDIA at CES 2014
PDF
Machine Learning for Self-Driving Cars
Silicom Ventures Talk Aug 2013 - GPUs and Parallel Programming create new opp...
CES 2012 Jen-Hsun Huang Press Conference
May 2021 Embedded Vision Summit Opening Remarks (May 28)
The IBM Software Story
H2O.ai - Road Ahead - keynote presentation by Sri Ambati
AWE Tel Aviv Startup Pitch: Mark Goldfarb with Sixdof Space
Junli Gu at AI Frontiers: Autonomous Driving Revolution
Preparing your team for a new XR platform; 7 key take-aways
Maximizing the Business Value of Connected Devices by Transforming the CIO's ...
Artificial Intelligence in Small Embedded System
Scott Montgomerie (Scope AR): AR’s Influence on the Workforce of Tomorrow: Jo...
GT C Tour 2018 Highlights
"A Practitioner’s Guide to Commercializing Applications of Computer Vision," ...
Find Your Passion and Make a Difference in Your Career
Technology and AI sharing - From 2016 to Y2017 and Beyond
T Bytes IoT & AR
ImageCon CTO keynote
Qualcomm: How to take advantage of XR over 5G in 2019: Understanding XR Viewers
Highlights from NVIDIA at CES 2014
Machine Learning for Self-Driving Cars
Ad

Viewers also liked (20)

PDF
Andres Rodriguez at AI Frontiers: Catalyzing Deep Learning's Impact in the En...
PDF
Nikko Ström at AI Frontiers: Deep Learning in Alexa
PDF
Lukasz Kaiser at AI Frontiers: How Deep Learning Quietly Revolutionized NLP
PDF
James Manyika at AI Frontiers: A Future That Works: Automation, Employment, a...
PDF
Adam Coates at AI Frontiers: AI for 100 Million People with Deep Learning
PDF
Jisheng Wang at AI Frontiers: Deep Learning in Security
PDF
Yangqing Jia at AI Frontiers: Towards Better DL Frameworks
PDF
Rajat Monga at AI Frontiers: Deep Learning with TensorFlow
PDF
Li Deng at AI Frontiers: Three Generations of Spoken Dialogue Systems (Bots)
PDF
Jeff Dean at AI Frontiers: Trends and Developments in Deep Learning Research
PDF
Soumith Chintala at AI Frontiers: A Dynamic View of the Deep Learning World
PDF
Alex Smola at AI Frontiers: Scalable Deep Learning Using MXNet
PDF
Hai Tao at AI Frontiers: Deep Learning For Embedded Vision System
PDF
Intelligent Chatbot on WeChat
PDF
Naghi Prasad at AI Frontiers: Building AI systems to automate enterprise proc...
PDF
Augmented Reality: The Next 20 Years
PPT
Augmented Reality ppt
PDF
Charles Fan at AI Frontiers: The New Era of AI Plus
PDF
Kevin Shaw at AI Frontiers: AI on the Edge: Bringing Intelligence to Small De...
PPTX
Augmented reality for the Enterprise
Andres Rodriguez at AI Frontiers: Catalyzing Deep Learning's Impact in the En...
Nikko Ström at AI Frontiers: Deep Learning in Alexa
Lukasz Kaiser at AI Frontiers: How Deep Learning Quietly Revolutionized NLP
James Manyika at AI Frontiers: A Future That Works: Automation, Employment, a...
Adam Coates at AI Frontiers: AI for 100 Million People with Deep Learning
Jisheng Wang at AI Frontiers: Deep Learning in Security
Yangqing Jia at AI Frontiers: Towards Better DL Frameworks
Rajat Monga at AI Frontiers: Deep Learning with TensorFlow
Li Deng at AI Frontiers: Three Generations of Spoken Dialogue Systems (Bots)
Jeff Dean at AI Frontiers: Trends and Developments in Deep Learning Research
Soumith Chintala at AI Frontiers: A Dynamic View of the Deep Learning World
Alex Smola at AI Frontiers: Scalable Deep Learning Using MXNet
Hai Tao at AI Frontiers: Deep Learning For Embedded Vision System
Intelligent Chatbot on WeChat
Naghi Prasad at AI Frontiers: Building AI systems to automate enterprise proc...
Augmented Reality: The Next 20 Years
Augmented Reality ppt
Charles Fan at AI Frontiers: The New Era of AI Plus
Kevin Shaw at AI Frontiers: AI on the Edge: Bringing Intelligence to Small De...
Augmented reality for the Enterprise
Ad

Similar to Liu Ren at AI Frontiers: Sensor-aware Augmented Reality (20)

PPTX
Opportunities and Challenges in Open Source AR in 2018
PDF
IT Sapiens-SAP ITS Mobile Solution With Responsive Design for Multi Device Ad...
PDF
Say Hello to Watson on the Shopfloor - AI and Cognitive Manufacturing in Action
PPTX
THE THIRD EYE-Presentation
PDF
Aurea Vision & Aurea Consulting: technology proposal
PDF
Panoptix-BBOX
PDF
Big Video Data Revolution, Challenges Unresolved
PDF
E-COMMERCE WEBSITE WITH AR-BASED TECHNOLOGY
PPTX
How the Bosch Group is making use of OSGi for IoT - Kai Hackbarth
PPTX
Enabling the Real Time Analytical Enterprise
PDF
The Internet Of Everything - How To Make It Smarter
PPTX
JAKT Portfolio
PDF
From Vision to Reality
PDF
From Business Idea to Successful Delivery by Serhiy Haziyev & Olha Hrytsay, S...
PPTX
Internet of Things (IoT): More Opportunity than Risk
DOC
ANIRBAN_GHOSH_RESUME
PPTX
VRSafety NETCO 2017
PDF
Datapalooza ibm 051916_final
PPTX
Smarter Retail
Opportunities and Challenges in Open Source AR in 2018
IT Sapiens-SAP ITS Mobile Solution With Responsive Design for Multi Device Ad...
Say Hello to Watson on the Shopfloor - AI and Cognitive Manufacturing in Action
THE THIRD EYE-Presentation
Aurea Vision & Aurea Consulting: technology proposal
Panoptix-BBOX
Big Video Data Revolution, Challenges Unresolved
E-COMMERCE WEBSITE WITH AR-BASED TECHNOLOGY
How the Bosch Group is making use of OSGi for IoT - Kai Hackbarth
Enabling the Real Time Analytical Enterprise
The Internet Of Everything - How To Make It Smarter
JAKT Portfolio
From Vision to Reality
From Business Idea to Successful Delivery by Serhiy Haziyev & Olha Hrytsay, S...
Internet of Things (IoT): More Opportunity than Risk
ANIRBAN_GHOSH_RESUME
VRSafety NETCO 2017
Datapalooza ibm 051916_final
Smarter Retail

More from AI Frontiers (20)

PPTX
Divya Jain at AI Frontiers : Video Summarization
PPTX
Training at AI Frontiers 2018 - LaiOffer Data Session: How Spark Speedup AI
PDF
Training at AI Frontiers 2018 - LaiOffer Self-Driving-Car-Lecture 1: Heuristi...
PDF
Training at AI Frontiers 2018 - Ni Lao: Weakly Supervised Natural Language Un...
PDF
Training at AI Frontiers 2018 - LaiOffer Self-Driving-Car-lecture 2: Incremen...
PDF
Training at AI Frontiers 2018 - Udacity: Enhancing NLP with Deep Neural Networks
PDF
Training at AI Frontiers 2018 - LaiOffer Self-Driving-Car-Lecture 3: Any-Angl...
PDF
Training at AI Frontiers 2018 - Lukasz Kaiser: Sequence to Sequence Learning ...
PDF
Percy Liang at AI Frontiers : Pushing the Limits of Machine Learning
PDF
Ilya Sutskever at AI Frontiers : Progress towards the OpenAI mission
PDF
Mark Moore at AI Frontiers : Uber Elevate
PPTX
Mario Munich at AI Frontiers : Consumer robotics: embedding affordable AI in ...
PPTX
Arnaud Thiercelin at AI Frontiers : AI in the Sky
PPTX
Anima Anandkumar at AI Frontiers : Modern ML : Deep, distributed, Multi-dimen...
PPTX
Wei Xu at AI Frontiers : Language Learning in an Interactive and Embodied Set...
PPTX
Sumit Gupta at AI Frontiers : AI for Enterprise
PPTX
Yuandong Tian at AI Frontiers : Planning in Reinforcement Learning
PPTX
Alex Ermolaev at AI Frontiers : Major Applications of AI in Healthcare
PPTX
Long Lin at AI Frontiers : AI in Gaming
PDF
Melissa Goldman at AI Frontiers : AI & Finance
Divya Jain at AI Frontiers : Video Summarization
Training at AI Frontiers 2018 - LaiOffer Data Session: How Spark Speedup AI
Training at AI Frontiers 2018 - LaiOffer Self-Driving-Car-Lecture 1: Heuristi...
Training at AI Frontiers 2018 - Ni Lao: Weakly Supervised Natural Language Un...
Training at AI Frontiers 2018 - LaiOffer Self-Driving-Car-lecture 2: Incremen...
Training at AI Frontiers 2018 - Udacity: Enhancing NLP with Deep Neural Networks
Training at AI Frontiers 2018 - LaiOffer Self-Driving-Car-Lecture 3: Any-Angl...
Training at AI Frontiers 2018 - Lukasz Kaiser: Sequence to Sequence Learning ...
Percy Liang at AI Frontiers : Pushing the Limits of Machine Learning
Ilya Sutskever at AI Frontiers : Progress towards the OpenAI mission
Mark Moore at AI Frontiers : Uber Elevate
Mario Munich at AI Frontiers : Consumer robotics: embedding affordable AI in ...
Arnaud Thiercelin at AI Frontiers : AI in the Sky
Anima Anandkumar at AI Frontiers : Modern ML : Deep, distributed, Multi-dimen...
Wei Xu at AI Frontiers : Language Learning in an Interactive and Embodied Set...
Sumit Gupta at AI Frontiers : AI for Enterprise
Yuandong Tian at AI Frontiers : Planning in Reinforcement Learning
Alex Ermolaev at AI Frontiers : Major Applications of AI in Healthcare
Long Lin at AI Frontiers : AI in Gaming
Melissa Goldman at AI Frontiers : AI & Finance

Recently uploaded (20)

PDF
Encapsulation theory and applications.pdf
PDF
Modernizing your data center with Dell and AMD
PPTX
Digital-Transformation-Roadmap-for-Companies.pptx
PDF
Blue Purple Modern Animated Computer Science Presentation.pdf.pdf
PPTX
MYSQL Presentation for SQL database connectivity
PDF
NewMind AI Monthly Chronicles - July 2025
PDF
TokAI - TikTok AI Agent : The First AI Application That Analyzes 10,000+ Vira...
PDF
How UI/UX Design Impacts User Retention in Mobile Apps.pdf
PDF
Mobile App Security Testing_ A Comprehensive Guide.pdf
PDF
Diabetes mellitus diagnosis method based random forest with bat algorithm
PDF
Build a system with the filesystem maintained by OSTree @ COSCUP 2025
PDF
Review of recent advances in non-invasive hemoglobin estimation
PPTX
PA Analog/Digital System: The Backbone of Modern Surveillance and Communication
PDF
NewMind AI Weekly Chronicles - August'25 Week I
PDF
Approach and Philosophy of On baking technology
PDF
Encapsulation_ Review paper, used for researhc scholars
PDF
Electronic commerce courselecture one. Pdf
PDF
Advanced methodologies resolving dimensionality complications for autism neur...
PDF
Agricultural_Statistics_at_a_Glance_2022_0.pdf
PDF
Shreyas Phanse Resume: Experienced Backend Engineer | Java • Spring Boot • Ka...
Encapsulation theory and applications.pdf
Modernizing your data center with Dell and AMD
Digital-Transformation-Roadmap-for-Companies.pptx
Blue Purple Modern Animated Computer Science Presentation.pdf.pdf
MYSQL Presentation for SQL database connectivity
NewMind AI Monthly Chronicles - July 2025
TokAI - TikTok AI Agent : The First AI Application That Analyzes 10,000+ Vira...
How UI/UX Design Impacts User Retention in Mobile Apps.pdf
Mobile App Security Testing_ A Comprehensive Guide.pdf
Diabetes mellitus diagnosis method based random forest with bat algorithm
Build a system with the filesystem maintained by OSTree @ COSCUP 2025
Review of recent advances in non-invasive hemoglobin estimation
PA Analog/Digital System: The Backbone of Modern Surveillance and Communication
NewMind AI Weekly Chronicles - August'25 Week I
Approach and Philosophy of On baking technology
Encapsulation_ Review paper, used for researhc scholars
Electronic commerce courselecture one. Pdf
Advanced methodologies resolving dimensionality complications for autism neur...
Agricultural_Statistics_at_a_Glance_2022_0.pdf
Shreyas Phanse Resume: Experienced Backend Engineer | Java • Spring Boot • Ka...

Liu Ren at AI Frontiers: Sensor-aware Augmented Reality

  • 1. Sensor-Aware Augmented Reality Addressing Real-World HMI Challenges Dr. Liu Ren Global Head and Chief Scientist, HMI Bosch Research North America Palo Alto, CA
  • 2. Sensor-Aware Augmented Reality Research and Technology Center North America | Liu Ren | 12/20/2016 © 2016 Robert Bosch LLC and affiliates. All rights reserved. 2 Bosch Overview Mobility Solutions Industrial Technology Energy and Building Technology Consumer Goods Bosch is one of the world’s leading international providers of technology and services • 375,0001 Bosch associates • More than 4401 subsidiary companies and regional subsidiaries in some 601 countries • Including its sales and service partners, Bosch is represented in some 1501 countries. 1 As of Dec. 2015 Home Appliance Personal Assistant Home Robots Garden Tools Smart Home Internet of Things (IoT) Thermothenology Security Systems Smart Cities Assembly Technology Industry 4.0 Packaging Technology Industrial Robots Car Infotainment Concept Car Autonomous Driving Automotive Aftermarket
  • 3. Sensor-Aware Augmented Reality Research and Technology Center North America | Liu Ren | 12/20/2016 © 2016 Robert Bosch LLC and affiliates. All rights reserved. 3 Industry 4.0Smart HomeRoboticsAftermarket Repair Shops Car Infotainment Highly Automated Driving Human Machine Interaction (HMI) Research in Bosch
  • 4. Industry 4.0Smart HomeRoboticsAftermarket Repair Shops Car Infotainment Highly Automated Driving Sensor-Aware Augmented Reality Research and Technology Center North America | Liu Ren | 12/20/2016 © 2016 Robert Bosch LLC and affiliates. All rights reserved. 4 Key Success Factors of HMI Products Intuitive Interactive Intelligent Human Machine Interaction
  • 5. • Global Head and Chief Scientist, HMI, Bosch Research • Ph.D. and M.Sc. in Computer Science, Carnegie Mellon University • B.Sc. in Computer Science, Zhejiang University, P.R. China Sensor-Aware Augmented Reality Research and Technology Center North America | Liu Ren | 12/20/2016 © 2016 Robert Bosch LLC and affiliates. All rights reserved. 5 Global HMI Research Team Renningen, Germany Shanghai, China Headquarters Palo Alto, USA Liu Ren Short Bio HMI teams in Bosch Research
  • 6. Sensor-Aware Augmented Reality Research and Technology Center North America | Liu Ren | 12/20/2016 © 2016 Robert Bosch LLC and affiliates. All rights reserved. 6 Real-World HMI Challenges for (Wearable) AR Hardware Form Factor SoftwareField-of-View Comfort Battery Life Context-Aware Visualization Scalable & Easy Content Generation Natural Interaction (Speech, Gesture, etc.) AI (Perception, Understanding, etc.)
  • 7. Sensor-Aware Augmented Reality Research and Technology Center North America | Liu Ren | 12/20/2016 © 2016 Robert Bosch LLC and affiliates. All rights reserved. 7 Bosch Product: CAP (Common Augmented Reality Platform) Software Context-Aware Visualization Scalable & Easy Content Generation Natural Interaction Bosch CAP enables implementation of complete enterprises AR solutions • Integrates the production of visual and digital content directly into the authoring process. • Existing CAD, image and video data were used and save the expense of creating new content. Bosch CAP Production and manufacturing Target/actual comparison and collision planning Plant and system planning Education and training Maintenance, service and repair Marketing, trade shows and distribution Technical doc. and digital operating instructions
  • 8. Sensor-Aware Augmented Reality Research and Technology Center North America | Liu Ren | 12/20/2016 © 2016 Robert Bosch LLC and affiliates. All rights reserved. 8 Our Sensor-Aware Solutions Software Context-Aware Visualization Scalable & Easy Content Generation Natural Interaction (Speech, Gesture, etc.) Dynamic occlusion handling ISMAR 2016[1] • Enhance realistic depth perception Robust visual tracking ISMAR 2016[2] • Improve tracking robustness and accuracy [1] Chao Du, Yen-Lin Chen, Mao Ye, and Liu Ren, “Edge Snapping-Based Depth Enhancement for Dynamic Occlusion Handling in Augmented Reality”, IEEE International Symposium on Mixed and Augmented Reality (ISMAR) 2016. [2] Benzun Wisely Babu, Soohwan Kim, Zhixin Yan, and Liu Ren, “σ-DVO: Sensor Noise Model Meets Dense Visual Odometry”, IEEE International Symposium on Mixed and Augmented Reality (ISMAR) 2016.
  • 9. Sensor-Aware Augmented Reality Research and Technology Center North America | Liu Ren | 12/20/2016 © 2016 Robert Bosch LLC and affiliates. All rights reserved. 9 Dynamic Occlusion Handling Software Context-Aware Visualization Scalable & Easy Content Generation Natural Interaction (Speech, Gesture, etc.) Robust visual tracking ISMAR 2016[2] • Improve tracking robustness and accuracy Dynamic occlusion handling ISMAR 2016[1] • Enhance realistic depth perception
  • 10. 1. Compact setup: single sensor 2. Dynamic occlusion handling Sensor-Aware Augmented Reality Research and Technology Center North America | Liu Ren | 12/20/2016 © 2016 Robert Bosch LLC and affiliates. All rights reserved. 10 Dynamic Occlusion Handling: Motivation One near-range RGBD sensor Optical see-through head- mounted display (HMD) Challenges • Performance requirements for real-time AR applications • Limited computational resources (e.g., on tablet) Goals
  • 11. Sensor-Aware Augmented Reality Research and Technology Center North America | Liu Ren | 12/20/2016 © 2016 Robert Bosch LLC and affiliates. All rights reserved. 11 Dynamic Occlusion Handling: Our Sensor-Aware Solution Target Object Boundary Boundary from Depth Align Object Boundary (Edge-Snapping)1 Depth Color Enhance Depth Map2 Raw Depth Map Enhanced Depth Map • Use color images as guidance • Snap object boundaries in depth data towards the edges in color images • Formulated as an optimization problem, efficiently solved via dynamic programming Edge Snapping-Based Algorithm Depth data not reliable at the object boundary • Structured Light/Stereo: matching is not accurate at the boundary • Time of Flight: light signal reaching object boundary barely bounce back to the sensor Knowledge on RGBD Sensor
  • 12. Sensor-Aware Augmented Reality Research and Technology Center North America | Liu Ren | 12/20/2016 © 2016 Robert Bosch LLC and affiliates. All rights reserved. 12 Dynamic Occlusion Handling: Experimental Results
  • 13. Sensor-Aware Augmented Reality Research and Technology Center North America | Liu Ren | 12/20/2016 © 2016 Robert Bosch LLC and affiliates. All rights reserved. 13 Robust Visual Tracking Software Context-Aware Visualization Scalable & Easy Content Generation Natural Interaction (Speech, Gesture, etc.) Dynamic occlusion handling ISMAR 2016[1] • Enhance realistic depth perception Robust visual tracking ISMAR 2016[2] • Improve tracking robustness and accuracy
  • 14. Sensor-Aware Augmented Reality Research and Technology Center North America | Liu Ren | 12/20/2016 © 2016 Robert Bosch LLC and affiliates. All rights reserved. 14 Robust Visual Tracking: Motivation 1. Visual tracking is an essential AR component • 6 DoF camera pose • Correctly place virtual objects in real world 2. Markerless Visual Tracking: • Visual SLAM (simultaneous localization and mapping) Background Challenges Textureless Blurry image Lighting condition change
  • 15. Sensor-Aware Augmented Reality Research and Technology Center North America | Liu Ren | 12/20/2016 © 2016 Robert Bosch LLC and affiliates. All rights reserved. 15 Robust Visual Tracking: Our Sensor-Aware Solution (σ-DVO) • Working well with textureless environments • Less sensitive to lighting condition changes • Noise of depth measurement grows quadratically as depth increases • Estimate the relative pose between two given frames based on residuals (front-end of visual SLAM) • Utilize all pixels from RGBD images RGBD Dense Visual OdometryKnowledge on RGBD Sensor Previous Frame − = Current Frame Relative Camera Pose Warped Current Frame Residuals Color Residual Depth Residual Non-linear Optimization Previous Frame Current Frame Relative Camera Pose x Color Weights Depth Weights Weights Weight decreases as noise of depth measurement grows Sensor-Aware Weighting • Incorporate sensor noise model to guide pose optimization
  • 16. Previous Frame − = Current Frame Warped Current Frame Non-linear Optimization Relative Camera Pose Residuals Color Residual Depth Residual Sensor-Aware Augmented Reality Research and Technology Center North America | Liu Ren | 12/20/2016 © 2016 Robert Bosch LLC and affiliates. All rights reserved. 16 Robust Visual Tracking: Our Sensor-Aware Solution (σ-DVO) • For better robustness and accuracy, we formulated the optimization problem in a Bayesian framework Bayesian Framework 𝑝 𝑝𝑜𝑠𝑒|𝑟𝑒𝑠𝑖𝑑𝑢𝑎𝑙𝑠 ∝ 𝒑 𝒓𝒆𝒔𝒊𝒅𝒖𝒂𝒍𝒔|𝒑𝒐𝒔𝒆 ∙ 𝑝 𝑝𝑜𝑠𝑒 • Assume uniform distribution of residuals  All the pixels share the same weight Early approaches • Find an empirical distribution via experiments  Weights only depends on residuals The state-of-the-art approach (DVO[1]) • Explore the source of residuals, especially based on sensor characteristics • Develop a sensor noise model to generate distribution of residuals  Decrease weights of pixels with either noisy sensor measurement or high residuals • Easily incorporate sensor-specific noise model for different sensors to customize pose optimization for best performance Our sensor-aware approach (σ-DVO) Sensor Measurement Noise One near-range RGBD sensor x Color Weights Depth Weights Weights [1] Christian Kerl, Jürgen Sturm, and Daniel Cremers. "Dense visual SLAM for RGB-D cameras." 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems. IEEE, 2013.
  • 17. Visual SLAM * Sensor-Aware Augmented Reality Research and Technology Center North America | Liu Ren | 12/20/2016 © 2016 Robert Bosch LLC and affiliates. All rights reserved. 17 Robust Visual Tracking: Experimental Results Dataset RGB-D SLAM[2] MRSM ap[3] Kintinu ous[4] ElasticFu sion[5] DVO SLAM[1] Our SLAM approach r1/desk 0.023 0.043 0.037 0.020 0.021 0.019 fr2/xyz 0.008 0.020 0.029 0.011 0.018 0.018 fr3/office 0.032 0.042 0.030 0.017 0.035 0.015 fr1/360 0.079 0.069 - - 0.083 0.061 Dataset DVO[1] Our approach (σ-DVO ) ATE RPE ATE RPE fr1/360 0.415 0.153 0.229 0.110 fr1/desk 0.109 0.048 0.067 0.039 fr1/desk2 0.261 0.074 0.088 0.065 fr1/floor 0.242 0.070 0.226 0.053 fr1/room 0.459 0.092 0.314 0.063 fr1/rpy 0.216 0.065 0.072 0.046 fr1/xyz 0.102 0.05 0.052 0.036 fr2/desk 0.561 0.038 0.184 0.016 fr2/large 4.370 0.240 0.724 0.134 fr2/rpy 0.501 0.039 0.188 0.012 fr2/xyz 0.497 0.030 0.188 0.010 fr3/office 0.485 0.044 0.164 0.014 average 0.684 0.067 0.208 0.050 Absolute Tracking Error [m] ATE: Absolute Tracking Error [m], RPE: Relative Pose Error [m/s] Visual Odometry • Our σ-DVO outperforms DVO significantly in all the datasets. On average, 70% reduction in ATE and 25% reduction in RPE • σ-DVO SLAM outperforms the state-of-the-art SLAM algorithms in most of the RGB-D datasets. On average 25% reduction in ATE * σ-DVO is extended to σ-DVO SLAM by combining the front end (visual odometry) with the backend (pose-graph optimization) [1] Kerl, et al. "Dense visual SLAM for RGB-D cameras." 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems. [2] Endres, et al. "An evaluation of the RGB-D SLAM system." Robotics and Automation (ICRA), 2012. [3] Stückler, et al. “Model Learning and Real-Time Tracking using Multi-Resolution Surfel Maps”. AAAI, 2012. [4] Whelan, et al. "Kintinuous: Spatially extended kinectfusion." Proc. Workshop RGB-D, Adv. Reason. Depth Cameras, 2012. [5] Whelan, et al. "ElasticFusion: Dense SLAM without a pose graph." Proc. Robotics: Science and Systems, 2015.
  • 18. Sensor-Aware Augmented Reality Research and Technology Center North America | Liu Ren | 12/20/2016 © 2016 Robert Bosch LLC and affiliates. All rights reserved. 18 Robust Visual Tracking: Experimental Results
  • 19. Sensor-Aware Augmented Reality Research and Technology Center North America | Liu Ren | 12/20/2016 © 2016 Robert Bosch LLC and affiliates. All rights reserved. 19 Deep Learning for Augmented Reality? Software Context-Aware Visualization Scalable & Easy Content Generation Natural Interaction (Speech, Gesture, etc.) ? • Require semantic understanding of the environments & context • Modern AI technologies, e.g., Deep Learning, could be effective approaches
  • 20. Sensor-Aware Augmented Reality Research and Technology Center North America | Liu Ren | 12/20/2016 © 2016 Robert Bosch LLC and affiliates. All rights reserved. 20 Summary and Outlook 1 The three “I”s (Intuitive, Interactive, Intelligent) are key success factors of Human Machine Interaction (HMI) solutions. 2 Sensor-aware approaches that leverage sensor knowledge and machine learning are effective to address real-world HMI challenges. 3 Using the right AI technology to address the right problem. Deep Learning, could be effective for core AR solutions.
  • 21. We’re looking for good research scientists and interns! (Send CV to rbvisualcomp@gmail.com) Thank You