SlideShare a Scribd company logo
Multi-Sensor Calibration by
Deep Learning
Yu Huang
Sunnyvale, California
Yu.huang07@gmail.com
Outline
• RegNet: Multimodal Sensor Registration Using Deep Neural Networks
• CalibNet: Self-Supervised Extrinsic Calibration using 3D Spatial
Transformer Networks
• RGGNet: Tolerance Aware LiDAR-Camera Online Calibration with
Geometric Deep Learning and Generative Model
• CalibRCNN: Calibrating Camera and LiDAR by Recurrent Convolutional
Neural Network and Geometric Constraints
• LCCNet: LiDAR and Camera Self-Calibration using Cost Volume Network
• CFNet: LiDAR-Camera Registration Using Calibration Flow Network
RegNet: Multimodal Sensor Registration
Using Deep Neural Networks
• RegNet, the deep CNN to infer a 6 DOF extrinsic calibration between multimodal sensors,
exemplified using a scanning LiDAR and a monocular camera.
• Compared to existing approaches, RegNet casts all 3 conventional calibration steps (feature
extraction, feature matching and global regression) into a single real-time capable CNN.
• It does not require any human interaction and bridges the gap between classical offline and
target-less online calibration approaches as it provides both a stable initial estimation as well as a
continuous online correction of the extrinsic parameters.
• During training, randomly decalibrate our system in order to train RegNet to infer the
correspondence between projected depth measurements and RGB image and finally regress the
extrinsic calibration.
• Additionally, with an iterative execution of multiple CNNs, that are trained on different
magnitudes of decalibration, it compares favorably to state-of-the-art methods in terms of a
mean calibration error of 0.28◦ for the rotational and 6 cm for the translation components even
for large decalibrations up to 1.5 m and 20◦ .
RegNet: Multimodal Sensor Registration
Using Deep Neural Networks
It estimates the calibration btw a depth and an RGB sensor. The depth points are projected on the RGB image using
an initial calibration Hinit. In the 1st and 2nd part of the network, use NiN blocks to extract rich features for
matching. The final part regresses decalibration by gathering global info. using two FCLs. During training φdecalib is
randomly permutated resulting in different projections of the depth points.
RegNet: Multimodal Sensor Registration
Using Deep Neural Networks
CalibNet: Self-Supervised Extrinsic Calibration
using 3D Spatial Transformer Networks
• CalibNet: a self-supervised deep network capable of automatically estimating the 6-DoF rigid
body transformation between a 3D LiDAR and a 2D camera in real-time.
• CalibNet alleviates the need for calibration targets, thereby resulting in significant savings in
calibration efforts.
• During training, the network only takes as input a LiDAR point cloud, the corresponding monocular
image, and the camera calibration matrix K.
• At train time, no impose direct supervision (i.e., no directly regress to the calibration parameters, for
example).
• Instead, train the network to predict calibration parameters that maximize the geometric and
photometric consistency of the input images and point clouds.
• CalibNet learns to iteratively solve the underlying geometric problem and accurately predicts
extrinsic calibration parameters for a wide range of mis-calibrations, without requiring retraining
or domain adaptation.
• Code: https://github.com/epiception/CalibNet.
CalibNet: Self-Supervised Extrinsic Calibration
using 3D Spatial Transformer Networks
Input RGB image (a), a raw LiDAR point cloud (b), and outputs a transformation T that best aligns the two inputs. (c)
the colorized point cloud output for a mis-calibrated setup, and (d) the output after calibration using CalibNet.
CalibNet: Self-Supervised Extrinsic Calibration
using 3D Spatial Transformer Networks
Network architecture
CalibNet: Self-Supervised Extrinsic Calibration
using 3D Spatial Transformer Networks
RGGNet: Tolerance Aware LiDAR-Camera Online Calibration
with Geometric Deep Learning and Generative Model
• With increasing popularity of deep learning (DL), a few recent efforts have
demonstrated the advantages of DL for feature extraction on this task.
• However, their reported performances are not sufficiently satisfying yet.
• One improvement can be the formulation with proper consideration of the
underneath geometry.
• Besides, existing online calibration methods focus on optimizing the
calibration error while overlooking the tolerance within the error bounds.
• To address the research gap, a DL-based LiDAR-camera calibration method,
named as the RGGNet, is proposed by considering the Riemannian
geometry and utilizing a deep generative model to learn an implicit
tolerance model.
• The code available at https://github.com/KleinYuan/RGGNet.
RGGNet: Tolerance Aware LiDAR-Camera Online Calibration
with Geometric Deep Learning and Generative Model
The architecture of the proposed RGGNet.
RGGNet: Tolerance Aware LiDAR-Camera Online Calibration
with Geometric Deep Learning and Generative Model
The architecture of the Tolerance Regularizer. VAE encoders contain four Conv-ReLU-BN layers with
depths as 64/128/64/32, following with a FC-ReLU-BN layer with 1024 units and one FC layer with 200
units; decoders contain two FC-ReLU-BN FC layers with units as 200 and (HW ) 8 , three Deconv-ReLU-BN
layers with depths as 32/64/128 and output Deconv-Sigmoid layer with 5 units. Kernel size of 4 and stride
size of 2 are set for all convolution and deconvolution layers.
RGGNet: Tolerance Aware LiDAR-Camera Online Calibration
with Geometric Deep Learning and Generative Model
The intensity depth map and the point map are
projected by the point clouds with the
calibration parameters (K, Y ). Each feature
map corresponds to an RGB image.
Examples of RGGNet predictions. Reference objects
for calibration are shown in bounding boxes.
RGGNet: Tolerance Aware LiDAR-Camera Online Calibration
with Geometric Deep Learning and Generative Model
One bad prediction example with the same input fromT2b test. a: initial de-calibrated input;
b: the ground-truth; c: the prediction from -RGGNet, yielding a 0.0550 se3 error; d: the
prediction from RGGNet, yielding a 0.060 se3 error. Images are cropped for demonstration.
RGGNet: Tolerance Aware LiDAR-Camera Online Calibration
with Geometric Deep Learning and Generative Model
TABLE I PERFORMANCE COMPARISONS WITH DNN BASED METHODS
CalibRCNN: Calibrating Camera and LiDAR by Recurrent
Convolutional Neural Network and Geometric Constraints
• Calibration Recurrent Convolutional Neural Network (CalibRCNN) to infer a 6
degrees of freedom (DOF) rigid body transformation between 3D LiDAR and 2D
camera.
• Different from the existing methods, 3D-2D CalibRCNN not only uses the LSTM
network to extract the temporal features between 3D point clouds and RGB
images of consecutive frames, but also uses the geometric loss and photometric
loss obtained by the interframe constraint to refine the calibration accuracy of
the predicted transformation parameters.
• The CalibRCNN aims at inferring the correspondence between projected depth
image and RGB image to learn the underlying geometry of 2D-3D calibration.
• Thus, the proposed calibration model achieves a good generalization ability to
adapt to unknown initial calibration error ranges, and other 3D LiDAR and 2D
camera pairs with different intrinsic parameters from the training dataset.
CalibRCNN: Calibrating Camera and LiDAR by Recurrent
Convolutional Neural Network and Geometric Constraints
Architecture of CalibRCNN. (a) the RGB images of consecutive frames; (b) depth maps of
consecutive frames generated from the mis-calibrated point clouds. By projecting 3D point
cloud to the corresponding 2D RGB image we can obtain images similar to (c) and (d), where
the color of the projected point represents its depth value. (c) is a 3D-2D projection image
converted by mis-calibration parameters, while (d) is the projection image after calibration
using our network. Red rectangles show the difference before and after calibration.
CalibRCNN: Calibrating Camera and LiDAR by Recurrent
Convolutional Neural Network and Geometric Constraints
Architecture of the proposed Calibration Recurrent Convolutional Neural Network.
CalibRCNN: Calibrating Camera and LiDAR by Recurrent
Convolutional Neural Network and Geometric Constraints
Pose transformation relationship diagram of
continuous frames of LiDAR and camera data
LCCNet: LiDAR and Camera Self-Calibration using
Cost Volume Network
• It is online LiDAR-Camera Self-calibration Network (LCCNet), different from the previous CNN-based
methods.
• LCCNet can be trained end-to-end and predict the extrinsic parameters in real-time.
• In the LCCNet, exploit the cost volume layer to express the correlation between the RGB image features and
the depth image projected from point clouds.
• Besides using the smooth L1-Loss of the predicted extrinsic calibration parameters as a supervised signal, an
additional self-supervised signal, point cloud distance loss, is applied during training.
• Instead of directly regressing the extrinsic parameters, predict the decalibrated deviation from initial
calibration to the ground truth.
• The calibration error decreases further with iterative refinement and the temporal filtering approach in the
inference stage.
• The execution time of the calibration process is 24ms for each iteration on a single GPU.
• LCCNet achieves a mean absolute calibration error of 0.297cm in translation and 0.017◦ in rotation with
miscalibration magnitudes of up to  1.5m and  20◦ on the KITTI-odometry dataset, which is better than the
state-of-the-art CNN-based calibration methods.
• The code available at https://github.com/LvXudong-HIT/LCCNet.
LCCNet: LiDAR and Camera Self-Calibration using
Cost Volume Network
The proposed LCCNet takes the RGB and the
projecte depth image as inputs to predict the
extrinsic parameters between the LiDAR and
the camera. The point clouds are reprojected b
the predicted extrinsic parameters. The
reprojected depth image and the RGB image
will be the subsequent inputs of the LCCNet
This process is called iterative refinement.
After five time iterative refinements, we obtain
the final extrinsic parameters estimation.
LCCNet: LiDAR and Camera Self-Calibration using
Cost Volume Network
The network takes an RGB image from a calibrated camera and a projected sparse depth image from
a mis-calibrated LiDAR as input. The output is a 6-DoF rigid-body transformation that represents the
deviation between the initial extrinsic and the ground truth extrinsic.
LCCNet: LiDAR and Camera Self-Calibration using
Cost Volume Network
(a) Initial Calibration, (b) Ground truth, (c) Calibration results.
LCCNet: LiDAR and Camera Self-Calibration using
Cost Volume Network
CFNet: LiDAR-Camera Registration Using
Calibration Flow Network
• An online LiDAR-camera extrinsic calibration algorithm that combines
the DL and the geometry methods.
• Define a two-channel image named calibration flow to illustrate the
deviation from the initial projection to the ground truth.
• EPnP algorithm within the RANSAC scheme is applied to estimate the
extrinsic parameters with 2D-3D correspondences constructed by the
calibration flow.
• A semantic initialization with the introduction of instance centroids
(ICs).
• The code available at https://github.com/LvXudong-HIT/CFNet.
CFNet: LiDAR-Camera Registration Using
Calibration Flow Network
(First Row) The 2D instance centroid (2D-
IC) extracted from images. (Second Row)
The 3D instance centroid (3D-IC)
extracted from LiDAR point clouds. (Third
Row) The initial calibration parameters
provided by semantic initialization
(registration of 2DIC and 3D-IC). (Fourth
Row) The final calibration parameters
predicted by CFNet. (Bottom Row) Three-
dimensional map generated by fusing the
sensors using calibration values provided
by our approach.
CFNet: LiDAR-Camera Registration Using
Calibration Flow Network
CFNet, an automatic online extrinsic calibration method that estimates
the transformation parameters between 3D LiDAR and 2D camera.
CFNet: LiDAR-Camera Registration Using
Calibration Flow Network
The Architecture of calibration network CFNet
CFNet: LiDAR-Camera Registration Using
Calibration Flow Network
CFNet: LiDAR-Camera Registration Using
Calibration Flow Network
CFNet: LiDAR-Camera Registration Using
Calibration Flow Network
Multi sensor calibration by deep learning

More Related Content

PDF
LiDAR-based Autonomous Driving III (by Deep Learning)
PDF
Deep VO and SLAM
PDF
Depth Fusion from RGB and Depth Sensors by Deep Learning
PDF
NeRF: Representing Scenes as Neural Radiance Fields for View Synthesis
PDF
fusion of Camera and lidar for autonomous driving I
PDF
3D Perception for Autonomous Driving - Datasets and Algorithms -
PDF
fusion of Camera and lidar for autonomous driving II
PDF
Introduction of slam
LiDAR-based Autonomous Driving III (by Deep Learning)
Deep VO and SLAM
Depth Fusion from RGB and Depth Sensors by Deep Learning
NeRF: Representing Scenes as Neural Radiance Fields for View Synthesis
fusion of Camera and lidar for autonomous driving I
3D Perception for Autonomous Driving - Datasets and Algorithms -
fusion of Camera and lidar for autonomous driving II
Introduction of slam

What's hot (20)

PDF
Unsupervised Anomaly Detection with Generative Adversarial Networks to Guide ...
PDF
object detection with lidar-camera fusion: survey (updated)
PPTX
SuperGlue; Learning Feature Matching with Graph Neural Networks (CVPR'20)
PDF
Mask R-CNN
PDF
SLAM入門 第2章 SLAMの基礎
PPTX
Depth estimation using deep learning
PDF
20180424 orb slam
PPTX
[DL輪読会]End-to-End Object Detection with Transformers
PDF
Lidar for Autonomous Driving II (via Deep Learning)
PDF
[DL輪読会]"CyCADA: Cycle-Consistent Adversarial Domain Adaptation"&"Learning Se...
PPTX
LiDARの原理などのまとめ
PPTX
Super Resolution
PDF
DeepLearningTutorial
PDF
Depth Fusion from RGB and Depth Sensors II
PDF
PR-302: NeRF: Representing Scenes as Neural Radiance Fields for View Synthesis
PDF
[PR12] You Only Look Once (YOLO): Unified Real-Time Object Detection
PDF
BA-Net: Dense Bundle Adjustment Network (3D勉強会@関東)
PDF
[PR12] image super resolution using deep convolutional networks
PDF
SSII2019TS: 実践カメラキャリブレーション ~カメラを用いた実世界計測の基礎と応用~
PDF
BEV Object Detection and Prediction
Unsupervised Anomaly Detection with Generative Adversarial Networks to Guide ...
object detection with lidar-camera fusion: survey (updated)
SuperGlue; Learning Feature Matching with Graph Neural Networks (CVPR'20)
Mask R-CNN
SLAM入門 第2章 SLAMの基礎
Depth estimation using deep learning
20180424 orb slam
[DL輪読会]End-to-End Object Detection with Transformers
Lidar for Autonomous Driving II (via Deep Learning)
[DL輪読会]"CyCADA: Cycle-Consistent Adversarial Domain Adaptation"&"Learning Se...
LiDARの原理などのまとめ
Super Resolution
DeepLearningTutorial
Depth Fusion from RGB and Depth Sensors II
PR-302: NeRF: Representing Scenes as Neural Radiance Fields for View Synthesis
[PR12] You Only Look Once (YOLO): Unified Real-Time Object Detection
BA-Net: Dense Bundle Adjustment Network (3D勉強会@関東)
[PR12] image super resolution using deep convolutional networks
SSII2019TS: 実践カメラキャリブレーション ~カメラを用いた実世界計測の基礎と応用~
BEV Object Detection and Prediction
Ad

Similar to Multi sensor calibration by deep learning (20)

PDF
3-d interpretation from single 2-d image IV
PDF
Depth Fusion from RGB and Depth Sensors IV
PDF
Driving behaviors for adas and autonomous driving xiv
PDF
Masters Thesis
PDF
Voxelized GICP for Fast and Accurate 3D Point Cloud Registration [ICRA2021]
PDF
3-d interpretation from stereo images for autonomous driving
PDF
Automatic Detection of Window Regions in Indoor Point Clouds Using R-CNN
PDF
Depth Fusion from RGB and Depth Sensors III
PDF
Fisheye based Perception for Autonomous Driving VI
PDF
3-d interpretation from single 2-d image for autonomous driving II
PDF
Decomposing image generation into layout priction and conditional synthesis
PDF
10.1109@ICCMC48092.2020.ICCMC-000167.pdf
PDF
Deep Learning’s Application in Radar Signal Data II
PPTX
Image Segmentation Using Deep Learning : A survey
PDF
A Robust Automatic Meter Reading System based on Mask-RCNN
PDF
CNN vs SIFT-based Visual Localization - Laura Leal-Taixé - UPC Barcelona 2018
PDF
Camera-Based Road Lane Detection by Deep Learning II
PPTX
QUALITY ASSESSMENT FOR LIDAR POINT CLOUD REGISTRATION USING IN-SITU CONJUGATE...
PPTX
[NS][Lab_Seminar_241118]Relation Matters: Foreground-aware Graph-based Relati...
PDF
“Introduction to Simultaneous Localization and Mapping (SLAM),” a Presentatio...
3-d interpretation from single 2-d image IV
Depth Fusion from RGB and Depth Sensors IV
Driving behaviors for adas and autonomous driving xiv
Masters Thesis
Voxelized GICP for Fast and Accurate 3D Point Cloud Registration [ICRA2021]
3-d interpretation from stereo images for autonomous driving
Automatic Detection of Window Regions in Indoor Point Clouds Using R-CNN
Depth Fusion from RGB and Depth Sensors III
Fisheye based Perception for Autonomous Driving VI
3-d interpretation from single 2-d image for autonomous driving II
Decomposing image generation into layout priction and conditional synthesis
10.1109@ICCMC48092.2020.ICCMC-000167.pdf
Deep Learning’s Application in Radar Signal Data II
Image Segmentation Using Deep Learning : A survey
A Robust Automatic Meter Reading System based on Mask-RCNN
CNN vs SIFT-based Visual Localization - Laura Leal-Taixé - UPC Barcelona 2018
Camera-Based Road Lane Detection by Deep Learning II
QUALITY ASSESSMENT FOR LIDAR POINT CLOUD REGISTRATION USING IN-SITU CONJUGATE...
[NS][Lab_Seminar_241118]Relation Matters: Foreground-aware Graph-based Relati...
“Introduction to Simultaneous Localization and Mapping (SLAM),” a Presentatio...
Ad

More from Yu Huang (20)

PDF
Embodied AI: Ushering in the Next Era of Intelligent Systems
PDF
GOSIM_China_2024_Embodied AI Data VLA World Model
PDF
Levels of AI Agents: from Rules to Large Language Models
PDF
Application of Foundation Model for Autonomous Driving
PDF
The New Perception Framework in Autonomous Driving: An Introduction of BEV N...
PDF
Data Closed Loop in Simulation Test of Autonomous Driving
PDF
Techniques and Challenges in Autonomous Driving
PDF
BEV Joint Detection and Segmentation
PDF
Fisheye/Omnidirectional View in Autonomous Driving V
PDF
Fisheye/Omnidirectional View in Autonomous Driving IV
PDF
Prediction,Planninng & Control at Baidu
PDF
Cruise AI under the Hood
PDF
LiDAR in the Adverse Weather: Dust, Snow, Rain and Fog (2)
PDF
Scenario-Based Development & Testing for Autonomous Driving
PDF
How to Build a Data Closed-loop Platform for Autonomous Driving?
PDF
Annotation tools for ADAS & Autonomous Driving
PDF
Simulation for autonomous driving at uber atg
PDF
Prediction and planning for self driving at waymo
PDF
Jointly mapping, localization, perception, prediction and planning
PDF
Data pipeline and data lake for autonomous driving
Embodied AI: Ushering in the Next Era of Intelligent Systems
GOSIM_China_2024_Embodied AI Data VLA World Model
Levels of AI Agents: from Rules to Large Language Models
Application of Foundation Model for Autonomous Driving
The New Perception Framework in Autonomous Driving: An Introduction of BEV N...
Data Closed Loop in Simulation Test of Autonomous Driving
Techniques and Challenges in Autonomous Driving
BEV Joint Detection and Segmentation
Fisheye/Omnidirectional View in Autonomous Driving V
Fisheye/Omnidirectional View in Autonomous Driving IV
Prediction,Planninng & Control at Baidu
Cruise AI under the Hood
LiDAR in the Adverse Weather: Dust, Snow, Rain and Fog (2)
Scenario-Based Development & Testing for Autonomous Driving
How to Build a Data Closed-loop Platform for Autonomous Driving?
Annotation tools for ADAS & Autonomous Driving
Simulation for autonomous driving at uber atg
Prediction and planning for self driving at waymo
Jointly mapping, localization, perception, prediction and planning
Data pipeline and data lake for autonomous driving

Recently uploaded (20)

PPTX
CH1 Production IntroductoryConcepts.pptx
DOCX
ASol_English-Language-Literature-Set-1-27-02-2023-converted.docx
PDF
Mitigating Risks through Effective Management for Enhancing Organizational Pe...
PDF
SM_6th-Sem__Cse_Internet-of-Things.pdf IOT
PPTX
Strings in CPP - Strings in C++ are sequences of characters used to store and...
PPTX
Lesson 3_Tessellation.pptx finite Mathematics
PPTX
Geodesy 1.pptx...............................................
PDF
Model Code of Practice - Construction Work - 21102022 .pdf
PDF
Mohammad Mahdi Farshadian CV - Prospective PhD Student 2026
PDF
keyrequirementskkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkk
PDF
PRIZ Academy - 9 Windows Thinking Where to Invest Today to Win Tomorrow.pdf
PPTX
FINAL REVIEW FOR COPD DIANOSIS FOR PULMONARY DISEASE.pptx
PDF
BMEC211 - INTRODUCTION TO MECHATRONICS-1.pdf
PPTX
Lecture Notes Electrical Wiring System Components
PDF
Evaluating the Democratization of the Turkish Armed Forces from a Normative P...
PPTX
Infosys Presentation by1.Riyan Bagwan 2.Samadhan Naiknavare 3.Gaurav Shinde 4...
PPTX
Welding lecture in detail for understanding
PPTX
Engineering Ethics, Safety and Environment [Autosaved] (1).pptx
PPTX
CYBER-CRIMES AND SECURITY A guide to understanding
PPT
Project quality management in manufacturing
CH1 Production IntroductoryConcepts.pptx
ASol_English-Language-Literature-Set-1-27-02-2023-converted.docx
Mitigating Risks through Effective Management for Enhancing Organizational Pe...
SM_6th-Sem__Cse_Internet-of-Things.pdf IOT
Strings in CPP - Strings in C++ are sequences of characters used to store and...
Lesson 3_Tessellation.pptx finite Mathematics
Geodesy 1.pptx...............................................
Model Code of Practice - Construction Work - 21102022 .pdf
Mohammad Mahdi Farshadian CV - Prospective PhD Student 2026
keyrequirementskkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkk
PRIZ Academy - 9 Windows Thinking Where to Invest Today to Win Tomorrow.pdf
FINAL REVIEW FOR COPD DIANOSIS FOR PULMONARY DISEASE.pptx
BMEC211 - INTRODUCTION TO MECHATRONICS-1.pdf
Lecture Notes Electrical Wiring System Components
Evaluating the Democratization of the Turkish Armed Forces from a Normative P...
Infosys Presentation by1.Riyan Bagwan 2.Samadhan Naiknavare 3.Gaurav Shinde 4...
Welding lecture in detail for understanding
Engineering Ethics, Safety and Environment [Autosaved] (1).pptx
CYBER-CRIMES AND SECURITY A guide to understanding
Project quality management in manufacturing

Multi sensor calibration by deep learning

  • 1. Multi-Sensor Calibration by Deep Learning Yu Huang Sunnyvale, California Yu.huang07@gmail.com
  • 2. Outline • RegNet: Multimodal Sensor Registration Using Deep Neural Networks • CalibNet: Self-Supervised Extrinsic Calibration using 3D Spatial Transformer Networks • RGGNet: Tolerance Aware LiDAR-Camera Online Calibration with Geometric Deep Learning and Generative Model • CalibRCNN: Calibrating Camera and LiDAR by Recurrent Convolutional Neural Network and Geometric Constraints • LCCNet: LiDAR and Camera Self-Calibration using Cost Volume Network • CFNet: LiDAR-Camera Registration Using Calibration Flow Network
  • 3. RegNet: Multimodal Sensor Registration Using Deep Neural Networks • RegNet, the deep CNN to infer a 6 DOF extrinsic calibration between multimodal sensors, exemplified using a scanning LiDAR and a monocular camera. • Compared to existing approaches, RegNet casts all 3 conventional calibration steps (feature extraction, feature matching and global regression) into a single real-time capable CNN. • It does not require any human interaction and bridges the gap between classical offline and target-less online calibration approaches as it provides both a stable initial estimation as well as a continuous online correction of the extrinsic parameters. • During training, randomly decalibrate our system in order to train RegNet to infer the correspondence between projected depth measurements and RGB image and finally regress the extrinsic calibration. • Additionally, with an iterative execution of multiple CNNs, that are trained on different magnitudes of decalibration, it compares favorably to state-of-the-art methods in terms of a mean calibration error of 0.28◦ for the rotational and 6 cm for the translation components even for large decalibrations up to 1.5 m and 20◦ .
  • 4. RegNet: Multimodal Sensor Registration Using Deep Neural Networks It estimates the calibration btw a depth and an RGB sensor. The depth points are projected on the RGB image using an initial calibration Hinit. In the 1st and 2nd part of the network, use NiN blocks to extract rich features for matching. The final part regresses decalibration by gathering global info. using two FCLs. During training φdecalib is randomly permutated resulting in different projections of the depth points.
  • 5. RegNet: Multimodal Sensor Registration Using Deep Neural Networks
  • 6. CalibNet: Self-Supervised Extrinsic Calibration using 3D Spatial Transformer Networks • CalibNet: a self-supervised deep network capable of automatically estimating the 6-DoF rigid body transformation between a 3D LiDAR and a 2D camera in real-time. • CalibNet alleviates the need for calibration targets, thereby resulting in significant savings in calibration efforts. • During training, the network only takes as input a LiDAR point cloud, the corresponding monocular image, and the camera calibration matrix K. • At train time, no impose direct supervision (i.e., no directly regress to the calibration parameters, for example). • Instead, train the network to predict calibration parameters that maximize the geometric and photometric consistency of the input images and point clouds. • CalibNet learns to iteratively solve the underlying geometric problem and accurately predicts extrinsic calibration parameters for a wide range of mis-calibrations, without requiring retraining or domain adaptation. • Code: https://github.com/epiception/CalibNet.
  • 7. CalibNet: Self-Supervised Extrinsic Calibration using 3D Spatial Transformer Networks Input RGB image (a), a raw LiDAR point cloud (b), and outputs a transformation T that best aligns the two inputs. (c) the colorized point cloud output for a mis-calibrated setup, and (d) the output after calibration using CalibNet.
  • 8. CalibNet: Self-Supervised Extrinsic Calibration using 3D Spatial Transformer Networks Network architecture
  • 9. CalibNet: Self-Supervised Extrinsic Calibration using 3D Spatial Transformer Networks
  • 10. RGGNet: Tolerance Aware LiDAR-Camera Online Calibration with Geometric Deep Learning and Generative Model • With increasing popularity of deep learning (DL), a few recent efforts have demonstrated the advantages of DL for feature extraction on this task. • However, their reported performances are not sufficiently satisfying yet. • One improvement can be the formulation with proper consideration of the underneath geometry. • Besides, existing online calibration methods focus on optimizing the calibration error while overlooking the tolerance within the error bounds. • To address the research gap, a DL-based LiDAR-camera calibration method, named as the RGGNet, is proposed by considering the Riemannian geometry and utilizing a deep generative model to learn an implicit tolerance model. • The code available at https://github.com/KleinYuan/RGGNet.
  • 11. RGGNet: Tolerance Aware LiDAR-Camera Online Calibration with Geometric Deep Learning and Generative Model The architecture of the proposed RGGNet.
  • 12. RGGNet: Tolerance Aware LiDAR-Camera Online Calibration with Geometric Deep Learning and Generative Model The architecture of the Tolerance Regularizer. VAE encoders contain four Conv-ReLU-BN layers with depths as 64/128/64/32, following with a FC-ReLU-BN layer with 1024 units and one FC layer with 200 units; decoders contain two FC-ReLU-BN FC layers with units as 200 and (HW ) 8 , three Deconv-ReLU-BN layers with depths as 32/64/128 and output Deconv-Sigmoid layer with 5 units. Kernel size of 4 and stride size of 2 are set for all convolution and deconvolution layers.
  • 13. RGGNet: Tolerance Aware LiDAR-Camera Online Calibration with Geometric Deep Learning and Generative Model The intensity depth map and the point map are projected by the point clouds with the calibration parameters (K, Y ). Each feature map corresponds to an RGB image. Examples of RGGNet predictions. Reference objects for calibration are shown in bounding boxes.
  • 14. RGGNet: Tolerance Aware LiDAR-Camera Online Calibration with Geometric Deep Learning and Generative Model One bad prediction example with the same input fromT2b test. a: initial de-calibrated input; b: the ground-truth; c: the prediction from -RGGNet, yielding a 0.0550 se3 error; d: the prediction from RGGNet, yielding a 0.060 se3 error. Images are cropped for demonstration.
  • 15. RGGNet: Tolerance Aware LiDAR-Camera Online Calibration with Geometric Deep Learning and Generative Model TABLE I PERFORMANCE COMPARISONS WITH DNN BASED METHODS
  • 16. CalibRCNN: Calibrating Camera and LiDAR by Recurrent Convolutional Neural Network and Geometric Constraints • Calibration Recurrent Convolutional Neural Network (CalibRCNN) to infer a 6 degrees of freedom (DOF) rigid body transformation between 3D LiDAR and 2D camera. • Different from the existing methods, 3D-2D CalibRCNN not only uses the LSTM network to extract the temporal features between 3D point clouds and RGB images of consecutive frames, but also uses the geometric loss and photometric loss obtained by the interframe constraint to refine the calibration accuracy of the predicted transformation parameters. • The CalibRCNN aims at inferring the correspondence between projected depth image and RGB image to learn the underlying geometry of 2D-3D calibration. • Thus, the proposed calibration model achieves a good generalization ability to adapt to unknown initial calibration error ranges, and other 3D LiDAR and 2D camera pairs with different intrinsic parameters from the training dataset.
  • 17. CalibRCNN: Calibrating Camera and LiDAR by Recurrent Convolutional Neural Network and Geometric Constraints Architecture of CalibRCNN. (a) the RGB images of consecutive frames; (b) depth maps of consecutive frames generated from the mis-calibrated point clouds. By projecting 3D point cloud to the corresponding 2D RGB image we can obtain images similar to (c) and (d), where the color of the projected point represents its depth value. (c) is a 3D-2D projection image converted by mis-calibration parameters, while (d) is the projection image after calibration using our network. Red rectangles show the difference before and after calibration.
  • 18. CalibRCNN: Calibrating Camera and LiDAR by Recurrent Convolutional Neural Network and Geometric Constraints Architecture of the proposed Calibration Recurrent Convolutional Neural Network.
  • 19. CalibRCNN: Calibrating Camera and LiDAR by Recurrent Convolutional Neural Network and Geometric Constraints Pose transformation relationship diagram of continuous frames of LiDAR and camera data
  • 20. LCCNet: LiDAR and Camera Self-Calibration using Cost Volume Network • It is online LiDAR-Camera Self-calibration Network (LCCNet), different from the previous CNN-based methods. • LCCNet can be trained end-to-end and predict the extrinsic parameters in real-time. • In the LCCNet, exploit the cost volume layer to express the correlation between the RGB image features and the depth image projected from point clouds. • Besides using the smooth L1-Loss of the predicted extrinsic calibration parameters as a supervised signal, an additional self-supervised signal, point cloud distance loss, is applied during training. • Instead of directly regressing the extrinsic parameters, predict the decalibrated deviation from initial calibration to the ground truth. • The calibration error decreases further with iterative refinement and the temporal filtering approach in the inference stage. • The execution time of the calibration process is 24ms for each iteration on a single GPU. • LCCNet achieves a mean absolute calibration error of 0.297cm in translation and 0.017◦ in rotation with miscalibration magnitudes of up to  1.5m and  20◦ on the KITTI-odometry dataset, which is better than the state-of-the-art CNN-based calibration methods. • The code available at https://github.com/LvXudong-HIT/LCCNet.
  • 21. LCCNet: LiDAR and Camera Self-Calibration using Cost Volume Network The proposed LCCNet takes the RGB and the projecte depth image as inputs to predict the extrinsic parameters between the LiDAR and the camera. The point clouds are reprojected b the predicted extrinsic parameters. The reprojected depth image and the RGB image will be the subsequent inputs of the LCCNet This process is called iterative refinement. After five time iterative refinements, we obtain the final extrinsic parameters estimation.
  • 22. LCCNet: LiDAR and Camera Self-Calibration using Cost Volume Network The network takes an RGB image from a calibrated camera and a projected sparse depth image from a mis-calibrated LiDAR as input. The output is a 6-DoF rigid-body transformation that represents the deviation between the initial extrinsic and the ground truth extrinsic.
  • 23. LCCNet: LiDAR and Camera Self-Calibration using Cost Volume Network (a) Initial Calibration, (b) Ground truth, (c) Calibration results.
  • 24. LCCNet: LiDAR and Camera Self-Calibration using Cost Volume Network
  • 25. CFNet: LiDAR-Camera Registration Using Calibration Flow Network • An online LiDAR-camera extrinsic calibration algorithm that combines the DL and the geometry methods. • Define a two-channel image named calibration flow to illustrate the deviation from the initial projection to the ground truth. • EPnP algorithm within the RANSAC scheme is applied to estimate the extrinsic parameters with 2D-3D correspondences constructed by the calibration flow. • A semantic initialization with the introduction of instance centroids (ICs). • The code available at https://github.com/LvXudong-HIT/CFNet.
  • 26. CFNet: LiDAR-Camera Registration Using Calibration Flow Network (First Row) The 2D instance centroid (2D- IC) extracted from images. (Second Row) The 3D instance centroid (3D-IC) extracted from LiDAR point clouds. (Third Row) The initial calibration parameters provided by semantic initialization (registration of 2DIC and 3D-IC). (Fourth Row) The final calibration parameters predicted by CFNet. (Bottom Row) Three- dimensional map generated by fusing the sensors using calibration values provided by our approach.
  • 27. CFNet: LiDAR-Camera Registration Using Calibration Flow Network CFNet, an automatic online extrinsic calibration method that estimates the transformation parameters between 3D LiDAR and 2D camera.
  • 28. CFNet: LiDAR-Camera Registration Using Calibration Flow Network The Architecture of calibration network CFNet
  • 29. CFNet: LiDAR-Camera Registration Using Calibration Flow Network
  • 30. CFNet: LiDAR-Camera Registration Using Calibration Flow Network
  • 31. CFNet: LiDAR-Camera Registration Using Calibration Flow Network