SlideShare une entreprise Scribd logo
1  sur  51
Télécharger pour lire hors ligne
Deep VO and SLAM II
Yu Huang
Yu.huang07@gmail.com
Sunnyvale, California
Outline
• Single View Stereo Matching, 3, 2018
• LEGO: Learning Edge with Geometry all at
Once by Watching Videos, 3, 2018
• Unsupervised Learning of Monocular
Depth Estimation and Visual Odometry
with Deep Feature Reconstruction, 4, 2018
• Unsupervised Learning of Depth and Ego-
Motion from Monocular Video Using 3D
Geometric Constraints, 6, 2018
• Look Deeper into Depth: Monocular Depth
Estimation with Semantic Booster and
Attention-Driven Loss, ECCV, 2018
• GeoNet: Geometric Neural Network, CVPR
2018
• GeoNet: Unsupervised Learning of Dense
Depth, Optical Flow and Camera Pose
• Learning Depth from Monocular Videos
using Direct Methods, CVPR, 2018
• CNN-SVO: Improving the Mapping in Semi-
Direct Visual Odometry Using Single-Image
Depth Prediction, 10, 2018
• Depth Prediction Without the Sensors:
Leveraging Structure for Unsupervised
Learning from Monocular Videos, 11, 2018
• Self-Supervised Learning of Depth and
Camera Motion from 360◦ Videos, 11, 2018
• Unsupervised Learning-based Depth
Estimation aided Visual SLAM Approach, 1,
2019
Single View Stereo Matching
• Previous monocular depth estimation methods take a single view and directly regress the
expected results.
• Though recent advances are made by applying geometrically inspired loss functions during
training, the inference procedure does not explicitly impose any geometrical constraint.
• Therefore these models purely rely on the quality of data and the effectiveness of learning to
generalize.
• This either leads to suboptimal results or the demand of huge amount of expensive ground truth
labelled data to generate reasonable results.
• This paper shows that the monocular depth estimation problem can be reformulated as two sub-
problems, a view synthesis procedure followed by stereo matching, with two intriguing
properties, namely i) geometrical constraints can be explicitly imposed during inference; ii)
demand on labelled depth data can be greatly alleviated.
• The whole pipeline can still be trained in an end-to-end fashion and this new formulation plays a
critical role in advancing the performance.
• The model also generalizes well to other monocular depth estimation benchmarks.
• It also discusses the implications and the advantages of solving monocular depth estimation using
stereo methods.
Single View Stereo Matching
Single View Stereo Matching
Single View Stereo Matching
“Unsupervised monocular depth estimation with left-right consistency”
LEGO: Learning Edge with Geometry all at
Once by Watching Videos
• Learning to estimate 3D geometry in a single image by watching unlabeled videos via deep
convolutional network is attracting significant attention.
• This paper introduces a "3D as-smooth-as-possible (3D-ASAP)" prior inside the pipeline, which enables
joint estimation of edges and 3D scene, yielding results with significant improvement in accuracy for
fine detailed structures.
• Specifically, it defines the 3D-ASAP prior by requiring that any two points recovered in 3D from an
image should lie on an existing planar surface if no other cues provided.
• They design an unsupervised framework that Learns Edges and Geometry (depth, normal) all at Once
(LEGO).
• The predicted edges are embedded into depth and surface normal smoothness terms, where pixels
without edges in-between are constrained to satisfy the prior.
• In the framework, the predicted depths, normals and edges are forced to be consistent all the time.
LEGO: Learning Edge with Geometry all at
Once by Watching Videos
LEGO: Learning Edge with Geometry all at
Once by Watching Videos
Unsupervised Learning of Depth and Ego-Motion from
Monocular Video Using 3D Geometric Constraints
• It is an approach for unsupervised learning of depth and ego-motion from monocular video.
• Unsupervised learning removes the need for separate supervisory signals (depth or ego-motion
ground truth, or multi-view video).
• Prior work in unsupervised depth learning uses pixel-wise or gradient-based losses, which only
consider pixels in small local neighborhoods.
• This idea is to explicitly consider the inferred 3D geometry of the whole scene, and enforce
consistency of the estimated 3D point clouds and ego-motion across consecutive frames.
• This is a challenging task and is solved by a novel (approximate) backpropagation algorithm for
aligning 3D structures.
• They combine this 3D-based loss with 2D losses based on photometric quality of frame
reconstructions using estimated depth and ego-motion from adjacent frames.
• It also incorporates validity masks to avoid penalizing areas in which no useful information exists.
• Because they only require a simple video, learning depth and ego-motion on large and varied
datasets becomes possible.
• Codes are available at https://sites.google.com/view/vid2depth
Unsupervised Learning of Depth and Ego-Motion from
Monocular Video Using 3D Geometric Constraints
Unsupervised Learning of Depth and Ego-Motion from
Monocular Video Using 3D Geometric Constraints
Unsupervised Learning of Depth and Ego-Motion from
Monocular Video Using 3D Geometric Constraints
Unsupervised Learning of Depth and Ego-Motion from
Monocular Video Using 3D Geometric Constraints
“Unsupervised cnn for single view depth
estimation: Geometry to the rescue”
“Unsupervised learning of depth
and ego-motion from video”
Look Deeper into Depth: Monocular Depth Estimation
with Semantic Booster and Attention-Driven Loss
• Monocular depth estimation benefits greatly from learning based techniques.
• By studying the training data, they observe that the per-pixel depth values in existing datasets
typically exhibit a long-tailed distribution.
• However, most previous approaches treat all the regions in the training data equally regardless
of the imbalanced depth distribution, which restricts the model performance particularly on
distant depth regions.
• This paper investigates the long tail property and delve deeper into the distant depth regions
(i.e. the tail part) to propose an attention driven loss for the network supervision.
• In addition, to better leverage the semantic information for monocular depth estimation, it
proposes a synergy network to automatically learn the information sharing strategies between
the two tasks.
• With the proposed attention-driven loss and synergy network, the depth estimation and
semantic labeling tasks can be mutually improved.
Look Deeper into Depth: Monocular Depth Estimation
with Semantic Booster and Attention-Driven Loss
Look Deeper into Depth: Monocular Depth Estimation
with Semantic Booster and Attention-Driven Loss
“Deep convolutional
neural fields for
depth estimation
from a single image”
“Depth map prediction from a single image
using a multi-scale deep network”
“Deeper depth prediction with fully
convolutional residual networks”
“Depth map prediction from a single
image using a multi-scale deep network”
GeoNet: Geometric Neural Network
18
• This paper proposes Geometric Neural Network (GeoNet) to jointly predict depth
and surface normal maps from a single image.
• Building on top of two-stream CNNs, the GeoNet incorporates geometric relation
between depth and surface normal via the new depth-to-normal and normal-to-
depth networks.
• Depth-to-normal network exploits the least square solution of surface normal
from depth and improves its quality with a residual module.
• Normal-to-depth network, contrarily, refines the depth map based on the
constraints from the surface normal through a kernel regression module, which
has no parameter to learn.
• These two networks enforce the underlying model to efficiently predict depth
and surface normal for high consistency and corresponding accuracy.
GeoNet: Geometric Neural Network
19
GeoNet: Geometric Neural Network
20
GeoNet: Geometric Neural Network
21
GeoNet: Geometric Neural Network
22
GeoNet--Unsupervised Learning of Dense
Depth, Optical Flow and Camera Pose
• GeoNet, a jointly unsupervised learning framework for monocular depth, optical
flow and ego-motion estimation from videos.
• The three components are coupled by the nature of 3D scene geometry, jointly
learned by our framework in an end-to-end manner.
• Specifically, geometric relationships are extracted over the predictions of
individual modules and then combined as an image reconstruction loss,
reasoning about static and dynamic scene parts separately.
• Furthermore, they propose an adaptive geometric consistency loss to increase
robustness towards outliers and non-Lambertian regions, which resolves
occlusions and texture ambiguities effectively.
23
GeoNet--Unsupervised Learning of Dense
Depth, Optical Flow and Camera Pose
24
GeoNet--Unsupervised Learning of Dense
Depth, Optical Flow and Camera Pose
25
GeoNet--Unsupervised Learning of Dense
Depth, Optical Flow and Camera Pose
26
Unsupervised Learning of Monocular Depth Estimation
and Visual Odometry with Deep Feature Reconstruction
• Despite learning based methods showing promising results in single view depth estimation and
visual odometry, most existing approaches treat the tasks in a supervised manner.
• Recent approaches to single view depth estimation explore the possibility of learning without full
supervision via minimizing photometric error.
• This paper explores the use of stereo sequences for learning depth and visual odometry.
• The use of stereo sequences enables the use of both spatial (between left-right pairs) and
temporal (forward backward) photometric warp error, and constrains the scene depth and
camera motion to be in a common, real world scale.
• At test time the framework is able to estimate single view depth and two-view odometry from a
monocular sequence.
• They improve on a standard photometric warp loss by considering a warp of deep features.
• The source code is available at https://github.com/Huangying-Zhan/Depth-VO-Feat.
Unsupervised Learning of Monocular Depth Estimation
and Visual Odometry with Deep Feature Reconstruction
Unsupervised Learning of Monocular Depth Estimation
and Visual Odometry with Deep Feature Reconstruction
Unsupervised Learning of Monocular Depth Estimation
and Visual Odometry with Deep Feature Reconstruction
Learning Depth from Monocular Videos using Direct
Methods
• The ability to predict depth from a single image - using recent advances in CNNs - is of
increasing interest to the vision community.
• Unsupervised strategies to learning are particularly appealing as they can utilize much
larger and varied monocular video datasets during learning without the need for ground
truth depth or stereo.
• In previous works, separate pose and depth CNN predictors had to be determined such
that their joint outputs minimized the photometric error.
• Inspired by recent advances in direct visual odometry (DVO), it argues that the depth
CNN predictor can be learned without a pose CNN predictor.
• Further, they demonstrate empirically that incorporation of a differentiable
implementation of DVO, along with a novel depth normalization strategy - substantially
improves performance over state of the art that use monocular videos for training.
• https://github.com/MightyChaos/LKVOLearner
Learning Depth from Monocular Videos using Direct
Methods
Unsupervised learning pipeline. The learning
algorithm takes 3 sequential images at a time. The
Depth-CNN produces 3 inverse depth maps for
the inputs, and the pose predictor estimates two
relative camera pose between the second image
and the other two. The appearance dissimilarity
loss is measured between the second image I2
and the inversely warped images of I1, I3; In
addition, the loss is evaluated in a reverse
direction- it is also measured between I1, I3 and
two warped images of I2. Lower part illustrates 3
architectures we evaluated for pose prediction: 1)
Pose-CNN, 2) use the proposed differentiable
Direct Visual Odometry (DDVO), the initialization
of pose is set as zero (identity transformation),
and 3) a hybrid of the above two - use pretrained
Pose-CNN to give a better initial pose for DDVO.
Learning Depth from Monocular Videos using Direct
Methods “Unsupervised learning of depth and ego-motion from video”
CNN-SVO: Improving the Mapping in Semi-Direct Visual
Odometry Using Single-Image Depth Prediction
• Reliable feature correspondence between frames is a critical step in visual odometry (VO) and
visual simultaneous localization and mapping (V-SLAM) algorithms.
• In comparison with existing VO and V-SLAM algorithms, semi-direct visual odometry (SVO) has
two main advantages that lead to state-of-the-art frame rate camera motion estimation: direct
pixel correspondence and efficient implementation of probabilistic mapping method.
• This paper improves the SVO mapping by initializing the mean and the variance of the depth at a
feature location according to the depth prediction from a single image depth prediction network.
• By significantly reducing the depth uncertainty of the initialized map point (i.e., small variance
centered about the depth prediction), the benefits are twofold: reliable feature correspondence
between views and fast convergence to the true depth in order to create new map points.
CNN-SVO: Improving the Mapping in Semi-Direct Visual
Odometry Using Single-Image Depth Prediction
map point initialization
CNN-SVO: Improving the Mapping in Semi-Direct Visual
Odometry Using Single-Image Depth Prediction
CNN-SVO: Improving the Mapping in Semi-Direct Visual
Odometry Using Single-Image Depth Prediction
Camera motion estimation in the HDR environment Camera trajectory and map points
CNN-SVO: Improving the Mapping in Semi-Direct Visual
Odometry Using Single-Image Depth Prediction
Depth Prediction Without the Sensors: Leveraging Structure for
Unsupervised Learning from Monocular Videos
• Learning to predict depth from RGB is challenging both for indoor and outdoor robot navigation.
• This work addresses unsupervised learning of scene depth and robot ego-motion where
supervision is provided by monocular videos, as cameras are the cheapest, least restrictive and
most ubiquitous sensor for robotics.
• Previous unsupervised image-to-depth learning has established strong baselines in the domain.
• This approach is able to model moving objects and is shown to transfer across data domains, e.g.
from outdoors to indoor scenes.
• The main idea is to introduce geometric structure in the learning process, by modeling the scene
and the individual objects; camera ego-motion and object motions are learned from monocular
videos as input.
• An online refinement method is introduced to adapt learning on the fly to unknown domains.
• The code can be found at https://sites.google.com/view/struct2depth.
Depth Prediction Without the Sensors: Leveraging Structure for
Unsupervised Learning from Monocular Videos
Depth Prediction Without the Sensors: Leveraging Structure for
Unsupervised Learning from Monocular Videos
Depth Prediction Without the Sensors: Leveraging Structure for
Unsupervised Learning from Monocular Videos
Self-Supervised Learning of Depth and Camera
Motion from 360◦ Videos
• As 360◦ cameras become prevalent in many autonomous systems (e.g., self-driving cars and
drones), efficient 360◦ perception becomes more and more important.
• This is a self-supervised learning approach for predicting the omnidirectional depth and camera
motion from a 360◦ video.
• In particular, starting from the SfMLearner, which is designed for cameras with normal field-of-
view, they introduce three key features to process 360 ◦ images efficiently.
• convert each image from equirectangular projection to cubic projection in order to avoid image
distortion. In each network layer, use Cube Padding (CP), which pads intermediate features from
adjacent faces, to avoid image boundaries.
• apply a “spherical” photometric consistency constraint on the whole viewing sphere. In this way, no
pixel will be projected outside the image boundary with normal field-of-view.
• rather than estimating 6 independent camera motions (i.e., SfM-Learner to each face on a cube), apply
camera pose consistency loss to ensure the estimated camera motions reaching consensus.
• They collect a PanoSUNCG dataset containing a large amount of 360◦ videos with ground truth
depth and camera motion.
Self-Supervised Learning of Depth and Camera
Motion from 360◦ Videos
Self-Supervised Learning of Depth and Camera
Motion from 360◦ Videos
Spherical photometric constraints.
Self-Supervised Learning of Depth and Camera
Motion from 360◦ Videos
Unsupervised Learning-based Depth Estimation
aided Visual SLAM Approach
• Recently, deep learning technologies have achieved great success in the visual SLAM area, which
can directly learn high-level features from the visual inputs and improve the estimation accuracy
of the depth information.
• Therefore, deep learning technologies maintain the potential to extend the source of the depth
information and improve the performance of the SLAM system.
• However, the existing deep learning-based methods are mainly supervised and require a large
amount of ground-truth depth data, which is hard to acquire because of the realistic constraints.
• This paper presents an unsupervised learning framework, which not only uses image
reconstruction for supervising but also exploits the pose estimation method to enhance the
supervised signal and add training constraints for the task of monocular depth and camera
motion estimation.
• Furthermore, it exploits the unsupervised learning framework to assist the traditional ORB-SLAM
system when the initialization module of ORB-SLAM method could not match enough features.
• Unsupervised learning framework could significantly accelerate the initialization process of ORB-
SLAM system and effectively improve the accuracy on environmental mapping in strong lighting
and weak texture scenes.
Unsupervised Learning-based Depth Estimation
aided Visual SLAM Approach
Unsupervised Learning-based Depth Estimation
aided Visual SLAM Approach
Unsupervised Learning-based Depth Estimation
aided Visual SLAM Approach
Deep vo and slam ii

Contenu connexe

Tendances

3-d interpretation from single 2-d image III
3-d interpretation from single 2-d image III3-d interpretation from single 2-d image III
3-d interpretation from single 2-d image IIIYu Huang
 
3-d interpretation from single 2-d image for autonomous driving II
3-d interpretation from single 2-d image for autonomous driving II3-d interpretation from single 2-d image for autonomous driving II
3-d interpretation from single 2-d image for autonomous driving IIYu Huang
 
Pedestrian behavior/intention modeling for autonomous driving V
Pedestrian behavior/intention modeling for autonomous driving VPedestrian behavior/intention modeling for autonomous driving V
Pedestrian behavior/intention modeling for autonomous driving VYu Huang
 
fusion of Camera and lidar for autonomous driving II
fusion of Camera and lidar for autonomous driving IIfusion of Camera and lidar for autonomous driving II
fusion of Camera and lidar for autonomous driving IIYu Huang
 
Depth Fusion from RGB and Depth Sensors III
Depth Fusion from RGB and Depth Sensors  IIIDepth Fusion from RGB and Depth Sensors  III
Depth Fusion from RGB and Depth Sensors IIIYu Huang
 
Depth Fusion from RGB and Depth Sensors IV
Depth Fusion from RGB and Depth Sensors  IVDepth Fusion from RGB and Depth Sensors  IV
Depth Fusion from RGB and Depth Sensors IVYu Huang
 
Deep learning for 3-D Scene Reconstruction and Modeling
Deep learning for 3-D Scene Reconstruction and Modeling Deep learning for 3-D Scene Reconstruction and Modeling
Deep learning for 3-D Scene Reconstruction and Modeling Yu Huang
 
3-d interpretation from single 2-d image for autonomous driving
3-d interpretation from single 2-d image for autonomous driving3-d interpretation from single 2-d image for autonomous driving
3-d interpretation from single 2-d image for autonomous drivingYu Huang
 
Deep VO and SLAM IV
Deep VO and SLAM IVDeep VO and SLAM IV
Deep VO and SLAM IVYu Huang
 
Deep learning for image video processing
Deep learning for image video processingDeep learning for image video processing
Deep learning for image video processingYu Huang
 
Depth Fusion from RGB and Depth Sensors II
Depth Fusion from RGB and Depth Sensors IIDepth Fusion from RGB and Depth Sensors II
Depth Fusion from RGB and Depth Sensors IIYu Huang
 
camera-based Lane detection by deep learning
camera-based Lane detection by deep learningcamera-based Lane detection by deep learning
camera-based Lane detection by deep learningYu Huang
 
Driving Behavior for ADAS and Autonomous Driving VII
Driving Behavior for ADAS and Autonomous Driving VIIDriving Behavior for ADAS and Autonomous Driving VII
Driving Behavior for ADAS and Autonomous Driving VIIYu Huang
 
Anchor free object detection by deep learning
Anchor free object detection by deep learningAnchor free object detection by deep learning
Anchor free object detection by deep learningYu Huang
 
fusion of Camera and lidar for autonomous driving I
fusion of Camera and lidar for autonomous driving Ifusion of Camera and lidar for autonomous driving I
fusion of Camera and lidar for autonomous driving IYu Huang
 
BEV Semantic Segmentation
BEV Semantic SegmentationBEV Semantic Segmentation
BEV Semantic SegmentationYu Huang
 
Lidar for Autonomous Driving II (via Deep Learning)
Lidar for Autonomous Driving II (via Deep Learning)Lidar for Autonomous Driving II (via Deep Learning)
Lidar for Autonomous Driving II (via Deep Learning)Yu Huang
 
Deep Learning’s Application in Radar Signal Data II
Deep Learning’s Application in Radar Signal Data IIDeep Learning’s Application in Radar Signal Data II
Deep Learning’s Application in Radar Signal Data IIYu Huang
 
Multi sensor calibration by deep learning
Multi sensor calibration by deep learningMulti sensor calibration by deep learning
Multi sensor calibration by deep learningYu Huang
 

Tendances (20)

3-d interpretation from single 2-d image III
3-d interpretation from single 2-d image III3-d interpretation from single 2-d image III
3-d interpretation from single 2-d image III
 
3-d interpretation from single 2-d image for autonomous driving II
3-d interpretation from single 2-d image for autonomous driving II3-d interpretation from single 2-d image for autonomous driving II
3-d interpretation from single 2-d image for autonomous driving II
 
Pedestrian behavior/intention modeling for autonomous driving V
Pedestrian behavior/intention modeling for autonomous driving VPedestrian behavior/intention modeling for autonomous driving V
Pedestrian behavior/intention modeling for autonomous driving V
 
fusion of Camera and lidar for autonomous driving II
fusion of Camera and lidar for autonomous driving IIfusion of Camera and lidar for autonomous driving II
fusion of Camera and lidar for autonomous driving II
 
Depth Fusion from RGB and Depth Sensors III
Depth Fusion from RGB and Depth Sensors  IIIDepth Fusion from RGB and Depth Sensors  III
Depth Fusion from RGB and Depth Sensors III
 
Depth Fusion from RGB and Depth Sensors IV
Depth Fusion from RGB and Depth Sensors  IVDepth Fusion from RGB and Depth Sensors  IV
Depth Fusion from RGB and Depth Sensors IV
 
Deep learning for 3-D Scene Reconstruction and Modeling
Deep learning for 3-D Scene Reconstruction and Modeling Deep learning for 3-D Scene Reconstruction and Modeling
Deep learning for 3-D Scene Reconstruction and Modeling
 
3-d interpretation from single 2-d image for autonomous driving
3-d interpretation from single 2-d image for autonomous driving3-d interpretation from single 2-d image for autonomous driving
3-d interpretation from single 2-d image for autonomous driving
 
Deep VO and SLAM IV
Deep VO and SLAM IVDeep VO and SLAM IV
Deep VO and SLAM IV
 
Deep learning for image video processing
Deep learning for image video processingDeep learning for image video processing
Deep learning for image video processing
 
Depth Fusion from RGB and Depth Sensors II
Depth Fusion from RGB and Depth Sensors IIDepth Fusion from RGB and Depth Sensors II
Depth Fusion from RGB and Depth Sensors II
 
camera-based Lane detection by deep learning
camera-based Lane detection by deep learningcamera-based Lane detection by deep learning
camera-based Lane detection by deep learning
 
Driving Behavior for ADAS and Autonomous Driving VII
Driving Behavior for ADAS and Autonomous Driving VIIDriving Behavior for ADAS and Autonomous Driving VII
Driving Behavior for ADAS and Autonomous Driving VII
 
Anchor free object detection by deep learning
Anchor free object detection by deep learningAnchor free object detection by deep learning
Anchor free object detection by deep learning
 
fusion of Camera and lidar for autonomous driving I
fusion of Camera and lidar for autonomous driving Ifusion of Camera and lidar for autonomous driving I
fusion of Camera and lidar for autonomous driving I
 
Survey 1 (project overview)
Survey 1 (project overview)Survey 1 (project overview)
Survey 1 (project overview)
 
BEV Semantic Segmentation
BEV Semantic SegmentationBEV Semantic Segmentation
BEV Semantic Segmentation
 
Lidar for Autonomous Driving II (via Deep Learning)
Lidar for Autonomous Driving II (via Deep Learning)Lidar for Autonomous Driving II (via Deep Learning)
Lidar for Autonomous Driving II (via Deep Learning)
 
Deep Learning’s Application in Radar Signal Data II
Deep Learning’s Application in Radar Signal Data IIDeep Learning’s Application in Radar Signal Data II
Deep Learning’s Application in Radar Signal Data II
 
Multi sensor calibration by deep learning
Multi sensor calibration by deep learningMulti sensor calibration by deep learning
Multi sensor calibration by deep learning
 

Similaire à Deep vo and slam ii

Deep VO and SLAM
Deep VO and SLAMDeep VO and SLAM
Deep VO and SLAMYu Huang
 
AN ENHANCEMENT FOR THE CONSISTENT DEPTH ESTIMATION OF MONOCULAR VIDEOS USING ...
AN ENHANCEMENT FOR THE CONSISTENT DEPTH ESTIMATION OF MONOCULAR VIDEOS USING ...AN ENHANCEMENT FOR THE CONSISTENT DEPTH ESTIMATION OF MONOCULAR VIDEOS USING ...
AN ENHANCEMENT FOR THE CONSISTENT DEPTH ESTIMATION OF MONOCULAR VIDEOS USING ...mlaij
 
Fisheye/Omnidirectional View in Autonomous Driving V
Fisheye/Omnidirectional View in Autonomous Driving VFisheye/Omnidirectional View in Autonomous Driving V
Fisheye/Omnidirectional View in Autonomous Driving VYu Huang
 
PS1_2014_2012B5A7521P_2012B5A7848P_2012B4A7958H
PS1_2014_2012B5A7521P_2012B5A7848P_2012B4A7958HPS1_2014_2012B5A7521P_2012B5A7848P_2012B4A7958H
PS1_2014_2012B5A7521P_2012B5A7848P_2012B4A7958HSaurabh Kumar
 
High quality single shot capture of facial geometry
High quality single shot capture of facial geometryHigh quality single shot capture of facial geometry
High quality single shot capture of facial geometryBrohi Aijaz Ali
 
Indoor scene understanding for autonomous agents
Indoor scene understanding for autonomous agentsIndoor scene understanding for autonomous agents
Indoor scene understanding for autonomous agentsVarun Bhaseen
 
Quality assessment of stereoscopic 3 d image compression by binocular integra...
Quality assessment of stereoscopic 3 d image compression by binocular integra...Quality assessment of stereoscopic 3 d image compression by binocular integra...
Quality assessment of stereoscopic 3 d image compression by binocular integra...Shakas Technologies
 
Depth Estimation from Defocused Images: a Survey
Depth Estimation from Defocused Images: a SurveyDepth Estimation from Defocused Images: a Survey
Depth Estimation from Defocused Images: a SurveyIJAAS Team
 
Event recognition image & video segmentation
Event recognition image & video segmentationEvent recognition image & video segmentation
Event recognition image & video segmentationeSAT Journals
 
AaSeminar_Template.pptx
AaSeminar_Template.pptxAaSeminar_Template.pptx
AaSeminar_Template.pptxManojGowdaKb
 
Action Genome: Action As Composition of Spatio Temporal Scene Graphs
Action Genome: Action As Composition of Spatio Temporal Scene GraphsAction Genome: Action As Composition of Spatio Temporal Scene Graphs
Action Genome: Action As Composition of Spatio Temporal Scene GraphsSangmin Woo
 
Automatic 3D view Generation from a Single 2D Image for both Indoor and Outdo...
Automatic 3D view Generation from a Single 2D Image for both Indoor and Outdo...Automatic 3D view Generation from a Single 2D Image for both Indoor and Outdo...
Automatic 3D view Generation from a Single 2D Image for both Indoor and Outdo...ijcsa
 
A Review On Single Image Depth Prediction with Wavelet Decomposition
A Review On Single Image Depth Prediction with Wavelet DecompositionA Review On Single Image Depth Prediction with Wavelet Decomposition
A Review On Single Image Depth Prediction with Wavelet DecompositionIRJET Journal
 
Deep Learning for Structure-from-Motion (SfM)
Deep Learning for Structure-from-Motion (SfM)Deep Learning for Structure-from-Motion (SfM)
Deep Learning for Structure-from-Motion (SfM)PetteriTeikariPhD
 
10.1109@ICCMC48092.2020.ICCMC-000167.pdf
10.1109@ICCMC48092.2020.ICCMC-000167.pdf10.1109@ICCMC48092.2020.ICCMC-000167.pdf
10.1109@ICCMC48092.2020.ICCMC-000167.pdfmokamojah
 
IEEE 2014 MATLAB IMAGE PROCESSING PROJECTS Seamless view synthesis through te...
IEEE 2014 MATLAB IMAGE PROCESSING PROJECTS Seamless view synthesis through te...IEEE 2014 MATLAB IMAGE PROCESSING PROJECTS Seamless view synthesis through te...
IEEE 2014 MATLAB IMAGE PROCESSING PROJECTS Seamless view synthesis through te...IEEEBEBTECHSTUDENTPROJECTS
 
IRJET- Exploring Image Super Resolution Techniques
IRJET- Exploring Image Super Resolution TechniquesIRJET- Exploring Image Super Resolution Techniques
IRJET- Exploring Image Super Resolution TechniquesIRJET Journal
 
Fusion of Multi-MAV Data
Fusion of Multi-MAV DataFusion of Multi-MAV Data
Fusion of Multi-MAV DataDariolakis
 
Whiteboard image reconstruction using matlab
Whiteboard image reconstruction using matlabWhiteboard image reconstruction using matlab
Whiteboard image reconstruction using matlabeSAT Publishing House
 

Similaire à Deep vo and slam ii (20)

Deep VO and SLAM
Deep VO and SLAMDeep VO and SLAM
Deep VO and SLAM
 
AN ENHANCEMENT FOR THE CONSISTENT DEPTH ESTIMATION OF MONOCULAR VIDEOS USING ...
AN ENHANCEMENT FOR THE CONSISTENT DEPTH ESTIMATION OF MONOCULAR VIDEOS USING ...AN ENHANCEMENT FOR THE CONSISTENT DEPTH ESTIMATION OF MONOCULAR VIDEOS USING ...
AN ENHANCEMENT FOR THE CONSISTENT DEPTH ESTIMATION OF MONOCULAR VIDEOS USING ...
 
Fisheye/Omnidirectional View in Autonomous Driving V
Fisheye/Omnidirectional View in Autonomous Driving VFisheye/Omnidirectional View in Autonomous Driving V
Fisheye/Omnidirectional View in Autonomous Driving V
 
PS1_2014_2012B5A7521P_2012B5A7848P_2012B4A7958H
PS1_2014_2012B5A7521P_2012B5A7848P_2012B4A7958HPS1_2014_2012B5A7521P_2012B5A7848P_2012B4A7958H
PS1_2014_2012B5A7521P_2012B5A7848P_2012B4A7958H
 
High quality single shot capture of facial geometry
High quality single shot capture of facial geometryHigh quality single shot capture of facial geometry
High quality single shot capture of facial geometry
 
Indoor scene understanding for autonomous agents
Indoor scene understanding for autonomous agentsIndoor scene understanding for autonomous agents
Indoor scene understanding for autonomous agents
 
Quality assessment of stereoscopic 3 d image compression by binocular integra...
Quality assessment of stereoscopic 3 d image compression by binocular integra...Quality assessment of stereoscopic 3 d image compression by binocular integra...
Quality assessment of stereoscopic 3 d image compression by binocular integra...
 
Depth Estimation from Defocused Images: a Survey
Depth Estimation from Defocused Images: a SurveyDepth Estimation from Defocused Images: a Survey
Depth Estimation from Defocused Images: a Survey
 
Event recognition image & video segmentation
Event recognition image & video segmentationEvent recognition image & video segmentation
Event recognition image & video segmentation
 
AaSeminar_Template.pptx
AaSeminar_Template.pptxAaSeminar_Template.pptx
AaSeminar_Template.pptx
 
Action Genome: Action As Composition of Spatio Temporal Scene Graphs
Action Genome: Action As Composition of Spatio Temporal Scene GraphsAction Genome: Action As Composition of Spatio Temporal Scene Graphs
Action Genome: Action As Composition of Spatio Temporal Scene Graphs
 
Automatic 3D view Generation from a Single 2D Image for both Indoor and Outdo...
Automatic 3D view Generation from a Single 2D Image for both Indoor and Outdo...Automatic 3D view Generation from a Single 2D Image for both Indoor and Outdo...
Automatic 3D view Generation from a Single 2D Image for both Indoor and Outdo...
 
A Review On Single Image Depth Prediction with Wavelet Decomposition
A Review On Single Image Depth Prediction with Wavelet DecompositionA Review On Single Image Depth Prediction with Wavelet Decomposition
A Review On Single Image Depth Prediction with Wavelet Decomposition
 
Deep Learning for Structure-from-Motion (SfM)
Deep Learning for Structure-from-Motion (SfM)Deep Learning for Structure-from-Motion (SfM)
Deep Learning for Structure-from-Motion (SfM)
 
10.1109@ICCMC48092.2020.ICCMC-000167.pdf
10.1109@ICCMC48092.2020.ICCMC-000167.pdf10.1109@ICCMC48092.2020.ICCMC-000167.pdf
10.1109@ICCMC48092.2020.ICCMC-000167.pdf
 
IEEE 2014 MATLAB IMAGE PROCESSING PROJECTS Seamless view synthesis through te...
IEEE 2014 MATLAB IMAGE PROCESSING PROJECTS Seamless view synthesis through te...IEEE 2014 MATLAB IMAGE PROCESSING PROJECTS Seamless view synthesis through te...
IEEE 2014 MATLAB IMAGE PROCESSING PROJECTS Seamless view synthesis through te...
 
IRJET- Exploring Image Super Resolution Techniques
IRJET- Exploring Image Super Resolution TechniquesIRJET- Exploring Image Super Resolution Techniques
IRJET- Exploring Image Super Resolution Techniques
 
Image segmentation using wvlt trnsfrmtn and fuzzy logic. ppt
Image segmentation using wvlt trnsfrmtn and fuzzy logic. pptImage segmentation using wvlt trnsfrmtn and fuzzy logic. ppt
Image segmentation using wvlt trnsfrmtn and fuzzy logic. ppt
 
Fusion of Multi-MAV Data
Fusion of Multi-MAV DataFusion of Multi-MAV Data
Fusion of Multi-MAV Data
 
Whiteboard image reconstruction using matlab
Whiteboard image reconstruction using matlabWhiteboard image reconstruction using matlab
Whiteboard image reconstruction using matlab
 

Plus de Yu Huang

Application of Foundation Model for Autonomous Driving
Application of Foundation Model for Autonomous DrivingApplication of Foundation Model for Autonomous Driving
Application of Foundation Model for Autonomous DrivingYu Huang
 
The New Perception Framework in Autonomous Driving: An Introduction of BEV N...
The New Perception Framework  in Autonomous Driving: An Introduction of BEV N...The New Perception Framework  in Autonomous Driving: An Introduction of BEV N...
The New Perception Framework in Autonomous Driving: An Introduction of BEV N...Yu Huang
 
Data Closed Loop in Simulation Test of Autonomous Driving
Data Closed Loop in Simulation Test of Autonomous DrivingData Closed Loop in Simulation Test of Autonomous Driving
Data Closed Loop in Simulation Test of Autonomous DrivingYu Huang
 
Techniques and Challenges in Autonomous Driving
Techniques and Challenges in Autonomous DrivingTechniques and Challenges in Autonomous Driving
Techniques and Challenges in Autonomous DrivingYu Huang
 
BEV Joint Detection and Segmentation
BEV Joint Detection and SegmentationBEV Joint Detection and Segmentation
BEV Joint Detection and SegmentationYu Huang
 
BEV Object Detection and Prediction
BEV Object Detection and PredictionBEV Object Detection and Prediction
BEV Object Detection and PredictionYu Huang
 
Fisheye based Perception for Autonomous Driving VI
Fisheye based Perception for Autonomous Driving VIFisheye based Perception for Autonomous Driving VI
Fisheye based Perception for Autonomous Driving VIYu Huang
 
Fisheye/Omnidirectional View in Autonomous Driving IV
Fisheye/Omnidirectional View in Autonomous Driving IVFisheye/Omnidirectional View in Autonomous Driving IV
Fisheye/Omnidirectional View in Autonomous Driving IVYu Huang
 
Prediction,Planninng & Control at Baidu
Prediction,Planninng & Control at BaiduPrediction,Planninng & Control at Baidu
Prediction,Planninng & Control at BaiduYu Huang
 
Cruise AI under the Hood
Cruise AI under the HoodCruise AI under the Hood
Cruise AI under the HoodYu Huang
 
LiDAR in the Adverse Weather: Dust, Snow, Rain and Fog (2)
LiDAR in the Adverse Weather: Dust, Snow, Rain and Fog (2)LiDAR in the Adverse Weather: Dust, Snow, Rain and Fog (2)
LiDAR in the Adverse Weather: Dust, Snow, Rain and Fog (2)Yu Huang
 
Scenario-Based Development & Testing for Autonomous Driving
Scenario-Based Development & Testing for Autonomous DrivingScenario-Based Development & Testing for Autonomous Driving
Scenario-Based Development & Testing for Autonomous DrivingYu Huang
 
How to Build a Data Closed-loop Platform for Autonomous Driving?
How to Build a Data Closed-loop Platform for Autonomous Driving?How to Build a Data Closed-loop Platform for Autonomous Driving?
How to Build a Data Closed-loop Platform for Autonomous Driving?Yu Huang
 
Annotation tools for ADAS & Autonomous Driving
Annotation tools for ADAS & Autonomous DrivingAnnotation tools for ADAS & Autonomous Driving
Annotation tools for ADAS & Autonomous DrivingYu Huang
 
Simulation for autonomous driving at uber atg
Simulation for autonomous driving at uber atgSimulation for autonomous driving at uber atg
Simulation for autonomous driving at uber atgYu Huang
 
Prediction and planning for self driving at waymo
Prediction and planning for self driving at waymoPrediction and planning for self driving at waymo
Prediction and planning for self driving at waymoYu Huang
 
Jointly mapping, localization, perception, prediction and planning
Jointly mapping, localization, perception, prediction and planningJointly mapping, localization, perception, prediction and planning
Jointly mapping, localization, perception, prediction and planningYu Huang
 
Data pipeline and data lake for autonomous driving
Data pipeline and data lake for autonomous drivingData pipeline and data lake for autonomous driving
Data pipeline and data lake for autonomous drivingYu Huang
 
Open Source codes of trajectory prediction & behavior planning
Open Source codes of trajectory prediction & behavior planningOpen Source codes of trajectory prediction & behavior planning
Open Source codes of trajectory prediction & behavior planningYu Huang
 
Lidar in the adverse weather: dust, fog, snow and rain
Lidar in the adverse weather: dust, fog, snow and rainLidar in the adverse weather: dust, fog, snow and rain
Lidar in the adverse weather: dust, fog, snow and rainYu Huang
 

Plus de Yu Huang (20)

Application of Foundation Model for Autonomous Driving
Application of Foundation Model for Autonomous DrivingApplication of Foundation Model for Autonomous Driving
Application of Foundation Model for Autonomous Driving
 
The New Perception Framework in Autonomous Driving: An Introduction of BEV N...
The New Perception Framework  in Autonomous Driving: An Introduction of BEV N...The New Perception Framework  in Autonomous Driving: An Introduction of BEV N...
The New Perception Framework in Autonomous Driving: An Introduction of BEV N...
 
Data Closed Loop in Simulation Test of Autonomous Driving
Data Closed Loop in Simulation Test of Autonomous DrivingData Closed Loop in Simulation Test of Autonomous Driving
Data Closed Loop in Simulation Test of Autonomous Driving
 
Techniques and Challenges in Autonomous Driving
Techniques and Challenges in Autonomous DrivingTechniques and Challenges in Autonomous Driving
Techniques and Challenges in Autonomous Driving
 
BEV Joint Detection and Segmentation
BEV Joint Detection and SegmentationBEV Joint Detection and Segmentation
BEV Joint Detection and Segmentation
 
BEV Object Detection and Prediction
BEV Object Detection and PredictionBEV Object Detection and Prediction
BEV Object Detection and Prediction
 
Fisheye based Perception for Autonomous Driving VI
Fisheye based Perception for Autonomous Driving VIFisheye based Perception for Autonomous Driving VI
Fisheye based Perception for Autonomous Driving VI
 
Fisheye/Omnidirectional View in Autonomous Driving IV
Fisheye/Omnidirectional View in Autonomous Driving IVFisheye/Omnidirectional View in Autonomous Driving IV
Fisheye/Omnidirectional View in Autonomous Driving IV
 
Prediction,Planninng & Control at Baidu
Prediction,Planninng & Control at BaiduPrediction,Planninng & Control at Baidu
Prediction,Planninng & Control at Baidu
 
Cruise AI under the Hood
Cruise AI under the HoodCruise AI under the Hood
Cruise AI under the Hood
 
LiDAR in the Adverse Weather: Dust, Snow, Rain and Fog (2)
LiDAR in the Adverse Weather: Dust, Snow, Rain and Fog (2)LiDAR in the Adverse Weather: Dust, Snow, Rain and Fog (2)
LiDAR in the Adverse Weather: Dust, Snow, Rain and Fog (2)
 
Scenario-Based Development & Testing for Autonomous Driving
Scenario-Based Development & Testing for Autonomous DrivingScenario-Based Development & Testing for Autonomous Driving
Scenario-Based Development & Testing for Autonomous Driving
 
How to Build a Data Closed-loop Platform for Autonomous Driving?
How to Build a Data Closed-loop Platform for Autonomous Driving?How to Build a Data Closed-loop Platform for Autonomous Driving?
How to Build a Data Closed-loop Platform for Autonomous Driving?
 
Annotation tools for ADAS & Autonomous Driving
Annotation tools for ADAS & Autonomous DrivingAnnotation tools for ADAS & Autonomous Driving
Annotation tools for ADAS & Autonomous Driving
 
Simulation for autonomous driving at uber atg
Simulation for autonomous driving at uber atgSimulation for autonomous driving at uber atg
Simulation for autonomous driving at uber atg
 
Prediction and planning for self driving at waymo
Prediction and planning for self driving at waymoPrediction and planning for self driving at waymo
Prediction and planning for self driving at waymo
 
Jointly mapping, localization, perception, prediction and planning
Jointly mapping, localization, perception, prediction and planningJointly mapping, localization, perception, prediction and planning
Jointly mapping, localization, perception, prediction and planning
 
Data pipeline and data lake for autonomous driving
Data pipeline and data lake for autonomous drivingData pipeline and data lake for autonomous driving
Data pipeline and data lake for autonomous driving
 
Open Source codes of trajectory prediction & behavior planning
Open Source codes of trajectory prediction & behavior planningOpen Source codes of trajectory prediction & behavior planning
Open Source codes of trajectory prediction & behavior planning
 
Lidar in the adverse weather: dust, fog, snow and rain
Lidar in the adverse weather: dust, fog, snow and rainLidar in the adverse weather: dust, fog, snow and rain
Lidar in the adverse weather: dust, fog, snow and rain
 

Dernier

MANUFACTURING PROCESS-II UNIT-1 THEORY OF METAL CUTTING
MANUFACTURING PROCESS-II UNIT-1 THEORY OF METAL CUTTINGMANUFACTURING PROCESS-II UNIT-1 THEORY OF METAL CUTTING
MANUFACTURING PROCESS-II UNIT-1 THEORY OF METAL CUTTINGSIVASHANKAR N
 
Top Rated Pune Call Girls Budhwar Peth ⟟ 6297143586 ⟟ Call Me For Genuine Se...
Top Rated  Pune Call Girls Budhwar Peth ⟟ 6297143586 ⟟ Call Me For Genuine Se...Top Rated  Pune Call Girls Budhwar Peth ⟟ 6297143586 ⟟ Call Me For Genuine Se...
Top Rated Pune Call Girls Budhwar Peth ⟟ 6297143586 ⟟ Call Me For Genuine Se...Call Girls in Nagpur High Profile
 
KubeKraft presentation @CloudNativeHooghly
KubeKraft presentation @CloudNativeHooghlyKubeKraft presentation @CloudNativeHooghly
KubeKraft presentation @CloudNativeHooghlysanyuktamishra911
 
(SHREYA) Chakan Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Esc...
(SHREYA) Chakan Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Esc...(SHREYA) Chakan Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Esc...
(SHREYA) Chakan Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Esc...ranjana rawat
 
Online banking management system project.pdf
Online banking management system project.pdfOnline banking management system project.pdf
Online banking management system project.pdfKamal Acharya
 
Russian Call Girls in Nagpur Grishma Call 7001035870 Meet With Nagpur Escorts
Russian Call Girls in Nagpur Grishma Call 7001035870 Meet With Nagpur EscortsRussian Call Girls in Nagpur Grishma Call 7001035870 Meet With Nagpur Escorts
Russian Call Girls in Nagpur Grishma Call 7001035870 Meet With Nagpur EscortsCall Girls in Nagpur High Profile
 
UNIT-III FMM. DIMENSIONAL ANALYSIS
UNIT-III FMM.        DIMENSIONAL ANALYSISUNIT-III FMM.        DIMENSIONAL ANALYSIS
UNIT-III FMM. DIMENSIONAL ANALYSISrknatarajan
 
(PRIYA) Rajgurunagar Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
(PRIYA) Rajgurunagar Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...(PRIYA) Rajgurunagar Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
(PRIYA) Rajgurunagar Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...ranjana rawat
 
(ANJALI) Dange Chowk Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
(ANJALI) Dange Chowk Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...(ANJALI) Dange Chowk Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
(ANJALI) Dange Chowk Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...ranjana rawat
 
MANUFACTURING PROCESS-II UNIT-2 LATHE MACHINE
MANUFACTURING PROCESS-II UNIT-2 LATHE MACHINEMANUFACTURING PROCESS-II UNIT-2 LATHE MACHINE
MANUFACTURING PROCESS-II UNIT-2 LATHE MACHINESIVASHANKAR N
 
The Most Attractive Pune Call Girls Budhwar Peth 8250192130 Will You Miss Thi...
The Most Attractive Pune Call Girls Budhwar Peth 8250192130 Will You Miss Thi...The Most Attractive Pune Call Girls Budhwar Peth 8250192130 Will You Miss Thi...
The Most Attractive Pune Call Girls Budhwar Peth 8250192130 Will You Miss Thi...ranjana rawat
 
result management system report for college project
result management system report for college projectresult management system report for college project
result management system report for college projectTonystark477637
 
Call Girls in Nagpur Suman Call 7001035870 Meet With Nagpur Escorts
Call Girls in Nagpur Suman Call 7001035870 Meet With Nagpur EscortsCall Girls in Nagpur Suman Call 7001035870 Meet With Nagpur Escorts
Call Girls in Nagpur Suman Call 7001035870 Meet With Nagpur EscortsCall Girls in Nagpur High Profile
 
Booking open Available Pune Call Girls Koregaon Park 6297143586 Call Hot Ind...
Booking open Available Pune Call Girls Koregaon Park  6297143586 Call Hot Ind...Booking open Available Pune Call Girls Koregaon Park  6297143586 Call Hot Ind...
Booking open Available Pune Call Girls Koregaon Park 6297143586 Call Hot Ind...Call Girls in Nagpur High Profile
 
VIP Call Girls Ankleshwar 7001035870 Whatsapp Number, 24/07 Booking
VIP Call Girls Ankleshwar 7001035870 Whatsapp Number, 24/07 BookingVIP Call Girls Ankleshwar 7001035870 Whatsapp Number, 24/07 Booking
VIP Call Girls Ankleshwar 7001035870 Whatsapp Number, 24/07 Bookingdharasingh5698
 
College Call Girls Nashik Nehal 7001305949 Independent Escort Service Nashik
College Call Girls Nashik Nehal 7001305949 Independent Escort Service NashikCollege Call Girls Nashik Nehal 7001305949 Independent Escort Service Nashik
College Call Girls Nashik Nehal 7001305949 Independent Escort Service NashikCall Girls in Nagpur High Profile
 
Porous Ceramics seminar and technical writing
Porous Ceramics seminar and technical writingPorous Ceramics seminar and technical writing
Porous Ceramics seminar and technical writingrakeshbaidya232001
 
Coefficient of Thermal Expansion and their Importance.pptx
Coefficient of Thermal Expansion and their Importance.pptxCoefficient of Thermal Expansion and their Importance.pptx
Coefficient of Thermal Expansion and their Importance.pptxAsutosh Ranjan
 

Dernier (20)

MANUFACTURING PROCESS-II UNIT-1 THEORY OF METAL CUTTING
MANUFACTURING PROCESS-II UNIT-1 THEORY OF METAL CUTTINGMANUFACTURING PROCESS-II UNIT-1 THEORY OF METAL CUTTING
MANUFACTURING PROCESS-II UNIT-1 THEORY OF METAL CUTTING
 
Top Rated Pune Call Girls Budhwar Peth ⟟ 6297143586 ⟟ Call Me For Genuine Se...
Top Rated  Pune Call Girls Budhwar Peth ⟟ 6297143586 ⟟ Call Me For Genuine Se...Top Rated  Pune Call Girls Budhwar Peth ⟟ 6297143586 ⟟ Call Me For Genuine Se...
Top Rated Pune Call Girls Budhwar Peth ⟟ 6297143586 ⟟ Call Me For Genuine Se...
 
KubeKraft presentation @CloudNativeHooghly
KubeKraft presentation @CloudNativeHooghlyKubeKraft presentation @CloudNativeHooghly
KubeKraft presentation @CloudNativeHooghly
 
(SHREYA) Chakan Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Esc...
(SHREYA) Chakan Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Esc...(SHREYA) Chakan Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Esc...
(SHREYA) Chakan Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Esc...
 
Online banking management system project.pdf
Online banking management system project.pdfOnline banking management system project.pdf
Online banking management system project.pdf
 
Russian Call Girls in Nagpur Grishma Call 7001035870 Meet With Nagpur Escorts
Russian Call Girls in Nagpur Grishma Call 7001035870 Meet With Nagpur EscortsRussian Call Girls in Nagpur Grishma Call 7001035870 Meet With Nagpur Escorts
Russian Call Girls in Nagpur Grishma Call 7001035870 Meet With Nagpur Escorts
 
UNIT-III FMM. DIMENSIONAL ANALYSIS
UNIT-III FMM.        DIMENSIONAL ANALYSISUNIT-III FMM.        DIMENSIONAL ANALYSIS
UNIT-III FMM. DIMENSIONAL ANALYSIS
 
(PRIYA) Rajgurunagar Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
(PRIYA) Rajgurunagar Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...(PRIYA) Rajgurunagar Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
(PRIYA) Rajgurunagar Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
 
(ANJALI) Dange Chowk Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
(ANJALI) Dange Chowk Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...(ANJALI) Dange Chowk Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
(ANJALI) Dange Chowk Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
 
MANUFACTURING PROCESS-II UNIT-2 LATHE MACHINE
MANUFACTURING PROCESS-II UNIT-2 LATHE MACHINEMANUFACTURING PROCESS-II UNIT-2 LATHE MACHINE
MANUFACTURING PROCESS-II UNIT-2 LATHE MACHINE
 
Roadmap to Membership of RICS - Pathways and Routes
Roadmap to Membership of RICS - Pathways and RoutesRoadmap to Membership of RICS - Pathways and Routes
Roadmap to Membership of RICS - Pathways and Routes
 
Water Industry Process Automation & Control Monthly - April 2024
Water Industry Process Automation & Control Monthly - April 2024Water Industry Process Automation & Control Monthly - April 2024
Water Industry Process Automation & Control Monthly - April 2024
 
The Most Attractive Pune Call Girls Budhwar Peth 8250192130 Will You Miss Thi...
The Most Attractive Pune Call Girls Budhwar Peth 8250192130 Will You Miss Thi...The Most Attractive Pune Call Girls Budhwar Peth 8250192130 Will You Miss Thi...
The Most Attractive Pune Call Girls Budhwar Peth 8250192130 Will You Miss Thi...
 
result management system report for college project
result management system report for college projectresult management system report for college project
result management system report for college project
 
Call Girls in Nagpur Suman Call 7001035870 Meet With Nagpur Escorts
Call Girls in Nagpur Suman Call 7001035870 Meet With Nagpur EscortsCall Girls in Nagpur Suman Call 7001035870 Meet With Nagpur Escorts
Call Girls in Nagpur Suman Call 7001035870 Meet With Nagpur Escorts
 
Booking open Available Pune Call Girls Koregaon Park 6297143586 Call Hot Ind...
Booking open Available Pune Call Girls Koregaon Park  6297143586 Call Hot Ind...Booking open Available Pune Call Girls Koregaon Park  6297143586 Call Hot Ind...
Booking open Available Pune Call Girls Koregaon Park 6297143586 Call Hot Ind...
 
VIP Call Girls Ankleshwar 7001035870 Whatsapp Number, 24/07 Booking
VIP Call Girls Ankleshwar 7001035870 Whatsapp Number, 24/07 BookingVIP Call Girls Ankleshwar 7001035870 Whatsapp Number, 24/07 Booking
VIP Call Girls Ankleshwar 7001035870 Whatsapp Number, 24/07 Booking
 
College Call Girls Nashik Nehal 7001305949 Independent Escort Service Nashik
College Call Girls Nashik Nehal 7001305949 Independent Escort Service NashikCollege Call Girls Nashik Nehal 7001305949 Independent Escort Service Nashik
College Call Girls Nashik Nehal 7001305949 Independent Escort Service Nashik
 
Porous Ceramics seminar and technical writing
Porous Ceramics seminar and technical writingPorous Ceramics seminar and technical writing
Porous Ceramics seminar and technical writing
 
Coefficient of Thermal Expansion and their Importance.pptx
Coefficient of Thermal Expansion and their Importance.pptxCoefficient of Thermal Expansion and their Importance.pptx
Coefficient of Thermal Expansion and their Importance.pptx
 

Deep vo and slam ii

  • 1. Deep VO and SLAM II Yu Huang Yu.huang07@gmail.com Sunnyvale, California
  • 2. Outline • Single View Stereo Matching, 3, 2018 • LEGO: Learning Edge with Geometry all at Once by Watching Videos, 3, 2018 • Unsupervised Learning of Monocular Depth Estimation and Visual Odometry with Deep Feature Reconstruction, 4, 2018 • Unsupervised Learning of Depth and Ego- Motion from Monocular Video Using 3D Geometric Constraints, 6, 2018 • Look Deeper into Depth: Monocular Depth Estimation with Semantic Booster and Attention-Driven Loss, ECCV, 2018 • GeoNet: Geometric Neural Network, CVPR 2018 • GeoNet: Unsupervised Learning of Dense Depth, Optical Flow and Camera Pose • Learning Depth from Monocular Videos using Direct Methods, CVPR, 2018 • CNN-SVO: Improving the Mapping in Semi- Direct Visual Odometry Using Single-Image Depth Prediction, 10, 2018 • Depth Prediction Without the Sensors: Leveraging Structure for Unsupervised Learning from Monocular Videos, 11, 2018 • Self-Supervised Learning of Depth and Camera Motion from 360◦ Videos, 11, 2018 • Unsupervised Learning-based Depth Estimation aided Visual SLAM Approach, 1, 2019
  • 3. Single View Stereo Matching • Previous monocular depth estimation methods take a single view and directly regress the expected results. • Though recent advances are made by applying geometrically inspired loss functions during training, the inference procedure does not explicitly impose any geometrical constraint. • Therefore these models purely rely on the quality of data and the effectiveness of learning to generalize. • This either leads to suboptimal results or the demand of huge amount of expensive ground truth labelled data to generate reasonable results. • This paper shows that the monocular depth estimation problem can be reformulated as two sub- problems, a view synthesis procedure followed by stereo matching, with two intriguing properties, namely i) geometrical constraints can be explicitly imposed during inference; ii) demand on labelled depth data can be greatly alleviated. • The whole pipeline can still be trained in an end-to-end fashion and this new formulation plays a critical role in advancing the performance. • The model also generalizes well to other monocular depth estimation benchmarks. • It also discusses the implications and the advantages of solving monocular depth estimation using stereo methods.
  • 6. Single View Stereo Matching “Unsupervised monocular depth estimation with left-right consistency”
  • 7. LEGO: Learning Edge with Geometry all at Once by Watching Videos • Learning to estimate 3D geometry in a single image by watching unlabeled videos via deep convolutional network is attracting significant attention. • This paper introduces a "3D as-smooth-as-possible (3D-ASAP)" prior inside the pipeline, which enables joint estimation of edges and 3D scene, yielding results with significant improvement in accuracy for fine detailed structures. • Specifically, it defines the 3D-ASAP prior by requiring that any two points recovered in 3D from an image should lie on an existing planar surface if no other cues provided. • They design an unsupervised framework that Learns Edges and Geometry (depth, normal) all at Once (LEGO). • The predicted edges are embedded into depth and surface normal smoothness terms, where pixels without edges in-between are constrained to satisfy the prior. • In the framework, the predicted depths, normals and edges are forced to be consistent all the time.
  • 8. LEGO: Learning Edge with Geometry all at Once by Watching Videos
  • 9. LEGO: Learning Edge with Geometry all at Once by Watching Videos
  • 10. Unsupervised Learning of Depth and Ego-Motion from Monocular Video Using 3D Geometric Constraints • It is an approach for unsupervised learning of depth and ego-motion from monocular video. • Unsupervised learning removes the need for separate supervisory signals (depth or ego-motion ground truth, or multi-view video). • Prior work in unsupervised depth learning uses pixel-wise or gradient-based losses, which only consider pixels in small local neighborhoods. • This idea is to explicitly consider the inferred 3D geometry of the whole scene, and enforce consistency of the estimated 3D point clouds and ego-motion across consecutive frames. • This is a challenging task and is solved by a novel (approximate) backpropagation algorithm for aligning 3D structures. • They combine this 3D-based loss with 2D losses based on photometric quality of frame reconstructions using estimated depth and ego-motion from adjacent frames. • It also incorporates validity masks to avoid penalizing areas in which no useful information exists. • Because they only require a simple video, learning depth and ego-motion on large and varied datasets becomes possible. • Codes are available at https://sites.google.com/view/vid2depth
  • 11. Unsupervised Learning of Depth and Ego-Motion from Monocular Video Using 3D Geometric Constraints
  • 12. Unsupervised Learning of Depth and Ego-Motion from Monocular Video Using 3D Geometric Constraints
  • 13. Unsupervised Learning of Depth and Ego-Motion from Monocular Video Using 3D Geometric Constraints
  • 14. Unsupervised Learning of Depth and Ego-Motion from Monocular Video Using 3D Geometric Constraints “Unsupervised cnn for single view depth estimation: Geometry to the rescue” “Unsupervised learning of depth and ego-motion from video”
  • 15. Look Deeper into Depth: Monocular Depth Estimation with Semantic Booster and Attention-Driven Loss • Monocular depth estimation benefits greatly from learning based techniques. • By studying the training data, they observe that the per-pixel depth values in existing datasets typically exhibit a long-tailed distribution. • However, most previous approaches treat all the regions in the training data equally regardless of the imbalanced depth distribution, which restricts the model performance particularly on distant depth regions. • This paper investigates the long tail property and delve deeper into the distant depth regions (i.e. the tail part) to propose an attention driven loss for the network supervision. • In addition, to better leverage the semantic information for monocular depth estimation, it proposes a synergy network to automatically learn the information sharing strategies between the two tasks. • With the proposed attention-driven loss and synergy network, the depth estimation and semantic labeling tasks can be mutually improved.
  • 16. Look Deeper into Depth: Monocular Depth Estimation with Semantic Booster and Attention-Driven Loss
  • 17. Look Deeper into Depth: Monocular Depth Estimation with Semantic Booster and Attention-Driven Loss “Deep convolutional neural fields for depth estimation from a single image” “Depth map prediction from a single image using a multi-scale deep network” “Deeper depth prediction with fully convolutional residual networks” “Depth map prediction from a single image using a multi-scale deep network”
  • 18. GeoNet: Geometric Neural Network 18 • This paper proposes Geometric Neural Network (GeoNet) to jointly predict depth and surface normal maps from a single image. • Building on top of two-stream CNNs, the GeoNet incorporates geometric relation between depth and surface normal via the new depth-to-normal and normal-to- depth networks. • Depth-to-normal network exploits the least square solution of surface normal from depth and improves its quality with a residual module. • Normal-to-depth network, contrarily, refines the depth map based on the constraints from the surface normal through a kernel regression module, which has no parameter to learn. • These two networks enforce the underlying model to efficiently predict depth and surface normal for high consistency and corresponding accuracy.
  • 23. GeoNet--Unsupervised Learning of Dense Depth, Optical Flow and Camera Pose • GeoNet, a jointly unsupervised learning framework for monocular depth, optical flow and ego-motion estimation from videos. • The three components are coupled by the nature of 3D scene geometry, jointly learned by our framework in an end-to-end manner. • Specifically, geometric relationships are extracted over the predictions of individual modules and then combined as an image reconstruction loss, reasoning about static and dynamic scene parts separately. • Furthermore, they propose an adaptive geometric consistency loss to increase robustness towards outliers and non-Lambertian regions, which resolves occlusions and texture ambiguities effectively. 23
  • 24. GeoNet--Unsupervised Learning of Dense Depth, Optical Flow and Camera Pose 24
  • 25. GeoNet--Unsupervised Learning of Dense Depth, Optical Flow and Camera Pose 25
  • 26. GeoNet--Unsupervised Learning of Dense Depth, Optical Flow and Camera Pose 26
  • 27. Unsupervised Learning of Monocular Depth Estimation and Visual Odometry with Deep Feature Reconstruction • Despite learning based methods showing promising results in single view depth estimation and visual odometry, most existing approaches treat the tasks in a supervised manner. • Recent approaches to single view depth estimation explore the possibility of learning without full supervision via minimizing photometric error. • This paper explores the use of stereo sequences for learning depth and visual odometry. • The use of stereo sequences enables the use of both spatial (between left-right pairs) and temporal (forward backward) photometric warp error, and constrains the scene depth and camera motion to be in a common, real world scale. • At test time the framework is able to estimate single view depth and two-view odometry from a monocular sequence. • They improve on a standard photometric warp loss by considering a warp of deep features. • The source code is available at https://github.com/Huangying-Zhan/Depth-VO-Feat.
  • 28. Unsupervised Learning of Monocular Depth Estimation and Visual Odometry with Deep Feature Reconstruction
  • 29. Unsupervised Learning of Monocular Depth Estimation and Visual Odometry with Deep Feature Reconstruction
  • 30. Unsupervised Learning of Monocular Depth Estimation and Visual Odometry with Deep Feature Reconstruction
  • 31. Learning Depth from Monocular Videos using Direct Methods • The ability to predict depth from a single image - using recent advances in CNNs - is of increasing interest to the vision community. • Unsupervised strategies to learning are particularly appealing as they can utilize much larger and varied monocular video datasets during learning without the need for ground truth depth or stereo. • In previous works, separate pose and depth CNN predictors had to be determined such that their joint outputs minimized the photometric error. • Inspired by recent advances in direct visual odometry (DVO), it argues that the depth CNN predictor can be learned without a pose CNN predictor. • Further, they demonstrate empirically that incorporation of a differentiable implementation of DVO, along with a novel depth normalization strategy - substantially improves performance over state of the art that use monocular videos for training. • https://github.com/MightyChaos/LKVOLearner
  • 32. Learning Depth from Monocular Videos using Direct Methods Unsupervised learning pipeline. The learning algorithm takes 3 sequential images at a time. The Depth-CNN produces 3 inverse depth maps for the inputs, and the pose predictor estimates two relative camera pose between the second image and the other two. The appearance dissimilarity loss is measured between the second image I2 and the inversely warped images of I1, I3; In addition, the loss is evaluated in a reverse direction- it is also measured between I1, I3 and two warped images of I2. Lower part illustrates 3 architectures we evaluated for pose prediction: 1) Pose-CNN, 2) use the proposed differentiable Direct Visual Odometry (DDVO), the initialization of pose is set as zero (identity transformation), and 3) a hybrid of the above two - use pretrained Pose-CNN to give a better initial pose for DDVO.
  • 33. Learning Depth from Monocular Videos using Direct Methods “Unsupervised learning of depth and ego-motion from video”
  • 34. CNN-SVO: Improving the Mapping in Semi-Direct Visual Odometry Using Single-Image Depth Prediction • Reliable feature correspondence between frames is a critical step in visual odometry (VO) and visual simultaneous localization and mapping (V-SLAM) algorithms. • In comparison with existing VO and V-SLAM algorithms, semi-direct visual odometry (SVO) has two main advantages that lead to state-of-the-art frame rate camera motion estimation: direct pixel correspondence and efficient implementation of probabilistic mapping method. • This paper improves the SVO mapping by initializing the mean and the variance of the depth at a feature location according to the depth prediction from a single image depth prediction network. • By significantly reducing the depth uncertainty of the initialized map point (i.e., small variance centered about the depth prediction), the benefits are twofold: reliable feature correspondence between views and fast convergence to the true depth in order to create new map points.
  • 35. CNN-SVO: Improving the Mapping in Semi-Direct Visual Odometry Using Single-Image Depth Prediction map point initialization
  • 36. CNN-SVO: Improving the Mapping in Semi-Direct Visual Odometry Using Single-Image Depth Prediction
  • 37. CNN-SVO: Improving the Mapping in Semi-Direct Visual Odometry Using Single-Image Depth Prediction Camera motion estimation in the HDR environment Camera trajectory and map points
  • 38. CNN-SVO: Improving the Mapping in Semi-Direct Visual Odometry Using Single-Image Depth Prediction
  • 39. Depth Prediction Without the Sensors: Leveraging Structure for Unsupervised Learning from Monocular Videos • Learning to predict depth from RGB is challenging both for indoor and outdoor robot navigation. • This work addresses unsupervised learning of scene depth and robot ego-motion where supervision is provided by monocular videos, as cameras are the cheapest, least restrictive and most ubiquitous sensor for robotics. • Previous unsupervised image-to-depth learning has established strong baselines in the domain. • This approach is able to model moving objects and is shown to transfer across data domains, e.g. from outdoors to indoor scenes. • The main idea is to introduce geometric structure in the learning process, by modeling the scene and the individual objects; camera ego-motion and object motions are learned from monocular videos as input. • An online refinement method is introduced to adapt learning on the fly to unknown domains. • The code can be found at https://sites.google.com/view/struct2depth.
  • 40. Depth Prediction Without the Sensors: Leveraging Structure for Unsupervised Learning from Monocular Videos
  • 41. Depth Prediction Without the Sensors: Leveraging Structure for Unsupervised Learning from Monocular Videos
  • 42. Depth Prediction Without the Sensors: Leveraging Structure for Unsupervised Learning from Monocular Videos
  • 43. Self-Supervised Learning of Depth and Camera Motion from 360◦ Videos • As 360◦ cameras become prevalent in many autonomous systems (e.g., self-driving cars and drones), efficient 360◦ perception becomes more and more important. • This is a self-supervised learning approach for predicting the omnidirectional depth and camera motion from a 360◦ video. • In particular, starting from the SfMLearner, which is designed for cameras with normal field-of- view, they introduce three key features to process 360 ◦ images efficiently. • convert each image from equirectangular projection to cubic projection in order to avoid image distortion. In each network layer, use Cube Padding (CP), which pads intermediate features from adjacent faces, to avoid image boundaries. • apply a “spherical” photometric consistency constraint on the whole viewing sphere. In this way, no pixel will be projected outside the image boundary with normal field-of-view. • rather than estimating 6 independent camera motions (i.e., SfM-Learner to each face on a cube), apply camera pose consistency loss to ensure the estimated camera motions reaching consensus. • They collect a PanoSUNCG dataset containing a large amount of 360◦ videos with ground truth depth and camera motion.
  • 44. Self-Supervised Learning of Depth and Camera Motion from 360◦ Videos
  • 45. Self-Supervised Learning of Depth and Camera Motion from 360◦ Videos Spherical photometric constraints.
  • 46. Self-Supervised Learning of Depth and Camera Motion from 360◦ Videos
  • 47. Unsupervised Learning-based Depth Estimation aided Visual SLAM Approach • Recently, deep learning technologies have achieved great success in the visual SLAM area, which can directly learn high-level features from the visual inputs and improve the estimation accuracy of the depth information. • Therefore, deep learning technologies maintain the potential to extend the source of the depth information and improve the performance of the SLAM system. • However, the existing deep learning-based methods are mainly supervised and require a large amount of ground-truth depth data, which is hard to acquire because of the realistic constraints. • This paper presents an unsupervised learning framework, which not only uses image reconstruction for supervising but also exploits the pose estimation method to enhance the supervised signal and add training constraints for the task of monocular depth and camera motion estimation. • Furthermore, it exploits the unsupervised learning framework to assist the traditional ORB-SLAM system when the initialization module of ORB-SLAM method could not match enough features. • Unsupervised learning framework could significantly accelerate the initialization process of ORB- SLAM system and effectively improve the accuracy on environmental mapping in strong lighting and weak texture scenes.
  • 48. Unsupervised Learning-based Depth Estimation aided Visual SLAM Approach
  • 49. Unsupervised Learning-based Depth Estimation aided Visual SLAM Approach
  • 50. Unsupervised Learning-based Depth Estimation aided Visual SLAM Approach