Title:
Scene flow for autonomous navigation

Thumbnail Image
Author(s)
Dedhia, Vaibhav
Authors
Advisor(s)
Vela, Patricio A.
AlRegib, Ghassan
Davenport, Mark A.
Advisor(s)
Editor(s)
Associated Organization(s)
Series
Supplementary to
Abstract
Today, there are various different paradigms for vision based autonomous navigation: mediated perception approaches that parse an entire scene to make driving decision, a direct perception approach that estimates the affordance of driving that maps an input image to small number of key perception indicators that directly relate to the affordance of road/traffic state for driving. Also, deep learning models trained for specific tasks such as obstacle classification, detecting drivable spaces have been used as modules for autonomous navigation of vehicles. Recent applications of deep learning to navigation have generated end-to-end navigation solutions whereby visual sensor input is mapped to control signals or to motion primitives. It is accepted that these solutions cannot provide the same level of performance as a global planner. However, it is less clear how such end-to-end systems should be integrated into a full navigation pipeline. We evaluate the typical end-to-end solution within a full navigation pipeline in order to expose its weaknesses. Doing so illuminates how to better integrate deep learning methods into the navigation pipeline. For the thesis, we evaluate global path planning using sampling based path planning algorithms. Global planners assume that the world is static and location of obstacle is known. However, for autonomous navigation scenerio, this assumption does not hold true. A need arises to be able to detect the obstacles in the scene, localize them and then make appropriate changes to the decisions for navigation. We train Convolutional Neural Network based deep networks for object recognition that are very effective for detecting the objects in the scene such as vehicles, pedestrian etc. We also propose methods to track the objects in the scene in three dimensions thus effectively localizing the objects in the scene.
Sponsor
Date Issued
2018-04-30
Extent
Resource Type
Text
Resource Subtype
Thesis
Rights Statement
Rights URI