Title:
Investigating the Impact of Estimated Modalities in Multi-Modal Activity Recognition

Thumbnail Image
Author(s)
Rajan, Rahul
Authors
Advisor(s)
Advisor(s)
Editor(s)
Associated Organization(s)
Organizational Unit
Organizational Unit
Supplementary to
Abstract
RGB-D data obtained from affordable depth-sensors, like the XBox Kinect has allowed for remarkable progress in the field of human activity recognition (HAR). Depth information has been found to significantly increase performance in HAR tasks, especially when it’s fused with other modalities like RGB and Optical flow. Unfortu- nately, the use of depth sensors limits where these models can be used since these sensors are often difficult to use in outdoor settings. Additionally, most videos available today are shot on traditional video cameras, which don’t provide depth information needed to run RGB-D based HAR models. Fortunately, deep learning has al- lowed us to estimate this depth data with high accuracy from just RGB video. This paper investigates the viability of directly using this estimated depth information in RGB-D models for HAR-related tasks.
Sponsor
Date Issued
2021-12
Extent
Resource Type
Text
Resource Subtype
Undergraduate Thesis
Rights Statement
Rights URI