Magnitude-Orientation Stream network and depth information applied to activity recognition
•Magnitude by the depth weighting to circumvent problems related to camera distance.•Detailed revision of the literature with recently published works.•A study regarding the behavior of Magnitude-Orientation Stream (MOS). The temporal component of videos provides an important clue for activity recog...
Saved in:
Published in | Journal of visual communication and image representation Vol. 63; p. 102596 |
---|---|
Main Authors | , , , , |
Format | Journal Article |
Language | English |
Published |
Elsevier Inc
01.08.2019
Elsevier |
Subjects | |
Online Access | Get full text |
ISSN | 1047-3203 1095-9076 |
DOI | 10.1016/j.jvcir.2019.102596 |
Cover
Loading…
Summary: | •Magnitude by the depth weighting to circumvent problems related to camera distance.•Detailed revision of the literature with recently published works.•A study regarding the behavior of Magnitude-Orientation Stream (MOS).
The temporal component of videos provides an important clue for activity recognition, as a number of activities can be reliably recognized based on the motion information. In view of that, this work proposes a novel temporal stream for two-stream convolutional networks based on images computed from the optical flow magnitude and orientation, named Magnitude-Orientation Stream (MOS), to learn the motion in a better and richer manner. Our method applies simple non-linear transformations on the vertical and horizontal components of the optical flow to generate input images for the temporal stream. Moreover, we also employ depth information to use as a weighting scheme on the magnitude information to compensate the distance of the subjects performing the activity to the camera. Experimental results, carried on two well-known datasets (UCF101 and NTU), demonstrate that using our proposed temporal stream as input to existing neural network architectures can improve their performance for activity recognition. Results demonstrate that our temporal stream provides complementary information able to improve the classical two-stream methods, indicating the suitability of our approach to be used as a temporal video representation. |
---|---|
ISSN: | 1047-3203 1095-9076 |
DOI: | 10.1016/j.jvcir.2019.102596 |