Understanding human motion and gestures for underwater human–robot collaboration

In this paper, we present a number of robust methodologies for an underwater robot to visually detect, follow, and interact with a diver for collaborative task execution. We design and develop two autonomous diver‐following algorithms, the first of which utilizes both spatial‐ and frequency‐domain f...

Full description

Saved in:
Bibliographic Details
Published inJournal of field robotics Vol. 36; no. 5; pp. 851 - 873
Main Authors Islam, Md Jahidul, Ho, Marc, Sattar, Junaed
Format Journal Article
LanguageEnglish
Published Hoboken Wiley Subscription Services, Inc 01.08.2019
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:In this paper, we present a number of robust methodologies for an underwater robot to visually detect, follow, and interact with a diver for collaborative task execution. We design and develop two autonomous diver‐following algorithms, the first of which utilizes both spatial‐ and frequency‐domain features pertaining to human swimming patterns to visually track a diver. The second algorithm uses a convolutional neural network‐based model for robust tracking‐by‐detection. In addition, we propose a hand gesture‐based human–robot communication framework that is syntactically simpler and computationally more efficient than the existing grammar‐based frameworks. In the proposed interaction framework, deep visual detectors are used to provide accurate hand gesture recognition; subsequently, a finite‐state machine performs robust and efficient gesture‐to‐instruction mapping. The distinguishing feature of this framework is that it can be easily adopted by divers for communicating with underwater robots without using artificial markers or requiring memorization of complex language rules. Furthermore, we validate the performance and effectiveness of the proposed methodologies through a number of field experiments in closed‐ and open‐water environments. Finally, we perform a user interaction study to demonstrate the usability benefits of our proposed interaction framework compared to the existing methods.
Bibliography:ObjectType-Article-1
SourceType-Scholarly Journals-1
ObjectType-Feature-2
content type line 14
ISSN:1556-4959
1556-4967
DOI:10.1002/rob.21837