Video description method for action detection based on part-of-speech information

The invention belongs to the field of computer vision, and provides a video description method for action detection based on part-of-speech information, which comprises the following steps of: (1) extracting video features, (2) extracting scene features, (3) constructing a progressive part-of-speech...

Full description

Saved in:
Bibliographic Details
Main Authors JIANG XINYI, REN DONG
Format Patent
LanguageChinese
English
Published 09.08.2022
Subjects
Online AccessGet full text

Cover

More Information
Summary:The invention belongs to the field of computer vision, and provides a video description method for action detection based on part-of-speech information, which comprises the following steps of: (1) extracting video features, (2) extracting scene features, (3) constructing a progressive part-of-speech decoder, (4) constructing a description generation decoder, and (5) generating a video description statement. According to the method, a part-of-speech information detection module composed of a graph convolutional neural network and a progressive part-of-speech decoder is utilized to judge the global dependency relationship among a plurality of objects in a video so as to determine subject objects, and more accurate part-of-speech information is obtained by combining subject categories and video dynamic feature prediction actions. And finally, a decoder is generated through grammar-guided description, and a more accurate scene description statement is generated. 本发明属于计算机视觉领域,提供了一种基于词性信息的动作检测的视频描述方法,包括:步骤(1)提取视频特征
Bibliography:Application Number: CN202210478744