Violence 4D: Violence detection in surveillance using 4D convolutional neural networks

As violence has increased around the world, surveillance cameras are everywhere, and they are only going to get more ubiquitous. Due to the massive volume of video footage, automatic activity detection systems must be used to create an online warning in the event of aberrant activity. A deep learnin...

Full description

Saved in:
Bibliographic Details
Published inIET computer vision Vol. 17; no. 3; pp. 282 - 294
Main Authors Magdy, Mai, Fakhr, Mohamed Waleed, Maghraby, Fahima A.
Format Journal Article
LanguageEnglish
Published Stevenage John Wiley & Sons, Inc 01.04.2023
Wiley
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:As violence has increased around the world, surveillance cameras are everywhere, and they are only going to get more ubiquitous. Due to the massive volume of video footage, automatic activity detection systems must be used to create an online warning in the event of aberrant activity. A deep learning architecture is presented in this study using four‐dimensional video‐level convolution neural networks. The proposed architecture includes residual blocks that are used with three‐Dimensional Convolution Neural Networks 3D (CNNs) to learn long‐term and short‐term spatiotemporal representation from the video as well as record inter‐clip interaction. ResNet50 is used as the backbone for three‐dimensional convolution networks and dense optical flow for the region of interest. The proposed architecture is applied on four benchmarks for violence and non‐violence videos, which are commonly used for violent detection. It obtained test accuracies of 94.67% on RWF2000, 97.29% on Crowd violence, 100% on Movie fight and 100% on the Hockey Fight dataset. These results outperform the previous methods used on RWF2000 datasets. Videos, which are commonly used for violent detection. A deep learning architecture is presented in this work using four‐dimensional video‐level convolution neural networks. The proposed architecture includes residual blocks that are used with three‐Dimensional Convolution Neural Networks 3D (CNNs) to learn long‐term and short‐term spatiotemporal representation from the video as well as record inter‐clip interaction.
Bibliography:ObjectType-Article-1
SourceType-Scholarly Journals-1
ObjectType-Feature-2
content type line 14
ISSN:1751-9632
1751-9640
DOI:10.1049/cvi2.12162