An intelligent correlation learning system for person Re-identification

Person re-identification (PRe-id) aims to retrieve a target person's images captured across multiple/single non-overlapping cameras. To this end, significant techniques have been implemented that extract handcrafted, deep, part-based, and ensemble features to get more refined patterns for match...

Full description

Saved in:
Bibliographic Details
Published inEngineering applications of artificial intelligence Vol. 128; p. 107213
Main Authors Khan, Samee Ullah, Khan, Noman, Hussain, Tanveer, Baik, Sung Wook
Format Journal Article
LanguageEnglish
Published Elsevier Ltd 01.02.2024
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:Person re-identification (PRe-id) aims to retrieve a target person's images captured across multiple/single non-overlapping cameras. To this end, significant techniques have been implemented that extract handcrafted, deep, part-based, and ensemble features to get more refined patterns for matching. But due to the limited focus on the multi-grained, view-consistent, and semantic correlation among different views, these approaches show low performance. Therefore, we present an attention-based multi-view correlation learning framework named (ACLS), which enables to learn multi-grained spatiotemporal features from individuals. The ACLS is mainly composed of three key steps: First, multi-view correlated visual features of pedestrians are extracted using a correlation vision transformer (CVIT) and a pyramid dilated network (PDN), followed by the person attention mechanism. Next, we employ convolutional long short-term memory (ConvLSTM) to extract spatiotemporal information from pedestrian images captured in different time frames. Finally, a deep fusion strategy is employed, which intelligently integrates features for the final matching task. Extensive evaluations are conducted over three famous datasets: Market-1501, DukeMCMT-reID, and CUHK03, results show tremendous ranking performance including 93.7%, 90.4%, and 85.7%. Thus, concluded remarks that our learning mechanism beats the current state-of-the-art (SOTA) methods.
ISSN:0952-1976
1873-6769
DOI:10.1016/j.engappai.2023.107213