Please use this identifier to cite or link to this item: http://bura.brunel.ac.uk/handle/2438/23428
Full metadata record
DC FieldValueLanguage
dc.contributor.authorZhang, C-
dc.contributor.authorChen, P-
dc.contributor.authorLei, T-
dc.contributor.authorWu, Y-
dc.contributor.authorMeng, H-
dc.date.accessioned2021-11-01T11:24:22Z-
dc.date.available2021-11-01T11:24:22Z-
dc.date.issued2021-10-07-
dc.identifierORCiD: Hongying Meng https://orcid.org/0000-0002-8836-1382-
dc.identifier.citationZhang, C. et al. (2022) 'What-Where-When Attention Network for video-based person re-identification', Neurocomputing, 468, pp. 33-47. doi: 10.1016/j.neucom.2021.10.018.en_US
dc.identifier.issn0925-2312-
dc.identifier.urihttps://bura.brunel.ac.uk/handle/2438/23428-
dc.description.abstractVideo-based person re-identification plays a critical role in intelligent video surveillance by learning temporal correlations from consecutive video frames. Most existing methods aim to solve the challenging variations of pose, occlusion, backgrounds and so on by using attention mechanism. They almost all draw attention to the occlusion and learn occlusion-invariant video representations by abandoning the occluded area or frames, while the other areas in these frames contain sufficient spatial information and temporal cues. To overcome these drawbacks, this paper proposes a comprehensive attention mechanism covering what, where, and when to pay attention in the discriminative spatial-temporal feature learning, namely What-Where-When Attention Network (W3AN). Concretely, W3AN designs a spatial attention module to focus on pedestrian identity and obvious attributes by the importance estimating layer (What and Where), and a temporal attention module to calculate the frame-level importance (when), which is embedded into a graph attention network to exploit temporal attention features rather than computing weighted average feature for video frames like existing methods. Moreover, the experiments on three widely-recognized datasets demonstrate the effectiveness of our proposed W3AN model and the discussion of major modules elaborates the contributions of this paper.-
dc.description.sponsorshipThis work was supported in part by the National Natural Science Foundation of China (61801437, 61871351, 61971381); Natural Science Foundation of Shanxi Province (201801D221206, 201801D221207); Scientific and Technological Innovation Programs of Higher Education Institutions in Shanxi (2020L0683); The National Natural Science Foundation of China under Grant 61461025, Grant 61871259, Grant 61811530325 (IECnNSFCn170396, Royal Society, U.K.), and Grant 61861024; The Key research and development plan of Luliang City (2020GXZDYF21).en_US
dc.format.extent33 - 47 (15)-
dc.format.mediumPrint-Electronic-
dc.language.isoen_USen_US
dc.publisherElsevieren_US
dc.rightsAttribution-NonCommercial-NoDerivatives 4.0 International-
dc.rights.urihttps://creativecommons.org/licenses/by-nc-nd/4.0/-
dc.subjectperson re-identificationen_US
dc.subjectwhat-where-whenen_US
dc.subjectattentionen_US
dc.subjectspatial-temporal featureen_US
dc.subjectgraph attention networken_US
dc.subjectattributeen_US
dc.subjectidentityen_US
dc.titleWhat-Where-When Attention Network for Video-based Person Re-identification Neurocomputingen_US
dc.typeArticleen_US
dc.identifier.doihttps://doi.org/10.1016/j.neucom.2021.10.018-
dc.relation.isPartOfNeurocomputing-
pubs.publication-statusPublished-
pubs.volume468-
dc.identifier.eissn1872-8286-
dc.rights.licensehttps://creativecommons.org/licenses/by-nc-nd/4.0/legalcode.en-
dcterms.dateAccepted2021-10-02-
dc.rights.holderElsevier B.V.-
Appears in Collections:Dept of Electronic and Electrical Engineering Research Papers

Files in This Item:
File Description SizeFormat 
FullText.pdfCopyright © 2021 Elsevier B.V. All rights reserved. This manuscript version is made available under the CC-BY-NC-ND 4.0 license https://creativecommons.org/licenses/by-nc-nd/4.0/ (see: https://www.elsevier.com/about/policies/sharing).32.95 MBAdobe PDFView/Open


This item is licensed under a Creative Commons License Creative Commons