Please use this identifier to cite or link to this item: http://bura.brunel.ac.uk/handle/2438/31532
Full metadata record
DC FieldValueLanguage
dc.contributor.authorSong, B-
dc.contributor.authorZhao, S-
dc.contributor.authorWang, Z-
dc.contributor.authorLiu, W-
dc.contributor.authorLiu, X-
dc.date.accessioned2025-07-10T14:11:13Z-
dc.date.available2025-07-10T14:11:13Z-
dc.date.issued2025-05-27-
dc.identifierORCiD: Baoye Song https://orcid.org/0000-0003-1631-5237-
dc.identifierORCiD: Zidong Wang https://orcid.org/0000-0002-9576-7401-
dc.identifierORCiD: Weibo Liu https://orcid.org/0000-0002-8169-3261-
dc.identifierORCiD: Xiaohui Liu https://orcid.org/0000-0003-1589-1267-
dc.identifierArticle number: 113760-
dc.identifier.citationSong, B. et al. (2025) 'DAF-DETR: A dynamic adaptation feature transformer for enhanced object detection in unmanned aerial vehicles', Knowledge Based Systems, 323, 113760, pp. 1 - 13. doi: 10.1016/j.knosys.2025.113760.en_US
dc.identifier.issn0950-7051-
dc.identifier.urihttps://bura.brunel.ac.uk/handle/2438/31532-
dc.descriptionData availability: Data will be made available on request.en_US
dc.description.abstractObject detection in complex environments is challenged by overlapping objects, complex spatial relationships, and dynamic variations in target scales. To address these challenges, the Dynamic Adaptation Feature DEtection TRansformer (DAF-DETR) is proposed as a novel transformer-based model optimized for real-time detection in spatially complex environments. The framework introduces four key innovations. First, a learnable position encoding mechanism is employed in place of fixed positional encoding, enhancing adaptability and flexibility when processing complex spatial layouts. Second, the Resynthetic Network (ResynNet) backbone, which consists of stacked Resynthetic Blocks (ResynBlocks) integrating ResBlock and FasterBlock feature extraction strategies, is designed to optimize multi-scale feature representation and improve computational efficiency. Third, an enhanced feature fusion module is incorporated to strengthen the detection of small, densely packed objects by integrating multi-scale contextual information. Fourth, a dynamic perception module is introduced, utilizing deformable attention to capture complex spatial relationships between overlapping objects. Extensive experiments conducted on the Vision meets Drone 2019 (VisDrone2019) and Tiny Object Detection in Aerial Images (AI-TOD) datasets demonstrate the superiority of DAF-DETR, achieving state-of-the-art detection accuracy while maintaining real-time efficiency. The results confirm its robustness in handling scale variations, occlusions, and spatial complexity, establishing it as a reliable solution for real-world applications such as aerial imagery and crowded scene analysis.en_US
dc.description.sponsorshipThis work was supported in part by the Natural Science Foundation of Shandong Province of China under Grant ZR2023MF067, the Royal Society of the UK, and the Alexander von Humboldt Foundation of Germany .en_US
dc.format.extent1 - 13-
dc.format.mediumPrint-Electronic-
dc.languageEnglish-
dc.language.isoen_USen_US
dc.publisherElsevieren_US
dc.rightsCreative Commons Attribution-NonCommercial-NoDerivatives 4.0 International-
dc.rights.urihttps://creativecommons.org/licenses/by-nc-nd/4.0/-
dc.subjectobject detectionen_US
dc.subjectcomplex environmentsen_US
dc.subjecttiny object detectionen_US
dc.subjecttransformeren_US
dc.subjectunmanned aerial vehiclesen_US
dc.titleDAF-DETR: A dynamic adaptation feature transformer for enhanced object detection in unmanned aerial vehiclesen_US
dc.typeArticleen_US
dc.date.dateAccepted2025-05-11-
dc.identifier.doihttps://doi.org/10.1016/j.knosys.2025.113760-
dc.relation.isPartOfKnowledge Based Systems-
pubs.publication-statusPublished-
pubs.volume323-
dc.identifier.eissn1872-7409-
dc.rights.licensehttps://creativecommons.org/licenses/by-nc-nd/4.0/legalcode.en-
dcterms.dateAccepted2025-05-11-
dc.rights.holderElsevier B.V.-
Appears in Collections:Dept of Computer Science Embargoed Research Papers

Files in This Item:
File Description SizeFormat 
FullText.pdfEmbargoed until 27 May 2026. Copyright © 2025 Elsevier B.V. All rights reserved. This manuscript version is made available under the CC-BY-NC-ND 4.0 license https://creativecommons.org/licenses/by-nc-nd/4.0/ (see: https://www.elsevier.com/about/policies/sharing).9.31 MBAdobe PDFView/Open


This item is licensed under a Creative Commons License Creative Commons