Please use this identifier to cite or link to this item: http://bura.brunel.ac.uk/handle/2438/29480
Full metadata record
DC FieldValueLanguage
dc.contributor.authorSun, R-
dc.contributor.authorLei, T-
dc.contributor.authorWan, Y-
dc.contributor.authorXia, Y-
dc.contributor.authorNandi, AK-
dc.date.accessioned2024-08-02T13:09:58Z-
dc.date.available2024-08-02T13:09:58Z-
dc.date.issued2023-12-20-
dc.identifierORCiD: Tao Lei https://orcid.org/0000-0002-2104-9298-
dc.identifierORCiD: Asoke K. Nandi https://orcid.org/0000-0001-6248-2875-
dc.identifierarXiv:2306.04086v3 [eess.IV]-
dc.identifier.citationSun, R. et al. (2023) 'TEC-Net: Vision Transformer Embrace Convolutional Neural Networks for Medical Image Segmentation', arXiv:2306.04086v3 [eess.IV] (preprint), pp. 1 - 12. doi: 10.48550/arXiv.2306.04086.en_US
dc.identifier.urihttps://bura.brunel.ac.uk/handle/2438/29480-
dc.descriptionThe code is publicly available at https://github.com/SR0920/TEC-Net .en_US
dc.descriptionComments: arXiv admin note: substantial text overlap with arXiv:2306.03373, https://arxiv.org/abs/2306.03373 .-
dc.description.abstractThe hybrid architecture of convolution neural networks (CNN) and Transformer has been the most popular method for medical image segmentation. However, the existing networks based on the hybrid architecture suffer from two problems. First, although the CNN branch can capture image local features by using convolution operation, the vanilla convolution is unable to achieve adaptive extraction of image features. Second, although the Transformer branch can model the global information of images, the conventional self-attention only focuses on the spatial self-attention of images and ignores the channel and cross-dimensional self-attention leading to low segmentation accuracy for medical images with complex backgrounds. To solve these problems, we propose vision Transformer embrace convolutional neural networks for medical image segmentation (TEC-Net). Our network has two advantages. First, dynamic deformable convolution (DDConv) is designed in the CNN branch, which not only overcomes the difficulty of adaptive feature extraction using fixed-size convolution kernels, but also solves the defect that different inputs share the same convolution kernel parameters, effectively improving the feature expression ability of CNN branch. Second, in the Transformer branch, a (shifted)-window adaptive complementary attention module ((S)W-ACAM) and compact convolutional projection are designed to enable the network to fully learn the cross-dimensional long-range dependency of medical images with few parameters and calculations. Experimental results show that the proposed TEC-Net provides better medical image segmentation results than SOTA methods including CNN and Transformer networks. In addition, our TEC-Net requires fewer parameters and computational costs and does not rely on pre-training.en_US
dc.format.extent1 - 12-
dc.format.mediumElectronic-
dc.languageEnglish-
dc.language.isoen_USen_US
dc.publisherCornell Universityen_US
dc.relation.urihttps://github.com/SR0920/TEC-Net-
dc.relation.urihttps://arxiv.org/abs/2306.04086-
dc.rightsCopyright © 2023 The Authors. This is a preprint made available under the arXiv.org - Non-exclusive license to distribute, see: https://arxiv.org/licenses/nonexclusive-distrib/1.0/license.html. This arXiv preprint (version 3) has been prepared for publication in a future issue of a journal, but has not been fully edited. Content may change prior to final publication. It has not been certified by peer review.-
dc.rights.urihttps://arxiv.org/licenses/nonexclusive-distrib/1.0/license.html-
dc.subjectimage and video processing (eess.IV);-
dc.subjectcomputer vision and pattern recognition (cs.CV)-
dc.titleTEC-Net: Vision Transformer Embrace Convolutional Neural Networks for Medical Image Segmentationen_US
dc.typeArticleen_US
dc.date.dateAccepted2023-06-07-
dc.identifier.doihttps://doi.org/10.48550/arXiv.2306.04086-
dc.relation.isPartOfarXiv-
dc.identifier.eissn2331-8422-
dc.rights.licensehttps://arxiv.org/licenses/nonexclusive-distrib/1.0/license.html-
dc.rights.holderThe Authors-
Appears in Collections:Dept of Electronic and Electrical Engineering Research Papers

Files in This Item:
File Description SizeFormat 
Preprint.pdfCopyright © 2023 The Authors. This is a preprint made available under the arXiv.org - Non-exclusive license to distribute, see: https://arxiv.org/licenses/nonexclusive-distrib/1.0/license.html. This arXiv preprint (version 3) has been prepared for publication in a future issue of a journal, but has not been fully edited. Content may change prior to final publication. It has not been certified by peer review.6.63 MBAdobe PDFView/Open


Items in BURA are protected by copyright, with all rights reserved, unless otherwise indicated.