Please use this identifier to cite or link to this item: http://bura.brunel.ac.uk/handle/2438/25756
Full metadata record
DC FieldValueLanguage
dc.contributor.authorWang, Z-
dc.contributor.authorZhou, Y-
dc.contributor.authorGan, L-
dc.contributor.authorChen, R-
dc.contributor.authorTang, X-
dc.contributor.authorLiu, H-
dc.date.accessioned2023-01-11T15:16:16Z-
dc.date.available2023-01-11T15:16:16Z-
dc.date.issued2022-10-25-
dc.identifierORCID iD: Lu Gan https://orcid.org/0000-0003-1056-7660-
dc.identifier.citationWang, Z. et al. (2022) 'DE-DPCTnet: Deep Encoder Dual-path Convolutional Transformer Network for Multi-channel Speech Separation', 2022 IEEE Workshop on Signal Processing Systems (SiPS), Rennes, France, 02-04 November, pp. 1 - 5. doi: 10.1109/SiPS55645.2022.9919247.en_US
dc.identifier.isbn978-1-6654-8524-1 (ebk)-
dc.identifier.isbn978-1-6654-8525-8 (PoD)-
dc.identifier.issn1520-6130-
dc.identifier.urihttps://bura.brunel.ac.uk/handle/2438/25756-
dc.description.abstractIn recent years, beamforming has been extensively investigated in multi-channel speech separation task. In this paper, we propose a deep encoder dual-path convolutional transformer network (DE-DPCTnet), which directly estimates the beamforming filters for speech separation task in time domain. In order to learn the signal repetitions correctly, nonlinear deep encoder module is proposed to replace the traditional linear one. The improved transformer is also developed by utilizing convolutions to capture long-time speech sequences. The ablation studies demonstrate that the deep encoder and improved transformer indeed benefit the separation performance. The comparisons show that the DE-DPCTnet outperforms the state-of-the-art filter-and-sum network with transform-average-concatenate module (FaSNet-TAC), even with a lower computational complexity.en_US
dc.format.extent1 - 5-
dc.format.mediumPrint-Electronic-
dc.language.isoen_USen_US
dc.publisherInstitute of Electrical and Electronics Engineers (IEEE)en_US
dc.rightsCopyright © 2022 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.-
dc.rights.urihttps://www.ieee.org/publications/rights/rights-policies.html-
dc.subjectspeech separationen_US
dc.subjectmulti-channelen_US
dc.subjectdeep encoderen_US
dc.subjectimproved transformeren_US
dc.subjectbeamformingen_US
dc.titleDE-DPCTnet: Deep Encoder Dual-path Convolutional Transformer Network for Multi-channel Speech Separationen_US
dc.typeConference Paperen_US
dc.identifier.doihttps://doi.org/10.1109/SiPS55645.2022.9919247-
dc.relation.isPartOf2022 IEEE Workshop on Signal Processing Systems (SiPS)-
pubs.publication-statusPublished-
pubs.volume2022-
dc.rights.holderInstitute of Electrical and Electronics Engineers (IEEE)-
Appears in Collections:Dept of Electronic and Electrical Engineering Research Papers

Files in This Item:
File Description SizeFormat 
FullText.pdfCopyright © 2022 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.448.5 kBAdobe PDFView/Open


Items in BURA are protected by copyright, with all rights reserved, unless otherwise indicated.