Please use this identifier to cite or link to this item: http://bura.brunel.ac.uk/handle/2438/31136
Full metadata record
DC FieldValueLanguage
dc.contributor.authorNykoniuk, M-
dc.contributor.authorBasystiuk, O-
dc.contributor.authorShakhovska, N-
dc.contributor.authorMelnykova, N-
dc.date.accessioned2025-05-04T09:33:48Z-
dc.date.available2025-05-04T09:33:48Z-
dc.date.issued2025-01-02-
dc.identifierORCiD: Oleh Basystiuk https://orcid.org/0000-0003-0064-6584-
dc.identifierORCiD: Nataliya Shakhovska https://orcid.org/0000-0002-6875-8534-
dc.identifierORCiD: Nataliia Melnykova https://orcid.org/0000-0002-2114-3436-
dc.identifierArticle number 9-
dc.identifier.citationNykoniuk, M. et al. (2025) 'Multimodal Data Fusion for Depression Detection Approach', Computation, 13 (1), 9, pp. 1 - 18. doi: 10.3390/computation13010009.en_US
dc.identifier.urihttps://bura.brunel.ac.uk/handle/2438/31136-
dc.descriptionData Availability Statement Dataset available on request from https://dcapswoz.ict.usc.edu/ (accessed on 28 November 2024).en_US
dc.descriptionAcknowledgments: The authors would like to thank the Armed Forces of Ukraine for providing the security to perform this work. This work was only possible because of the resilience and courage of the Ukrainian Army. The third author would like to acknowledge the financial support from the British Academy for this research (RaR\100727).-
dc.description.abstractDepression is one of the most common mental health disorders in the world, affecting millions of people. Early detection of depression is crucial for effective medical intervention. Multimodal networks can greatly assist in the detection of depression, especially in situations where in patients are not always aware of or able to express their symptoms. By analyzing text and audio data, such networks are able to automatically identify patterns in speech and behavior that indicate a depressive state. In this study, we propose two multimodal information fusion networks: early and late fusion. These networks were developed using convolutional neural network (CNN) layers to learn local patterns, a bidirectional LSTM (Bi-LSTM) to process sequences, and a self-attention mechanism to improve focus on key parts of the data. The DAIC-WOZ and EDAIC-WOZ datasets were used for the experiments. The experiments compared the precision, recall, f1-score, and accuracy metrics for the cases of using early and late multimodal data fusion and found that the early information fusion multimodal network achieved higher classification accuracy results. On the test dataset, this network achieved an f1-score of 0.79 and an overall classification accuracy of 0.86, indicating its effectiveness in detecting depression.en_US
dc.description.sponsorshipThis research received no external funding.en_US
dc.format.extent1 - 18-
dc.format.mediumElectronic-
dc.language.isoen_USen_US
dc.publisherMDPIen_US
dc.subjectdepression detectionen_US
dc.subjectmultimodal networksen_US
dc.subjectearly fusionen_US
dc.subjectlate fusionen_US
dc.subjectmental healthen_US
dc.subjectdeep learningen_US
dc.titleMultimodal Data Fusion for Depression Detection Approachen_US
dc.typeArticleen_US
dc.date.dateAccepted2024-12-25-
dc.identifier.doihttps://doi.org/10.3390/computation13010009-
pubs.finish-date1-
pubs.volume13-
dc.identifier.eissn2079-3197-
dcterms.dateAccepted2024-12-25-
Appears in Collections:Dept of Civil and Environmental Engineering Research Papers

Files in This Item:
File Description SizeFormat 
FullText.pdfCopyright © 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).2.81 MBAdobe PDFView/Open


Items in BURA are protected by copyright, with all rights reserved, unless otherwise indicated.