Please use this identifier to cite or link to this item: http://bura.brunel.ac.uk/handle/2438/32316
Full metadata record
DC FieldValueLanguage
dc.contributor.authorWu, H-
dc.contributor.authorWang, Q-
dc.contributor.authorLuo, X-
dc.contributor.authorWang, Z-
dc.date.accessioned2025-11-07T16:19:47Z-
dc.date.available2025-11-07T16:19:47Z-
dc.date.issued2025-10-03-
dc.identifierORCiD: Zidong Wang https://orcid.org/0000-0002-9576-7401-
dc.identifier.citationWu, H. et al. (2025) 'Learning Accurate Representation to Nonstandard Tensors via a Mode-Aware Tucker Network', IEEE Transactions on Knowledge and Data Engineering, 0 (early access), pp. 1 - 14. doi: 10.1109/TKDE.2025.3617894.en_US
dc.identifier.issn1041-4347-
dc.identifier.urihttps://bura.brunel.ac.uk/handle/2438/32316-
dc.description.abstractA nonstandard tensor is frequently adopted to model a large-sale complex dynamic network. A Tensor Representation Learning (TRL) model enables extracting valuable knowledge form a dynamic network via learning low-dimensional representation of a target nonstandard tensor. Nevertheless, the representation learning ability of existing TRL models are limited for a nonstandard tensor due to its inability to accurately represent the specific nature of the nonstandard tensor, i.e., mode imbalance, high-dimension, and incompleteness. To address this issue, this study innovatively proposes a Mode-Aware Tucker Networkbased Tensor Representation Learning (MTN-TRL) model with three-fold ideas: a) designing a mode-aware Tucker network to accurately represent the imbalanced mode of a nonstandard tensor, b) building an MTN-based high-efficient TRL model that fuses both data density-oriented modeling principle and adaptive parameters learning scheme, and c) theoretically proving the MTN-TRL model's convergence. Extensive experiments on eight nonstandard tensors generating from real-world dynamic networks demonstrate that MTN-TRL significantly outperforms state-of-the-art models in terms of representation accuracy.en_US
dc.description.sponsorshipThis work is supported by the National Natural Science Foundation of China under grant 62302402, 62272078, and the Science and Technology Re- search Program of Chongqing Municipal Education Commission under grant KJQN202403017, KJZD-K202400209.en_US
dc.format.extent1 - 14-
dc.format.mediumPrint-Electronic-
dc.language.isoen_USen_US
dc.publisherInstitute of Electrical and Electronics Engineers (IEEE)en_US
dc.rightsCopyright © 2025 Institute of Electrical and Electronics Engineers (IEEE). Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works ( https://journals.ieeeauthorcenter.ieee.org/become-an-ieee-journal-author/publishing-ethics/guidelines-and-policies/post-publication-policies/ ).-
dc.rights.urihttps://journals.ieeeauthorcenter.ieee.org/become-an-ieee-journal-author/publishing-ethics/guidelines-and-policies/post-publication-policies/-
dc.subjecttensor representation learningen_US
dc.subjectnonstandard tensoren_US
dc.subjecttensor networken_US
dc.subjectdynamic network representationen_US
dc.titleLearning Accurate Representation to Nonstandard Tensors via a Mode-Aware Tucker Networken_US
dc.typeArticleen_US
dc.identifier.doihttps://doi.org/10.1109/TKDE.2025.3617894-
dc.relation.isPartOfIEEE Transactions on Knowledge and Data Engineering-
pubs.issue0-
pubs.publication-statusPublished-
pubs.volume00-
dc.identifier.eissn1558-2191-
dc.rights.holderInstitute of Electrical and Electronics Engineers (IEEE)-
Appears in Collections:Dept of Computer Science Research Papers

Files in This Item:
File Description SizeFormat 
FullText.pdfCopyright © 2025 Institute of Electrical and Electronics Engineers (IEEE). Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works ( https://journals.ieeeauthorcenter.ieee.org/become-an-ieee-journal-author/publishing-ethics/guidelines-and-policies/post-publication-policies/ ).5.07 MBAdobe PDFView/Open


Items in BURA are protected by copyright, with all rights reserved, unless otherwise indicated.