Please use this identifier to cite or link to this item: http://bura.brunel.ac.uk/handle/2438/33125
Full metadata record
DC FieldValueLanguage
dc.contributor.authorHe, Y-
dc.contributor.authorWu, H-
dc.contributor.authorLiu, W-
dc.contributor.authorLuo, X-
dc.date.accessioned2026-04-09T16:50:28Z-
dc.date.available2026-04-09T16:50:28Z-
dc.date.issued2026-03-25-
dc.identifierORCiD: Yaping He https://orcid.org/0009-0000-4882-1631-
dc.identifierORCiD: Hao Wu https://orcid.org/0000-0002-4138-1239-
dc.identifierORCiD: Weibo Liu https://orcid.org/0000-0002-8169-3261-
dc.identifierORCiD: Xin Luo https://orcid.org/0000-0002-1348-5305-
dc.identifier.citationHe, Y. et al. (2026) 'A survey of latent factorization of tensor-based model compression: Algorithms, toolboxes and future directions', Neurocomputing, 682, 133455, pp. 1–20. doi: 10.1016/j.neucom.2026.133455.en-US
dc.identifier.issn0925-2312-
dc.identifier.urihttps://bura.brunel.ac.uk/handle/2438/33125-
dc.descriptionData availability: No data was used for the research described in the article.en-US
dc.description.abstractModern neural networks (NNs), while effective at learning representations from given samples and handling downstream pattern recognition tasks, typically contain tens to hundreds of millions of parameters. The growth in NN size motivates ongoing research on effective network compression with the purpose of reducing the computational burden without significantly sacrificing the model performance. It is especially critical when deploying NNs on resource-constrained devices where computation and storage efficiency are of high concern. A promising and currently popular solution to model compression is to replace the NN weight matrix with its low-rank tensor approximation, i.e., implementing an efficient latent factorization of tensors (LFT) process on the NNs parameters. Based on thorough investigations into the state-of-the-art LFT-based model compression methods, this survey 1) provides a comprehensive review of the latest research progress on LFT-based model compression methods for various NNs (e.g., Convolutional NNs, Recurrent NNs, and Transformers); 2) summarizes a number of widely-used LFT toolboxes; 3) evaluates LFT methods for model compression on a variety of main-stream NN backbones; and 4) discusses the development trends of LFT-based model compression techniques. This survey aims to provide a systematic and comprehensive overview of LFT-based model compression methods to artificial intelligence researchers and engineers, thereby promoting further research development in this crucial field.en-US
dc.description.sponsorshipThis work was supported in part by the Science and Technology Innovation Key R&D Program of Chongqing under Grant CSTB2025TIAD-STX0032, the National Key Research and Development Program of China under Grant 2024YFF0908200, the Chongqing Technology Innovation and Application Development Special Key Project under Grant CSTB2024TIAD-KPX0018, the Royal Society of the UK under Grant IES\R3\243021, and the Southwest University Graduate Student Research Innovation Grant SWUB24051.en-US
dc.format.extent1–20-
dc.format.mediumPrint-Electronic-
dc.languageen-USen-US
dc.language.isoenen-US
dc.publisherElsevieren-US
dc.rightsCreative Commons Attribution 4.0 International-
dc.rights.urihttps://creativecommons.org/licenses/by/4.0/-
dc.subjectlatent factorization of tensoren-US
dc.subjectmodel compressionen-US
dc.subjectresource-constrained devicesen-US
dc.subjectconvolutional neural networken-US
dc.subjectrecurrent neural networken-US
dc.subjecttransformeren-US
dc.titleA survey of latent factorization of tensor-based model compression: Algorithms, toolboxes and future directionsen-US
dc.typeArticleen-US
dc.date.dateAccepted2026-03-24-
dc.identifier.doihttps://doi.org/10.1016/j.neucom.2026.133455-
dc.relation.isPartOfNeurocomputing-
pubs.publication-statusPublished online-
pubs.volume682-
dc.identifier.eissn1872-8286-
dc.rights.licensehttps://creativecommons.org/licenses/by/4.0/legalcode.en-
dcterms.dateAccepted2026-03-24-
dc.rights.holderThe Authors-
dc.contributor.orcidHe, Yaping [0009-0000-4882-1631]-
dc.contributor.orcidWu, Hao [0000-0002-4138-1239]-
dc.contributor.orcidLiu, Weibo[0000-0002-8169-3261]-
dc.contributor.orcidLuo, Xin [0000-0002-1348-5305]-
dc.identifier.number133455-
Appears in Collections:Department of Computer Science Research Papers

Files in This Item:
File Description SizeFormat 
FullText.pdfCopyright © 2026 The Authors. Published by Elsevier B.V. This is an open access article under the CC BY license ( https://creativecommons.org/licenses/by/4.0/ ).4.03 MBAdobe PDFView/Open


This item is licensed under a Creative Commons License Creative Commons