Please use this identifier to cite or link to this item: http://bura.brunel.ac.uk/handle/2438/33043
Title: Scalable Semi-supervised Learning with Discriminative Label Propagation and Correction
Authors: Jiang, B
Wen, J
Wang, Z
Sheng, W
Yu, Z
Chen, H
Ding, W
Keywords: discriminative label propagation;multi-view learning;semi-supervised classification;similarity graph learning
Issue Date: 19-Jan-2026
Publisher: Institute of Electrical and Electronics Engineers (IEEE)
Citation: Jiang, B. et al. (2026 'Scalable Semi-supervised Learning with Discriminative Label Propagation and Correction', IEEE Transactions on Pattern Analysis and Machine Intelligence, 0 (early access), pp. 1–17. doi: 10.1109/tpami.2026.3655456
Abstract: Semi-supervised learning can leverage both labeled and unlabeled samples simultaneously to improve performance. However, existing methods often present the following issues: (1) The emphasis of learning is put on either the similarity structures or the regression losses of data, neglecting the interaction between them. (2) The similarity structures among boundary samples might be unreliable, which misleads label propagation and impairs the performance of models on out-of-sample data. (3) They often involve the inverses of high-order matrices, making them inefficient in computation. To overcome these issues, we propose a scalable semi-supervised learning framework with Discriminative Label Propagation and Correction (DLPC), which collaboratively exploits the regression losses and similarity structures of data. Particularly, each sample is projected onto the independent class labels associated with nonnegative adjustment vectors rather than the propagated labels, such that the distances between samples from different classes are naturally enlarged, making regression losses more effective for boundary samples. Benefiting from this, the regression losses can guide the propagation of labels in boundary areas. Thus, the label information is first propagated through dynamically optimized graph structures and then corrected by the regression losses, effectively improving the quality of labels and facilitating feature projection learning. Furthermore, an accelerated solution has been developed to reduce the computational costs of DLPC on sample scales, thereby making it scalable to relatively large-scale problems. Moreover, the proposed DLPC can not only be applied to single-view scenarios but also extended to multi-view tasks. Additionally, an optimization strategy with fast convergence has been presented for DLPC, and extensive experiments demonstrate the effectiveness and superiority of DLPC over state-of-the-art competitors.
URI: https://bura.brunel.ac.uk/handle/2438/33043
DOI: https://doi.org/10.1109/tpami.2026.3655456
ISSN: 0162-8828
Other Identifiers: ORCiD: Bingbing Jiang https://orcid.org/0009-0006-3153-6125
ORCiD: Jie Wen https://orcid.org/0000-0001-9554-2379
ORCiD: Zidong Wang https://orcid.org/0000-0002-9576-7401
ORCiD: Weiguo Sheng https://orcid.org/0000-0001-9680-5126
ORCiD: Zhiwen Yu https://orcid.org/0000-0002-0935-5890
ORCiD: Huanhuan Chen https://orcid.org/0000-0002-3918-384X
Appears in Collections:Department of Computer Science Research Papers

Files in This Item:
File Description SizeFormat 
FullText.pdfFor the purpose of open access, the author has applied a Creative Commons Attribution (CC BY) licence to any Author Accepted Manuscript version arising.7.54 MBAdobe PDFView/Open


This item is licensed under a Creative Commons License Creative Commons