Please use this identifier to cite or link to this item:
http://bura.brunel.ac.uk/handle/2438/30985
Title: | Visual Language Model based Cross-modal Semantic Communication Systems |
Authors: | Jiang, F Tang, C Dong, L Wang, K Yang, K Pan, C |
Keywords: | semantic communication;knowledge base;vision language model;large language model;continual learning |
Issue Date: | 4-Mar-2025 |
Publisher: | Institute of Electrical and Electronics Engineers (IEEE) |
Citation: | Jiang, F. et al. (2025) 'Visual Language Model based Cross-modal Semantic Communication Systems', IEEE Transactions on Wireless Communications, 0 (early access), pp. 1 - 13. doi: 10.1109/TWC.2025.3539526. |
Abstract: | Semantic Communication (SC) has emerged as a novel communication paradigm in recent years. Nevertheless, extant Image Semantic Communication (ISC) systems face several challenges in dynamic environments, including low information density, catastrophic forgetting, and uncertain Signal-to-Noise Ratio (SNR). To address these challenges, we propose a novel Vision-Language Model-based Cross-modal Semantic Communication (VLM-CSC) system. The VLM-CSC comprises three novel components: (1) Cross-modal Knowledge Base (CKB) is used to extract high-density textual semantics from the semantically sparse image at the transmitter and reconstruct the original image based on textual semantics at the receiver. The transmission of high-density semantics contributes to alleviating bandwidth pressure. (2) Memory-assisted Encoder and Decoder (MED) employ a hybrid long/short-term memory mechanism, enabling the semantic encoder and decoder to overcome catastrophic forgetting in dynamic environments when there is a drift in the distribution of semantic features. (3) Noise Attention Module (NAM) employs attention mechanisms to adaptively adjust the semantic coding and the channel coding based on SNR, ensuring the robustness of the CSC system. The experimental simulations validate the effectiveness, adaptability, and robustness of the CSC system. |
URI: | https://bura.brunel.ac.uk/handle/2438/30985 |
DOI: | https://doi.org/10.1109/TWC.2025.3539526 |
ISSN: | 1536-1276 |
Other Identifiers: | ORCiD: Feibo Jiang https://orcid.org/0000-0002-0235-0253 ORCiD: Li Dong https://orcid.org/0000-0002-0127-8480 ORCiD: Kezhi Wang https://orcid.org/0000-0001-8602-0800 ORCiD: Kun Yang https://orcid.org/0000-0002-6782-6689 ORCiD: Cunhua Pan https://orcid.org/0000-0001-5286-7958 |
Appears in Collections: | Dept of Computer Science Research Papers |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
FullText.pdf | Copyright © 2025 Institute of Electrical and Electronics Engineers (IEEE). Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works. See: https://journals.ieeeauthorcenter.ieee.org/become-an-ieee-journal-author/publishing-ethics/guidelines-and-policies/post-publication-policies/. | 4.62 MB | Adobe PDF | View/Open |
Items in BURA are protected by copyright, with all rights reserved, unless otherwise indicated.