Please use this identifier to cite or link to this item: http://bura.brunel.ac.uk/handle/2438/32664
Full metadata record
DC FieldValueLanguage
dc.contributor.authorPeng, Y-
dc.contributor.authorXiang, L-
dc.contributor.authorYang, K-
dc.contributor.authorWang, K-
dc.contributor.authorDebbah, M-
dc.date.accessioned2026-01-16T17:37:39Z-
dc.date.available2026-01-16T17:37:39Z-
dc.date.issued2025-12-22-
dc.identifierORCiD: Kezhi Wang https://orcid.org/0000-0001-8602-0800-
dc.identifier.citationPeng, Y. et al. (2025) 'Semantic Communications With Computer Vision Sensing for Edge Video Transmission', IEEE Transactions on Mobile Computing, 0 (early access), pp. 1 - 14. doi: 10.1109/TMC.2025.3646710.en_US
dc.identifier.issn1536-1233-
dc.identifier.urihttps://bura.brunel.ac.uk/handle/2438/32664-
dc.description.abstractDespite the widespread adoption of vision sensors in edge applications, such as surveillance, video transmission consumes substantial spectrum resources. Semantic communication (SC) offers a solution by extracting and compressing information at the semantic level, but traditional SC without sensing capabilities faces inefficiencies due to the repeated transmission of static frames in edge videos. To address this challenge, we propose an SC with computer vision sensing (SCCVS) framework for edge video transmission. The framework first introduces a compression ratio (CR) adaptive SC (CRSC) model, capable of adjusting CR based on whether the frames are static or dynamic, effectively conserving spectrum resources. Simultaneously, we present a knowledge distillation (KD)-based approach to ensure the efficient learning of the CRSC model. Additionally, we implement a computer vision (CV)-based sensing model (CVSM) scheme, which intelligently perceives the scene changes by detecting the movement of the sensing targets. Therefore, CVSM can assess the significance of each frame through in-context analysis and provide CR prompts to the CRSC model based on real-time sensing results. Moreover, both CRSC and CVSM are designed as lightweight models, ensuring compatibility with resource-constrained sensors commonly used in practical edge applications. Experimental results show that SCCVS improves transmission accuracy by approximately 70% and reduces transmission latency by about 89% compared with baselines. We also deploy this framework on an NVIDIA Jetson Orin NX Super, achieving an inference speed of 14 ms per frame with TensorRT acceleration and demonstrating its real-time capability and effectiveness in efficient semantic video transmission.en_US
dc.description.sponsorship...en_US
dc.format.extent1 - 14-
dc.format.mediumPrint-Electronic-
dc.languageEnglish-
dc.language.isoen_USen_US
dc.publisherInstitute of Electrical and Electronics Engineers (IEEE)en_US
dc.rightsCreative Commons Attribution 4.0 International-
dc.rights.urihttps://creativecommons.org/licenses/by/4.0/-
dc.subjectsemantic communicationen_US
dc.subjectcomputer visionen_US
dc.subjectvideo transmissionen_US
dc.subjectintelligence sensingen_US
dc.titleSemantic Communications With Computer Vision Sensing for Edge Video Transmissionen_US
dc.typeArticleen_US
dc.identifier.doihttps://doi.org/10.1109/TMC.2025.3646710-
dc.relation.isPartOfIEEE Transactions on Mobile Computing-
pubs.issue0-
pubs.publication-statusPublished-
pubs.volume00-
dc.identifier.eissn1558-0660-
dc.rights.licensehttps://creativecommons.org/licenses/by/4.0/legalcode.en-
dc.rights.holderThe Author(s)-
dc.contributor.orcidWang, Kezhi [0000-0001-8602-0800]-
Appears in Collections:Dept of Computer Science Research Papers

Files in This Item:
File Description SizeFormat 
FullText.pdfCopyright © 2025 The Author(s). This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/4.07 MBAdobe PDFView/Open


This item is licensed under a Creative Commons License Creative Commons