Please use this identifier to cite or link to this item:
http://bura.brunel.ac.uk/handle/2438/32664Full metadata record
| DC Field | Value | Language |
|---|---|---|
| dc.contributor.author | Peng, Y | - |
| dc.contributor.author | Xiang, L | - |
| dc.contributor.author | Yang, K | - |
| dc.contributor.author | Wang, K | - |
| dc.contributor.author | Debbah, M | - |
| dc.date.accessioned | 2026-01-16T17:37:39Z | - |
| dc.date.available | 2026-01-16T17:37:39Z | - |
| dc.date.issued | 2025-12-22 | - |
| dc.identifier | ORCiD: Kezhi Wang https://orcid.org/0000-0001-8602-0800 | - |
| dc.identifier.citation | Peng, Y. et al. (2025) 'Semantic Communications With Computer Vision Sensing for Edge Video Transmission', IEEE Transactions on Mobile Computing, 0 (early access), pp. 1 - 14. doi: 10.1109/TMC.2025.3646710. | en_US |
| dc.identifier.issn | 1536-1233 | - |
| dc.identifier.uri | https://bura.brunel.ac.uk/handle/2438/32664 | - |
| dc.description.abstract | Despite the widespread adoption of vision sensors in edge applications, such as surveillance, video transmission consumes substantial spectrum resources. Semantic communication (SC) offers a solution by extracting and compressing information at the semantic level, but traditional SC without sensing capabilities faces inefficiencies due to the repeated transmission of static frames in edge videos. To address this challenge, we propose an SC with computer vision sensing (SCCVS) framework for edge video transmission. The framework first introduces a compression ratio (CR) adaptive SC (CRSC) model, capable of adjusting CR based on whether the frames are static or dynamic, effectively conserving spectrum resources. Simultaneously, we present a knowledge distillation (KD)-based approach to ensure the efficient learning of the CRSC model. Additionally, we implement a computer vision (CV)-based sensing model (CVSM) scheme, which intelligently perceives the scene changes by detecting the movement of the sensing targets. Therefore, CVSM can assess the significance of each frame through in-context analysis and provide CR prompts to the CRSC model based on real-time sensing results. Moreover, both CRSC and CVSM are designed as lightweight models, ensuring compatibility with resource-constrained sensors commonly used in practical edge applications. Experimental results show that SCCVS improves transmission accuracy by approximately 70% and reduces transmission latency by about 89% compared with baselines. We also deploy this framework on an NVIDIA Jetson Orin NX Super, achieving an inference speed of 14 ms per frame with TensorRT acceleration and demonstrating its real-time capability and effectiveness in efficient semantic video transmission. | en_US |
| dc.description.sponsorship | ... | en_US |
| dc.format.extent | 1 - 14 | - |
| dc.format.medium | Print-Electronic | - |
| dc.language | English | - |
| dc.language.iso | en_US | en_US |
| dc.publisher | Institute of Electrical and Electronics Engineers (IEEE) | en_US |
| dc.rights | Creative Commons Attribution 4.0 International | - |
| dc.rights.uri | https://creativecommons.org/licenses/by/4.0/ | - |
| dc.subject | semantic communication | en_US |
| dc.subject | computer vision | en_US |
| dc.subject | video transmission | en_US |
| dc.subject | intelligence sensing | en_US |
| dc.title | Semantic Communications With Computer Vision Sensing for Edge Video Transmission | en_US |
| dc.type | Article | en_US |
| dc.identifier.doi | https://doi.org/10.1109/TMC.2025.3646710 | - |
| dc.relation.isPartOf | IEEE Transactions on Mobile Computing | - |
| pubs.issue | 0 | - |
| pubs.publication-status | Published | - |
| pubs.volume | 00 | - |
| dc.identifier.eissn | 1558-0660 | - |
| dc.rights.license | https://creativecommons.org/licenses/by/4.0/legalcode.en | - |
| dc.rights.holder | The Author(s) | - |
| dc.contributor.orcid | Wang, Kezhi [0000-0001-8602-0800] | - |
| Appears in Collections: | Dept of Computer Science Research Papers | |
Files in This Item:
| File | Description | Size | Format | |
|---|---|---|---|---|
| FullText.pdf | Copyright © 2025 The Author(s). This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/ | 4.07 MB | Adobe PDF | View/Open |
This item is licensed under a Creative Commons License