Please use this identifier to cite or link to this item:
http://bura.brunel.ac.uk/handle/2438/31732
Title: | LAMBO: Large AI Model Empowered Edge Intelligence |
Authors: | Dong, L Jiang, F Peng, Y Wang, K Yang, K Pan, C Schober, R |
Keywords: | large AI model;edge intelligence;encoder-decoder architecture;reinforcement learning;active learning |
Issue Date: | 22-Oct-2024 |
Publisher: | Institute of Electrical and Electronics Engineers (IEEE) |
Citation: | Dong, L. et al. (2025) 'LAMBO: Large AI Model Empowered Edge Intelligence', IEEE Communications Magazine, 63 (4), pp. 88 - 94. doi: 10.1109/MCOM.001.2400076. |
Abstract: | Next-generation edge intelligence is anticipated to benefit various applications via offloading techniques. However, traditional offloading architectures face several issues, including heterogeneous constraints, partial perception, uncertain generalization, and lack of tractability. In this article, we propose a large AI model-based offloading (LAMBO) framework with over one billion parameters for solving these problems. We first use input embedding (IE) to achieve normalized feature representation with heterogeneous constraints and task prompts. Then, we introduce a novel asymmetric encoder-decoder (AED) as the decision-making model, which is an improved transformer architecture consisting of a deep encoder and a shallow decoder for global perception and decision. Next, actor-critic learning (ACL) is used to pre-train the AED for different optimization tasks under corresponding prompts, enhancing the AED's generalization in multi-task scenarios. Finally, we propose an active learning from expert feedback (ALEF) method to fine-tune the decoder of the AED for tracking changes in dynamic environments. Our simulation results validate the advantages of the proposed LAMBO framework. |
Description: | The article archived on this institutional repository is a preprint version, available at arXiv:2308.15078v2 [cs.AI], https://arxiv.org/abs/2308.15078 . It has not been certified by peer review. Please consult the published version at: https://doi.org/10.1109/MCOM.001.2400076 . |
URI: | http://bura.brunel.ac.uk/handle/2438/31732 |
DOI: | https://doi.org/10.1109/MCOM.001.2400076 |
ISSN: | 0163-6804 |
Other Identifiers: | ORCiD: Kezhi Wang https://orcid.org/0000-0001-8602-0800 |
Appears in Collections: | Dept of Computer Science Research Papers |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
FullText.pdf | Copyright © 2024 The Author(s). The URI http://arxiv.org/licenses/nonexclusive-distrib/1.0/ is used to record the fact that the submitter granted the following license to arXiv.org on submission of an article: I grant arXiv.org a perpetual, non-exclusive license to distribute this article. I certify that I have the right to grant this license. I understand that submissions cannot be completely removed once accepted. I understand that arXiv.org reserves the right to reclassify or reject any submission. [v2] Sat, 3 Aug 2024 13:43:01 UTC (1,544 KB) | 2.02 MB | Adobe PDF | View/Open |
Items in BURA are protected by copyright, with all rights reserved, unless otherwise indicated.