TY - JOUR
T1 - EdgeNetLLM
T2 - Cloud–Edge Collaborative Adaptation of Large Language Models for Mobile Networking
AU - Zheng, Xixi
AU - Li, You
AU - Zheng, Baokun
AU - Zhang, Chuan
AU - Zhu, Liehuang
N1 - Publisher Copyright:
© 2013 IEEE.
PY - 2026
Y1 - 2026
N2 - With the development of deep learning (DL), many applications in mobile networks increasingly rely on deep neural networks (DNNs) optimized for specific tasks. However, existing DNN-based methods require labor-intensive model design and computationally intensive centralized training, resulting in high costs, poor scalability, and limited generalization. Inspired by large language models (LLMs), some works have utilized the powerful pre-trained knowledge of LLMs as foundation models to adapt to a variety of mobile networking tasks. However, due to the inherent nature of LLMs, fine-tuning and deploying these architectures in real-world mobile network environments is a challenging task. To address this challenge, we propose EdgeNetLLM, a novel framework that enables efficient LLM adaptation in mobile network scenarios through cloud-edge collaboration. EdgeNetLLM offloads the heavy computation and fine-tuning tasks to the cloud via cloud-edge cooperation, and adapts to the corresponding mobile network tasks with minimal cost through one-time pruning, supporting lightweight inference. Additionally, on the edge side, multimodal input processing and task specialization can also be adapted to different tasks with minimal cost, reducing interaction latency between the cloud and the edge. We evaluate EdgeNetLLM on the representative mobile network tasks of adaptive bitrate streaming (ABR) and cluster job scheduling (CJS). The results demonstrate that it significantly reduces adaptation costs while maintaining task performance. Specifically, a model with approximately 40% sparsity achieves 89% and 96.8% of the task performance compared to state-of-the-art method.
AB - With the development of deep learning (DL), many applications in mobile networks increasingly rely on deep neural networks (DNNs) optimized for specific tasks. However, existing DNN-based methods require labor-intensive model design and computationally intensive centralized training, resulting in high costs, poor scalability, and limited generalization. Inspired by large language models (LLMs), some works have utilized the powerful pre-trained knowledge of LLMs as foundation models to adapt to a variety of mobile networking tasks. However, due to the inherent nature of LLMs, fine-tuning and deploying these architectures in real-world mobile network environments is a challenging task. To address this challenge, we propose EdgeNetLLM, a novel framework that enables efficient LLM adaptation in mobile network scenarios through cloud-edge collaboration. EdgeNetLLM offloads the heavy computation and fine-tuning tasks to the cloud via cloud-edge cooperation, and adapts to the corresponding mobile network tasks with minimal cost through one-time pruning, supporting lightweight inference. Additionally, on the edge side, multimodal input processing and task specialization can also be adapted to different tasks with minimal cost, reducing interaction latency between the cloud and the edge. We evaluate EdgeNetLLM on the representative mobile network tasks of adaptive bitrate streaming (ABR) and cluster job scheduling (CJS). The results demonstrate that it significantly reduces adaptation costs while maintaining task performance. Specifically, a model with approximately 40% sparsity achieves 89% and 96.8% of the task performance compared to state-of-the-art method.
KW - Large language models
KW - cloud-edge collaboration
KW - deep learning
KW - mobile networking
KW - model adaptation
UR - https://www.scopus.com/pages/publications/105019925534
U2 - 10.1109/TNSE.2025.3624100
DO - 10.1109/TNSE.2025.3624100
M3 - Article
AN - SCOPUS:105019925534
SN - 2327-4697
VL - 13
SP - 3928
EP - 3943
JO - IEEE Transactions on Network Science and Engineering
JF - IEEE Transactions on Network Science and Engineering
ER -