TY - JOUR
T1 - Adaptive ensemble optimization for memory-related hyperparameters in retraining DNN at edge
AU - Xu, Yidong
AU - Han, Rui
AU - Zuo, Xiaojiang
AU - Ouyang, Junyan
AU - Liu, Chi Harold
AU - Chen, Lydia Y.
N1 - Publisher Copyright:
© 2024 Elsevier B.V.
PY - 2025/3
Y1 - 2025/3
N2 - Edge applications are increasingly empowered by deep neural networks (DNN) and face the challenges of adapting or retraining models for the changes in input data domains and learning tasks. The existing techniques to enable DNN retraining on edge devices are to configure the memory-related hyperparameters, termed m-hyperparameters, via batch size reduction, parameter freezing, and gradient checkpoint. While those methods show promising results for static DNNs, little is known about how to online and opportunistically optimize all their m-hyperparameters, especially for retraining tasks of edge applications. In this paper, we propose, MPOptimizer, which jointly optimizes an ensemble of m-hyperparameters according to the input distribution and available edge resources at runtime. The key feature of MPOptimizer is to easily emulate the execution of retraining tasks under different m-hyperparameters and thus effectively estimate their influence on task performance. We implement MPOptimizer on prevalent DNNs and demonstrate its effectiveness against state-of-the-art techniques, i.e. successfully find the best configuration that improves model accuracy by an average of 13% (up to 25.3%) while reducing memory and training time by 4.1x and 5.3x under the same model accuracies.
AB - Edge applications are increasingly empowered by deep neural networks (DNN) and face the challenges of adapting or retraining models for the changes in input data domains and learning tasks. The existing techniques to enable DNN retraining on edge devices are to configure the memory-related hyperparameters, termed m-hyperparameters, via batch size reduction, parameter freezing, and gradient checkpoint. While those methods show promising results for static DNNs, little is known about how to online and opportunistically optimize all their m-hyperparameters, especially for retraining tasks of edge applications. In this paper, we propose, MPOptimizer, which jointly optimizes an ensemble of m-hyperparameters according to the input distribution and available edge resources at runtime. The key feature of MPOptimizer is to easily emulate the execution of retraining tasks under different m-hyperparameters and thus effectively estimate their influence on task performance. We implement MPOptimizer on prevalent DNNs and demonstrate its effectiveness against state-of-the-art techniques, i.e. successfully find the best configuration that improves model accuracy by an average of 13% (up to 25.3%) while reducing memory and training time by 4.1x and 5.3x under the same model accuracies.
KW - Deep neural networks (DNN)
KW - Edge computing
KW - Memory-related hyperparameters
KW - Model retraining
UR - http://www.scopus.com/inward/record.url?scp=85208761444&partnerID=8YFLogxK
U2 - 10.1016/j.future.2024.107600
DO - 10.1016/j.future.2024.107600
M3 - Article
AN - SCOPUS:85208761444
SN - 0167-739X
VL - 164
JO - Future Generation Computer Systems
JF - Future Generation Computer Systems
M1 - 107600
ER -