TY - JOUR
T1 - Privacy-Preserving Machine Learning Training in IoT Aggregation Scenarios
AU - Zhu, Liehuang
AU - Tang, Xiangyun
AU - Shen, Meng
AU - Gao, Feng
AU - Zhang, Jie
AU - Du, Xiaojiang
N1 - Publisher Copyright:
© 2014 IEEE.
PY - 2021/8/1
Y1 - 2021/8/1
N2 - In developing smart city, the growing popularity of machine learning (ML) that appreciates high-quality training data sets generated from diverse Internet-of-Things (IoT) devices raises natural questions about the privacy guarantees that can be provided in such settings. Privacy-preserving ML training in an aggregation scenario enables a model demander to securely train ML models with the sensitive IoT data gathered from IoT devices. The existing solutions are generally server aided, cannot deal with the collusion threat between the servers or between the servers and data owners, and do not match the delicate environments of IoT. We propose a privacy-preserving ML training framework named Heda that consists of a library of building blocks based on partial homomorphic encryption, which enables constructing multiple privacy-preserving ML training protocols for the aggregation scenario without the assistance of untrusted servers, and defending the security under collusion situations. Rigorous security analysis demonstrates the proposed protocols can protect the privacy of each participant in the honest-but-curious model and guarantee the security under most collusion situations. Extensive experiments validate the efficiency of Heda, which achieves privacy-preserving ML training without losing the model accuracy.
AB - In developing smart city, the growing popularity of machine learning (ML) that appreciates high-quality training data sets generated from diverse Internet-of-Things (IoT) devices raises natural questions about the privacy guarantees that can be provided in such settings. Privacy-preserving ML training in an aggregation scenario enables a model demander to securely train ML models with the sensitive IoT data gathered from IoT devices. The existing solutions are generally server aided, cannot deal with the collusion threat between the servers or between the servers and data owners, and do not match the delicate environments of IoT. We propose a privacy-preserving ML training framework named Heda that consists of a library of building blocks based on partial homomorphic encryption, which enables constructing multiple privacy-preserving ML training protocols for the aggregation scenario without the assistance of untrusted servers, and defending the security under collusion situations. Rigorous security analysis demonstrates the proposed protocols can protect the privacy of each participant in the honest-but-curious model and guarantee the security under most collusion situations. Extensive experiments validate the efficiency of Heda, which achieves privacy-preserving ML training without losing the model accuracy.
KW - Homomorphic encryption
KW - Internet-of-Things (IoT) data
KW - machine learning (ML)
KW - modular sequential composition
KW - secure two-party computation
UR - http://www.scopus.com/inward/record.url?scp=85101775417&partnerID=8YFLogxK
U2 - 10.1109/JIOT.2021.3060764
DO - 10.1109/JIOT.2021.3060764
M3 - Article
AN - SCOPUS:85101775417
SN - 2327-4662
VL - 8
SP - 12106
EP - 12118
JO - IEEE Internet of Things Journal
JF - IEEE Internet of Things Journal
IS - 15
M1 - 9359659
ER -