TY - JOUR
T1 - Privacy-preserving Sparse Generalized Eigenvalue Problem
AU - Hu, Lijie
AU - Xiang, Zihang
AU - Liu, Jiabin
AU - Wang, Di
N1 - Publisher Copyright:
Copyright © 2023 by the author(s)
PY - 2023
Y1 - 2023
N2 - In this paper we study the (sparse) Generalized Eigenvalue Problem (GEP), which arises in a number of modern statistical learning models, such as principal component analysis (PCA), canonical correlation analysis (CCA), Fisher's discriminant analysis (FDA) and sliced inverse regression (SIR). We provide the first study on GEP in the differential privacy (DP) model under both deterministic and stochastic settings. In the low dimensional case, we provide a ρ- Concentrated DP (CDP) method namely DP-Rayleigh Flow and show if the initial vector is close enough to the optimal vector, its output has an ℓ2-norm estimation error of Õ(n/d + d/n2ρ) (under some mild assumptions), where d is the dimension and n is the sample size. Next, we discuss how to find such a initial parameter privately. In the high dimensional sparse case where d ≫ n, we propose the DP-Truncated Rayleigh Flow method whose output could achieve an error of Õ(s log d/n + s log d/n2ρ) for various statistical models, where s is the sparsity of the underlying parameter. Moreover, we show that these errors in the stochastic setting are optimal up to a factor of Poly(log n) by providing the lower bounds of PCA and SIR under statistical setting and in the CDP model. Finally, to give a separation between ∊-DP and ρ-CDP for GEP, we also provide the lower bound Ω(d/n + d2/n2 ∊2) and Ω(s log d/n + s2 log2d/n2∊2) of private minimax risk for PCA, under the statistical setting and ∊-DP model, in low and high dimensional sparse case respectively.
AB - In this paper we study the (sparse) Generalized Eigenvalue Problem (GEP), which arises in a number of modern statistical learning models, such as principal component analysis (PCA), canonical correlation analysis (CCA), Fisher's discriminant analysis (FDA) and sliced inverse regression (SIR). We provide the first study on GEP in the differential privacy (DP) model under both deterministic and stochastic settings. In the low dimensional case, we provide a ρ- Concentrated DP (CDP) method namely DP-Rayleigh Flow and show if the initial vector is close enough to the optimal vector, its output has an ℓ2-norm estimation error of Õ(n/d + d/n2ρ) (under some mild assumptions), where d is the dimension and n is the sample size. Next, we discuss how to find such a initial parameter privately. In the high dimensional sparse case where d ≫ n, we propose the DP-Truncated Rayleigh Flow method whose output could achieve an error of Õ(s log d/n + s log d/n2ρ) for various statistical models, where s is the sparsity of the underlying parameter. Moreover, we show that these errors in the stochastic setting are optimal up to a factor of Poly(log n) by providing the lower bounds of PCA and SIR under statistical setting and in the CDP model. Finally, to give a separation between ∊-DP and ρ-CDP for GEP, we also provide the lower bound Ω(d/n + d2/n2 ∊2) and Ω(s log d/n + s2 log2d/n2∊2) of private minimax risk for PCA, under the statistical setting and ∊-DP model, in low and high dimensional sparse case respectively.
UR - http://www.scopus.com/inward/record.url?scp=85165177202&partnerID=8YFLogxK
M3 - Conference article
AN - SCOPUS:85165177202
SN - 2640-3498
VL - 206
SP - 5052
EP - 5062
JO - Proceedings of Machine Learning Research
JF - Proceedings of Machine Learning Research
T2 - 26th International Conference on Artificial Intelligence and Statistics, AISTATS 2023
Y2 - 25 April 2023 through 27 April 2023
ER -