TY - GEN
T1 - Mean Estimation in High-Dimensional Binary Markov Gaussian Mixture Models
AU - Zhang, Yihan
AU - Weinberger, Nir
N1 - Publisher Copyright: © 2022 Neural information processing systems foundation. All rights reserved.
PY - 2022
Y1 - 2022
N2 - We consider a high-dimensional mean estimation problem over a binary hidden Markov model, which illuminates the interplay between memory in data, sample size, dimension, and signal strength in statistical inference. In this model, an estimator observes n samples of a d-dimensional parameter vector θ∗ ∈ Rd, multiplied by a random sign Si (1 ≤ i ≤ n), and corrupted by isotropic standard Gaussian noise. The sequence of signs {Si}i∈[n] ∈ {−1, 1}n is drawn from a stationary homogeneous Markov chain with flip probability δ ∈ [0, 1/2]. As δ varies, this model smoothly interpolates two well-studied models: the Gaussian Location Model for which δ = 0 and the Gaussian Mixture Model for which δ = 1/2. Assuming that the estimator knows δ, we establish a nearly minimax optimal (up to logarithmic factors) estimation error rate, as a function of ∥θ∗∥, δ, d, n. We then provide an upper bound to the case of estimating δ, assuming a (possibly inaccurate) knowledge of θ∗. The bound is proved to be tight when θ∗ is an accurately known constant. These results are then combined to an algorithm which estimates θ∗ with δ unknown a priori, and theoretical guarantees on its error are stated.
AB - We consider a high-dimensional mean estimation problem over a binary hidden Markov model, which illuminates the interplay between memory in data, sample size, dimension, and signal strength in statistical inference. In this model, an estimator observes n samples of a d-dimensional parameter vector θ∗ ∈ Rd, multiplied by a random sign Si (1 ≤ i ≤ n), and corrupted by isotropic standard Gaussian noise. The sequence of signs {Si}i∈[n] ∈ {−1, 1}n is drawn from a stationary homogeneous Markov chain with flip probability δ ∈ [0, 1/2]. As δ varies, this model smoothly interpolates two well-studied models: the Gaussian Location Model for which δ = 0 and the Gaussian Mixture Model for which δ = 1/2. Assuming that the estimator knows δ, we establish a nearly minimax optimal (up to logarithmic factors) estimation error rate, as a function of ∥θ∗∥, δ, d, n. We then provide an upper bound to the case of estimating δ, assuming a (possibly inaccurate) knowledge of θ∗. The bound is proved to be tight when θ∗ is an accurately known constant. These results are then combined to an algorithm which estimates θ∗ with δ unknown a priori, and theoretical guarantees on its error are stated.
UR - http://www.scopus.com/inward/record.url?scp=85163142257&partnerID=8YFLogxK
M3 - منشور من مؤتمر
T3 - Advances in Neural Information Processing Systems
BT - Advances in Neural Information Processing Systems 35 - 36th Conference on Neural Information Processing Systems, NeurIPS 2022
A2 - Koyejo, S.
A2 - Mohamed, S.
A2 - Agarwal, A.
A2 - Belgrave, D.
A2 - Cho, K.
A2 - Oh, A.
T2 - 36th Conference on Neural Information Processing Systems, NeurIPS 2022
Y2 - 28 November 2022 through 9 December 2022
ER -