TY - GEN
T1 - OPTIMIZATION GUARANTEES FOR ISTA AND ADMM BASED UNFOLDED NETWORKS
AU - Pu, Wei
AU - Eldar, Yonina C.
AU - Rodrigues, Miguel R.D.
N1 - Publisher Copyright: © 2022 IEEE
PY - 2022
Y1 - 2022
N2 - Recently, unfolding techniques have been widely utilized to solve the inverse problems in various applications. In this paper, we study optimization guarantees for two popular unfolded networks, i.e., unfolded networks derived from iterative soft thresholding algorithms (ISTA) and derived from Alternating Direction Method of Multipliers (ADMM). Our guarantees - leveraging the Polyak-Lojasiewicz* (PL*) condition - state that the training (empirical) loss decreases to zero with the increase in the number of gradient descent epochs provided that the number of training samples is less than some threshold that depends on various quantities underlying the desired information processing task. Our guarantees also show that this threshold is larger for unfolded ISTA in comparison to unfolded ADMM, suggesting that there are certain regimes of number of training samples where the training error of unfolded ADMM does not converge to zero whereas the training error of unfolded ISTA does. A number of numerical results are provided backing up our theoretical findings.
AB - Recently, unfolding techniques have been widely utilized to solve the inverse problems in various applications. In this paper, we study optimization guarantees for two popular unfolded networks, i.e., unfolded networks derived from iterative soft thresholding algorithms (ISTA) and derived from Alternating Direction Method of Multipliers (ADMM). Our guarantees - leveraging the Polyak-Lojasiewicz* (PL*) condition - state that the training (empirical) loss decreases to zero with the increase in the number of gradient descent epochs provided that the number of training samples is less than some threshold that depends on various quantities underlying the desired information processing task. Our guarantees also show that this threshold is larger for unfolded ISTA in comparison to unfolded ADMM, suggesting that there are certain regimes of number of training samples where the training error of unfolded ADMM does not converge to zero whereas the training error of unfolded ISTA does. A number of numerical results are provided backing up our theoretical findings.
KW - Algorithm unfolding
KW - Polyak-Lojasiewicz (PL) condition
KW - optimization guarantee
UR - http://www.scopus.com/inward/record.url?scp=85131252523&partnerID=8YFLogxK
U2 - 10.1109/ICASSP43922.2022.9746860
DO - 10.1109/ICASSP43922.2022.9746860
M3 - منشور من مؤتمر
T3 - ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings
SP - 8687
EP - 8691
BT - 2022 IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP 2022 - Proceedings
T2 - 2022 IEEE International Conference on Acoustics, Speech and Signal Processing, ICASSP 2022
Y2 - 22 May 2022 through 27 May 2022
ER -