TY - GEN
T1 - Failures of gradient-based deep learning
AU - Shalev-Shwartz, Shai
AU - Shamir, Ohad
AU - Shammah, Shaked
N1 - Publisher Copyright: Copyright © 2017 by the author(s).
PY - 2017
Y1 - 2017
N2 - In recent years, Deep Learning has become the go-to solution for a broad range of applications, often outperforming state-of-the-art. However, it is important, for both theoreticians and practitioners, to gain a deeper understanding of the difficulties and limitations associated with common approaches and algorithms. We describe four types of simple problems, for which the gradientbased algorithms commonly used in deep learning either fail or suffer from significant difficulties. We illustrate the failures through practical experiments, and provide theoretical insights explaining their source, and how they might be remedied.
AB - In recent years, Deep Learning has become the go-to solution for a broad range of applications, often outperforming state-of-the-art. However, it is important, for both theoreticians and practitioners, to gain a deeper understanding of the difficulties and limitations associated with common approaches and algorithms. We describe four types of simple problems, for which the gradientbased algorithms commonly used in deep learning either fail or suffer from significant difficulties. We illustrate the failures through practical experiments, and provide theoretical insights explaining their source, and how they might be remedied.
UR - http://www.scopus.com/inward/record.url?scp=85048576310&partnerID=8YFLogxK
M3 - Conference contribution
T3 - 34th International Conference on Machine Learning, ICML 2017
SP - 4694
EP - 4708
BT - 34th International Conference on Machine Learning, ICML 2017
T2 - 34th International Conference on Machine Learning, ICML 2017
Y2 - 6 August 2017 through 11 August 2017
ER -