Implicit Bias of Gradient Descent on Linear Convolutional Networks

Suriya Gunasekar, Jason D. Lee, Daniel Soudry, Nathan Srebro

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

We show that gradient descent on full width linear convolutional networks of depth L converges to a linear predictor related to the `2/L bridge penalty in the frequency domain. This is in contrast to fully connected linear networks, where regardless of depth, gradient descent converges to the `2 maximum margin solution.

Original languageEnglish
Title of host publication32nd Conference on Neural Information Processing Systems, NeurIPS 2018
Pages9461-9471
Number of pages11
StatePublished - 2018
Event32nd Conference on Neural Information Processing Systems, NeurIPS 2018 - Montreal, Canada
Duration: 2 Dec 20188 Dec 2018

Conference

Conference32nd Conference on Neural Information Processing Systems, NeurIPS 2018
Country/TerritoryCanada
CityMontreal
Period2/12/188/12/18

All Science Journal Classification (ASJC) codes

  • Computer Networks and Communications
  • Information Systems
  • Signal Processing

Fingerprint

Dive into the research topics of 'Implicit Bias of Gradient Descent on Linear Convolutional Networks'. Together they form a unique fingerprint.

Cite this