A Mean Field Theory of Quantized Deep Networks: The Quantization-Depth Trade-Off: The quantization-depth trade-off

Yaniv Blumenfeld, Dar Gilboa, Daniel Soudry

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

Reducing the precision of weights and activation functions in neural network training, with minimal impact on performance, is essential for the deployment of these models in resource-constrained environments. We apply mean field techniques to networks with quantized activations in order to evaluate the degree to which quantization degrades signal propagation at initialization. We derive initialization schemes which maximize signal propagation in such networks, and suggest why this is helpful for generalization. Building on these results, we obtain a closed form implicit equation for Lmax, the maximal trainable depth (and hence model capacity), given N, the number of quantization levels in the activation function. Solving this equation numerically, we obtain asymptotically: Lmax ? N1.82,.

Original languageEnglish
Title of host publication33rd Annual Conference on Neural Information Processing Systems, NeurIPS 2019
StatePublished - 2019
Event33rd Annual Conference on Neural Information Processing Systems, NeurIPS 2019 - Vancouver, Canada
Duration: 8 Dec 201914 Dec 2019

Conference

Conference33rd Annual Conference on Neural Information Processing Systems, NeurIPS 2019
Country/TerritoryCanada
CityVancouver
Period8/12/1914/12/19

All Science Journal Classification (ASJC) codes

  • Computer Networks and Communications
  • Information Systems
  • Signal Processing

Fingerprint

Dive into the research topics of 'A Mean Field Theory of Quantized Deep Networks: The Quantization-Depth Trade-Off: The quantization-depth trade-off'. Together they form a unique fingerprint.

Cite this