Two Regimes of Generalization for Non-Linear Metric Learning

Mark Kozdoba, Shie Mannor

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

A common approach to metric learning is to seek an embedding of the input data that behaves well with respect to the labels. While generalization bounds for linear embeddings are known, the non-linear case is not well understood. In this work we fill this gap by
providing uniform generalization guarantees for the case where the metric is induced by a neural network type embedding of the data. Specifically, we discover and analyze two regimes of behavior of the networks, which are roughly related to the sparsity of the last layer. The bounds corresponding to the first regime are based on the spectral and
-norms of the weight matrices, while the second regime bounds use the
-norm at the last layer, and are significantly stronger when the last layer is dense. In addition, we empirically evaluate the behavior of the bounds for networks trained with SGD on the MNIST and 20newsgroups datasets. In particular, we demonstrate that both regimes occur naturally on realistic data.
Original languageEnglish
Title of host publicationTenth International Conference on Learning Representations
Number of pages11
StatePublished - 2022
EventTenth International Conference on Learning Representations - Virtual
Duration: 25 Apr 202229 Apr 2022
Conference number: 10th
https://iclr.cc/Conferences/2022

Conference

ConferenceTenth International Conference on Learning Representations
Abbreviated titleICLR
Period25/04/2229/04/22
Internet address

Cite this