Non-isotropy Regularization for Proxy-based Deep Metric Learning

Karsten Roth, Oriol Vinyals, Zeynep Akata

IEEE Conference on Computer Vision and Pattern Recognition, CVPR

2022

Abstract

Deep Metric Learning (DML) aims to learn representation spaces on which semantic relations can simply be expressed through predefined distance metrics. Best performing approaches commonly leverage class proxies as sample stand-ins for better convergence and generalization. However, these proxy-methods solely optimize for sample-proxy distances. Given the inherent non-bijectiveness of used distance functions, this can induce locally isotropic sample distributions, leading to crucial semantic context being missed due to difficulties resolving local structures and intraclass relations between samples. To alleviate this problem, we propose non-isotropy regularization (NIR) for proxy-based Deep Metric Learning. By leveraging Normalizing Flows, we enforce unique translatability of samples from their respective class proxies. This allows us to explicitly induce a non-isotropic distribution of samples around a proxy to optimize for. In doing so, we equip proxy-based objectives to better learn local structures. Extensive experiments highlight consistent generalization benefits of NIR while achieving competitive and state-of-the-art performance on the standard benchmarks CUB200-2011, Cars196 and Stanford Online Products. In addition, we find the superior convergence properties of proxy-based methods to still be retained or even improved, making NIR very attractive for practical usage.

Visual similarity plays a crucial role for applications in image & video retrieval, clustering, face re-identification or general supervised and unsupervised contrastive representation learning. A majority of approaches used in these fields employ or can be derived from Deep Metric Learning (DML). DML aims to learn highly nonlinear distance metrics parametrized by deep networks that span a representation space in which semantic relations between images are expressed as distances between respective representations.

In the field of DML, methods utilizing proxies have shown to provide among the most consistent and highest performances in addition to fast convergence. While other methods introduce ranking tasks over samples for the network to solve, proxy-based methods require the network to contrast samples against a proxy representation, commonly approximating generic class prototypes. Their utilization addresses sampling complexity issues inherent to purely sample-based approaches, resulting in improved convergence and benchmark performance.

However, there is no free lunch. Relying on sample-proxy relations, relations between samples within a class can not be explicitly captured. This is exacerbated by proxy-based objectives optimizing for distances between samples and proxies using non-bijective distance functions. This means, for a particular proxy, that alignment to a sample is non-unique - as long as the angle between sample and proxy is retained, i.e. samples being aligned isotropically around a proxy (see thumbnail figure), their distances and respective loss remain the same. This means that samples lie on a hypersphere centered around a proxy with same distance and thus incurring the same training loss. This incorporates an undesired prior over sample-proxy distributions which doesn't allow local structures to be resolved well.

By incorporating multiple classes and proxies (which is automatically done when applying proxy-based losses such as to training data with multiple classes), this is extended to a mixture of sample distributions around proxies. While this offers an implicit workaround to address isotropy around modes by incorporating relations of samples to proxies from different classes, relying only on other unrelated proxies potentially far away makes fine-grained resolution of local structures difficult. Furthermore, as training progresses and proxies move further apart. As a consequence, the distribution of samples around proxies, which proxy-based objectives optimize for, comprises modes with high affinity towards local isotropy. This introduces semantic ambiguity, as semantic relations between samples within a class are not resolved well, and intraclass semantics can not be addressed. However, a lot of recent work has shown that understanding and incorporating these non-discriminative relations drives generalization performance.

To address this, we propose a novel method that

- introduces a principled
**Non-Isotropy Regularization**-module ($\mathbb{NIR}$) for proxy-based DML on the basis of class-conditioned normalizing flows, - avoids the use of sample-to-sample relations, thus maintaining fast convergence speeds of proxy-based methods,
- has minimal impact of overall training time while signficantly and reliably improving generalization performance of proxy-based DML,
- consistently improves upon structures (such as Uniformity or feature variety) often related to better out-of-distribution generalization performance, as shown in multiple ablational studies.

To address the non-learnability of intraclass context in proxy-based DML, we have to address the inherent issue of local isotropy in the learned sample-proxy distribution $p(\psi|\rho)$.

**However**, to retain the convergence (and generalization) benefits of proxy-based methods, we have to find a $p(\psi|\rho)$, whose optimization better resolves the distribution of sample representations $\psi$ around our proxy $\rho$, **without** resorting to additional augmentations that move the overall objective away from a purely proxy-based one.

Assuming feature representations $\psi_i\in\Psi$ taken from the (pseudo-)metric presentation space $\Psi$ spanned by a deep neural network $\psi$ and respective class proxy representations $\rho$, we argue the following: Non-isotropy can be achieved by breaking down the fundamental issue of non-bijectivity in the used similarity measure $s(\psi, \rho)$ (or equivalently the distance function $d(\cdot, \cdot)$), which (on its own) introduces non-unique sample-proxy relations.

Consequently, we propose to enforce that each sample representation $\psi$ can be mapped by a **bijective** and thus **invertible** (deterministic) translation $\psi = \tau(\zeta|\rho)$ which, given some residual $\zeta$ from some prior distribution $q(\zeta)$, allows to uniquely translate from the respective proxy $\rho$ to $\psi$.

Such invertible, ideally non-linear translations $\tau$ are naturally expressed through Normalizing Flows (NF), which can be generally seen as a transformation between two probability distributions, most commonly between simple, well-defined ones and complex multimodal ones. Normalizing Flows introduce a sequence of non-linear, but still invertible coupling operations as showcased via $\leftrightharpoons$ in our setup figure. Given some input representation $\psi$, a coupling block splits $\psi$ into $\psi_1$ and $\psi_2$, which are scaled and translated in succession with non-linear scaling and translation networks $\eta_1$ and $\eta_2$, respectively. Note that following \cite{glow}, each network $\eta_i$ provides both scaling $\eta_i^s$ and translation values $\eta_i^t$. Successive application of different $\eta_i$ then gives our non-linear invertible transformation $\tau$ from some prior distribution over residuals $q(\zeta)$ with explicit density and CDF (for sampling) to our target distribution.

By conditioning $\tau$ on respective classproxies $\rho$, we can induce a new sample representation distribution $p(\tau(\zeta|\rho)|\rho)$ as *pushforward* from our prior distribution of residuals $q(\zeta)$ which accounts for unique sample-to-proxy relations, and which we wish to impose over our learned sample distribution $p(\psi|\rho)$. This introduces our Non-Isotropy Regularization $\mathbb{NIR}$

In more detail, $\mathbb{NIR}$ can be naturally approached through maximization of the expected log-likelihood $\mathbb{E}_{x, \rho_{y_x}}\left[\log p\left(\psi(x)|\rho_{y_x}\right)\right]$ over sample-proxy pairs $(x, \rho_{y_x})$, but under the constraint that each distribution of samples around a respective proxy, $p(\psi|\rho)$, is a *pushforward* of $\tau$ from our residual distribution $q(\zeta)$. This gives

$\mathcal{L}_\mathbb{NIR} = -\mathbb{E}_{x, \rho_{y_x}}[\log q\left(\tau^{-1}(\psi(x)|\rho_{y_x})\right) + \log|\det J_{\tau^{-1}}(\tau^{-1}(\psi(x)|\rho_{y_x})|\rho_{y_x})|]$

with Jacobian $J$ for translation $\tau^{-1}$ and proxies $\rho_{y_x}$, where $y_x$ denotes the class of sample $x$. To arrive at above equation, we simply leveraged the change of variables formula

$p(\psi|\rho) = q(\tau^{-1}(\psi|\rho))|\det J_{\tau^{-1}}(\tau^{-1}(\psi|\rho)|\rho)|$

In practice, by setting our prior $q(\zeta)$ to be a standard zero-mean unit-variance normal distribution $\mathcal{N}(0, 1)$, we get

$\mathcal{L}_\mathbb{NIR} = \frac{1}{|\mathcal{B}|}\sum_{(x, \rho_{y_x})\sim\mathcal{B}}\left\Vert \tau^{-1}(\psi(x)|\rho_{y_x})\right\Vert^2_2 - \log |\det J_{\tau^{-1}}(\tau^{-1}(\psi(x)|\rho_{y_x})|\rho_{y_x})|$

i.e. given sample representations $\psi(x)$, we project them onto our residual space $\zeta$ via $\tau^{-1}$ and compute our $\mathbb{NIR}$-objective. By selecting suitable normalizing flows such as GLOW, we make sure that the Jacobian is cheap to compute.

*In applying $\mathbb{NIR}$, we effectively enforce that all sample representations around each proxies are defineable through unique translations. This means that our underlying Normalizing Flow retains implicit knowledge about class-internal structures without our overall setup having to ever rely on sample-to-sample relations!*

By applying $\mathbb{NIR}$ to various strong proxy-based DML objectives such as *ProxyNCA*, *ProxyAnchor* or *SoftTriplet*, we find consistent improvements in generalization performance across all methods and in particular across all kinds of benchmarks.

In addition to that, these performance improvements come at no relevant impact to training time, as $\mathbb{NIR}$ only operates in the much lower-dimensional representation space. Furthermore, and importantly, convergence speeds are retained and in parts even improved!

Finally, when evaluating $\mathbb{NIR}$-regularized objectives, we find consistent increases in feature space uniformity ($G_2$) and feature variety ($\rho$), reduced overclustering ($\pi_\text{density}$) as well as higher variability in class-cluster sizes $\sigma_\kappa^2$. While the former three are commonly linked to better generalization performance, the latter provides nice additional support on $\mathbb{NIR}$ encouraging more class-specific sample-distributions to be learned!

(c) 2023 Explainable Machine Learning Munich Impressum