Next Article in Journal
A New Strategy in Boosting Information Spread
Next Article in Special Issue
A Method for Unsupervised Semi-Quantification of Inmunohistochemical Staining with Beta Divergences
Previous Article in Journal
Stochastic Density Functional Theory on Lane Formation in Electric-Field-Driven Ionic Mixtures: Flow-Kernel-Based Formulation
Previous Article in Special Issue
An Information-Theoretic Analysis of the Cost of Decentralization for Learning and Inference under Privacy Constraints
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Discrete Infomax Codes for Supervised Representation Learning

1
Stanford AI Lab, Stanford University, Stanford, CA 94305, USA
2
NAVER AI Lab, Seongnam 13561, Korea
3
Standigm, Seoul 06234, Korea
4
Intellicode & BARO AI Academy, Seoul 06367, Korea
*
Author to whom correspondence should be addressed.
Entropy 2022, 24(4), 501; https://doi.org/10.3390/e24040501
Submission received: 26 February 2022 / Revised: 26 March 2022 / Accepted: 29 March 2022 / Published: 2 April 2022
(This article belongs to the Special Issue Theory and Applications of Information Processing Algorithms)

Abstract

:
For high-dimensional data such as images, learning an encoder that can output a compact yet informative representation is a key task on its own, in addition to facilitating subsequent processing of data. We present a model that produces discrete infomax codes (DIMCO); we train a probabilistic encoder that yields k-way d-dimensional codes associated with input data. Our model maximizes the mutual information between codes and ground-truth class labels, with a regularization which encourages entries of a codeword to be statistically independent. In this context, we show that the infomax principle also justifies existing loss functions, such as cross-entropy as its special cases. Our analysis also shows that using shorter codes reduces overfitting in the context of few-shot classification, and our various experiments show this implicit task-level regularization effect of DIMCO. Furthermore, we show that the codes learned by DIMCO are efficient in terms of both memory and retrieval time compared to prior methods.

1. Introduction

Metric learning and few-shot classification are two problem settings that test a model’s ability to classify data from classes that were unseen during training. Such problems are also commonly interpreted as testing meta-learning ability, since the process of constructing a classifier with examples from new classes can be seen as learning. Many recent works [1,2,3,4] tackled this problem by learning a continuous embedding ( x ˜ R n ) of datapoints. Such models compare pairs of embeddings using, e.g., Euclidean distance to perform nearest neighbor classification. However, it remains unclear whether such models effectively utilize the entire space of R n .
Information theory provides a framework for effectively asking such questions about representation schemes. In particular, the information bottleneck principle [5,6] characterizes the optimality of a representation. This principle states that the optimal representation X ˜ is one that maximally compresses the input X while also being predictive of labels Y. From this viewpoint, we see that the previous methods which map data to R n focus on being predictive of labels Y without considering the compression of X.
The degree of compression of an embedding is the number of bits it reflects about the original data. Note that for continuous embeddings, each of the n numbers in a n-dimensional embedding requires 32 bits. It is unlikely that unconstrained optimization of such embeddings use all of these 32 n bits effectively. We propose to resolve this limitation by instead using discrete embeddings and controlling the number of bits in each dimension via hyperparameters. To this end, we propose a model that produces discrete infomax codes (DIMCO) via an end-to-end learnable neural network encoder.
This work’s primary contributions are as follows. We usw mutual information as an objective for learning embeddings, and propose an efficient method of estimating it in the discrete case. We experimentally demonstrate that learned discrete embeddings are more memory and time-efficient compared to continuous embeddings. Our experiments also show that using discrete embeddings help meta-generalization by acting as an information bottleneck. We also provide theoretical support for this connection through an information-theoretic probably approximately correct (PAC) bound that shows the generalization properties of learned discrete codes.
This paper is organized as follows. We propose our model for learning discrete codes in Section 2. We justify our loss function and also provide a generalization bound for our setup in Section 3. We compare our method to related work in Section 4, and present experimental results in Section 5. Finally, we conclude our paper in Section 6.

2. Discrete Infomax Codes (Dimco)

We present our model which produces discrete infomax codes (DIMCO). A deep neural network is trained end-to-end to learn k-way d-dimensional discrete codes that maximally preserve the information on labels. We outline the training procedure in Algorithm 1, and we also illustrate the overall structure in the case of 4-way 3-dimensional codes ( k = 4 , d = 3 ) in Figure 1.
Algorithm 1 DIMCO training procedure.
  • Initialize network parameters θ
  • repeat
  •    Sample a batch { ( x n , y n ) }
  •    Compute logits { l i , j n } = f ( x n ; θ )             via (1)
  •    Compute probs { p i , j n } = softmax ( { l i , j n } )     via (2)
  •    Update θ to minimize loss L ( { p i , j n } , { y n } )   via (10)
  • until converged

2.1. Learnable Discrete Codes

Suppose that we are given a set of labeled examples, which are realizations of random variables ( X , Y ) p ( x , y ) , where X is the continuous input, and its corresponding discrete label is Y. Realizations of X and Y are denoted by x R D and y { 1 , , c } . The codebook X ˜ serves as a compressed representation of X.
We constructed a probabilistic encoder p ( x ˜ | x ) —which is implemented by a deep neural network—that maps an input x to a k-way d-dimensional code x ˜ { 1 , 2 , , k } d . That is, each entry of x ˜ takes on one of k possible values, and the cardinality of X ˜ is | X ˜ | = k d . Special cases of this coding scheme include k-way class labels ( d = 1 ), d-dimensional binary codes ( k = 2 ), and even fixed-length decimal integers ( k = 10 ).
We now describe our model which produces discrete infomax codes. A neural network encoder f ( x ; θ ) outputs k-dimensional categorical distributions, Cat ( p i , 1 , , p i , k ) . Here, p i , j ( x ) represents the probability that output variable i takes on value j, consuming x as an input, for i = 1 , , d and j = 1 , , k . The encoder takes x as an input to produce logits l i , j = f ( x ) i , j , which form a d × k matrix:
l 1 , 1 l 1 , 2 l 1 , 3 l 1 , k l 2 , 1 l 2 , 2 l 2 , 3 l 2 , k l d , 1 l d , 2 l d , 3 l d , k .
These logits undergo softmax functions to yield
p i , j = exp ( l i , j ) j = 1 k exp ( l i , j ) .
Each example x in the training set is assigned a codeword x ˜ = [ x ˜ 1 , , x ˜ d ] , each entry of which is determined by one of k events that is most probable; i.e.,
x ˜ = arg max j p 1 j x ˜ 1 , , arg max j p d j x ˜ d .
While the stochastic encoder p ( x ˜ | x ) induces a soft partitioning of input data, codewords assigned by the rule in (3) yield a hard partitioning of X.

2.2. Loss Function

The i-th symbol is assumed to be sampled from the resulting categorical distribution Cat ( p i 1 , , p i k ) . We denote the resulting distribution over codes as X ˜ and a code as x ˜ . Instead of sampling x ˜ X ˜ during training, we use a loss function that optimizes the expected performance of the entire distribution X ˜ .
We train the encoder by maximizing the mutual information between the distributions of codes X ˜ and labels Y. The mutual information is a symmetric quantity that measures the amount of information shared between two random variables. It is defined as
I ( X ˜ ; Y ) = H ( X ˜ ) H ( X ˜ | Y ) .
Since X ˜ and Y are discrete, their mutual information is bounded from both above and below as 0 I ( X ˜ ; Y ) log | X ˜ | = d log k . To optimize the mutual information, the encoder directly computes empirical estimates of the two terms on the right-hand side of (4). Note that both terms consist of entropies of categorical distributions, which have the general closed-form formula:
H ( p 1 , , p n ) = i = 1 n p i log p i .
Let p ¯ i j be the empirical average of p i j calculated using data points in a batch. Then, p ¯ i j is an empirical estimate of the marginal distribution p ( x ˜ ) . We compute the empirical estimate of H ( X ˜ ) by adding its entropy estimate for each dimension.
H ^ ( X ˜ ) = i = 1 d H ( p ¯ i 1 , , p ¯ i k ) .
We can also compute
H ^ ( X ˜ | Y ) = y = 1 c p ( Y = y ) H ^ ( X ˜ | Y = y ) ,
where c is the number of classes. The marginal probability p ( Y = y ) is the frequency of class y in the minibatch, and H ( X ˜ | Y = y ) can be computed by computing (6) using only datapoints which belong to class y. We emphasize that such a closed-form estimation of I ( X ˜ ; Y ) is only possible because we are using discrete codes. If X ˜ were instead a continuous variable, we would only be able to maximize an approximation of I ( X ˜ ; Y ) (e.g., Belghazi et al. [7]).
We briefly examine the loss function (4) to see why maximizing it results in discriminative X ˜ . Maximizing H ( X ˜ ) encourages the distribution of all codes to be as dispersed as possible, and minimizing H ( X ˜ | Y ) encourages the average embedding of each class to be as concentrated as possible. Thus, the overall loss I ( X ˜ ; Y ) imposes a partitioning problem on the model: it learns to split the entire probability space into regions with minimal overlap between different classes. As this problem is intractable for the large models considered in this work, we seek to find a local minima via stochastic gradient descent (SGD). We provide a further analysis of this loss function in Section 3.1.

2.3. Similarity Measure

Suppose that all data points in the training set are assigned their codewords according to the rule (3). Now we introduce how to compute a similarity between a query datapoint x ( q ) and a support datapoint x ( s ) for information retrieval or few-shot classification, where the superscripts ( q ) , ( s ) stand for query and support, respectively. Denote by x ˜ ( s ) the codeword associated with x ( s ) , constructed by (3). For the test data x ( q ) , the encoder yields p i , j ( x ( q ) ) for i = 1 , , d and j = 1 , , k . As a similarity measure between x ( q ) and x ( s ) , we calculate the following log probability.
i = 1 d log p i , x ˜ i ( s ) ( x ( q ) ) .
The probabilistic quantity (8) indicates that x ( q ) and x ( s ) become more similar when the encoder’s output—when x ( q ) is provided—is well aligned with x ˜ ( s ) .
We can view our similarity measure (8) as a probabilistic generalization of the Hamming distance [8]. The Hamming distance quantifies the similarity between two strings of equal length as the number of positions at which the corresponding symbols are equal. As we have access to a distribution over codes, we use (8) to directly compute the log probability of having the same symbol at each position.
We use (8) as a similarity metric for both few-shot classification and image retrieval. We perform few-shot classification by computing a codeword for each class via (3) and classifying each test image by choosing the class that has the highest value of (8). We similarly perform image retrieval by mapping each support image to its most likely code (3) and for each query image retrieving the support image that has the highest (8).
While we have described the operations in (3) and (8) for a single pair ( x ( q ) , x ( s ) ) , one can easily parallelize our evaluation procedure, since it is an argmax followed by a sum. Furthermore, x ˜ typically requires little memory, as it consists of discrete values, allowing us to compare against large support sets in parallel. Experiments in Section 5.4 investigate the degree of DIMCO’s efficiency in terms of both time and memory.

2.4. Regularizing by Enforcing Independence

One way of interpreting the code distribution X ˜ is as a group of d separate code distributions x ˜ 1 , , x ˜ d . Note that the similarity measure described in (8) can be seen as ensemble of the similarity measures of these d models. A classic result in ensemble learning is that using more diverse learners increases ensemble performance [9]. In a similar spirit, we used an optional regularizer which promotes pairwise independence between each pair in these d codes. Using this regularizer stabilized training, especially in more large-scale problems.
Specifically, we randomly sample pairs of indices i 1 , i 2 from 1 , , d during each forward pass. Note that x ˜ i 1 x ˜ i 2 and ( x ˜ i 1 , x ˜ i 2 ) are both categorical distributions with support size k 2 , and that we can estimate the two different distributions within each batch. We minimize their KL divergence to promote independence between these two distributions:
L ind = D K L ( x ˜ i 1 x ˜ i 2 x ˜ i 1 , x ˜ i 2 ) .
We compute (9) for a fixed number of random pairs of indices for each batch. The cost of computing this regularization term is miniscule compared to that of other components such as feeding data through the encoder.
Using this regularizer in conjunction with the learning objective (4) yields the following regularized loss:
L = I ( X ˜ ; Y ) + λ L ind = H ( X ˜ ) + H ( X ˜ | Y ) + λ L ind .
We fix λ = 1 in all experiments, as we found that DIMCO’s performance was not particularly sensitive to this hyperparameter. We emphasize that while this optional regularizer stabilizes training, our learning objective is the mutual information I ( X ˜ ; Y ) in (4).

2.5. Visualization of Codes

In Figure 2, we show images retrieved using our similarity measure (8). We trained a DIMCO model ( k = 16 , d = 4 ) on the CIFAR100 dataset. We selected specific code locations and plotted the top 10 test images according to our similarity measure. For example, the top (leftmost) image for code ( · , j 2 , · , j 4 ) would be computed as
arg max n { 1 , , N } log p 2 , j 2 ( x n ) + log p 4 , j 4 ( x n ) ,
where N is the number of test images.
We visualize two different combinations of codes in Figure 2. The two examples show that using codewords together results in their respective semantic concepts being combined: (man + fish = man holding fish), (round + warm color = orange). While we visualized combinations of 2 codewords for clarity, DIMCO itself uses a combination of d such codewords. The regularizer described in Section 2.4 further encourages each of these d codewords to represent different concepts. The combinatorially many ( k d ) combinations in which DIMCO can assemble such codewords gives DIMCO sufficient expressive power to solve challenging tasks.

3. Analysis

3.1. Is Mutual Information a Good Objective?

Our learning objective for DIMCO (4) is the mutual information between codes and labels. In this subsection, we justify this choice by showing that many previous objectives are closely related to mutual information. Due to space constraints, we only show high-level connections here and provide a more detailed exposition in Appendix A.

3.1.1. Cross-Entropy

The de facto loss for classification is the cross-entropy loss, which is defined as
xent ( Y , X ) = E y Y , x X log q ( Y ^ = y | X ˜ ( x ) ) ,
where Y ^ is the model’s prediction of Y. Using the observation that the final layer q ( · ) acts a parameterized approximation of the true conditional distribution p ( Y | X ˜ ) , we write this as
xent ( Y , X ) E y Y , x X log p ( y | x ˜ ) = I ( X ˜ ; Y ) + H ( Y ) .
The H ( Y ) term can be ignored since it is not affected by model parameters. Therefore, minimizing cross-entropy is approximately equivalent to maximizing mutual information. The two objectives become completely equivalent when the final linear layer q ( · ) perfectly represents the conditional distribution q ( y | x ˜ ) . Note that for discrete x ˜ , we cannot use a linear layer to parameterize q ( y | x ˜ ) , and therefore, cannot directly optimize the cross-entropy loss. We can therefore view our loss as a necessary modification of the cross-entropy loss for our setup of using discrete embeddings.

3.1.2. Contrastive Losses

Many metric learning methods [1,2,10,11,12] use a contrastive learning objective to learn a continuous embedding ( X ˜ ). Such contrastive losses consist of (1) a positive term that encourages an embedding to move closer to that of other relevant embeddings and (2) a negative term that encourages it to move away from irrelevant embeddings. The positive term approximately minimizes log p ( X ˜ | y ) , and the negative term as approximately minimizes log p ( X ˜ ) . Together, these terms have the combined effect of maximizing
E log p ( x ˜ | y ) log p ( x ˜ )       = H ( X ˜ | Y ) + H ( X ˜ ) = I ( X ˜ ; Y ) .
We show such equivalences in detail in Appendix A.
In addition to these direct connections to previous loss functions, we show empirically in Section 5.1 that the mutual information strongly correlates with both the top-1 accuracy metric for classification and the Recall@1 metric for retrieval.

3.2. Does Using Discrete Codes Help Generalization?

In Section 1, we have provided motivation for the use of discrete codes through the regularization effect of an information bottleneck. In this subsection, we theoretically analyze whether learning discrete codes by maximizing mutual information leads to better generalization. In particular, we study how the mutual information on the test set is affected by the choice of input dataset structure and code hyperparameters k and d through a PAC learning bound.
We analyze DIMCO’s characteristics at the level of minibatches. Following related meta-learning works [13,14], we call each batch a “task”. We note that this is only a difference in naming convention, and our analysis applies equally well to the metric learning setup: we can view each batch consisting of support and query points as a task.
Define a task T to be a distribution over Z = X × Y . Let tasks T 1 , , T n be sampled i.i.d. from a distribution of tasks τ . Each task T consists of a fixed-size dataset D T = z T 1 , , z T m = ( x T 1 , y T 1 ) , , ( x T m , y T m ) , which is a set of m i.i.d. samples from the data distribution ( z T j T ). Let θ be the parameters of DIMCO. Let X , Y , X ˜ be the random variables for data, labels, and codes, respectively. Recall that our objective is the expected mutual information between labels and codes:
L ( τ , θ ) = E T τ I ( X ˜ ( X T , θ ) ; Y T ) .
The loss that we actually optimize (Equations (6) and (7)) is the empirical loss:
L ^ ( T 1 : n , θ ) = 1 n i = 1 n I ^ ( X ˜ ( X T i , θ ) ; Y T i ) .
The following theorem bounds the difference between the expected loss L and the empirical loss L ^ .
Theorem 1.
Let d Θ be the VC dimension of the encoder X ˜ ( · ) . The following inequality holds with high probability:
L ( τ , θ ) L ^ ( T 1 : n , θ ) O d Θ n log n d Θ + O | X ˜ | log ( m ) m + O | X ˜ | | Y | m .
Proof. 
We use VC dimension bounds and a finite sample bound for mutual information [15]. We defer a detailed statement and proof to Appendix B.    □
First note that all three terms in our generalization gap (A11) converge to zero as n , m . This shows that training a model by maximizing empirical mutual information, as in Equations (6) and (7), generalizes perfectly in the limit of infinite data.
Theorem 1 also shows how the generalization gap is affected differently by dataset size m and number of datasets n. A large n directly compensates for using a large backbone ( d Θ ), and a large m compensates for using a large final representation ( X ˜ ). Put differently, to effectively learn from small datasets (m), one should use a small representation ( X ˜ ) . The number of datasets n is typically less of a problem because the number of different ways to sample datasets is combinatorially large (e.g., n > 10 10 for miniImagenet 5-way 1-shot tasks). Recall that DIMCO has | X ˜ | = d log k , meaning that we can control the latter two terms using our hyperparameters d , k . We have explained the use of discrete codes through the information bottleneck effect of small codes X ˜ , and Theorem 1 confirms this intuition.

4. Related Work

Information bottleneck. DIMCO and Theorem 1 are both close in spirit to the information bottleneck (IB) principle [5,6,16]. IB finds a set of compact representatives X ˜ while maintaining sufficient information about Y, minimizing the following objective function:
J ( p ( x ˜ | x ) ) = I ( X ˜ ; X ) β I ( X ˜ ; Y ) ,
subject to x ˜ p ( x ˜ | x ) = 1 . Equivalently, it can be stated that one maximizes I ( X ˜ ; Y ) while simultaneously minimizing I ( X ˜ ; X ) . Similarly, our objective (15) is information maximization I ( X ˜ ; Y ) , and our bound (A11) suggests that the representation capacity | X ˜ | should be low for generalization. In the deterministic information bottleneck [17], I ( X ˜ ; X ) is replaced by H ( X ˜ ) . These three approaches to generalization are related via the chain of inequalities I ( X ˜ ; X ) H ( X ˜ ) log | X ˜ | , which is tight in the limit of X ˜ being imcompressible. For any finite representation, i.e., | X ˜ | = N , the limit β in (18) yields a hard partitioning of X into N disjoint sets. DIMCO uses the infomax principle to learn N = k d such representatives, which are arranged by k-way d-dimensional discrete codes for compact representation with sufficient information on Y.
Regularizing meta-learning. Previous meta-learning methods have restricted task-specific learning by learning only a subset of the network [18], learning on a low-dimensional latent space [19], learning on a meta-learned prior distribution of parameters [20], and learning context vectors instead of model parameters [21]. Our analysis in Theorem 1 suggests that reducing the expressive power of the task-specific learner has a meta-regularizing effect, indirectly giving theoretical support for previous works that benefited from reducing the expressive power of task-specific learners.
Discrete representations. Discrete representations have been thoroughly studied in information theory [22]. Recent deep learning methods directly learn discrete representations by learning generative models with discrete latent variables [23,24,25] or maximizing the mutual information between representation and data [26]. DIMCO is related to but differs from these works, as it assumes a supervised meta-learning setting and performs infomax using labels instead of data.
A standard approach to learning label-aware discrete codes is to first learn continuous embeddings and then quantize it using an objective that maximally preserves its information [27,28,29]. DIMCO can be seen as an end-to-end alternative to quantization which directly learns discrete codes. Jeong and Song [30] similarly learns a sparse binary code in an end-to-end fashion by solving a minimum cost flow problem with respect to labels. Their method differs from DIMCO, which learns a dense discrete code by optimizing I ( X ˜ ; Y ) , which we estimate with a closed-form formula.
Metric learning. The structure and loss function of DIMCO are closely related to those of metric learning methods [1,11,12,31]. We show that the loss functions of these methods can be seen as approximations of the mutual information ( I ( X ˜ ; Y ) ) in Section 2.2, and provide more in-depth exposition in Appendix A. While all of these previous methods require a support/query split within each batch, DIMCO simply optimizes an information-theoretic quantity of each batch, removing the need for such structured batch construction.
Information theory and representation learning. Many works have applied information-theoretic principles to unsupervised representation learning: to derive an objective for GANs to learn disentangled features [32], to analyze the evidence lower bound (ELBO) [33,34], and to directly learn representations [35,36,37,38,39]. Related also are previous methods that enforce independence within an embedding [40,41]. DIMCO is also an information-theoretic representation learning method, but we instead assume a supervised learning setup where the representation must reflect ground-truth labels. We also used previous results from information theory to prove a generalization bound for our representation learning method.

5. Experiments

In our experiments, we used datasets with varying degrees of complexity: CIFAR10/100 [42], miniImageNet [31], CUB200 [43], Cars196 [44], and ImageNet (ILSVRC-2012-CLS, Deng et al. [45]). We used standard train/test splits for each dataset unless stated otherwise. We also used various network architectures: 4-layer convnet [31] and ResNet12/20/50 [46,47]. We followed previously reported experimental setups as closely as possible, and provide minor experiment details in Appendix C.

5.1. Correlation of Metrics

We have shown in Section 3.1 that the mutual information I ( X ˜ ; Y ) is strongly connected to previous loss functions for classification and retrieval. In this subsection, we show experiments performed to verify whether I ( X ˜ ; Y ) is a good metric that quantitatively shows the quality of the representation X ˜ . We trained DIMCO on the miniImageNet dataset with k = d = 64 for 20 epochs. We plot the pairwise correlations between five different metrics: ( 5 , 10 , 20 )-way 1-shot accuracy, Recall @ 1 , and I ( X ˜ ; Y ) . The results in Figure 3 show that all five metrics are very strongly correlated. We observed similar trends when training with loss functions other than I ( X ˜ ; Y ) as well; we show these experiments in Appendix C due to space constraints.

5.2. Label-Aware Compression

We applied DIMCO to compressing feature vectors of trained classifier networks. We obtained penultimate embeddings of ResNet20 networks each trained on CIFAR10 and CIFAR100. The two networks had top-1 accuracies of 91.65 and 66.61 , respectively. We trained on embeddings for the train set of each dataset, and measured top-1 accuracy of the test set using the training set as support. We compare DIMCO to product quantization (PQ, Jegou et al. [28]), which similarly compresses a given embededing to a k-way d-dimensional code. We compare the two methods in Table 1 with the same range of k , d hyperparameters. We performed the same experiment on the larger ImageNet dataset with a ResNet50 network which had a top-1 accuracy of 76.00 . We compare DIMCO to both adaptive scalar quantization (SQ) and PQ in Table 2. We show extended experiments for all three datasets in Appendix A.
The results in Table 1 and Table 2 demonstrate that DIMCO consistently outperforms PQ, and is especially efficient when d is low. Furthermore, the ImageNet experiment (Table 2) shows that DIMCO even outperforms SQ, which has a much lower compression rate compared to the embedding sizes we consider for DIMCO. These results are likely due to DIMCO performing label-aware compression, where it compresses the embedding while taking the label into account, whereas PQ and SQ only compress the embeddings themselves.
Figure 3. Pairwise correlations between MI = I ( X ; X ˜ ) and previous metrics. Best viewed zoomed in.
Figure 3. Pairwise correlations between MI = I ( X ; X ˜ ) and previous metrics. Best viewed zoomed in.
Entropy 24 00501 g003

5.3. Few-Shot Classification

We evaluated DIMCO’s few-shot classification performance on the miniImageNet dataset. We compare our method against the following previous works: Snell et al. [3], Vinyals et al. [31], Liu et al. [48], Ye et al. [49], Ravi and Larochelle [50], Sung et al. [51], Bertinetto et al. [52], Lee et al. [53]. All methods use the standard four-layer convnet with 64 filters per layer. While some methods used more filters, we used 64 for a fair comparison. We used the data augmentation scheme proposed by Lee et al. [53] and used balanced batches of 100 images consisting of 10 different classes. We evaluated both 5-way 1-shot and 5-way 5-shot learning, and report 95 % confidence intervals of 1000 random episodes on the test split.
Results are shown in Table 3, and we provide an extended table with an alternative backbone in Appendix A. Figure Table 3 shows that DIMCO outperforms previous works on the 5-way 1-shot benchmark. DIMCO’s 5-way 5-shot performance is relatively low, likely because the similarity metric (Section 2.3) handles support datapoints individually instead of aggregating them, similarly to Matching Nets [31]. Additionally, other methods are explicitly trained to optimize 5-shot performance, whereas DIMCO’s training procedure is the same regardless of task structure.

5.4. Image Retrieval

We conducted image retrieval experiments using two standard benchmark datasets: CUB-200-2011 and Cars-196. As baselines, we used three widely adopted metric learning methods: Binomial Deviance [54], Triplet loss [1], and Proxy-NCA [2]. The backbone for all methods was a ResNet-50 network pretrained on the ImageNet dataset. We trained DIMCO on various combinations of ( p , d ) , and set the embedding dimension of the baseline methods to 128. We measured the time per query for each method on a Xeon E5-2650 CPU without any parallelization. We note that computing the retrieval time using a parallel implementation would skew the results even more in favor of DIMCO, since DIMCO’s evaluation is simply one memory access followed by a sum.
Results presented in Table 4 show that DIMCO outperforms all three baseline, and that the compact code of DIMCO takes roughly an order of magnitude less memory, and requires less query time as well. This experiment also demonstrates that discrete representations can outperform modern methods that use continuous embeddings, even on this relatively large-scale task. Additionally, this experiment shows that DIMCO can train using large backbones without significantly overfitting.

6. Discussion

We introduced DIMCO, a model that learns a discrete representation of data by directly optimizing the mutual information with the label. To evaluate our initial intuition that shorter representations generalize better between tasks, we provided generalization bounds that get tighter as the representation gets shorter. Our experiments demonstrated that DIMCO is effective at both compressing a continuous embedding, and also at learning a discrete embedding from scratch in an end-to-end manner. The discrete embeddings of DIMCO outperformed recent continuous feature extraction methods while also being more efficient in terms of both memory and time. We believe the tradeoff between discrete and continuous embeddings is an exciting area for future research.
DIMCO was motivated by concepts such as the minimum description length (MDL) principle and the information bottleneck: compact task representations should have less room to overfit. Interestingly, Yin et al. [55] reports that doing the opposite—regularizing the task-general parameters—prevents meta-overfitting by discouraging the meta-learning model from memorizing the given set of tasks. In future work, we will investigate the common principle underlying these seemingly contradictory approaches for a fuller understanding of meta-generalization.

Author Contributions

Conceptualization, Y.L., W.K. and S.C.; methodology, Y.L., W.K. and S.C.; software, Y.L., W.K. and W.P.; validation, Y.L., W.K., W.P. and S.C.; formal analysis, Y.L. and S.C.; investigation, Y.L., W.K., W.P. and S.C.; resources, Y.L.; writing—original draft preparation, Y.L.; writing—review and editing, Y.L., W.K., W.P. and S.C.; visualization, Y.L., W.K. and W.P.; supervision, S.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Previous Loss Functions Are Approximations of Mutual Information

Appendix A.1. Cross-Entropy Loss

The cross-entropy loss has directly been used for few-shot classification [3,31].
Let q ( y | x ˜ ; ϕ ) be a parameterized prediction of y given x ˜ , which tries to approximate the true conditional distribution q ( y | x ˜ ) . Typically, in a classification network, ϕ is the parameters of a learned projection matrix and q ( · ) is the final linear layer. The expected cross-entropy loss can be written as
xent ( Y , X ˜ ) = E y Y , x ˜ X ˜ log q ( y | x ˜ , ϕ ) .
Assuming that the approximate distribution q ( · ) is sufficiently close to p ( y | x ˜ ) , minimizing (A1) can be seen as
arg min xent ( Y , X ˜ ) arg min E y Y , x ˜ X ˜ log p ( y | x ˜ )
= arg min H ( Y | X ˜ ) = arg max I ( X ˜ ; Y ) ,
where the last equality uses the fact that H ( Y ) is independent of model parameters. Therefore, cross-entropy minimization is approximate maximization of the mutual information between representation X ˜ and labels Y.
The approximation is that we parameterized q ( y | x ˜ ; ϕ ) as a linear projection. This structure cannot generalize to new classes because the parameters ϕ are specific to the labels y seen during training. For a model to generalize to unseen classes, one must amortize the learning of this approximate conditional distribution. [3,31] sidestepped this issue by using the embeddings for each class as ϕ .

Appendix A.2. Triplet Loss

The triplet loss [1] is defined as
L triplet = x ˜ q x ˜ p 2 2 x ˜ q x ˜ n 2 2 ,
where x ˜ q , x ˜ p , x ˜ n R d are the embedding vectors of query, positive, and negative images. Let y q denote the label of the query data. Recall that the pdf function of a unit Gaussian is log N ( x ˜ | μ , 1 ) = c 1 c 2 x ˜ μ 2 2 , where c 1 , c 2 are constants. Let p p ( x ˜ ) = N ( x ˜ p , 1 ) and p n ( x ˜ ) = N ( x ˜ n , 1 ) be unit Gaussian distributions centered at x ˜ p , x ˜ n , respectively. We have
E L triplet E log p p ( x ˜ ) log p n ( x ˜ )
E log p p ( x ˜ ) log p ( x ˜ )
= H ( X ˜ | Y ) + H ( X ˜ ) = I ( X ˜ ; Y ) .
Two approximations were made in the process. We first assumed that the embedding distribution of images not in y q is equal to the distribution of all embeddings. This is reasonable when each class only represents a small fraction of the full data. We also approximated the embedding distributions p ( x ˜ | y ) , p ( x ˜ ) with unit Gaussian distributions centered at single samples from each.

Appendix A.3. N-Pair Loss

Multiclass N-pair loss [11] was proposed as an alternative to Triplet loss. This loss function requires one positive embedding x ˜ + and multiple negative embeddings x ˜ 1 , , x ˜ N 1 , and takes the form  
log exp ( x ˜ x ˜ + ) exp ( x ˜ x ˜ + ) + i = 1 N 1 exp ( x ˜ x ˜ i ) .
This can be seen as the cross-entropy loss applied to softmax ( x ˜ x ˜ + , x ˜ x ˜ 1 , , x ˜ x ˜ N 1 ) .
Following the same logic as the cross-entropy loss, this is also an approximation of I ( X ˜ ; Y ) . This objective should have less variance than Triplet loss since it approximates p ( x ˜ ) using more examples.

Appendix A.4. Adversarial Metric Learning

Deep Adversarial Metric Learning [12] tackles the problem of most negative exmples being uninformative by directly generating meaningful negative embeddings. This model employs a generator which takes as input the embeddings of anchor, positive, and negative images. The generator then outputs a “synthetic negative” embedding that is hard to distinguish from a positive embedding while being close to the negative embedding.
This can be seen as optimizing
E log p p ( x ˜ ) log p ( x ˜ ) = I ( X ˜ ; Y )
by estimating p ( x ˜ ) using a generative network rather than directly from samples. Rather than modelling the marginal distribution p ( x ˜ ) , this method conditionally models p ( x ˜ ; x ˜ q , x ˜ p , x ˜ n ) so that x ˜ is hard to distinguish from x ˜ p while sufficiently close to both x ˜ q and x ˜ n .

Appendix B. Proof of Theorem 1

We restate and prove our main theorem.
Theorem A1.
Let d Θ be the VC dimension of the encoder X ˜ ( · ) . Let I ^ ( X ˜ ( X T , θ ) ; Y T ) be the empirical estimate of the mutual information using finite dataset D T , and define empirical loss as
L ^ ( T 1 : n , θ ) = 1 n i = 1 n I ^ ( X ˜ ( X T i , θ ) ; Y T i ) .
The following inequality holds with high probability:
L ( τ , θ ) L ^ ( T 1 : n , θ ) O d Θ n log n d Θ + O | X ˜ | log ( m ) m + O | X ˜ | | Y | m
Proof. 
We use the following lemma from [15], which we restate using our notation.
Lemma A1.
Let X ˜ be a random mapping of X. Let D be a sample of size m drawn from the joint probability distribution p ( X , Y ) . Denote the empirical mutual information observed from D between X ˜ and Y as I ^ ( X ˜ ; Y ) . For any δ ( 0 , 1 ) , the following holds with probability at least 1 δ :
| I ( X ˜ ; Y ) I ^ ( X ˜ ; Y ) | ( 3 | X ˜ | + 2 ) log ( m ) log ( 4 / δ ) 2 m + ( | Y | + 1 ) ( | X ˜ | + 1 ) 4 m
We simplify this and plug in our specific quantities of interest ( X ˜ ( X T , θ ) , Y T ):
| I X ˜ ( X T , θ ) ; Y T I ^ X ˜ ( X T , θ ) ; Y T | O | X ˜ | log ( m ) m + O | X ˜ | | Y | m .
We similarly bound the error caused by estimating L with a finite number of tasks sampled from τ . Denote the finite sample estimate of L as  
L ^ ( τ , θ ) = 1 n i = 1 n I ( X ˜ ( X T i , θ ) ; Y T i ) .
Let the mapping X X ˜ be parameterized by θ Θ and let this model have VC dimension d Θ . Using d Θ , we can state that with high probability,
| L ( τ , θ ) L ^ ( τ , θ ) | O d Θ n log n d Θ ,
where d Θ is the VC dimension of hypothesis class Θ .
Combining equations (A15, A13), we have with high probability
| L ( τ , θ ) 1 n i = 1 n I ^ ( X ˜ ( X T i , θ ) ; Y T i ) |
| L ( τ , θ ) L ^ ( τ , θ ) | + O | X ˜ | log ( m ) m + O | X ˜ | | Y | m
O d Θ n log n d Θ + O | X ˜ | log ( m ) m + O | X ˜ | | Y | m .
   □

Appendix C. Experiments

Appendix C.1. Parameterizing the Code Layer

Recall that each discrete code is parameterized by a R k × d matrix. A problem with a naive implementation of DIMCO is that simply using a linear layer that maps R D to R k × d takes D k d parameters in that single layer. This can be prohibitively expensive for large embeddings, e.g., d = 4096 . We therefore, parameterize this code layer as the product of two matrices, which sequentially map R D R n R k × d . The total number of parameters required for this is n D + n k d . We fix all n = 128 . While more complicated tricks could reduce the parameter count even further, we found that this simple structure was sufficient to produce the results in this paper.

Appendix C.2. Correlation of Metrics

We collected statistics from 8 different independent runs, and report the averages of 500 batches of 1-shot accuracies, Recall@1, and mutual information. I ( X ˜ ; Y ) was computed using balanced batches of 16 images each from 5 different classes. In addition to the experiment in the paper, we measured the correlations between 1-shot accuracies, Recall @ 1 , and NMI using three previously proposed losses (triplet, npair, protonet). Figure A1 shows that even for other methods for which mutual information is not the objective, mutual information strongly correlates with all other previous metrics.
Figure A1. Correlation between few-shot accuracy and retrieval metrics.
Figure A1. Correlation between few-shot accuracy and retrieval metrics.
Entropy 24 00501 g0a1

Appendix C.3. Code Visualizations

We provide additional visualizations of codes in Figure A2. These examples consistently show that each code encodes a semantic concept, and that such concepts can be but are not necessarily tied to a particular class.
Figure A2. Visualization of codes of a DIMCO model ( k = 16 , d = 4 ) trained on CIFAR100. Each of the subfigures (ad) correspond to one of the d = 4 dimensions. For each of the k d = 64 codewords, we show the top 10 images from the test set that assign it the highest marginal probability in each row.
Figure A2. Visualization of codes of a DIMCO model ( k = 16 , d = 4 ) trained on CIFAR100. Each of the subfigures (ad) correspond to one of the d = 4 dimensions. For each of the k d = 64 codewords, we show the top 10 images from the test set that assign it the highest marginal probability in each row.
Entropy 24 00501 g0a2
Table A1. Top-1 accuracies under various compressed embedding sizes ( k , d ), with best results for each setting in bold. The original embeddings were continuous 64-dimensional embeddings, which corresponds to k = 2 32 , d = 64 .
Table A1. Top-1 accuracies under various compressed embedding sizes ( k , d ), with best results for each setting in bold. The original embeddings were continuous 64-dimensional embeddings, which corresponds to k = 2 32 , d = 64 .
CIFAR-10CIFAR-100
d = 2d = 4d = 8d = 16d = 2d = 4d = 8d = 16
Product Quantization k = 2 21.8850.3878.6887.242.847.7715.2432.47
k = 4 60.7086.8689.5390.927.6918.7835.6252.02
k = 8 90.4690.4490.9891.4915.2831.0450.7458.82
k = 16 91.2991.0291.2791.5225.7143.1655.8861.00
DIMCO (ours) k = 2 36.3764.4383.1088.853.6610.0319.3133.46
k = 4 62.9288.4790.7891.0411.8325.737.453.36
k = 8 90.7791.2991.4191.4531.946.2252.7758.84
k = 16 91.4991.6891.4691.5746.1757.8361.1162.49

Appendix C.4. Label-Aware Bit Compression

We computed top-1 accuracies using a kNN classifier on each type of embedding with k = 200 . We present extended results comparing PQ and DIMCO on ImageNet embeddings in Table A2. For CIFAR-10 and CIFAR-100 pretrained ResNet20, we used pretrained weights of open-sourced repository (https://github.com/chenyaofo/CIFAR-pretrained-models, accessed on 15 March 2021), and for ImageNet pretrained ResNet50, we used torchvision. We optimized the probablistic encoder with Adam optimizer [56] with learning rate of 1 × 10 2 for CIFAR-100 and ImageNet, and 3 × 10 3 for CIFAR-10.
Table A2. Top-1 accuracies of compressed embeddings under various sizes ( k , d ) on the ImageNet dataset. The original embeddings were continuous 2048-dimensional embeddings, which corresponds to k = 2 32 , d = 2048 .
Table A2. Top-1 accuracies of compressed embeddings under various sizes ( k , d ) on the ImageNet dataset. The original embeddings were continuous 2048-dimensional embeddings, which corresponds to k = 2 32 , d = 2048 .
Product QuantizationDIMCO
d = 2d = 4d = 8d = 16d = 2d = 4d = 8d = 16
k = 2 0.210.311.005.210.140.421.475.50
k = 4 0.200.924.8621.450.672.6811.8026.96
k = 8 0.822.5711.9135.431.7112.8233.2848.11
k = 16 1.995.6620.3444.934.9026.2244.0155.34

Appendix C.5. Small Train Set

We performed an experiment to see how well DIMCO can generalize to new datasets after training with a small number of datasets. We trained each model using { 1 , 4 , 16 , 64 } samples from each class in the miniImageNet dataset. For example, 4 samples means that we trained on (64 classes × 4 images) instead of the full (64 classes × 600 images). We compare our method against three methods which use continuous embeddings for each datapoint: Triplet Nets [1], multiclass N-pair loss [11], and ProtoNets [3].
Figure A3 shows that DIMCO learns much more effectively compared to previous methods when the number of examples per class is low. We attribute this to DIMCO’s tight generalization gap. Since DIMCO uses fewer bits to describe each datapoint, the codes act as an implicit regularizer that helps generalization to unseen datasets. We additionally note that DIMCO is the only method in Figure A3 that can train using a dataset consisting of 1 example per class. DIMCO has this capability because, unlike other methods, DIMCO requires no support/query (also called train/test) split and maximizes the mutual information within a given batch. In contrast, other methods require at least one support and one query example per class within each batch.
For this experiment, we used the Adam optimizer and performed a log-uniform hyperparameter sweep for learning rate ∈ [1 × 10 7 , 1 × 10 3 ] For DIMCO, we swept p [ 32 , 128 ] and d [ 16 , 32 ] . For other methods, we made the embedding dimension [ 16 , 32 ] . For each combination of loss and number of training examples per class, we ran the experiment 64 times and reported the mean and standard deviation of the top 5.
Figure A3. Performance of methods trained using subsets of miniImageNet of varying size. The lowermost y axis value for each metric corresponds to the expected performance of random guessing. We show the mean and standard deviation of top 5 runs from a hyperparameter sweep of 64 runs per configuration.
Figure A3. Performance of methods trained using subsets of miniImageNet of varying size. The lowermost y axis value for each metric corresponds to the expected performance of random guessing. We show the mean and standard deviation of top 5 runs from a hyperparameter sweep of 64 runs per configuration.
Entropy 24 00501 g0a3

Appendix C.6. Few-Shot Classification

For this experiment, we built on the code released by Lee et al. [53] (https://github.com/kjunelee/MetaOptNet, accessed on 15 March 2021) with minimal adjustments. We used the repository’s default datasets, augmentation, optimizer, and backbones. The only difference was our added module for outputting discrete codes. We show an extended table with citations in Table A3.
Table A3. Few-shot classification accuracies on the miniImageNet benchmark, with best results for each setting in bold. Grouped according to backbone architecture. denotes transductive methods, which are more expressive by taking unlabeled examples into account.
Table A3. Few-shot classification accuracies on the miniImageNet benchmark, with best results for each setting in bold. Grouped according to backbone architecture. denotes transductive methods, which are more expressive by taking unlabeled examples into account.
Method5-Way 1-Shot5-Way 5-Shot
   ConvNet (64-64-64-64)
TPN [48]55.51 ± 0.8669.86 ± 0.65
FEAT [49]55.75 ± 0.2072.17 ± 0.16
MetaLSTM [50]43.44 ± 0.7760.60 ± 0.71
MatchingNet [31]43.56 ± 0.8455.31 ± 0.73
ProtoNet [3]49.42 ± 0.7868.20 ± 0.66
RelationNet [51]50.44 ± 0.8265.32 ± 0.70
R2D2 [52]51.2 ± 0.668.8 ± 0.1
MetaOptNet-SVM [53]52.87 ± 0.5768.76 ± 0.48
DIMCO ( 64 , 64 )47.33 ± 0.4661.59 ± 0.52
DIMCO ( 64 , 128 )53.29 ± 0.4764.79 ± 0.57
   ResNet-12
TPN [48]59.4675.65
FEAT [49]62.60 ± 0.2078.06 ± 0.15
SNAIL [47]55.71 ± 0.9968.88 ± 0.92
AdaResNet [57]56.88 ± 0.6271.94 ± 0.57
TADAM [4]58.50 ± 0.3076.70 ± 0.30
MetaOptNet-SVM [53]62.64 ± 0.6178.63 ± 0.46
DIMCO ( 128 , 64 )54.57 ± 0.4765.45 ± 0.31
DIMCO ( 128 , 128 )57.24 ± 0.4469.31 ± 0.38

References

  1. Hoffer, E.; Ailon, N. Deep metric learning using triplet network. In International Workshop on Similarity-Based Pattern Recognition; Springer: Berlin/Heidelberg, Germany, 2015; pp. 84–92. [Google Scholar]
  2. Movshovitz-Attias, Y.; Toshev, A.; Leung, T.K.; Ioffe, S.; Singh, S. No fuss distance metric learning using proxies. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 360–368. [Google Scholar]
  3. Snell, J.; Swersky, K.; Zemel, R. Prototypical networks for few-shot learning. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 4077–4087. [Google Scholar]
  4. Oreshkin, B.; López, P.R.; Lacoste, A. Tadam: Task dependent adaptive metric for improved few-shot learning. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 3–8 December 2018; pp. 721–731. [Google Scholar]
  5. Tishby, N.; Pereira, F.C.; Bialek, W. The information bottleneck method. arXiv 2000, arXiv:physics/0004057. [Google Scholar]
  6. Shwartz-Ziv, R.; Tishby, N. Opening the black box of deep neural networks via information. arXiv 2017, arXiv:1703.00810. [Google Scholar]
  7. Belghazi, M.I.; Baratin, A.; Rajeswar, S.; Ozair, S.; Bengio, Y.; Courville, A.; Hjelm, R.D. MINE: Mutual information neural estimation. In Proceedings of the International Conference on Machine Learning, Stockholm, Sweden, 10–15 July 2018. [Google Scholar]
  8. Hamming, R.W. Error detecting and error correcting codes. Bell Syst. Tech. J. 1950, 29, 147–160. [Google Scholar] [CrossRef]
  9. Kuncheva, L.I.; Whitaker, C.J. Measures of diversity in classifier ensembles and their relationship with the ensemble accuracy. Mach. Learn. 2003, 51, 181–207. [Google Scholar] [CrossRef]
  10. Koch, G.; Zemel, R.; Salakhutdinov, R. Siamese neural networks for one-shot image recognition. In Proceedings of the ICML Deep Learning Workshop, Lille, France, 6–11 July 2015; Volume 2. [Google Scholar]
  11. Sohn, K. Improved deep metric learning with multi-class n-pair loss objective. In Proceedings of the Advances in Neural Information Processing Systems, Barcelona, Spain, 5–10 December 2016; pp. 1857–1865. [Google Scholar]
  12. Duan, Y.; Zheng, W.; Lin, X.; Lu, J.; Zhou, J. Deep Adversarial Metric Learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 2780–2789. [Google Scholar]
  13. Amit, R.; Meir, R. Meta-learning by Adjusting Priors Based on Extended PAC-Bayes Theory. In Proceedings of the International Conference on Machine Learning, Sydney, Australia, 6–11 August 2017. [Google Scholar]
  14. Ravi, S.; Beatson, A. Amortized Bayesian Meta-Learning. In Proceedings of the International Conference on Learning Representations, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  15. Shamir, O.; Sabato, S.; Tishby, N. Learning and generalization with the information bottleneck. Theor. Comput. Sci. 2010, 411, 2696–2711. [Google Scholar] [CrossRef] [Green Version]
  16. Tishby, N.; Zaslavsky, N. Deep learning and the information bottleneck principle. In Proceedings of the 2015 IEEE Information Theory Workshop (ITW), Jerusalem, Israel, 26 April–1 May 2015; pp. 1–5. [Google Scholar]
  17. Strouse, D.; Schwab, D.J. The deterministic information bottleneck. Neural Comput. 2017, 29, 1611–1630. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  18. Lee, Y.; Choi, S. Gradient-Based Meta-Learning with Learned Layerwise Metric and Subspace. In Proceedings of the International Conference on Machine Learning, Stockholm, Sweden, 10–15 July 2018. [Google Scholar]
  19. Rusu, A.A.; Rao, D.; Sygnowski, J.; Vinyals, O.; Pascanu, R.; Osindero, S.; Hadsell, R. Meta-learning with latent embedding optimization. In Proceedings of the International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  20. Kim, T.; Yoon, J.; Dia, O.; Kim, S.; Bengio, Y.; Ahn, S. Bayesian model-agnostic meta-learning. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 3–8 December 2018. [Google Scholar]
  21. Zintgraf, L.M.; Shiarlis, K.; Kurin, V.; Hofmann, K.; Whiteson, S. CAML: Fast Context Adaptation via Meta-Learning. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 10–15 June 2019. [Google Scholar]
  22. Shannon, C.E. A mathematical theory of communication. Bell Syst. Tech. J. 1948, 27, 379–423. [Google Scholar] [CrossRef] [Green Version]
  23. Rolfe, J.T. Discrete variational autoencoders. In Proceedings of the International Conference on Learning Representations, Toulon, France, 24–26 April 2017. [Google Scholar]
  24. Van den Oord, A.; Vinyals, O. Neural discrete representation learning. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 6306–6315. [Google Scholar]
  25. Razavi, A.; Oord, A.v.d.; Vinyals, O. Generating Diverse High-Fidelity Images with VQ-VAE-2. In Proceedings of the Advances in Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019. [Google Scholar]
  26. Hu, W.; Miyato, T.; Tokui, S.; Matsumoto, E.; Sugiyama, M. Learning discrete representations via information maximizing self-augmented training. In Proceedings of the 34th International Conference on Machine Learning, Sydney, Australia, 6–11 August 2017; Volume 70, pp. 1558–1567. [Google Scholar]
  27. Gray, R.M.; Neuhoff, D.L. Quantization. IEEE Trans. Inf. Theory 1998, 44, 2325–2383. [Google Scholar] [CrossRef]
  28. Jegou, H.; Douze, M.; Schmid, C. Product quantization for nearest neighbor search. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 33, 117–128. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  29. Gong, Y.; Lazebnik, S.; Gordo, A.; Perronnin, F. Iterative quantization: A procrustean approach to learning binary codes for large-scale image retrieval. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 35, 2916–2929. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  30. Jeong, Y.; Song, H.O. Efficient end-to-end learning for quantizable representations. In Proceedings of the International Conference on Machine Learning, Stockholm, Sweden, 10–15 July 2018. [Google Scholar]
  31. Vinyals, O.; Blundell, C.; Lillicrap, T.; Wierstra, D. Matching networks for one shot learning. In Proceedings of the Advances in Neural Information Processing Systems, Barcelona, Spain, 5–10 December 2016; pp. 3630–3638. [Google Scholar]
  32. Chen, X.; Duan, Y.; Houthooft, R.; Schulman, J.; Sutskever, I.; Abbeel, P. Infogan: Interpretable representation learning by information maximizing generative adversarial nets. In Proceedings of the Advances in Neural Information Processing Systems, Barcelona, Spain, 5–10 December 2016; pp. 2172–2180. [Google Scholar]
  33. Alemi, A.A.; Poole, B.; Fischer, I.; Dillon, J.V.; Saurous, R.A.; Murphy, K. Fixing a Broken ELBO. In Proceedings of the International Conference on Machine Learning, Sydney, Australia, 6–11 August 2017. [Google Scholar]
  34. Chen, T.Q.; Li, X.; Grosse, R.B.; Duvenaud, D.K. Isolating sources of disentanglement in variational autoencoders. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 3–8 December 2018; pp. 2610–2620. [Google Scholar]
  35. Alemi, A.A.; Fischer, I.; Dillon, J.V.; Murphy, K. Deep Variational Information Bottleneck. In Proceedings of the International Conference on Learning Representations, Toulon, France, 24–26 April 2017. [Google Scholar]
  36. Hjelm, R.D.; Fedorov, A.; Lavoie-Marchildon, S.; Grewal, K.; Trischler, A.; Bengio, Y. Learning deep representations by mutual information estimation and maximization. In Proceedings of the International Conference on Learning Representations, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  37. Oord, A.v.d.; Li, Y.; Vinyals, O. Representation learning with contrastive predictive coding. arXiv 2018, arXiv:1807.03748. [Google Scholar]
  38. Grover, A.; Ermon, S. Uncertainty autoencoders: Learning compressed representations via variational information maximization. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 10–15 June 2019. [Google Scholar]
  39. Choi, K.; Tatwawadi, K.; Grover, A.; Weissman, T.; Ermon, S. Neural Joint Source-Channel Coding. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 10–15 June 2019; pp. 1182–1192. [Google Scholar]
  40. Achille, A.; Soatto, S. Information dropout: Learning optimal representations through noisy computation. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 2897–2905. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  41. Kim, H.; Mnih, A. Disentangling by factorising. In Proceedings of the International Conference on Machine Learning, Stockholm, Sweden, 10–15 July 2018; pp. 2649–2658. [Google Scholar]
  42. Krizhevsky, A. Learning Multiple Layers of Features from Tiny Images; Technical Report; University of Tront: Toronto, ON, Canada, 2009. [Google Scholar]
  43. Wah, C.; Branson, S.; Welinder, P.; Perona, P.; Belongie, S. The Caltech-UCSD Birds-200-2011 Dataset; Technical Report; California Institute of Technology: Pasadena, CA, USA, 2011. [Google Scholar]
  44. Krause, J.; Stark, M.; Deng, J.; Fei-Fei, L. 3D Object Representations for Fine-Grained Categorization. In Proceedings of the 4th International IEEE Workshop on 3D Representation and Recognition (3dRR-13), Sydney, Australia, 2–8 December 2013. [Google Scholar]
  45. Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Fei-Fei, L. ImageNet: A Large-Scale Hierarchical Image Database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009. [Google Scholar]
  46. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HA, USA, 21–26 July 2016; pp. 770–778. [Google Scholar]
  47. Mishra, N.; Rohaninejad, M.; Chen, X.; Abbeel, P. A simple neural attentive meta-learner. In Proceedings of the International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  48. Liu, Y.; Lee, J.; Park, M.; Kim, S.; Yang, Y. Transductive propagation network for few-shot learning. In Proceedings of the International Conference on Learning Representations, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  49. Ye, H.J.; Hu, H.; Zhan, D.C.; Sha, F. Learning embedding adaptation for few-shot learning. arXiv 2018, arXiv:1812.03664. [Google Scholar]
  50. Ravi, S.; Larochelle, H. Optimization as a model for few-shot learning. In Proceedings of the International Conference on Learning Representations, Toulon, France, 24–26 April 2017. [Google Scholar]
  51. Sung, F.; Yang, Y.; Zhang, L.; Xiang, T.; Torr, P.H.; Hospedales, T.M. Learning to compare: Relation network for few-shot learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 1199–1208. [Google Scholar]
  52. Bertinetto, L.; Henriques, J.F.; Torr, P.H.; Vedaldi, A. Meta-learning with differentiable closed-form solvers. In Proceedings of the International Conference on Learning Representations, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  53. Lee, K.; Maji, S.; Ravichandran, A.; Soatto, S. Meta-learning with differentiable convex optimization. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 10657–10665. [Google Scholar]
  54. Yi, D.; Lei, Z.; Liao, S.; Li, S.Z. Deep metric learning for person re-identification. In Proceedings of the 2014 22nd International Conference on Pattern Recognition, Stockholm, Sweden, 24–28 August 2014; pp. 34–39. [Google Scholar]
  55. Yin, M.; Tucker, G.; Zhou, M.; Levine, S.; Finn, C. Meta-Learning without Memorization. In Proceedings of the International Conference on Learning Representations, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  56. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. In Proceedings of the International Conference on Learning Representations, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  57. Munkhdalai, T.; Yuan, X.; Mehri, S.; Trischler, A. Rapid adaptation with conditionally shifted neurons. In Proceedings of the International Conference on Machine Learning, Stockholm, Sweden, 10–15 July 2018. [Google Scholar]
Figure 1. A graphical overview of discrete infomax codes (DIMCO). (a) Discrete codes are produced by a probabilistic encoder that maps each datapoint to a distribution over k-way d-dimensional discrete codes. The encoder is trained to maximize the mutual information between code distribution and label distribution. (b) Given a query image, we compare it against a support set of discrete codes and corresponding labels. Our similarity metric is the query’s log probability for each discrete code.
Figure 1. A graphical overview of discrete infomax codes (DIMCO). (a) Discrete codes are produced by a probabilistic encoder that maps each datapoint to a distribution over k-way d-dimensional discrete codes. The encoder is trained to maximize the mutual information between code distribution and label distribution. (b) Given a query image, we compare it against a support set of discrete codes and corresponding labels. Our similarity metric is the query’s log probability for each discrete code.
Entropy 24 00501 g001
Figure 2. Compositionality of codes in a DIMCO model ( k = 16 , d = 4 ) trained on the CIFAR-100 dataset. Each row shows the top 10 images that assign highest marginal probability to specific codes (shown on left). Composed codewords retrieve images with combined semantic meaning, such as man + fish = man holding fish. We additionally visualize all k d codewords in Appendix A.
Figure 2. Compositionality of codes in a DIMCO model ( k = 16 , d = 4 ) trained on the CIFAR-100 dataset. Each row shows the top 10 images that assign highest marginal probability to specific codes (shown on left). Composed codewords retrieve images with combined semantic meaning, such as man + fish = man holding fish. We additionally visualize all k d codewords in Appendix A.
Entropy 24 00501 g002
Table 1. Top-1 accuracies under various compressed CIFAR-10/100 embeddings with size ( k , d ), with best results for each setting in bold. The original embeddings were continuous 64-dimensional embeddings ( k 2 32 , d = 64 ).
Table 1. Top-1 accuracies under various compressed CIFAR-10/100 embeddings with size ( k , d ), with best results for each setting in bold. The original embeddings were continuous 64-dimensional embeddings ( k 2 32 , d = 64 ).
CIFAR-10CIFAR-100
d = 4d = 16d = 4d = 16
PQ k = 2 50.3887.247.7732.47
k = 4 86.8690.9218.7852.02
k = 8 90.4491.4931.0458.82
k = 16 91.0291.5243.1661.00
DIMCO k = 2 64.4388.8510.0333.46
(ours) k = 4 88.4791.0425.753.36
k = 8 91.2991.4546.2258.84
k = 16 91.6891.5757.8362.49
Table 2. Top-1 accuracies and compression rates of ImageNet embeddings under various compressed embedding sizes ( k , d ), with best results in bold. The compression rate is the ratio between uncompressed and compressed sizes; it is calculated as 32 · 2048 d log k .
Table 2. Top-1 accuracies and compression rates of ImageNet embeddings under various compressed embedding sizes ( k , d ), with best results in bold. The compression rate is the ratio between uncompressed and compressed sizes; it is calculated as 32 · 2048 d log k .
Method(k,d)Compression RateAccuracy
SQ( 2 , 2048 )320.10
SQ( 4 , 2048 )160.28
SQ( 8 , 2048 )1012.36
SQ( 16 , 2048 )857.80
PQ( 2 , 2 )32,7680.21
PQ( 4 , 4 )81920.92
PQ( 8 , 8 )273011.91
PQ( 16 , 16 )102444.93
DIMCO( 2 , 2 )32,7680.14
DIMCO( 4 , 4 )81922.68
DIMCO( 8 , 8 )273033.28
DIMCO( 16 , 16 )102463.64
Table 3. Few-shot classification accuracies on the miniImageNet benchmark, with best results for each setting in bold. denotes transductive methods, which are more expressive by taking unlabeled examples into account.
Table 3. Few-shot classification accuracies on the miniImageNet benchmark, with best results for each setting in bold. denotes transductive methods, which are more expressive by taking unlabeled examples into account.
Method5-Way 1-Shot5-Way 5-Shot
TPN55.51 ± 0.8669.86 ± 0.65
FEAT55.75 ± 0.2072.17 ± 0.16
MetaLSTM43.44 ± 0.7760.60 ± 0.71
MatchingNet43.56 ± 0.8455.31 ± 0.73
ProtoNet49.42 ± 0.7868.20 ± 0.66
RelationNet50.44 ± 0.8265.32 ± 0.70
R2D251.2 ± 0.668.8 ± 0.1
MetaOptNet-SVM52.87 ± 0.5768.76 ± 0.48
DIMCO ( 64 , 64 )47.33 ± 0.4661.59 ± 0.52
DIMCO ( 64 , 128 )53.29 ± 0.4764.79 ± 0.57
Table 4. Image retrieval performance on CUB-200-2011 and Cars-196, measured by Recall@1, with best results for each setting in bold. Memory is the number of bits that an embedding vector of each image uses. Time is seconds taken to retrieve a single query from database (5924 and 8131 images for CUB-200-2011 and Cars-196, respectively).
Table 4. Image retrieval performance on CUB-200-2011 and Cars-196, measured by Recall@1, with best results for each setting in bold. Memory is the number of bits that an embedding vector of each image uses. Time is seconds taken to retrieve a single query from database (5924 and 8131 images for CUB-200-2011 and Cars-196, respectively).
CUB-200-2011Cars-196
Method(k,d)Memory [Bits]Recall@1Time [s]Recall@1Time [s]
Binomial Deviance(-, 128)409657.2516.3772.5321.86
Triplet(-, 128)409656.8016.3773.7921.86
Proxy-NCA(-, 128)409656.1916.3775.9421.86
DIMCO (ours)(32, 32)16051.041.4863.442.64
(64, 64)38455.782.8572.066.01
(128, 128)89658.055.8176.0411.92
(256, 256)204858.9012.2077.3216.04
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lee, Y.; Kim, W.; Park, W.; Choi, S. Discrete Infomax Codes for Supervised Representation Learning. Entropy 2022, 24, 501. https://doi.org/10.3390/e24040501

AMA Style

Lee Y, Kim W, Park W, Choi S. Discrete Infomax Codes for Supervised Representation Learning. Entropy. 2022; 24(4):501. https://doi.org/10.3390/e24040501

Chicago/Turabian Style

Lee, Yoonho, Wonjae Kim, Wonpyo Park, and Seungjin Choi. 2022. "Discrete Infomax Codes for Supervised Representation Learning" Entropy 24, no. 4: 501. https://doi.org/10.3390/e24040501

APA Style

Lee, Y., Kim, W., Park, W., & Choi, S. (2022). Discrete Infomax Codes for Supervised Representation Learning. Entropy, 24(4), 501. https://doi.org/10.3390/e24040501

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop