AET vs. AED: Unsupervised Representation Learning by Auto-Encoding Transformations rather than Data
Abstract
The success of deep neural networks often relies on a large amount of labeled examples, which can be difficult to obtain in many real scenarios. To address this challenge, unsupervised methods are strongly preferred for training neural networks without using any labeled data. In this paper, we present a novel paradigm of unsupervised representation learning by Auto-Encoding Transformation (AET) in contrast to the conventional Auto-Encoding Data (AED) approach. Given a randomly sampled transformation, AET seeks to predict it merely from the encoded features as accurately as possible at the output end. The idea is the following: as long as the unsupervised features successfully encode the essential information about the visual structures of original and transformed images, the transformation can be well predicted. We will show that this AET paradigm allows us to instantiate a large variety of transformations, from parameterized, to non-parameterized and GAN-induced ones. Our experiments show that AET greatly improves over existing unsupervised approaches, setting new state-of-the-art performances being greatly closer to the upper bounds by their fully supervised counterparts on CIFAR-10, ImageNet and Places datasets.
1 Introduction


The great success in applying deep neural networks to image classification, object detection and semantic segmentation has inspired us to explore their full ability in a wide variety of computer vision tasks. Unfortunately, training deep neural networks often requires a large amount of labeled data to learn adequate feature representations for visual understanding tasks. This greatly limits the applicability of deep neural networks when only a limited amount of labeled data is available for training the networks. Therefore, there has been an increasing interest in literature to learn deep feature representations in an unsupervised fashion to solve emerging visual understanding tasks with insufficient labeled data.
Among the efforts on unsupervised learning methods, the most representative ones are Auto-Encoders and Generative Adversarial Nets (GANs) [11]. The former trains an encoder network to output feature representations with sufficient information to reconstruct input images by a paired decoder. Many variants of auto-encoders [12, 15] have been proposed in literature but all of them stick to essentially the same idea of reconstructing input data at the output end, and thus we classify them into the Auto-Encoding Data (AED) paradigm illustrated in Figure 1(a).
On the other hand, GANs learn the feature representation in an unsupervised fashion by generating images from input noises with a pair of adversarially trained generator and discriminator. The input noises into the generator can be viewed as the feature representations of its output, since they contain necessary information to produce the corresponding images through the generator. To obtain the “noise” feature representation for each image, an encoder can be trained to form an auto-encoder architecture with the generator as the decoder. In this way, given an input image, the encoder can directly output its noise representation producing the original image through the generator [6, 8]. This combines the strength of both AED and GAN models. Recently, these models become a popular alternative to auto-encoders in many unsupervised and semi-supervised tasks, as they can generate the distribution of photo-realistic images as a whole so that better feature representations can be derived from the trained generator.
Besides auto-encoders and GANs, various paradigms of self-supervised learning methods exist without using manually labeled data. These methods create self-supervised objectives to train the networks. For example, Doersch et al. [5] propose to train neural networks by predicting the relative positions of two randomly sampled patches. Mehdi and Favaro [19] report to train a convolutional neural network by solving Jigsaw puzzles. Image colorization has also been used as a self-supervised task to train convolutional networks in literature [31, 17]. Instead, Dosovitskiy et al. [7] train neural networks by discriminating among a set of surrogate classes artificially formed by applying various transformations to image patches, while Gidaris et al. [10] attempt to classify image rotations of four discrete angles. These approaches explore supervisory signals at various levels of visual structures to train networks without manually labeling data. Unsupervised features have also been extracted from videos by estimating the self-motion of moving objects between consecutive frames [1].
In contrast, we are motivated to learn unsupervised feature representations by Auto-Encoding Transformations (AET) rather than the data themselves. Specifically, by sampling some operators to transform images, we seek to train auto-encoders that can directly reconstruct these operators from the learned feature representations between original and transformed images. We believe as long as the trained features are sufficiently informative, we can decode the transformations from the features that well encode visual structures of images. As compared with the conventional paradigm of Auto-Encoding Data (AED) in Figure 1, AET focuses on exploring dynamics of feature representations under different transformations, thereby revealing not only static visual structures but also how they would change by applying different transformations. Moreover, there is no restriction on the form of transformations applicable in the proposed AET framework. This allows us to flexibly explore a large variety of transformations, ranging from simple image warping to any parametric and non-parametric transformations. We will demonstrate the AET representations outperform the other unsupervised models in experiments, greatly pushing the state-of-the-art unsupervised method much closer to the upper bound set by the fully supervised counterparts.
The remainder of the paper is organized as follows. We first review the related work in Section 2, and then formally present the proposed AET model in Section 3. We conduct experiments in Section 4 to compare its performances with the other state-of-the-art unsupervised models. Finally, we summarize conclusions in Section 5.
2 Related Work
Auto-Encoders. The use of auto-encoder architecture in learning representations in an unsupervised fashion has been extensively studied in literature [13, 14, 28]. These existing auto-encoders are all based on reconstructing the input data at the output end through a pair of encoder and decoder. The encoder acts as an extractor of features usually compactly representing the most essential information about input data, while a decoder is jointly trained to recover the input data upon the extracted features. The idea is that a good feature representation should contain sufficient information to reconstruct the input data. A wide spectrum of auto-encoders have been proposed following this paradigm of auto-encoding data (AED). For example, the variational auto-encoder [15] explicitly introduces probabilistic assumption about the distribution of features extracted from data. Denoising auto-encoder [28] aims to learn more robust representation by reconstructing original inputs from noise-corrupted inputs. Contrastive Auto-Encoder [27] penalizes abrupt changes of representations around given data, thus encouraging representation invariance to small perturbation on input data. Zhang et al. [30] present a cross-channel auto-encoder by reconstructing a subset of data channels from another subset with the cross-channel features being concatenated as data representation. Hinton et al. [12] propose a transforming auto-encoder in the context of capsule nets, which is still trained in the AED fashion by minimizing the discrepancy between the reconstructed and target images. Conceptually, this differs from the proposed AET that aims to learn unsupervised features by directly minimizing the input and output transformations in an end-to-end auto-encoder architecture.
Generative Adversarial Nets. Besides the auto-encoders, Generative Adversarial Nets (GANs) become popular for training network representations of data in an unsupervised fashion. Unlike the auto-encoders, GANs attempt to directly generate data from noises drawn from a random distribution. By viewing the sampled noises as the coordinates over the manifold of real data, one can use them as the features to represent data. For this purpose, one usually needs to train a data encoder to find the noise that can generate the input images through the GAN generator. This can be implemented by jointly training a pair of mutually inverse generator and encoder [6, 8]. A prominent characteristic of GANs that make them different from auto-encoders is they do not rely on one-to-one reconstruction of input data at the output end. Instead, they focus on discovering and generating the entire distribution of data over the underlying manifold. Recent progress has shown the promising generalization ability of regularized GANs in generating unseen data based on the Lipschitz assumption on the real data distribution [24, 2], and this shows great potential of GANs in providing expressive representation of images [6, 8, 9].
Self-Supervised Representation Learning. In addition to auto-encoders and GANs, other unsupervised learning methods explore various self-supervised signals to train deep neural networks. These self-supervised signals can be directly derived from data themselves without having to be manually labeled. For example, Doersch et al. [5] use the relative positions of two randomly sampled patches from an image as self-supervised information to train the model. Mehdi and Favaro [19] propose to train a convolutional neural network by solving Jigsaw puzzles. Noroozi et al. [20] learn counting features that satisfy equivalence relations between downsampled and tiled images, and Gidaris et al. [10] train neural networks by classifying image rotations in a discrete set. Dosovitskiy et al. [7] train CNNs by classifying a set of surrogate classes, each of which is formed by applying various transformations to an individual image. However, the resultant features could over-discriminate visually similar images as they always belong to different surrogate classes, and the training cost is much more expensive as every training example results in an individual surrogate class. Luo et al. [18] introduce the concept of image-transform bootstrapping using transforms in the image space to augment training, testing, and both. The idea has also been employed to train feature representations for videos through the self-motion of moving objects [1]. In summary, this type of approaches train networks using various self-supervised objectives instead of manually labeled data.
3 AET: The Proposed Approach
We elaborate on the proposed paradigm of auto-encoding transformations (AET) in this section. First, we will formally present the formulation of AET in Section 3.1. Then we will instantiate AET with different genres of transformations in Section 3.2.
3.1 The Formulation
Suppose that we sample a transformation from a distribution (e.g., image warping, projective transformation and even GAN-induced transformation, c.f. Section 3.2 for more details). It is applied to an image drawn from a data distribution , resulting in the transformed version of .
Our goal is to learn an encoder , which aims to extract the representation for a sample . Meanwhile, we wish to learn a decoder , which gives an estimate of input transformation by decoding from the encoded representations of original and transformed images. Since the prediction on the input transformation is made through the encoded features rather than the original and transformed images, it forces the model to extract expressive features as a proxy to represent images.
The learning problem of Auto-Encoding Transformations (AET) now boils down to jointly training the feature encoder and the transformation decoder . To this end, let us choose a loss function that quantifies the difference between a transformation and its estimate . Then the AET can be solved by minimizing this loss as
where the transformation estimate is a function of the encoder and the decoder such that
and the expectation is taken over the sampled transformations and data. Like in training other deep neural networks, the network parameters of and are jointly updated over mini-batches by back-propagating the gradient of the loss .
3.2 The AET Family
A large variety of transformations can be easily incorporated into the AET formulation. Here we discuss three genres, parameterized, GAN-induced and non-parameterized transformations, to instantiate the AET models.
Parameterized Transformations. Suppose that we have a family of transformations with their parameters sampled from a distribution . This equivalently defines a distribution of parameterized transformations, where each transformation can be represented by its parameter and the loss between transformations can be captured by the difference in terms of their parameters. For example, many transformations such as affine and projective transformations can be represented by a parameterized matrix between homogeneous coordinates of images before and after transformations. Such a matrix captures the change of geometric structures caused by a given transformation, and thus it is straightforward to define to model the difference between the target and the estimated transformations. In the experiments, we will compare different instances of parameterized transformations in this category and demonstrate they can yield competitive performances on training AET.
GAN-Induced Transformations. One can choose other forms of transformations without explicit geometric implications like the affine and the projective transformations. Let us consider a GAN generator that transforms an input over the manifold of real images. For example, in [25], a local generator is learned with a sampled random noise that parameterizes the underlying transformation around a given image . This effectively defines a GAN-induced transformation such that with the transformation parameter . One can directly choose the loss between noise parameters, and train a network to decode the parameter from the features and by the encoder network . Compared with the classical transformations that change low-level appearance and geometric structures in images, the GAN-induced transformations can change high-level semantics in images. For example, the GANs have demonstrated their ability of manipulating attributes such as ages, hairs, genders and wearing glasses in facial images as well as changing the furniture layout in bedroom images [26]. This enables AET to explore a richer family of transformations to learn more expressive representations.
Non-Parametric Transformations. Even if a transformation is hard to parameterize, we can still define the loss by measuring the average difference between the transformations of randomly sampled images. Formally,
(1) |
where is a distance between two transformed images, and the expectation is taken over random samples. For an input non-parametric transformation , we also need a decoder network that outputs a transformation to estimate the input transformation. This can be done by choosing a parameterized transformation as to estimate . Although the non-parametric may not fall in the space of parameterized transformations, such an approximation should be enough for unsupervised learning since our ultimate goal is not to obtain an accurate estimate of input transformation; instead, we aim at learning a good feature representation to give us the best estimate that can be achieved in the parameterized transformation space.
Note that parameterized transformations can also be plugged into Eq. (1) to train the corresponding AET by minimizing this loss function. However, in experiments, we find the performance is not as good as the AET trained with the parameter-based loss. This is probably caused by the fact that the loss (1) cannot accurately reflect the actual difference between transformations unless a sufficiently large number of images are sampled. Thus, we suggest using the parameter-based loss for the AET with parameterized transformations.
We have shown that a wide spectrum of transformations can be adopted in training AET. In this paper, we will focus on the parameterized transformations as they do not involve training extra models like GAN-induced transformations, or require choosing auxiliary transformations to approximate non-parametric forms. This allows us to make a straightforward and fair comparison with the unsupervised methods in literature as shown in the experiments. Moreover, the GAN-induced transformations greatly rely on the quality of transformed images, but existing GAN models are still unable to generate high-quality images with fine-grained details at a high resolution. Thus, we leave it in future to study the GAN-induced and non-parametric transformations for training the AET representations.
4 Experiments
In this section, we evaluate the proposed AET model on the CIFAR-10, ImageNet and Places datasets by comparing it against different unsupervised methods. Unsupervised learning is usually evaluated indirectly based on the classification performance by using the learned representations. For the sake of fair comparison, we follow the test protocols widely adopted in literature.
4.1 CIFAR-10 Experiments
First, we evaluate the AET model on the CIFAR-10 dataset. We consider two different transformations – affine and projective transformations – to train AET, and name the resultant models AET-affine and AET-project for brevity, respectively.

4.1.1 Architecture and Implementation Details
To make a fair and direct comparison with existing unsupervised models, we adopt the Network-In-Network (NIN) architecture that has shown competitive performance previously on the CIFAR-10 dataset for the unsupervised learning task [10]. As illustrated in the top of Figure 2, the NIN consists of four convolutional blocks, each of which contains three convolutional layers. AET has two NIN branches, each taking the original and the transformed images as its input, respectively. The output features of the forth block of two branches are concatenated and average-pooled to form a -d feature vector. Then an output layer follows to predict the parameters of input transformation. The two branches share the same network weights, and are used as the encoder network producing the feature representations for input images.
The AET networks are trained by SGD with a batch size of original images and their transformed counterparts. Momentum and weight decay are set to and . The learning rate is initialized to and scheduled to drop by a factor of after , , , and epochs. The model is trained for epochs in total. For AET-affine, the affine transformation is a composition of a random rotation with , a random translation by of image height and width in both vertical and horizontal directions, and a random scaling factor of , along with a random shearing of degree. For the AET-projective, the projective transformation is formed by randomly translating four corners of an image in both horizontal and vertical directions by of its height and width, after it is randomly scaled by and rotated by or . We compare the results for both models below, and demonstrate both outperform the other existing models and AET-project performs better than AET-affine.
4.1.2 Evaluation Protocol
To evaluate the quality of the representation by an unsupervised model, a classifier is usually trained upon the learned features. Specifically, in our experiments on CIFAR-10, we follow the existing evaluation protocols [22, 7, 26, 21, 10] by building a classifier on top of the second convolutional block. See the bottom of Figure 2, where the first two blocks are frozen while the classifier on top of them is trained with labeled examples.
We evaluate the classification results by using the AET features with both model-based and model-free classifiers. For the model-based classifier, we follow [10] by training a non-linear classifier with three Fully-Connected (FC) layers – each of the two hidden layers has neurons with batch-normalization and ReLU activations, and the output layer is a soft-max layer with ten neurons each for an image class. Alternatively, we also test a convolutional classifier upon the unsupervised features by adding a third NIN block whose output feature map is averaged pooled and connected to a linear soft-max classifier.
Moreover, we also test the model-free KNN classifier based on the averaged-pooled output features from the second convolutional block. The KNN classifier has an advantage without need to train a model with labeled examples. This makes a more direct evaluation on the quality of unsupervised feature representation at the evaluation stage.
4.1.3 Results
Method | Error rate |
---|---|
Supervised NIN (Lower Bound) | 7.20 |
Random Init. + conv (Upper Bound) | 27.50 |
Roto-Scat + SVM [22] | 17.7 |
ExamplarCNN [7] | 15.7 |
DCGAN [26] | 17.2 |
Scattering [21] | 15.3 |
RotNet + FC [10] | 10.94 |
RotNet + conv [10] | 8.84 |
(Ours) AET-affine + FC | 9.77 |
(Ours) AET-affine + conv | 8.05 |
(Ours) AET-project + FC | 9.41 |
(Ours) AET-project + conv | 7.82 |
KNN | 1-FC | 2-FC | 3-FC | conv | |
RotNet baseline [10] | 24.97 | 18.21 | 11.34 | 10.94 | 8.84 |
AET-affine | 23.07 (7.6%) | 17.16 ( 5.8%) | 9.77 ( 13.8%) | 10.16 ( 7.1%) | 8.05( 8.9%) |
AET-project | 22.39 ( 10.3%) | 16.65 ( 8.6%) | 9.41 ( 17.0%) | 9.92 ( 9.3%) | 7.82( 11.5%) |

In Table 1, we compare the AET models with both fully supervised and unsupervised methods on CIFAR-10. First, we note that the unsupervised AET-project with the convolutional classifier almost achieves the same error rate as its fully supervised NIN counterpart with four convolutional blocks ( vs. ). This is a remarkable result demonstrating AET is capable of training unsupervised features with a much narrower gap of performance to its supervised counterpart on CIFAR-10.
Moreover, the AETs outperform the other unsupervised methods in Table 1. For example, ExamplarCNN also applies various transformations to images, including rotations, translations, scaling and even more such as manipulating contrasts and colors. Then it trains unsupervised CNNs by classifying the resultant surrogate classes each containing all transformed versions of an individual images. Compared with ExamplarCNN [7], AET still has a significant lead in error rate, implying it can explore the image transformations more effectively in training unsupervised networks.
It is worth pointing out on CIFAR-10, the other reported methods [22, 7, 26, 21, 10] are usually based on different unsupervised networks and supervised classifiers for evaluation, making it difficult to make a direct comparison between them. The results still suggest that the state-of-the-art performances can be reached by AETs, as their error rates are very close to the pre-assumptive lower bound set by the fully supervised counterpart.
Indeed, one can choose the RotNet in Table 1 as the baseline for comparison as it is trained with the same network and classifier as the AETs. Thus we can make a fair comparison directly. From the results, AETs successfully beat the RotNet with both fully connected (FC) and convolutional classifiers on top of the learned representations. We also compare AETs with this baseline when they are trained with the KNN classifier and varying FC layers in Table 2. The results show that AET-project can consistently achieve the smallest errors no matter which classifiers are used. In Figure 3, we also compare the KNN results with varying number of nearest neighbors. Again, AET-project performs the best without involving any labeled examples. The model-free KNN results suggest the AET model has an advantage when no labels are available in training classifiers upon the unsupervised features.
For the following ImageNet experiments, many existing methods have been compared in literature with the same unsupervised AlexNet architecture as well as the classifiers upon it for the evaluation. We will make a fair comparison directly, and the results show that AET still greatly outperforms the other unsupervised methods.
4.2 ImageNet Experiments
Method | Conv4 | Conv5 |
---|---|---|
ImageNet Labels [3](Upper Bound) | 59.7 | 59.7 |
Random [20] (Lower Bound) | 27.1 | 12.0 |
Tracking [29] | 38.8 | 29.8 |
Context [5] | 45.6 | 30.4 |
Colorization [31] | 40.7 | 35.2 |
Jigsaw Puzzles [19] | 45.3 | 34.6 |
BiGAN [6] | 41.9 | 32.2 |
NAT [3] | - | 36.0 |
DeepCluster [4] | - | 44.0 |
RotNet [10] | 50.0 | 43.8 |
(Ours) AET-project | 53.2 | 47.0 |
Method | Conv1 | Conv2 | Conv3 | Conv4 | Conv5 |
ImageNet Labels (Upper Bound) [10] | 19.3 | 36.3 | 44.2 | 48.3 | 50.5 |
Random (Lower Bound)[10] | 11.6 | 17.1 | 16.9 | 16.3 | 14.1 |
Random rescaled [16](Lower Bound) | 17.5 | 23.0 | 24.5 | 23.2 | 20.6 |
Context [5] | 16.2 | 23.3 | 30.2 | 31.7 | 29.6 |
Context Encoders [23] | 14.1 | 20.7 | 21.0 | 19.8 | 15.5 |
Colorization[31] | 12.5 | 24.5 | 30.4 | 31.5 | 30.3 |
Jigsaw Puzzles [19] | 18.2 | 28.8 | 34.0 | 33.9 | 27.1 |
BiGAN [6] | 17.7 | 24.5 | 31.0 | 29.9 | 28.0 |
Split-Brain [30] | 17.7 | 29.3 | 35.4 | 35.2 | 32.8 |
Counting [20] | 18.0 | 30.6 | 34.3 | 32.5 | 25.7 |
RotNet [10] | 18.8 | 31.7 | 38.7 | 38.2 | 36.5 |
(Ours) AET-project | 19.2 | 32.8 | 40.6 | 39.7 | 37.7 |
DeepCluster* [4] | 13.4 | 32.3 | 41.0 | 39.6 | 38.2 |
(Ours) AET-project* | 19.3 | 35.4 | 44.0 | 43.6 | 42.4 |
We further evaluate the performance by AET on the ImageNet dataset. The AlexNet is used as the backbone to learn the unsupervised features. As shown by the results on CIFAR-10, the projective transformation has better performance on training the AET model, and thus we report the AET-project results here.
Architectures and Training Details. Two AlexNet branches with shared parameters are created with original and transformed images as inputs respectively to train unsupervised AET-project. The -d output features from the second last fully connected layer in two branches are concatenated and fed into the output layer producing eight projective transformation parameters. We still use SGD to train the network, with a batch size of images and their corresponding transformed version, a momentum of , a weight decay of . The initial learning rate is set to , and it is dropped by a factor of at epoch 100 and 150. AET is trained for epochs in total. Finally, the projective transformations applied are randomly sampled in the same fashion as on CIFAR-10.
Results. First we report the Top-1 accuracies of compared methods in Table 3 on ImageNet by following the evaluation protocol in [19]. Two settings are adopted for evaluation – Conv4 and Conv5 denote to train the remaining part of AlexNet on top of Conv4 and Conv5 with the labeled data, while all the bottom convolutional layers up to Conv4 and Conv5 are frozen after they are trained in an unsupervised fashion. For example, in the Conv4 setting, Conv5 and three fully connected layers are trained on the labeled examples, including the last -way output layer. From the results, in both settings, the AET model successfully beats the other compared unsupervised models. In particular, among the compared models is the BiGAN [6] that trains a GAN-based unsupervised model, and learns a data-based auto-encoder as well to map an image to an unsupervised representation. Thus, it can be seen as combing the strengths of both GAN and AED models. The results show AET outperforms BiGAN by a significant lead, suggesting its advantage over the GAN and AED paradigms at least in this experiment setting.
We also compare with the fully supervised models that give the upper bounded performance by training the entire AlexNet with all labeled data. The classifiers of random models are trained on top of Conv4 and Conv5 with randomly sampled weights, and they set up the lower bounded performance. From the comparison, the AET models greatly narrow the performance gap to the upper bound – the gap to the upper bound Top-1 accuracy has been decreased from and by RotNet and DeepCluster on Conv4 and Conv5, respectively, to and by AET, which is relatively narrowed by and , respectively.
4.3 Places Experiments
Method | Conv1 | Conv2 | Conv3 | Conv4 | Conv5 |
---|---|---|---|---|---|
Places labels [32] | 22.1 | 35.1 | 40.2 | 43.3 | 44.6 |
ImageNet labels | 22.7 | 34.8 | 38.4 | 39.4 | 38.7 |
Random | 15.7 | 20.3 | 19.8 | 19.1 | 17.5 |
Random rescaled [16] | 21.4 | 26.2 | 27.1 | 26.1 | 24.0 |
Context [5] | 19.7 | 26.7 | 31.9 | 32.7 | 30.9 |
Context Encoders [23] | 18.2 | 23.2 | 23.4 | 21.9 | 18.4 |
Colorization[31] | 16.0 | 25.7 | 29.6 | 30.3 | 29.7 |
Jigsaw Puzzles [19] | 23.0 | 31.9 | 35.0 | 34.2 | 29.3 |
BiGAN [6] | 22.0 | 28.7 | 31.8 | 31.3 | 29.7 |
Split-Brain [30] | 21.3 | 30.7 | 34.0 | 34.1 | 32.5 |
Counting [30] | 23.3 | 33.9 | 36.3 | 34.7 | 29.6 |
RotNet [10] | 21.5 | 31.0 | 35.1 | 34.6 | 33.7 |
(Ours) AET-project | 22.1 | 32.9 | 37.1 | 36.2 | 34.7 |
We also conduct experiments on the Places dataset. As shown in Table 5, we evaluate unsupervised models that are pretrained on the ImageNet dataset. Then a single-layer logistic regression classifier is trained on top of different layers of feature maps with Places labels. Thus, we assess the generalizability of unsupervised features from one dataset to another. Our models are still based on AlexNet variants like those used in the ImageNet experiments. We also compare with the fully supervised models trained with the Places labels and ImageNet labels,as well as the random networks. The results show the AET models outperform the other unsupervised models in most of cases, except on Conv1 and Conv2, Counting [30] performs slightly better.
4.4 Analysis of Predicated Transformations


Although our ultimate goal is to learn good representations of images, it is insightful to look into the accuracy of predicting transformations and its relation with the supervised classification performance. As illustrated in Figure 4, the trend of transformation prediction loss (i.e. the AET loss being minimized to train the model) is well aligned with that of classification error and Top-1 accuracy on CIFAR-10 and ImageNet. This suggests that better prediction of transformations is a good surrogate of better classification result by using the learned features. This justifies our choice of AET to supervise the learning of feature representations.

In Figure 5, we also compare some examples of original images, along with the transformed images at the input and the output ends of the AET model. These examples show how well the model can decode the transformations from the encoded image features, thereby delivering unsupervised representations that offer competitive performances on classifying images in our experiments.
5 Conclusions
In this paper, we present a novel Auto-Encoding Transformation (AET) paradigm for unsupervised training of neural networks in contrast to the conventional Auto-Encoding Data (AED) approach. By estimating randomly sampled transformations at output end, AET forces the encoder to learn good representations so that they contain sufficient information about visual structures of both the original and transformed images. We demonstrate that a wide variety of transformations can be easily incorporated into this framework and the experiment results demonstrate substantial improvements over the state-of-the-art performances, significantly narrowing the gap with the fully supervised counterparts in literature.
6 Acknowledgement
This work was done during Liheng Zhang was interning at Huawei Cloud, Seattle WA, while the idea was conceived and formulated by Guo-Jun Qi.
References
- [1] P. Agrawal, J. Carreira, and J. Malik. Learning to see by moving. In Proceedings of the IEEE International Conference on Computer Vision, pages 37–45, 2015.
- [2] M. Arjovsky, S. Chintala, and L. Bottou. Wasserstein gan. arXiv preprint arXiv:1701.07875, 2017.
- [3] P. Bojanowski and A. Joulin. Unsupervised learning by predicting noise. arXiv preprint arXiv:1704.05310, 2017.
- [4] M. Caron, P. Bojanowski, A. Joulin, and M. Douze. Deep clustering for unsupervised learning of visual features. arXiv preprint arXiv:1807.05520, 2018.
- [5] C. Doersch, A. Gupta, and A. A. Efros. Unsupervised visual representation learning by context prediction. In Proceedings of the IEEE International Conference on Computer Vision, pages 1422–1430, 2015.
- [6] J. Donahue, P. Krähenbühl, and T. Darrell. Adversarial feature learning. arXiv preprint arXiv:1605.09782, 2016.
- [7] A. Dosovitskiy, J. T. Springenberg, M. Riedmiller, and T. Brox. Discriminative unsupervised feature learning with convolutional neural networks. In Advances in Neural Information Processing Systems, pages 766–774, 2014.
- [8] V. Dumoulin, I. Belghazi, B. Poole, O. Mastropietro, A. Lamb, M. Arjovsky, and A. Courville. Adversarially learned inference. arXiv preprint arXiv:1606.00704, 2016.
- [9] M. Edraki and G.-J. Qi. Generalized loss-sensitive adversarial learning with manifold margins. In Proceedings of European Conference on Computer Vision (ECCV 2018), 2018.
- [10] S. Gidaris, P. Singh, and N. Komodakis. Unsupervised representation learning by predicting image rotations. arXiv preprint arXiv:1803.07728, 2018.
- [11] I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville, and Y. Bengio. Generative adversarial nets. In Advances in neural information processing systems, pages 2672–2680, 2014.
- [12] G. E. Hinton, A. Krizhevsky, and S. D. Wang. Transforming auto-encoders. In International Conference on Artificial Neural Networks, pages 44–51. Springer, 2011.
- [13] G. E. Hinton and R. S. Zemel. Autoencoders, minimum description length and helmholtz free energy. In Advances in neural information processing systems, pages 3–10, 1994.
- [14] N. Japkowicz, S. J. Hanson, and M. A. Gluck. Nonlinear autoassociation is not equivalent to pca. Neural computation, 12(3):531–545, 2000.
- [15] D. P. Kingma and M. Welling. Auto-encoding variational bayes. arXiv preprint arXiv:1312.6114, 2013.
- [16] P. Krähenbühl, C. Doersch, J. Donahue, and T. Darrell. Data-dependent initializations of convolutional neural networks. arXiv preprint arXiv:1511.06856, 2015.
- [17] G. Larsson, M. Maire, and G. Shakhnarovich. Learning representations for automatic colorization. In European Conference on Computer Vision, pages 577–593. Springer, 2016.
- [18] J. Luo, M. Boutell, R. T. Gray, and C. Brown. Image transform bootstrapping and its applications to semantic scene classification. IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics), 35(3):563–570, 2005.
- [19] M. Noroozi and P. Favaro. Unsupervised learning of visual representations by solving jigsaw puzzles. In European Conference on Computer Vision, pages 69–84. Springer, 2016.
- [20] M. Noroozi, H. Pirsiavash, and P. Favaro. Representation learning by learning to count. In The IEEE International Conference on Computer Vision (ICCV), 2017.
- [21] E. Oyallon, E. Belilovsky, and S. Zagoruyko. Scaling the scattering transform: Deep hybrid networks. In International Conference on Computer Vision (ICCV), 2017.
- [22] E. Oyallon and S. Mallat. Deep roto-translation scattering for object classification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 2865–2873, 2015.
- [23] D. Pathak, P. Krahenbuhl, J. Donahue, T. Darrell, and A. A. Efros. Context encoders: Feature learning by inpainting. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 2536–2544, 2016.
- [24] G.-J. Qi. Loss-sensitive generative adversarial networks on lipschitz densities. arXiv preprint arXiv:1701.06264, 2017.
- [25] G.-J. Qi, L. Zhang, H. Hu, M. Edraki, J. Wang, and X.-S. Hua. Global versus localized generative adversarial nets. In Proceedings of IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2018.
- [26] A. Radford, L. Metz, and S. Chintala. Unsupervised representation learning with deep convolutional generative adversarial networks. arXiv preprint arXiv:1511.06434, 2015.
- [27] S. Rifai, P. Vincent, X. Muller, X. Glorot, and Y. Bengio. Contractive auto-encoders: Explicit invariance during feature extraction. In Proceedings of the 28th International Conference on International Conference on Machine Learning, pages 833–840. Omnipress, 2011.
- [28] P. Vincent, H. Larochelle, Y. Bengio, and P.-A. Manzagol. Extracting and composing robust features with denoising autoencoders. In Proceedings of the 25th international conference on Machine learning, pages 1096–1103. ACM, 2008.
- [29] X. Wang and A. Gupta. Unsupervised learning of visual representations using videos. In Proceedings of the IEEE International Conference on Computer Vision, pages 2794–2802, 2015.
- [30] R. Zhang, P. Isola, and A. A. Efros. Split-brain autoencoders: Unsupervised learning by cross-channel prediction.
- [31] R. Zhang, P. Isola, and A. A. Efros. Colorful image colorization. In European Conference on Computer Vision, pages 649–666. Springer, 2016.
- [32] B. Zhou, A. Lapedriza, J. Xiao, A. Torralba, and A. Oliva. Learning deep features for scene recognition using places database. In Advances in neural information processing systems, pages 487–495, 2014.