This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Texture-guided Coding for Deep Features

Lei Xiong, Xin Luo, Zihao Wang, Chaofan He, Shuyuan Zhu, , and Bing Zeng Lei Xiong, Xin Luo, Zihao Wang, Shuyuan Zhu, and Bing Zeng are with the School of Information and Communication Engineering, University of Electronic Science and Technology of China, Chengdu 611731, China.Chaofan He is with the Peng Cheng Laboratory, Shenzhen 518066, China.
Abstract

With the rapid development of machine vision technology in recent years, many researchers have begun to focus on feature compression that is better suited for machine vision tasks. The target of feature compression is deep features, which arise from convolution in the middle layer of a pre-trained convolutional neural network. However, due to the large volume of data and high level of abstraction of deep features, their application is primarily limited to machine-centric scenarios, which poses significant constraints in situations requiring human-computer interaction. This paper investigates features and textures and proposes a texture-guided feature compression strategy based on their characteristics. Specifically, the strategy comprises feature layers and texture layers. The feature layers serve the machine, including a feature selection module and a feature reconstruction network. With the assistance of texture images, they selectively compress and transmit channels relevant to visual tasks, reducing feature data while providing high-quality features for the machine. The texture layers primarily serve humans and consist of an image reconstruction network. This image reconstruction network leverages features and texture images to reconstruct preview images for humans. Our method fully exploits the characteristics of texture and features. It eliminates feature redundancy, reconstructs high-quality preview images for humans, and supports decision-making. The experimental results demonstrate excellent performance when employing our proposed method to compress the deep features.

Index Terms:
Deep feature compression, human–machine vision, machine vision

I Introduction

In recent years, with the rapid development of machine vision technology, intelligent applications based on machine vision have been widely deployed in various fields. In these applications, machines must analyze and process massive amounts of image data, significantly challenging storage and transmission. Traditional image compression algorithms are mainly designed for human visual perception, aiming to minimize image distortion and achieve better visual quality. However, the requirements of machine vision for image compression are different from those of human visual perception. Machine vision mainly deals with features extracted from images, and the quality of features directly affects the performance of visual tasks. Existing image compression algorithms cannot satisfy the needs of machine vision. To address this issue, researchers have proposed a series of deep feature compression methods for machine vision in recent years. In this compression, the deep features are firstly extracted from the uncompressed image and output from the shallow network layers. Then, they are compressed and transmitted to the machine for vision tasks. However, how to compress intermediate deep features targeting high representation ability and low bit cost is still a challenging problem.

The deep features are composed of several data channels, and each channel’s data is like a video frame. Thus, the classical video codec, such as high-efficiency video coding (HEVC) [1], can compress deep features. For example, in [2], the HEVC range extensions [3] were used to compress features extracted from YOLO [4], where the data of all the feature channels were combined to form a single image-like data. Different from [2], the data of each feature channel was processed as one video frame in [5] where the HEVC intra coding was employed to compress features extracted from the shallow layer of VGG [6] and ResNet [7]. Although using the HEVC intra-coding reduces the intra-channel redundancy of features, inter-channel redundancy still exists between features. To address this problem, [8] proposed to remove this redundancy by tiling all features into one frame and applying intra-prediction on the composed frame. Based on this work, [9] further divided the feature channels into groups according to the similarity and then tiled features of each group into a single frame for the compression with the HEVC inter-frame coding, removing both intra-channel and inter-channel redundancies.

The video-based solution reduces the redundancy of features; however, it typically assigns the same quantization parameter (QP) to different feature channels. This oversight ignores the importance of the feature and limits the feature’s ability to represent effectively. To address this issue, the sensitivity-aware bit allocation [10] was introduced. This approach defines the sensitivity of each feature channel based on its impact on the fidelity of the computer vision task. By assigning different QPs to feature channels with varying sensitivities, adaptive bit allocation is achieved.

Although the methods above effectively reduce the feature data, they still face some challenges. The feature maps obtained after a series of convolutions on images are highly abstract and can only be used by machines. Humans cannot directly understand their content, which has excellent limitations in scenarios that require human intervention. To address this issue, many researchers[11, 12, 13, 14] have proposed compression methods for humans and machines. However, these methods do not fully utilize human and machine vision characteristics and lack research on the collaborative work of texture and features.

In this work, we propose a texture-guided feature compression method. It can simultaneously provide natural images and deep features for humans and machines, respectively, and fully utilizes texture images to remove feature redundancy. Specifically, we construct a two-layer encoding framework consisting of a feature layer and a texture layer. The feature layer mainly serves machine vision tasks and can also assist in image reconstruction of the texture layer; the texture layer serves humans, providing preview images for humans and reducing the amount of feature data transmitted. During feature encoding, we use a channel selection module to select task-related channels for transmission to reduce the amount of feature data. The missing channels are filled with texture images at the decoding end. In the texture image encoding stage, we downsample it to reduce the amount of data. At the decoding end, we use our proposed image reconstruction neural network to reconstruct the image by combining texture and features.

The main contributions of this paper are as follows:

  1. 1.

    To the best of our knowledge, we are the first to attempt to use texture images to guide feature compression. We propose a texture-guided feature compression method that can fully utilize texture images to remove feature redundancy while improving the compression rate and maintaining the performance of visual tasks.

  2. 2.

    We propose an image compression framework that can serve machines and humans. The framework consists of a feature layer and a texture layer. The feature layer mainly serves machine vision tasks, while the texture layer mainly serves humans.

  3. 3.

    We propose a feature reconstruction neural network that fully utilizes texture and features to achieve high-quality feature reconstruction.

  4. 4.

    We propose an image reconstruction neural network that fuses features and textures. The network can fully utilize texture and features to complete image reconstruction.

The rest of this paper is organized as follows. Section II reviews some relevant works. Section III first introduces our framework for compression, followed by a description of the feature compression neural network serving machine vision tasks and the image reconstruction neural network serving humans. Section IV demonstrates the experimental results, and Section V concludes this paper.

II Related Work

II-A Deep Feature Compression with Codec

In recent years, there has been an increasing amount of literature on deep feature compression. Sufficiently compressing features without significantly degrading the performance of CV tasks is a common goal. In this subsection, we roughly divide the compression of depth features into lossless and lossy compression.

Lossless compression treats depth features as generic data and compresses them using GZIP[16], ZLIB[17], BZIP2[18] or LZMA[19]. Although lossless compression achieves the best performance for the task, it has a limited compression ratio and does not adequately compress the deep feature data.

In lossy compression, deep features are compressed by classical video codec, such as high-efficiency video coding (HEVC)[1]. In [2], the HEVC range extensions[3] were applied to compress the deep features which were extracted from YOLO [4], where all the channels were combined to form a single image [2]. Different from [2], the feature of each channel was processed as the video frame in [5] where the HEVC intra[20] coding was used to compress all the feature channels extracted from the shallow layer of VGG [6], ResNet [7], etc. Deep features have similarities not only within each channel but also between channels. In Paper [8], the authors explored techniques for eliminating redundancy among feature channels. They tested three methods, including two inter-frame[21] approaches and one intra-frame[20] technique that tiled all channels into a single frame. The results showed that the tiling method performs best. Suzuki et al.[9] also investigated methods for removing redundancy among feature channels. They rearrange all channels of a feature into two or three large feature maps and then compress them using HEVC inter-mode.

The video-based solution reduces the redundancy of features. However, it usually assigns the same quantization parameter (QP) to different feature channels, which ignores the importance of the feature and cannot achieve a high representation ability for the feature. To solve this problem, the sensitivity-aware bit allocation was proposed[10], where the sensitivity of each feature channel was defined by its influence on the fidelity to the CV task. Moreover, different QP is assigned to the feature channel with different sensitivity, which achieves the adaptive bit allocation for features. Specifically, the feature channel with higher sensitivity would be allocated more bits to achieve high fidelity. In comparison, the channel with lower sensitivity would be allocated with fewer bits to guarantee that the overall bits do not exceed the given bit cost. In addition, Wang et al.[22] investigated the quantization of deep features. Unlike [2, 23, 5, 9, 24], which use uniform or logarithmic quantizers to reduce the data volume, it assigns different quantization intervals to each feature channel to achieve minimal network output errors.

These codec-based approaches achieve a good balance between task efficiency and computational cost when performing machine tasks. However, these methods only consider machine vision and do not consider human vision; in other words, they do not provide a texture image that humans can view.

II-B Deep Feature Compression with Leaning-based methods

In [25], they propose an end-to-end feature compression system that incorporates a task-specific loss function into the training process, resulting in better analytics performance with low bit rates. In addition to providing intra-frame patterns, the system introduces inter-frame coding schemes.

Since deep features can be larger than the original image size, [26] and [27] use an auto-encoder(AE) to reduce the feature dimension and size. Similarly, [28] introduces a Variational Auto-Encoder (VAE) model to reduce feature size while preserving task performance.

Like the AE,[29] and [30] achieve feature size reduction and recovery using a front-end encoding module and a back-end decoding module, respectively. Choi et al.[13] also use feature reduction to decrease the amount of data, focusing on reducing feature channels rather than directly manipulating the 2-D feature size. For analysis tasks, they can use a subset of the necessary features to complete.

End-to-end compression and feature size reduction methods reduce the required bit rate for feature transmission. However, both methods primarily focus on the analysis task and do not fully consider the reconstruction task. A hierarchical coding approach has been proposed to process both tasks jointly. Wang et al.[12] introduced a coding framework that integrates feature and texture, consisting of a base layer and an enhancement layer. In this framework, the base layer transmits the features server for the analysis task, and the enhancement layer transmits the residuals of the image reconstructed by the base layer from the original image. In [11], they proposed a similar hierarchical framework using an end-to-end model that includes a viewable analysis layer and a high-quality reconstruction layer. The analysis layer is used for the analysis task, while the reconstruction layer transmits the residuals of the features used for image reconstruction.

Hierarchical coding effectively utilizes texture and material information, enabling image reconstruction and analysis tasks to be performed simultaneously. However, as mentioned in previous methods, the residuals transmitted by the enhancement layer depend on the quality of the reconstructed image in the base layer. In features with low texture information, such as convolutional layers near the classifier, the quality of the reconstructed image is often inadequate, and in some cases, reconstruction may be impossible. Therefore, a more general layer coding framework is necessary to address these issues.

III Proposed Method

Refer to caption
Figure 1: Overview framework of proposed method. $\uparrow$⃝ is up-sample, $\downarrow$⃝ is down-sample, Q is quantitation, and DQ is de-quantitation.

III-A Overview

The overall pipeline of our proposed scheme is illustrated in Fig. 1. Our coding scheme is composed of two distinct branches. The first is the texture compression (TC) branch, and the second is the Feature Compression (FC) branch. The TC branch primarily serves humans and is dedicated to providing high-quality images to people. The FC branch serves machines and focuses on enhancing machine performance for task completion.

The TC branch utilizes the Neural Network-based(NN) codec for image compression. Features contain highly abstract semantic information that is challenging for humans to comprehend. In conventional feature compression frameworks, it is challenging to satisfy human and machine perception requirements simultaneously. To address human visual perception, we introduce a TC branch. This branch provides preview images for humans and assists the FC branch in feature reconstruction. The workflow is as follows: at the encoding end, the source image II is downsampled and fed into a NN encoder to generate the texture stream 𝒮𝒯\mathcal{S}_{\mathcal{T}}, where downsampling is performed to reduce the cost of image transmission. At the decoding end, the received 𝒮𝒯\mathcal{S}_{\mathcal{T}} is decoded to obtain the down-sampled image I^\hat{I}_{\downarrow}. Then I^\hat{I}_{\downarrow} and the decoded features of the FC branch are jointly sent to the Image Reconstruction Network to obtain the reconstructed image I~\tilde{I} for human.

The FC branch is primarily used for feature compression and reconstruction. It consists of an encoder and a decoder of the Feature Compression Neural Network (FCNN). At the encoding stage, the source image II is fed into a feature extractor to obtain features 𝒬\mathcal{F}_{\mathcal{H}\mathcal{Q}}, which consist of CC channels. However, the importance of these CC channels varies, and not every channel needs to be compressed and transmitted. To reduce the data volume of the features, we use a channel selection module in the FCNN to determine which channels to keep. This module outputs the channel selection information consisting of CC values, where 1 indicates channel retention, and 0 indicates channel removal. The retained channels are then used to construct a feature subset ^m\hat{\mathcal{F}}{m} with MM(MCM\leq C) channels. Finally, the quantized ^m\hat{\mathcal{F}}_{m} is flattened and passed through the HEVC encoder for compression and transmission. The quantization parameters and channel selection information are also transmitted. At the decoding stage, the flattened features are first decoded from the bitstream 𝒮\mathcal{S}_{\mathcal{F}} and then rearranged to reconstruct the feature subset ^m\hat{\mathcal{F}}{m}. The missing channels are filled with zeroes. However, ^m\hat{\mathcal{F}}{m} contains only partial channel information and cannot be used directly for the task at hand. To reconstruct the full feature, the downsampled image reconstruction II{\downarrow} from the TC branch is upsampled and passed through the feature extractor to obtain the features LQ\mathcal{F}_{LQ}. By incorporating the channel selection information, the missing channel information is extracted from LQ\mathcal{F}_{LQ} and combined with ^m\hat{\mathcal{F}}{m} to reconstruct the feature ^HQ\hat{\mathcal{F}}_{HQ}. The process of feature recombination is explained in the next section. Finally, this feature is fed into the Feature Reconstruction Module(FRM) of FCNN to recover the features for machine vision tasks.

III-B Feature Compression Neural Network(FCNN)

Neural network design often involves gradually reducing feature dimensions while rapidly increasing convolutional kernel numbers to capture high-level semantic information for visual tasks. For instance, in conventional neural networks, the initial convolutional layer transforms a 3-channel RGB input image into 32 or 64 channels, halving the feature size. Consequently, the number of channels in intermediate convolutional feature maps increases more rapidly than the size reduction, leading to a substantial amount of feature data. Some researchers opt to reduce feature data by discarding channels, which may degrade task performance. To maintain task performance, researchers directly utilize the network to repair lost channels, mitigating the impact of dropped channels. However, regenerating channels filled with zeros poses a significant challenge for the network. To address this issue, we employ features extracted from texture images to substitute for the lost channels. Nevertheless, the disparity between features extracted from low-quality texture images and those from uncompressed images is considerable, leading to unsatisfactory performance improvements for visual tasks. Thus, applying an enhancement network to the reorganized features becomes necessary to enhance channel quality.

Based on this observation, we propose the Feature Compression Neural Network, which consists of a Channel Selection Module and a Feature Reconstruction Module. In our approach, we use the channel selection module to transmit the channels that contribute significantly to the task while filtering out the channels with low contribution. This allows us to achieve compression. To ensure task performance, we incorporate both the features of low-quality images and the high-contribution features into the Feature Reconstruction Module to recover high-quality features.

III-B1 Feature extraction

Feature extraction serves as the initial step in feature compression. Typically, providing the input image II to the pre-trained network TaskNetTaskNet yields the inference results, as demonstrated below:

Result=TaskNet(I,θ),Result=TaskNet(I,\theta), (1)

where θ\theta represents the pre-trained network parameters. In the process of feature compression, the task network is divided into two networks, TaskNet1TaskNet_{1} and TaskNet2TaskNet_{2}, with weights represented by θ1\theta_{1} and θ2\theta_{2}, respectively. It is noteworthy that θ1θ2=θ\theta_{1}\cup\theta_{2}=\theta. Consequently, the inference process can be redefined as follows:

=TaskNet1(I,θ1)\mathcal{F}=TaskNet_{1}(I,\theta_{1}) (2)
Result=TaskNet2(,θ2)Result=TaskNet_{2}(\mathcal{F},\theta_{2}) (3)

The inference results of Eq. (1) and Eq. (3) are indeed equivalent. In Eq. (2), the \mathcal{F} represents the extracted features obtained from the image II. In FCNN, we extract features HQ\mathcal{F}_{HQ} and LQ\mathcal{F}_{LQ} from the source image II and the compressed low-resolution image ILRI_{LR}, respectively, as follows:

HQ=TaskNet1(I,θ1)LQ=TaskNet1((ILR),θ1),\begin{split}\mathcal{F}_{HQ}&=TaskNet_{1}(I,\theta_{1})\\ \mathcal{F}_{LQ}&=TaskNet_{1}(\uparrow(I_{LR}),\theta_{1}),\end{split} (4)

where \uparrow denotes upsampling.

Refer to caption
Figure 2: Overall architecture of the Feature Compression Neural Network, including Channel Selection Module and Feature Reconstruction Module.

III-B2 Channel Selection

Channel selection aims to retain task-relevant channels and filter out unimportant ones. The feature maps, denoted as 𝒬\mathcal{F}_{\mathcal{H}\mathcal{Q}} and 𝒬\mathcal{F}_{\mathcal{L}\mathcal{Q}}, have dimensions of C×H×WC\times H\times W, representing CC channels, each with a size of H×WH\times W. Each channel contains different information (e.g., some channels may be all zeros while others have higher values overall). Additionally, different channels contribute differently to visual tasks. Treating all channels as equally important would lead to significant redundancy during compression. We employ a Channel Selection Module (CSM) to compress the features to select task-relevant channels efficiently. Inspired by the work of Xu et al.[31], the structure of the CSM is illustrated in Fig. 2.

The input 𝒬\mathcal{F}_{\mathcal{H}\mathcal{Q}} undergoes average pooling, reducing its size to C×1×1C\times 1\times 1. It then goes through a series of operations, including a 1×11\times 1 convolution, Batch Normalization, ReLU activation, and another 1×11\times 1 convolution, resulting in a channel importance matrix of size 2C×1×12C\times 1\times 1. The importance matrix consists of two vectors, each with a length of CC. The first vector represents the probabilities of selecting the current channel, while the second vector represents the probabilities of not selecting the current channel. Finally, the importance matrix is fed into the Gumbel-Softmax[32], which samples channels based on their importance probabilities. The output is a binary vector of length CC, with 1 indicating the selection of the corresponding channel and 0 indicating non-selection.

For example, if C=4C=4 and the output vector is 10101010, it means the first and third channels are selected. The channel selection process can be expressed as:

Cone=CSM(𝒬),C_{one}=\text{CSM}(\mathcal{F}_{\mathcal{H}\mathcal{Q}}), (5)

where ConeC_{one} is a one-hot vector of size C×1C\times 1, representing the retained channels of 𝒬\mathcal{F}_{\mathcal{H}\mathcal{Q}}. Multiplying ConeC_{one} with 𝒬\mathcal{F}_{\mathcal{H}\mathcal{Q}} channel-wise filters out the unimportant channels, resulting in the feature m\mathcal{F}_{m} that retains task-relevant channels, as shown below:

mi=Conei×𝒬i,i=1,2,,C.\mathcal{F}^{i}_{m}=C^{i}_{one}\times\mathcal{F}_{\mathcal{H}\mathcal{Q}}^{i},i=1,2,...,C. (6)

Note that during the training of the FCNN, we do not compress m\mathcal{F}_{m} to expedite the training process. After training the FCNN, during inference, we proceed to quantize and compress m\mathcal{F}_{m} as illustrated in Fig .2.

Simultaneously, by treating ConeC_{one} as a binary number and performing bitwise NOT operation on it, as shown in Eq. (7), a vector of the same length, CzeroC_{zero}, is obtained. This vector represents the discarded channels in 𝒬\mathcal{F}_{\mathcal{H}\mathcal{Q}}. During the feature reconstruction process, we will utilize CzeroC_{zero} along with the features 𝒬\mathcal{F}_{\mathcal{L}\mathcal{Q}} to reconstruct the channels that were discarded in Eq. (6).

Czeroi=NOT(Conei),i=1,2,,CC^{i}_{zero}=NOT(C^{i}_{one}),i=1,2,...,C (7)

III-B3 Feature Reconstruction

Feature reconstruction aims to restore the channels that were discarded during the feature selection process. After processing through Eq. (6), numerous channels in the feature m\mathcal{F}_{m} become zero, thereby improving the efficiency of feature compression but potentially reducing the performance in visual tasks. To enable TaskNet2TaskNet_{2} to perform tasks efficiently, we utilize the features 𝒬\mathcal{F}_{\mathcal{L}\mathcal{Q}} to fill in the missing channels in m\mathcal{F}_{m}. Specifically, the missing channel information in m\mathcal{F}{m} can be obtained by multiplying CzeroC{zero} with F𝒬{F}_{\mathcal{L}\mathcal{Q}}, as shown below:

~mi=Czeroi×𝒬i,i=1,2,,C.\tilde{\mathcal{F}}_{m}^{i}=C_{zero}^{i}\times\mathcal{F}_{\mathcal{L}\mathcal{Q}}^{i},i=1,2,…,C. (8)

Then, by adding m\mathcal{F}_{m} and ~m\tilde{\mathcal{F}}_{m}, the reconstructed feature ^𝒬\hat{\mathcal{F}}_{\mathcal{H}\mathcal{Q}} can be obtained:

^𝒬=m+~m.\hat{\mathcal{F}}_{\mathcal{H}\mathcal{Q}}=\mathcal{F}_{m}+\tilde{\mathcal{F}}_{m}. (9)

III-B4 Feature Enhancement

Feature enhancement is designed to improve the quality of the reconstructed feature ^𝒬\hat{\mathcal{F}}_{\mathcal{H}\mathcal{Q}} and further enhance task performance. The feature enhancement module consists of a Spatial feature transform (SFT)[33] Layer, a residual block, a 3×33\times 3 convolution, Leaky ReLU, and an Attention Block. The SFT layer is employed to further fuse the features ^𝒬\hat{\mathcal{F}}_{\mathcal{H}\mathcal{Q}} and ^𝒬\hat{\mathcal{F}}_{\mathcal{L}\mathcal{Q}}. Specifically, ^𝒬\hat{\mathcal{F}}_{\mathcal{L}\mathcal{Q}} is fed into the SFT layer, which generates a set of affine parameters (γi,βi)(\gamma_{i},\beta_{i}) through a non-linear mapping. Then, the feature ^𝒬\hat{\mathcal{F}}_{\mathcal{H}\mathcal{Q}} undergoes channel-wise affine transformation to obtain the fused new feature fusion\mathcal{F}_{fusion}, as shown below:

fusion=SFT(^𝒬,^𝒬)=γi×^𝒬i+βi\mathcal{F}_{fusion}=\text{SFT}(\hat{\mathcal{F}}_{\mathcal{H}\mathcal{Q}},\hat{\mathcal{F}}_{\mathcal{L}\mathcal{Q}})=\gamma_{i}\times\hat{\mathcal{F}}_{\mathcal{L}\mathcal{Q}}^{i}+\beta_{i} (10)

Next, the fused feature fusion\mathcal{F}_{fusion} is passed through a residual block and a 3×33\times 3 convolution layer, followed by Leaky ReLU activation, to further enhance feature representation. The convolutional result is then fed into an Attention Block, which weights each channel. The final enhanced feature ^\hat{\mathcal{F}} is obtained after the channel-wise weighting. Finally, ^\hat{\mathcal{F}} is fed into the weight-fixed TaskNet2TaskNet_{2} to obtain the ultimate prediction result.

III-B5 Loss Function

The loss function of FCNN comprises task performance loss and perceptual loss. The first part, task performance loss (LTL_{T}), aims to ensure optimal task performance with minimal rate and is defined as:

LT=T(^)+λmean(Cone),L_{T}=T(\hat{\mathcal{F}})+\lambda\cdot mean(C_{one}), (11)

where T(^)T(\hat{\mathcal{F}}) represents the performance loss of TaskNet2TaskNet_{2} in completing the visual task using features ^\hat{\mathcal{F}}. The mean(Cone)\text{mean}(C_{one}) is a sparse loss term, indicating the number of channels to be transmitted; a higher value implies a need for more channels and a higher required bitrate. ConeC_{one} consists of C binary values, indicating selected channel positions. The calculation of mean(Cone)\text{mean}(C_{one}) is given by:

mean(Cone)=1Ci=1CConeimean(C_{one})=\frac{1}{C}\sum_{i=1}^{C}{C_{one}^{i}} (12)

The parameter λ\lambda is used to balance the trade-off between bitrate and task quality.

The second part is perceptual loss (LdistL_{dist}), which compensates for the performance degradation caused by discarding most channels in the features 𝒬\mathcal{F_{\mathcal{H}{\mathcal{Q}}}}, extracted from the source image II 𝒬\mathcal{F_{\mathcal{H}{\mathcal{Q}}}}. To improve task performance, the perceptual loss is formulated as:

Ldist=MSE(𝒬,^).L_{dist}=\text{MSE}(\mathcal{F_{\mathcal{H}{\mathcal{Q}}}},\hat{\mathcal{F}}). (13)

where ^\hat{\mathcal{F}} is the features obtained by fusing partial channels of 𝒬\mathcal{F_{\mathcal{H}{\mathcal{Q}}}} and ILQI_{LQ} after processing through the feature enhancement module.

Combining the task loss and perceptual loss results in the final loss used for training the network:

L=LT+αLdistL=L_{T}+\alpha\cdot L_{dist} (14)

Where α\alpha is a weight parameter used to balance the task loss and the perceptual loss.

Refer to caption
Figure 3: Detailed architecture of the proposed IRNN model including DCU, DSU, USU and USB.

III-C Image Reconstruction Neural Network(IRNN)

Low-quality images lose semantic information due to downsampling and image compression operations. To enhance the quality of the reconstructed image, we fuse feature maps containing semantic information with the low-quality image. However, the size of the depth features and the low-quality map differ and cannot be fused directly. Moreover, the size of depth features extracted from different locations varies, posing texture and feature fusion challenges.

Through observation, we find that the size difference between different feature layers is always an integer multiple of 2. Thus, it is possible to simultaneously transform the features and texture to the same size by either up-sampling the features or down-sampling the low-quality image. Building on this observation, we propose a generalized feature fusion network named Image Reconstruction Neural Network(IRNN) using the UNet[35] framework. The IRNN maximizes the utilization of information from both the feature and texture domains to accomplish image reconstruction. Guided by texture information, reconstructing the image from the feature domain becomes more efficient. The IRNN takes two inputs, features and texture, and outputs an enhanced image for human viewing. As Fig. 3 shows, the IRNN is constructed by the double conv unit (DCU), the down-sampling unit (DSU), the up-sampling unit (USU), and the up-sampling block (USB), as shown in Fig. 3, where the kernel size of the convolution layer of each unit is 3×33\times 3 and stride = 1.

More specifically, the IRNN consists of two processes: top-down and bottom-up. The top-down process serves feature extraction and fusion and is composed of DCU and DSU modules. DCU is used to extract features from textured images and consists of two 3×33\times 3 convolutions and two Leaky ReLU activations. Additionally, DCU is embedded in DSU, USU, and UPU modules to adjust feature channels. DSU, in the top-down process, adjusts feature channels and sizes and is composed of a max-pooling layer and DCU. The number of DSU modules in IRNN is determined by the input feature dimensions. The bottom-up process is used to reconstruct images from features and comprises USU, USB, and a 3×33\times 3 convolution. USU consists of upsampling and DCU. Upsampling employs bilinear interpolation to reduce network parameters. The features of USU processes are residual-connected to the convolution results of the same layer and serve as input to the next USU. USB consists of a 3×33\times 3 convolution and PixelShuffer for feature upsampling. The convolution result from USB, after adjusting the channels with a final 3×33\times 3 convolution, is added to the upsampled low-resolution image I^\hat{I}_{\downarrow} to output the reconstructed image I~\tilde{I}. The L2 loss is adopted as the loss function in our proposed IRNN:

L=1W×Hx=1Wy=1H(I(x,y)I~(x,y))2,L=\frac{1}{W\times H}\sum_{x=1}^{W}\sum_{y=1}^{H}(I_{(x,y)}-\tilde{I}_{(x,y)})^{2}, (15)

where II represents the target image, I~\tilde{I} denotes the image reconstructed by the network, and HH and WW correspond to the image’s width and height, respectively. The xx and yy represent the positions of pixels.

IV Experiments

IV-A Experimental Detail

In computer vision, VGGNet[6] is widely used for feature extraction in the back-end of various visual tasks. Our work chooses the general VGG16 as the back-end for classification tasks. VGG16 consists of 13 convolutional layers, five max-pooling layers, and three fully connected layers. We disconnect the VGG16 network at the 4-th pooling layer and split it into two subnetworks, one for feature extraction and the other for classification tasks. It is worth noting that the VGG16 network used is pre-trained, and its weights are provided by PyTorch[36]. To maintain consistency with the pre-trained weights, the weights of the feature extraction network and the classification network are frozen.

IV-A1 Datasets

We selected 50,000 images from the ILSVRC2012[37] validation dataset to train our feature compression neural network(FCNN) and image reconstruction neural network(IRNN). The training set consists of 44,100 images, and the validation set consists of 4,900 images. We also randomly selected one image from each of the 1,000 classification categories as the test set to verify the classification performance.During training, we resized the images to 256×256256\times 256 and applied normalization.

IV-A2 Compress Methods

In our work, we adopt classic video compression algorithms and neural network compression algorithms to compress features and images separately. Specifically, for the feature layers in FCNN, we first rearrange the filtered feature channels into a two-dimensional feature map and fill the missing parts with 0. Then, we use HM-16.12 to compress the quantized features. For the texture layer, we downsample the input image by 2 times and use the Cheng2020-Attn[34] model for compression for easy training.

IV-A3 Training Settings

To accelerate training and enable backpropagation of gradients, we simplify the feature compression process to an identity function during FCNN training, i.e., no quantization or compression of features is performed. Quantization and compression operations are only performed on features during the testing phase. For texture images, we use the Chen2022-Attn model with quality=4\text{quality}=4 provided by CompressAI[38] for compression in both the training and testing phases of FCNN and IRNN models. Additionally, the learning rate and epoch for training FCNN and IRNN are the same, which are 1e-4 and 300, respectively.

IV-B Effectiveness Verification for Image Classification Task

To verify the visual task performance of our method, we evaluate the classification performance on the ImageNet 2012 dataset. The dataset contains 1000 categories. For fairness, we randomly select one image from each category for testing. For comparison, we compress the original image using the classic video compression algorithm HEVC, denoted as Image Anchor. In addition, we extract features from the original image and then tile all feature channels into a two-dimensional grayscale image, which is then compressed using HEVC, denoted as Feature Anchor. In addition, three more popular end-to-end image compression models bmshj2018-hyperprior[39], mbt2018[40] and cheng-2020-anchor[34] are selected for comparison.

During compression, we use the intra-configuration of HEVC to compress features and images. We selected some test points with close bitrates for model comparison. It is worth noting that features need to be quantized before HEVC can compress them. Therefore, we adopt the quantization method of Chen et al.[5] to perform 8-bit quantization on the features, and the quantization and dequantization formulas are as follows:

F^i=round(log(Fimin(Fi)+1)max(log(Fimin(Fi)+1)))255\hat{F}_{i}=round(\frac{log(F_{i}-min(F_{i})+1)}{max(log(F_{i}-min(F_{i})+1))})\cdot 255 (16)
F~i=2F^imax(log(Fimin(Fi)+1))255+min(Fi)1\tilde{F}_{i}=2^{\frac{\hat{F}_{i}\cdot max({log(F_{i}-min(F_{i})+1)})}{255}}+min(F_{i})-1 (17)

Where F^i\hat{F}_{i} represents the quantized feature of the ii-th channel, and F~i\tilde{F}_{i} denotes the dequantized ii-th channel.

Refer to caption
Figure 4: Rate-Accuracy curve of the proposed method and other methods for image classification.

We use rate-accuracy curves to evaluate the performance of our proposed method. The rate is measured by bits per pixel (bpp), and the accuracy is measured by top-1 accuracy(ACC). The calculation formula of bpp is as follows:

bpp=total_bitsH×Wbpp=\frac{total\_bits}{H\times W} (18)

where WW and HH represent the width and height of the original image, and total_bitstotal\_bits denotes the total consumed bits. For our method, total_bitstotal\_bits includes not only the bits of the feature stream and the texture stream, but also the bits consumed by the quantization parameters.

The experimental results are shown in Fig. 4. We set the weight parameters for our method to α=0.5\alpha=0.5 and λ=3\lambda=3. From the figures, it can be observed that our method achieves better task performance at the same bpp under low bit rates. Compared to Image Anchor, our algorithm’s BD-Accuracy can reach 69.46%-69.46\%, saving 69.85%-69.85\% bits at the same ACC. Fig. 4 also indicates that using texture images alone leads to poor task performance. However, task performance significantly improves when combined with the transmission of partial channels from the feature layer. This suggests the effectiveness of our collaborative strategy between texture and feature layers. Additionally, it is noteworthy that the performance improvement of features at high bit rates is limited, possibly due to quantization errors.

In addition to assessing the Rate-Accuracy performance, we evaluated the compression performance of our algorithm on various depth feature layers. The experimental results are summarized in Table I. The feature layers in the table represent features extracted from different positions of VGG16, where ”Pool1” denotes a shallow feature layer with the largest data volume, and ”Pool4” represents a deep feature layer with the smallest data volume. The compression rate in the table is computed by dividing the compressed data volume by the uncompressed data volume. A lower value indicates better compression performance. The calculation formula is given by:

Compression rate=Compressed Data VolumeUncompressed Data Volume×100%\text{Compression rate}=\frac{\text{Compressed Data Volume}}{\text{Uncompressed Data Volume}}\times 100\% (19)

The table reveals that, under the same accuracy, our feature compression algorithm achieves a lower compression rate compared to directly using HEVC to compress features, with an average compression performance improvement of approximately 3 times. This demonstrates that our algorithm not only achieves excellent compression performance on deep feature layers but also maintains strong performance on shallow feature layers, highlighting the generality of our method.

Refer to caption

Refer to caption

Refer to caption

(a) Ground Truth

Refer to caption

PSNR=29.322 dB

Refer to caption

PSNR=21.542 dB

Refer to caption

PSNR=28.447 dB

(b) Bicubic

Refer to caption

PSNR=29.996 dB

Refer to caption

PSNR=21.780 dB

Refer to caption

PSNR=29.095.78 dB

(c) Texture-SR

Refer to caption

PSNR=30.171 dB

Refer to caption

PSNR=21.802 dB

Refer to caption

PSNR=29.223 dB

(d) Texture-Feature-SR

Figure 5: Comparison of visual results. The three images have bpp values of 0.078, 0.114, and 0.114, respectively. (a) is the original uncompressed image. (b) is the image reconstructed by bicubic upsampling of the compressed texture image. (c) is the image reconstructed by IRNN with the texture image. (d) is the image reconstructed by IRNN with the texture and feature images.
TABLE I: Rate-Accuracy performance of different feature layers
       Feature Layer Target Accuracy[%] Method BPP Compression Rate[%]
Pool1 \approx 62 HEVC 0.999 26.24
Proposed 0.177 4.66
\approx 65 HEVC 1.495 39.27
Proposed 0.245 6.43
Pool2 \approx 62 HEVC 0.665 17.47
Proposed 0.215 5.65
\approx 65 HEVC 1.075 28.24
Proposed 0.517 13.59
Pool4 \approx 62 HEVC 0.130 3.41
Proposed 0.040 1.06
\approx 65 HEVC 0.205 5.38
Proposed 0.050 1.30
 

IV-C Quality Enhancement for Compressed Image

Guided by the texture image, our method achieves better machine vision task performance at the same bitrate. It provides humans with preview images at the decoding end to assist decision-making, meeting the ”dual vision” requirements of human eye visual perception and machine vision cognition. Moreover, the features can also be applied to texture images to facilitate image reconstruction.

We conducted experiments with three strategies to validate the performance of IRNN and the auxiliary role of features in image reconstruction. The first method applies Bicubic upsampling to the compressed low-resolution texture image I^\hat{I}_{\downarrow} to restore the original resolution with a 2×2\times factor, denoted bicubic. The second method feeds I^\hat{I}_{\downarrow} solely into the IRNN network for image reconstruction, denoted Texture-SR. The third method takes features and I^\hat{I}_{\downarrow} as inputs for joint reconstruction by IRNN, denoted Texture-Feature-SR. The experimental setup remains the same as the previous section, with 1000 images used for evaluation. The feature’s QP and texture quality are set to 11 and 4, respectively.

The experimental results are as follows: The average PSNR for the first method is 23.254dB, for the second method is 23.949dB, and for the third method is 24.009dB. Comparing the results of the first and second methods, our IRNN reconstruction network improves performance by 0.695dB. A comparison between Bicubic and Texture-SR demonstrates that our IRNN reconstruction network can improve performance by 0.695 dB. Furthermore, a comparison between Methods Texture-SR and Texture-Feature-SR reveals that features can guide image reconstruction. As shown in Fig. 5, the target of the preview image reconstructed by the joint features and texture is slightly clearer. However, the improvement is not significant compared to Texture-SR. This could be attributed to our feature selection strategy, as depicted in Fig. 6. Since the feature layer close to the classifier inherently contains less texture information, and the feature selection module tends to prioritize features relevant to machine vision tasks, the improvement is not substantial.

Refer to caption
(a)
Refer to caption
(b)
Refer to caption
(c)
Figure 6: Visualization of feature channels. (a), (b), and (c) correspond to the feature maps of the three images in Figure 5, respectively. These feature maps are used for image reconstruction together with the texture images.

IV-D Ablation Study

TABLE II: Ablation experiment setup
       Method Texture FRM
1. FCNN w/o texture and FRM
2. FCNN w/o FRM
3. FCNN w/o Texture
4. FCNN
 
Refer to caption
Figure 7: Rate-Accuracy results of image classification for different comparison methods

In order to assess the impact of two strategies on detection performance, we conducted ablation experiments. The strategies in question are direct channel discarding and using low-quality features to fill the missing channels. The experimental setups are detailed in Table II. The ”Texture” parameter indicates whether low-quality features extracted from the texture are used to replace the lost channels. When ”Texture” is enabled, it means that features ^HQ\hat{\mathcal{F}}_{HQ} from Fig. 2 are utilized for subsequent training; otherwise, the network is trained using HQm\mathcal{F}^{m}_{HQ}. The ”FRM” parameter indicates whether the FRM is employed to reconstruct the features ^HQ\hat{\mathcal{F}}_{HQ} or HQm\mathcal{F}^{m}_{HQ} for enhancement.

The experimental results, presented in Fig. 7, reveal comparisons between Method 1 and Method 2. It is observed that the classification performance when directly discarding feature channels and when utilizing low-quality features to fill in the lost channels is comparable, leading to a degradation in classification performance. Moreover, upon closer examination, filling the lost channels in Method 1 directly with low-quality feature channels results in a slight decrease in classification performance, while replacing the filled low-quality graphs in Method 2 with zeros maintains their classification performance. These findings suggest that the gap between low-quality and high-quality features is too significant to directly fill in the missing feature channels, thus limiting the potential improvement in visual task performance.

Additionally, a comparison between Method 1 and Method 3 demonstrates that, with the same QP compression and the assistance of the FRM module, Method 1 can further reduce the amount of data for the features—i.e., it can discard more channels—but at the cost of decreased classification performance.

Furthermore, contrasting Method 2 with Method 4 reveals that, with the assistance of the FRM module, it can further reduce the amount of data for the features, and the performance of the visual task remains unchanged. This underscores the efficient ability of FRM to reconstruct and enhance features.

V Conclusion

In this paper, we propose a texture-guided deep feature compression method. The proposed method consists of a two-layer encoding structure: the feature layer and the texture layer. The feature layer is designed to provide efficient deep features for machine vision tasks, while the texture layer aims to provide humans with preview images for decision-making assistance. Specifically, the feature layer consists of a channel selection module and a feature reconstruction module. During encoding, the channel selection module directly discards task-irrelevant channels to save bits, and the missing feature channels are recovered using the texture image during decoding to ensure the quality of machine vision tasks. The texture layer consists of an image reconstruction network that utilizes the features from the feature layer to guide the image reconstruction process.

Experimental results demonstrate that the proposed method can perform better vision tasks at lower bitrate consumption. Under texture guidance, the proposed method can effectively remove channel redundancy in features, thereby reducing the amount of feature data. Meanwhile, the features can also assist image reconstruction, which addresses the needs of both human and machine vision.

In the future, we will explore using end-to-end networks to replace traditional video codecs for generating bitstreams. This will enable end-to-end training and reduce the impact of different quantization parameters on tasks.

References

  • [1] G. J. Sullivan, J. Ohm, W. Han, and T. Wiegand, “Overview of the High Efficiency Video Coding (HEVC) standard,” IEEE Trans. Circuits Syst. Video Technol., vol. 22, no. 12, pp. 1649-1668, Dec. 2012.
  • [2] H. Choi and I. V. Bajic, “Deep feature compression for collaborative object detection,” in Proc. IEEE Int. Conf. Image Process.(ICIP), Oct. 2018.
  • [3] D. Flynn et al., “High Efficiency Video Coding (HEVC) range extensions text specification: draft 7,” document JCTVC-Q1005, Apr. 2014.
  • [4] J. Redmon and A. Farhadi, “YOLO9000: Better, faster, stronger,” in Proc. Conf. Comput. Vis. Pattern Recognit., Jul. 2017.
  • [5] Z. Chen, K. Fan, S. Wang, L. Duan, W. Lin, and A. C. Kot, “Toward intelligent sensing: Intermediate deep feature compression,” IEEE Trans. Image Process., vol. 29, pp. 2230-2243, Sep. 2019.
  • [6] K. Simonyan and A. Zisserman, “Very deep convolutional networks for large-scale image recognition,” in Proc. Int. Conf. Learn. Representations, May 2015.
  • [7] K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in Proc. Conf. Comput. Vis. Pattern Recognit., Dec. 2016.
  • [8] Z. Chen, L. -Y. Duan, S. Wang, W. Lin and A. C. Kot, “Data Representation in Hybrid Coding Framework for Feature Maps Compression,” 2020 IEEE International Conference on Image Processing (ICIP), Abu Dhabi, United Arab Emirates, 2020, pp. 3094-3098.
  • [9] S. Suzuki, S. Takeda, M. Takagi, R. Tanida, H. Kimata and H. Shouno, “Deep Feature Compression Using Spatio-Temporal Arrangement Toward Collaborative Intelligent World,” IEEE Trans. Circuits Syst. Video Technol., vol. 32, no. 6, pp. 3934-3946, June 2022.
  • [10] Y. Hu, S. Xia, W. Yang, and J. Liu, “Sensitivity-aware bit allocation for intermediate deep feature compression,” in Proc. IEEE Visual Commun. Image Process.(VCIP), Dec. 2020.
  • [11] Z. Wang, F. Li, J. Xu and P. C. Cosman, ”Human–Machine Interaction-Oriented Image Coding for Resource-Constrained Visual Monitoring in IoT,” in IEEE Internet of Things Journal, vol. 9, no. 17, pp. 16181-16195, 1 Sept.1, 2022.
  • [12] S. Wang et al., “Towards Analysis-Friendly Face Representation With Scalable Feature and Texture Compression,” in IEEE Transactions on Multimedia, vol. 24, pp. 3169-3181, 2022.
  • [13] H. Choi and I. V. Baji?, “Scalable Image Coding for Humans and Machines,” in IEEE Transactions on Image Processing, vol. 31, pp. 2739-2754, 2022.
  • [14] X. Fang, Y. Duan, Q. Du, X. Tao and F. Li, “Sketch Assisted Face Image Coding for Human and Machine Vision: A Joint Training Approach,” in IEEE Transactions on Circuits and Systems for Video Technology, vol. 33, no. 10, pp. 6086-6100, Oct. 2023, doi: 10.1109/TCSVT.2023.3262251.
  • [15] J. Seppälä et al., “Enhancing Image Coding for Machines with Compressed Feature Residuals,” 2021 IEEE International Symposium on Multimedia (ISM), Naple, Italy, 2021, pp. 217-225, doi: 10.1109/ISM52913.2021.00044.
  • [16] J. Gailly and M. Adler. “Gzip,” Accessed: Feb. 9, 2023. [Online]. Available: http://www.gzip.org/
  • [17] Jean-loup Gailly, Mark Adler, “Zlib,” Accessed: Feb. 9, 2023. [Online]. Available: https://zlib.net/
  • [18] Julian Seward, “Bzip2,” Accessed: Feb. 9, 2023. [Online]. [Online]. Available: https://en.wikipedia.org/wiki/Bzip2
  • [19] Igor Pavlov, “Lempel–ziv–markov chain algorithm,” Feb. 9, 2023. [Online]. Available: http://en.wikipedia.org/wiki/LZMA
  • [20] J. Lainema, F. Bossen, W.-J. Han, J. Min, and K. Ugur, “Intra coding of the HEVC standard,” IEEE Trans. Circuits Syst. Video Technol., vol. 22, no. 12, pp. 1792–1801, Dec. 2012.
  • [21] B. Bross, P. Helle, H. Lakshman, and K. Ugur, “Inter-picture prediction in HEVC,” in High Efficiency Video Coding (HEVC): Algorithms and Architectures. Cham, Switzerland: Springer, 2014, pp. 113–140
  • [22] W. Wang, Z. Chen, Z. Wang, J. Lin, L. Xu and W. Lin, “Channel-Wise Bit Allocation for Deep Visual Feature Quantization,” in Proc. IEEE Visual Commun. Image Process.(VCIP), Bordeaux, France, 2022, pp. 3978-3982.
  • [23] H. Choi and I. V. Baji?, “Near-Lossless Deep Feature Compression for Collaborative Intelligence,” 2018 IEEE 20th International Workshop on Multimedia Signal Processing (MMSP), Vancouver, BC, Canada, 2018, pp. 1-6.
  • [24] Zhuo Chen, Kui Fan, Shiqi Wang, Ling-Yu Duan, Weisi Lin, and Alex Kot, “Lossy intermediate deep learning feature compression and evaluation,” in Proceedings of the 27th ACM International Conference on Multimedia, 2019, pp. 2414–2422.
  • [25] K. Misra, T. Ji, A. Segall and F. Bossen, “Video Feature Compression for Machine Tasks,” 2022 IEEE International Conference on Multimedia and Expo (ICME), Taipei, Taiwan, 2022, pp. 1-6.
  • [26] M. Yamazaki, Y. Kora, T. Nakao, X. Lei and K. Yokoo, “Deep Feature Compression using Rate-Distortion Optimization Guided Autoencoder,” in Proc. IEEE Int. Conf. Image Process.(ICIP),Bordeaux, France, 2022, pp. 1216-1220.
  • [27] G. E. Hinton and R. R. Salakhutdinov, “Reducing the dimensionality of data with neural networks,” Science, vol. 313, no. 5786, pp. 504–507, 2006.
  • [28] Y. Mei, F. Li, L. Li and Z. Li, “Learn A Compression for Objection Detection - VAE with a Bridge,” Proc. IEEE Visual Commun. Image Process.(VCIP), Munich, Germany, 2021, pp. 1-5.
  • [29] Y. Matsubara, D. Callegaro, S. Singh, M. Levorato and F. Restuccia, “BottleFit: Learning Compressed Representations in Deep Neural Networks for Effective and Efficient Split Computing,” 2022 IEEE 23rd International Symposium on a World of Wireless, Mobile and Multimedia Networks (WoWMoM), Belfast, United Kingdom, 2022, pp. 337-346.
  • [30] Z. Yuan, S. Rawlekar, S. Garg, E. Erkip and Y. Wang, “Feature Compression for Rate Constrained Object Detection on the Edge,” 2022 IEEE 5th International Conference on Multimedia Information Processing and Retrieval (MIPR), CA, USA, 2022, pp. 1-68.
  • [31] K. Xu, M. Qin, F. Sun, Y. Wang, Y. -K. Chen and F. Ren, “Learning in the Frequency Domain,” 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 2020, pp. 1737-1746, doi: 10.1109/CVPR42600.2020.00181.
  • [32] E. Jang, S. Gu, and B. Poole, “Categorical reparameterization with gumbel-softmax,” in International Conference on Learning Representations, 2017.
  • [33] Xintao Wang, Ke Yu, Chao Dong, and Chen Change Loy, “Recovering realistic texture in image superresolution by deep spatial feature transform,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 606–615.
  • [34] Z. Cheng, H. Sun, M. Takeuchi and J. Katto, “Learned Image Compression With Discretized Gaussian Mixture Likelihoods and Attention Modules,” 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 2020, pp. 7936-7945, doi: 10.1109/CVPR42600.2020.00796.
  • [35] Ronneberger, Olaf, Philipp Fischer, and Thomas Brox. “U-net: Convolutional networks for biomedical image segmentation.” Medical image computing and computer-assisted intervention–MICCAI 2015: 18th international conference, Munich, Germany, October 5-9, 2015, proceedings, part III 18. Springer International Publishing, 2015.
  • [36] A. Paszke et al., “Pytorch: An imperative style, high-performance deep learning library,” in Proc. Adv. Neural Inf. Process. Syst., Dec. 2019.
  • [37] Russakovsky et al., “ImageNet large scale visual recognition challenge,” Int. J. Comput. Vis., vol. 115, no. 3, pp. 211-252, Dec. 2015.
  • [38] Bégaint, Jean, et al. “Compressai: a pytorch library and evaluation platform for end-to-end compression research.” arXiv preprint arXiv:2011.03029 (2020).
  • [39] Ballé, Johannes, et al. “Variational image compression with a scale hyperprior.” arXiv preprint arXiv:1802.01436 (2018).
  • [40] Minnen, David, Johannes Ballé, and George D. Toderici. “Joint autoregressive and hierarchical priors for learned image compression.” Advances in neural information processing systems 31 (2018).