This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

L1BSR: Exploiting Detector Overlap for Self-Supervised Single-Image Super-Resolution of Sentinel-2 L1B Imagery

Ngoc Long Nguyen1,∗ Jérémy Anger1,2,∗ Axel Davy1 Pablo Arias1 Gabriele Facciolo1
1 Université Paris-Saclay, CNRS, ENS Paris-Saclay, Centre Borelli, France       2 Kayrros SAS
These authors contributed equally to this work.
https://centreborelli.github.io/L1BSR/
Abstract

High-resolution satellite imagery is a key element for many Earth monitoring applications. Satellites such as Sentinel-2 feature characteristics that are favorable for super-resolution algorithms such as aliasing and band-misalignment. Unfortunately the lack of reliable high-resolution (HR) ground truth limits the application of deep learning methods to this task. In this work we propose L1BSR, a deep learning-based method for single-image super-resolution and band alignment of Sentinel-2 L1B 10m bands. The method is trained with self-supervision directly on real L1B data by leveraging overlapping areas in L1B images produced by adjacent CMOS detectors, thus not requiring HR ground truth. Our self-supervised loss is designed to enforce the super-resolved output image to have all the bands correctly aligned. This is achieved via a novel cross-spectral registration network (CSR) which computes an optical flow between images of different spectral bands. The CSR network is also trained with self-supervision using an Anchor-Consistency loss, which we also introduce in this work. We demonstrate the performance of the proposed approach on synthetic and real L1B data, where we show that it obtains comparable results to supervised methods.

[Uncaptioned image][Uncaptioned image][Uncaptioned image][Uncaptioned image][Uncaptioned image][Uncaptioned image][Uncaptioned image][Uncaptioned image][Uncaptioned image][Uncaptioned image]
Figure 1: L1BSR produces a 5m high-resolution (HR) output with all bands correctly registered from a single 10m low-resolution (LR) Sentinel-2 L1B image with misaligned bands. Note that our method is trained on real data with self-supervision, i.e. without any ground truth HR targets.

1 Introduction

Earth observation (EO) satellites play a crucial role in our understanding of the Earth systems including climate, natural resources, ecosystems, and natural and human-induced disasters. The Sentinel-2 mission, which is a part of the Copernicus Programme by the European Space Agency (ESA), is considered a significant EO effort alongside other missions such as Landsat. Sentinel-2 provides optical images of Earth’s surface in 13 spectral bands, 4 bands at 10m resolution, 6 bands at 20m, and 3 bands at 60m. The blue (B), green (G), red (R), and near-infrared (N) bands at a ground sample distance (GSD) of 10m/pixel are particularly useful for a variety of applications, including land cover classification, vegetation monitoring, and urban mapping [9]. However, for certain tasks, such as identifying small objects or analyzing fine-scale features, this spatial resolution is still inadequate. To address this limitation, super-resolution (SR) techniques can be used to achieve a GSD better than 10m for the RGBN bands.

SR approaches can be broadly classified into multi-image super-resolution (MISR) and single-image super-resolution (SISR). MISR aims at reconstructing a high-resolution (HR) image from a set of low-resolution (LR) images, typically captured with different viewpoints [1, 29, 30, 21] or at different satellite passes [27, 3]. If the LR images contain alias and sub-pixel misalignment, they present a perfect scenario for MISR to leverage complementary information in different frames and to recover the true details in the HR output [29]. SISR, on the other hand, is often considered an ill-posed problem due to the potential loss or corruption of high-frequency information caused by factors such as noise, blur, or compression. Nonetheless, a recent study [31] demonstrates the possibility of SISR for Sentinel-2 10m bands thanks to its unique sensor specifications, namely the inter-band shift and aliasing. The misaligned bands sample the ground at different positions. Since they are correlated each band obtains complementary information from the other bands, in a situation similar to a demosaicing problem.

Deep learning (DL) SISR methods currently outperform traditional model-based approaches by a large margin [42]. To date, all learning-based methods for SISR of Sentinel-2 10m bands have used supervised training, which penalizes a loss between the HR image predicted by the network and a ground truth HR image. Some studies attempt to train a SISR model on a simulated dataset where LR images are generated using a pre-defined degradation model [34]. However, the performance may drop substantially if the real low-resolution input deviates from the simulated degradation model. Other works use real HR images acquired by other satellites to directly supervise the SR of Sentinel-2 [12, 31]. However, obtaining the HR ground truth images can be costly. In addition, the use of HR images from different satellites introduces challenges such as spectral response discrepancies, and acquisition viewpoint and time differences, which complicate the process of dataset creation and negatively impact performance.

Refer to caption
Figure 2: Sensor layout of the Sentinel-2 MSI (figure adapted from [14]). The push-broom acquisition is done in the vertical direction. The overlap between detectors provides two near-simultaneous observations of the scene.

A promising direction is to use self-supervised learning techniques, which have been applied to multi-image restoration tasks such as video/burst denoising and demosaicing [11, 10, 8, 45, 37], and recently to MISR in the context of push-frame satellites [29, 30] that acquire bursts of images at high frame rate. These techniques exploit redundant information from multiple observations: one of the degraded frames in the input sequence is withheld from the network and used as label instead of the ground truth. As a consequence they require at least two degraded observations from the same HR signal.

In this work we leverage a unique feature of the Sentinel-2 hardware design that enables self-supervised training of single-image super-resolution (Fig.1). Sentinel-2 is equipped with a MultiSpectral Instrument (MSI) that has 12 detectors capturing information in the in the visible and near infrared (VNIR) wavelength range.These detectors operate in a push-broom fashion, scanning the image line-by-line as the satellite moves over the ground (as illustrated in Fig.2). Of note, adjacent detectors share a 2 km overlap area across the track (120 to 200 pixels), which offers opportunities for self-supervised image restoration techniques.

These overlapping regions are only available in early products in the Sentinel-2 processing pipeline, such as the level-1B (L1B) products, which present a significant inter-band parallax due to the hardware design of the detectors. This parallax, while beneficial for the super-resolution task, is undesirable for human interpreters, which is why it is removed later in the pipeline (e.g. L1C products) using camera calibration information to align the different bands.

Contributions.

In this work we propose L1BSR, a novel self-supervised method for SISR and band alignment of Sentinel-2 L1B RGBN bands. The method is trained directly on real L1B data using image crops contained in the detector overlap regions. One of the overlapping L1B crops is given as input to our network, and the other is used as target in the loss. The network is tasked to generate a super-resolved image such that, when properly aligned and downsampled, matches the target 10m LR crop [29, 30]. It should be noted that once trained, as a SISR method, our network has the capability to produce 5m HR images throughout the image domain, rather than just at the overlapping regions.

Our self-supervised loss is designed to enforce the super-resolution network to output an HR image with the bands correctly aligned. This is achieved by aligning all bands of the target image with the green channel of the super-resolved output.

To that aim, as a second contribution of our work, we present a novel cross-spectral registration (CSR) method which allows to compute an optical flow between images of different spectral bands. To train the CSR network we propose Anchor-Consistency, a simple yet effective self-supervised loss for cross-spectral registration. Our self-supervised cross-spectral registration simultaneously learns to handle all possible band combinations. We use our CSR network only during training of the SR network as part of the self-supervised loss. Once the reconstruction network is trained, the cross-spectral registration network is no longer required. The reconstruction network can directly generate high-quality HR images with correctly registered bands (as shown in Fig.1) without requiring an explicit alignment step, nor calibration information.

We validate our contributions with an empirical study on a synthetic dataset obtained from L1C products (Sec. 4), designed to model the main characteristics of Sentinel-2 L1B data. We show that our L1BSR network as well as our cross-spectral registration module trained with the proposed self-supervision strategy attain performance on par with those obtained with supervised training.

We train our self-supervised method on a dataset of 3740 pairs of L1B RGBN overlapping crops (Sec. 4.4) and compare with a supervised method designed for Sentinel-2 L1C [31] in Sec. 4.5. It is worth noting that our training dataset, which can find applications in various image restoration and cross-spectral registration tasks, will be soon available on our project website.

2 Related Work

SISR for Sentinel-2

Early research on SISR of Sentinel-2 images primarily focused on pan-sharpening the lower-resolution (20m and 60m) bands to create a uniform 10m GSD data cube [22, 13]. Recent years have seen an increased interest in enhancing the resolution of the Sentinel-2 10m bands. Some studies  [34] generated synthetic LR-HR pairs to train SISR models, but these models tend to suffer from generalization issues [5]. Another trending approach is to directly supervise the SISR of Sentinel-2 using another high-resolution satellite such as PlanetScope [31, 12, 47], VENµS [28], or WorldView [36]. However, creating the training dataset for these approaches requires significant engineering work due to the radiometric and geometric differences between the two constellations.

Self-supervised SR

Self-supervised and unsupervised learning are promising approaches to avoid the need for large labeled datasets in SR tasks. ZSSR [38] and MZSR [39] have been proposed to model image-specific LR-HR relations during the testing phase using example pairs generated from the LR test image and its degraded version. Although the idea is interesting, it may not be practical to train on each test image. Another approach is to use cycle-consistency and adversarial losses [46, 19, 26, 41] to train a neural network without requiring pairs of LR-HR images. However, these GAN-based models are prone to producing hallucinations, which may not be acceptable for certain applications.

Our SISR method is both fully self-supervised and free from hallucinations. We drew inspiration from the Noise2Noise framework [24], which introduced a pioneering approach to train a neural network for image denoising task using pairs of noisy images instead of pairs of noisy-clean images. The key idea behind Noise2Noise is that when comparing pairs of noisy images with independent noise realizations, the network learns to identify the underlying patterns in the noise and removes them accordingly. Similarly by comparing pairs of overlapping L1B images, our network can learn to recognize the aliasing patterns in each LR image and leverage them to recover the high-frequency details in the HR. The closest work to ours is the DSA method [29], which addresses MISR for SkySat imagery. During training, DSA hides the LR reference image and asks the network to produce a HR image from the other n1n-1 images such that after downsampling, it coincides with the reference image. Our work can be seen as the SISR version of DSA. However, our network also learns to perform implicit cross-spectral registration at inference time, which is a challenging and compelling task by itself.

Cross-spectral registration

Cross-spectral registration refers to the process of aligning two or more images that are captured using different sensors or imaging modalities. Cross-spectral registration has become increasingly important in various fields such as remote sensing [44, 33], medical imaging [23], and computer vision [2] as it allows for the integration of information from different spectral bands, thereby yielding richer scene representations. While increasing efforts have been made in the past few years to improve the performance of cross-spectral registration, this still remains an open problem [18]. Feature-based methods [44, 25] involve identifying distinctive features, such as edges or corners, in both images and then matching them to establish correspondences. Intensity-based methods [6, 48] rely on the similarity of the pixel values in both images. Examples of intensity-based methods include normalized cross-correlation, mutual information, and phase correlation. In recent years, deep learning-based methods have also been explored and achieved state-of-the-art (SOTA) performance. Most DL studies including  [33, 2, 40, 43] have employed image-to-image translation [17] techniques to map two images to the same image space and then register them accordingly. However, these methods require extensive work in designing models and sophisticated training losses. In contrast, we propose Anchor-Consistency a novel and straightforward loss for training a cross-modal registration network. Notwithstanding its simplicity, our method provides a strong baseline for the task. In addition, our loss can also be easily integrated into existing frameworks to improve consistency or used as a quantitative metric for evaluating cross-modal registration quality.

3 Proposed Method

Our primary aim is to leverage detector overlaps in Sentinel-2 L1B images to learn to recover high-frequency details hidden in its misaligned bands. Note that the maximum attainable resolution is capped by the spectral decay of the blur kernel resulting from the sensor’s pixel integration and the camera optics, which imposes a frequency cutoff beyond which there is no usable high frequency information [4]. For this reason, our aim in this work is to increase the resolution by a factor 2. In this section, we first present an overview of our proposed L1BSR framework (Sec. 3.1). Then, we describe our self-supervised losses in Sec. 3.2 and provide details about the training in Sec. 3.3.

Throughout the text, we denote by It,t{0,1}I_{t},t\in\{0,1\} the two 4-channel overlapping images. We refer to I0I_{0} as the input (or reference image) for the SR task and I1I_{1} as the target for our self-supervised losses, which are explained in more detail in Sec. 3.2. It,iI_{t,i} is the grayscale image extracted from the channel ii of ItI_{t}, where i{b,g,r,n}i\in\{b,g,r,n\} and bb, gg, rr, and nn stand for the blue, green, red, and near-infrared channels, respectively.

3.1 Architecture

Our proposed L1BSR framework (Fig. 3) consists of two main components: a cross-spectral registration network (CSR) and a reconstruction network (REC). The CSR module computes dense correspondences between I0,gI_{0,g} and all the bands of I1I_{1}. By utilizing these motion fields during training, the REC network learns to produce a HR output I^0HR\widehat{I}_{0}^{HR} with all four channels aligned with I0,gI_{0,g}. Of note, the CSR module is not used at test time. Instead, the REC network performs cross-channel registration implicitly.

Reconstruction Network

Our REC network is built on the Residual Channel Attention Networks (RCAN) architecture [49], which has been shown to achieve state-of-the-art performance on many image restoration tasks. We chose this architecture mainly due to its channel attention mechanism, which can be viewed as a weighting function that enables REC to selectively focus on informative channels in the feature space and disregard irrelevant ones.

The reconstruction network takes a LR image I0H×W×4I_{0}\in\mathbb{R}^{H\times W\times 4} with misaligned bands as input and produces a super-resolved output by a factor of two I^0HR\widehat{I}_{0}^{HR} with all four bands aligned with the green channel of the input image

I^0HR=REC(I0;ΘREC)2H×2W×4,\widehat{I}_{0}^{HR}=\textbf{REC}(I_{0};\Theta_{\textbf{REC}})\in\mathbb{R}^{2H\times 2W\times 4}, (1)

where ΘREC\Theta_{\textbf{REC}} denotes the network parameters. We opt for the default RCAN configuration to strike a balance between computational efficiency and performance. Overall, the REC network contains 10 residual groups, each with 20 residual channel attention blocks (RCAB), and a global skip connection. Each RCAB is a combination of a residual block and a channel attention layer implemented using a “squeeze-and-excitation” technique [16]. The number of feature channels is fixed to 64 across all layers.

It is important to highlight that the task our REC network must accomplish is particularly challenging, as it involves both super-resolution and cross-spectral registration at the same time. To tackle this problem, we incorporate a dedicated CSR module into the training process, which enables the REC network to learn efficiently the task in a self-supervised way.

Refer to caption
Figure 3: Overview of our proposed self-supervised L1BSR framework for Sentinel-2 L1B at training time. The depicted loss represents the self-supervision term Self-SR\ell_{\text{Self-SR}} (3). Note that at inference time, only one input and the reconstruction module are required.

Cross-Spectral Registration Network

The CSR module is instrumental during training. We use it to train the REC network to produce an HR output where all bands are aligned to the green one (as justified in Sec. 4.2). By having to align the channels, it becomes easier for the REC network to learn inter-band correlations, and thereby to leverage the complementary information in each band.

The cross-spectral registration network takes any two spectral bands of Sentinel-2 L1B images I,iI_{\cdot,i} and I,jI_{\cdot,j}, with i,j{b,r,g,n}i,j\in\{b,r,g,n\} as input and produces a dense correspondence between them

FI,jI,i=CSR(I¯,i,I¯,j;ΘCSR)[R,R]H×W×2,F_{I_{\cdot,j}\to I_{\cdot,i}}\!=\textbf{CSR}(\bar{I}_{\cdot,i},\bar{I}_{\cdot,j};\Theta_{\textbf{CSR}})\in[-R,R]^{H\times W\times 2}, (2)

where ΘCSR\Theta_{\textbf{CSR}} denotes the parameters of CSR, and I¯\bar{I} is the normalization of image II according to its mean and standard deviation. The network is trained with a maximum motion range of [R,R]2[-R,R]^{2} (with R=10R=10 pixels). Note that I,iI_{\cdot,i} and I,jI_{\cdot,j} should represent the same scene and be extracted either from the same image or from two overlapping images. The CSR follows a simple U-Net architecture [35] with 4 scales to increase the receptive field of the convolutions.

Refer to caption
Figure 4: Training setup of our proposed cross-spectral registration (CSR) module. The motion from I0,iI_{0,i} to I1,iI_{1,i} via the anchor It,jI_{t,j} should represent the direct motion between them. The depicted loss represents the Anchor-Consistency loss (7).

3.2 Self-supervised learning

Our framework is trained in a fully self-supervised manner, i.e. without requiring ground truth. In this section, we describe our self-supervised losses that are used to train the REC and the CSR modules.

Self-SR Loss

By utilizing the motion fields between the green band of I0I_{0} and all bands of I1I_{1} computed by CSR, REC can produce a HR output I^0HR\widehat{I}_{0}^{HR} with all bands correctly registered to I0,gI_{0,g}. We achieve this by minimizing the Self-SR loss:

Self-SR=Π2ω2(I^0HR,FI1I0,g)I11,\ell_{\text{Self-SR}}=\|\Pi_{2}\omega_{2}(\widehat{I}_{0}^{HR},F_{I_{1}\to I_{0,g}})-I_{1}\|_{1}, (3)

where Π2\Pi_{2} denotes the subsampling operator and ω2(,F)\omega_{2}(-,F) computes a bicubic sampling (Pullback) in the HR domain using the LR flow FF. FF denotes actually 4 optical flows, one for each band in I1I_{1}. The operator ω2\omega_{2} is equivalent to a backward warping with an upscaled version of the flow 2F2F.

The Self-SR loss forces ω2(I^0HR,FI1I0,g)\omega_{2}(\widehat{I}_{0}^{HR},F_{I_{1}\to I_{0,g}}) to be aligned to I1I_{1}, resulting in the requirement that all bands of the output I^0HR\widehat{I}_{0}^{HR} be registered with I0,gI_{0,g}. Following the work of [29], we can also incorporate a blur kernel kk into the Self-SR loss to directly produce a sharp HR image

Self-SR=Π2ω2(I^0HRk,FI1I0,g)I11.\ell_{\text{Self-SR}}^{*}=\|{\Pi_{2}}\omega_{2}(\widehat{I}_{0}^{HR}\ast k,F_{I_{1}\to I_{0,g}})-I_{1}\|_{1}. (4)

During training, we randomly choose the reference and the target between the two overlapping images. At inference time, we only need one single LR input and the REC network to obtain a high-quality HR output.

Anchor-Consistency Loss

The CSR network is trained with self-supervision. Fig. 4 illustrates our training setup for the CSR. For that we need 3 images I0,iI_{0,i}, It,jI_{t,j}, and I1,iI_{1,i}. The image It,jI_{t,j} serves as an anchor image extracted from either the I0I_{0} or I1I_{1} (i.e. t{0,1}t\in\{0,1\}) but may come from different spectral band than the two other images (i.e. jij\neq i). We compute the motion fields between these 3 images in two steps:

FIt,jI0,i\displaystyle F_{I_{t,j}\to I_{0,i}} =CSR(I¯0,i,I¯t,j),\displaystyle=\textbf{CSR}(\bar{I}_{0,i},\bar{I}_{t,j}), (5)
FI1,iIt,j\displaystyle F_{I_{1,i}\to I_{t,j}} =CSR(I¯t,j,I¯1,i).\displaystyle=\textbf{CSR}(\bar{I}_{t,j},\bar{I}_{1,i}).

These two motion fields should be consistent in such a way that their composition enables alignment between I0,iI_{0,i} and I1,iI_{1,i}:

F^I1,iI0,i=FI1,iIt,jFIt,jI0,i.\widehat{F}_{I_{1,i}\to I_{0,i}}=F_{I_{1,i}\to I_{t,j}}\circ F_{I_{t,j}\to I_{0,i}}. (6)

The Anchor-Consistency loss constrains that the motion from I0,iI_{0,i} to I1,iI_{1,i} via the anchor should represent the direct motion between them.

Anchor-Consistency=ω(I0,i,F^I1,iI0,i)I1,i1,\ell_{\text{Anchor-Consistency}}=\|\omega(I_{0,i},\widehat{F}_{I_{1,i}\to I_{0,i}})-I_{1,i}\|_{1}, (7)

where ω\omega is the classic pullback operator.

During training, we randomly choose the reference and the target between the two overlapping images. The spectra ii and jj are also picked arbitrarily in {b,r,g,n}\{b,r,g,n\}. It is important to note that the case where ii and jj are identical is also considered for CSR to learn to register images of the same band.

3.3 Training details

We train first the CSR network, as it is important to ensure that the CSR output can be effectively utilized by the REC network. We employ the Anchor-Consistency loss (7) to train the network, with weights initialized using Xavier’s initialization [15]. We set the batch size to 6464 and used Adam [20] with the default PyTorch parameters and a learning rate of 5e55e-5 to optimize the loss. The training converged after 200k iterations and took approximately 24 hours on a single NVIDIA V100 GPU.

The second phase consists of training the REC network using the Self-SR loss (3) and the trained CSR. We train REC on LR crops of size 96×96×496\times 96\times 4 pixels and validate on LR images of size 256×256×4256\times 256\times 4 pixels. We set the batch size to 16 and optimize the loss using the Adam optimizer with default parameters. Our learning rate is initialized to 5e55e-5 and decayed by a factor of 0.6 every 12k iterations. The training converges after 60k iterations and takes about 20 hours to complete on a single NVIDIA V100 GPU. We apply data augmentation (DA) techniques such as flips and rotations. The CSR network is fixed during the training of the REC network.

4 Experiments

In this section, we present experimental results that demonstrate the effectiveness of our fully self-supervised approach for Sentinel-2 SISR. To this aim, we conduct experiments on both real Sentinel-2 L1B data and a simulated dataset that we generated from Sentinel-2 L1C products. Through extensive ablation studies and quantitative analyses, we aim to demonstrate the efficacy of our method in addressing the challenges posed by the cross-spectral registration and SISR tasks in Sentinel-2 imagery. Additionally, we compare our self-supervised approach to a state-of-the-art supervised SISR method.

4.1 Simulated dataset

The simulated dataset used in our experiments was generated from 20 Sentinel-2 L1C products, with 18 used for training and 2 for testing. The products were extracted from 5 different continents in both summer and winter seasons to ensure geographic and radiometric diversity. For the training set, we selected 6,998 crops, each with a size of 512×512×4512\times 512\times 4 pixels. For the testing set, we selected 184 crops of the same size. Provided that Sentinel-2 imagery contains significant alias [31] which is unsuitable to use as ground truth HR, we first applied a Gaussian kernel with σ=0.7\sigma=0.7 to each crop to remove some aliasing, approximating the effect of an optical blur. The ground truth images I0HRI_{0}^{HR} and I1HRI_{1}^{HR} were then generated by applying 2 random homography transformations 0\mathcal{H}_{0} and 1\mathcal{H}_{1} to the blurred HR image (denoted as BHRB^{HR}). These ground truth HR should be aligned to I0,gI_{0,g} and I1,gI_{1,g}, respectively. We also simulated band-misalignment in the LR by applying a small homography transformation, where the translation component is dominant. Additionally, a little Gaussian noise (0.1%) was added to the LR to match the Sentinel-2 noise level. Overall, the simulation process can be summarized as follows:

ItHR\displaystyle I_{t}^{HR} =t(BHR),t{0,1}\displaystyle=\mathcal{H}_{t}(B^{HR}),\qquad t\in\{0,1\} (8)
It,g\displaystyle I_{t,g} =Π2(It,gHR)+ng,\displaystyle=\Pi_{2}(I_{t,g}^{HR})+n_{g},
It,i\displaystyle I_{t,i} =Π2((t,it)(BiHR))+ni,ig,\displaystyle=\Pi_{2}\left((\mathcal{H}_{t,i}\circ\mathcal{H}_{t})(B_{i}^{HR})\right)+n_{i},\quad i\neq g,

where t,i\mathcal{H}_{t,i} (i{b,r,n}i\in\{b,r,n\}) is a translation-dominant homography modeling the band-misalignment between It,gI_{t,g} and It,iI_{t,i}. nin_{i} models the noise in the Sentinel-2 L1B. The largest distortion between 2 bands of 2 images can be up to 10 pixels. To enable diverse ablation studies for both the CSR and the REC networks, the homographies are stored as ground truth flows.

The Sentinel-2 L1C products are derived from the L1B products by the Ground Segment. During this process, the bands are aligned and resampled using camera altitude and geometric models. However, due to imperfect parameter estimation, there may be residual shifts between the bands. These shifts are typically less than 0.3 pixels [14].

4.2 Cross-spectral registration

Table 1 reports the mean absolute pixel error of the self-supervised and supervised CSR networks on our test set when aligning the reference bands I0,iI_{0,i} to the target bands I1,jI_{1,j} with i,j{b,g,r,n}i,j\in\{b,g,r,n\}. The first half of the table shows the performance of the self-supervised CSR network, where the diagonal entries correspond to the same-band registration. The self-supervised CSR network performs exceptionally well for the RGB bands, in particular for the green band, exhibiting low mis-registration (less than 0.04 pixel), indicating a high correlation between the RGB bands. However, the registration between NIR and RGB is much more challenging, with an error around 0.1 pixel, which is twice as large as that of the RGB bands, suggesting a much lower correlation between NIR and RGB bands.

To validate the effectiveness of our self-supervised approach, we also performed supervised training of CSR on the same training set, where we penalized the error between the output flows and the ground truth flows obtained from the homographies. The second half of Table 1 presents the results of the supervised model over the test set. The table shows a small gap between the performance of the two models, with a maximum mean error of 0.03 pixels for RGB and 0.09 pixels for NIR in the supervised setting, compared to 0.04 pixels and 0.11 pixels for self-supervision. Overall, the self-supervised CSR approach performs well for many applications, without requiring any ground truth flows, knowledge of the optical instrument or scene modeling.

Table 1: Cross-spectral registration error (in pixel) of our self-supervised and supervised CSR networks over the synthetic test set (184×256×256×4184\times 256\times 256\times 4 pixels). The score of same-band registration is highlighted in bold.
Target bands
Ref. bands B G R N
Self-supervised B 0.026 0.035 0.039 0.106
G 0.034 0.026 0.038 0.092
R 0.035 0.037 0.026 0.104
N 0.100 0.086 0.098 0.027
Supervised B 0.016 0.028 0.029 0.088
G 0.027 0.014 0.027 0.076
R 0.029 0.027 0.017 0.090
N 0.083 0.072 0.081 0.017

4.3 Multi-band super-resolution

We conducted ablation experiments using the proposed synthetic L1C dataset to evaluate the performance of our self-supervised reconstruction network. We found that the residual misalignment in the L1C product affects PSNR measurements: a super-resolved result with well-aligned bands will be slightly misaligned with respect to the ground truth. To avoid this, we align each band of the ground truth to the corresponding band of the super-resolved output before computing the PSNR (note that this alignment is between images of the same band). We use a classical TV-L1 optical flow [32], setting the weight of the data attachment term to 0.3. The PSNRs shown in Tables 2 and 3 were computed in this way.

First, we studied the impact of the number of input bands on the reconstruction quality. Table 2 shows the PSNR results for four different networks trained with different input bands. We observed a significant improvement in performance as we increased the number of bands provided to the REC network, consistent with previous findings [31].

Furthermore, Fig. 5 illustrates the performance improvements in restoring the G band by providing different bands as input. As more bands are used, the network is able to restore aliased patterns into the true pattern and reach higher signal-to-noise ratio.

Secondly, we compared the performance of our self-supervised framework against a supervised training approach using the synthetic L1C dataset. The supervised training minimized an L1L_{1} loss between the restoration and the ground truth HR image. Table 3 shows the per-band mean PSNRs over the test set. We observe a significant PSNR gap in favour of the proposed self-supervised L1BSR method (0.70.7dB for the visible bands and 0.30.3 for the near-infrared). This is rather unexpected. Self-supervised training is at best equivalent to supervised training [24, 11, 8, 29]. The worse performance of the supervised method can be explained by the residual misalignment of the L1C images (see Sec. 4.1) used as ground truth during training, resulting in blurry super-resolved images. The self-supervised method does not suffer from this problem: during training each band of the target image is aligned to the green band of the super-resolved image by the CSR module. It should be possible to compensate the misalignment of the ground truth in a supervised setting by aligning the ground truth bands to the super-resolved prediction (see for example [7]). We did not explore this approach here as our interest lies mainly in the self-supervised training.

Refer to captionRefer to caption
LR (Green)
Refer to captionRefer to caption
G 42.90dB
Refer to captionRefer to caption
BG 46.71dB
Refer to captionRefer to caption
BGR 48.24dB
Refer to captionRefer to caption
BGRN 49.42dB
Refer to captionRefer to caption
HR (Green)
Refer to captionRefer to caption
LR (Green)
Refer to captionRefer to caption
G 40.62dB
Refer to captionRefer to caption
BG 44.20dB
Refer to captionRefer to caption
BGR 45.85dB
Refer to captionRefer to caption
BGRN 46.77dB
Refer to captionRefer to caption
HR (Green)
Figure 5: SR reconstruction of the green band when trained jointly with different input spectral bands. We observe that the other bands (B, R, N) provide valuable information for improving the reconstruction of the green band.
Table 2: Multi-band super-resolution with the Green band being the reference. PSNRs obtained after aligning GT bands to SR bands. Best PSNR in bold.
Testing bands
Training bands B G R N
G 46.39
BG 50.87 49.77
BGR 51.62 51.09 48.51
BGRN 51.80 51.67 48.82 41.34
Table 3: Comparison with supervised training. PSNRs obtained after aligning GT bands to SR bands. Best PSNR in bold.
Testing bands
Methods B G R N
Supervised 50.90 51.04 48.14 40.98
L1BSR 51.80 51.67 48.82 41.34

4.4 Real L1B dataset

The Sentinel-2 MSI includes 12 detectors for the VNIR bands arranged on a focal plane. The L1B product is composed of individual rasters, one per detector and per band. By design, two successive detectors acquire the scene with significant overlap, i.e. 120-200 pixels for the RGBN bands, which allows us to train our model on real L1B data.

However, due to the sensor layout, there is a noticeable vertical offset between consecutive detectors. To prepare the training and testing datasets, we pre-register the bands from different detectors using an integer translation. This process involves estimating a coarse translation between detectors using a SIFT-based matching method, refining the offset for each crop using an optical flow method, and regressing an integer translation. Since the bands of a given detector are acquired almost simultaneously, it is possible to ensure a registration precision up to a few pixels. For overlaps, registration is less precise due to the time delay, induced parallax and viewpoint changes, but typically remains less than 10 pixels. Overall, a refined registration is not required for our framework since the CSR network accurately captures residual shifts.

We used only two Sentinel-2 L1B products in this paper due to their current scarcity, and extracted training data around overlaps, consisting of 3740 pairs of height 400 pixels and width depending on the overlap width between detectors for RGBN bands.111Note that L1B products will be systematically available through the Copernicus Data Space Ecosystem starting October 2023. For testing we extracted crops outside of the overlapping regions, typically near the center of the detectors.

4.5 Qualitative analysis

We train the CSR and REC networks sequentially as described in Sec. 3 on the L1B dataset. For REC, the Self-SR\ell_{\text{Self-SR}}^{*} loss is used. The reconstruction results over the images of the test set are shown in Fig. 1. The REC network is able to successfully restore a high-quality HR image with aligned bands and fine details recovered.

Refer to caption
(a) L1B
Refer to caption
(b) L1BSR
Refer to caption
(c) L1C
Refer to caption
(d) L1C-based SR [31]
Figure 6: SR (x2) reconstruction by our proposed method 6(b) using L1B and the method in [31] 6(d) using L1C imagery. 6(a) and 6(c) are from the same acquisition.

To evaluate our self-supervised method, we compare it against another ×2\times 2 SISR method [31]. The method in [31] relies on ground truth data obtained from PlanetScope imagery. It is trained with a L1L_{1} loss, and is designed to work with Sentinel-2 L1C products. For comparison, we identified the L1C products from the same acquisition as the L1B samples and extracted the corresponding crops. Fig. 6 shows the SR results of the proposed methods on L1B, and of [31] on L1C. We found that the recovered details are very similar, yet our approach does not suffer from the gap between Sentinel-2 and PlanetScope images, which results in radiometric and geometric degradation by the network.

5 Conclusion

To conclude, this paper presents a novel self-supervised method, L1BSR, for single-image super-resolution of Sentinel-2 L1B 10m bands. Leveraging the unique hardware design of Sentinel-2, the proposed method achieves remarkable performance without the need for HR ground truth images. By training a cross-spectral registration module using an innovative self-supervised loss, Anchor-Consistency, L1BSR is capable of reconstructing high-resolution outputs with all bands correctly registered. To the best of our knowledge, our work is the first to explore the overlapping regions between detectors of Sentinel-2 and their potential for self-supervised SISR. Through an ablation study on a synthetic dataset and comparison with other supervised SISR methods for Sentinel-2 L1C images, we have demonstrated that L1BSR achieves performance on par with that of supervised training. The availability of our training dataset on the project website makes it a valuable resource for various image restoration and cross-spectral registration tasks.

Acknowledgements

This work was supported by a grant from Région Île-de-France. It was partly financed by the DGA Astrid Maturation project SURECAVI ANR-21-ASM3-0002. It was performed using HPC resources from GENCI–IDRIS (grants 2023-AD011012453R2, 2023-AD011012458R2, 2023-AD011012472R2). We thank CNES and ESA for providing the two Sentinel-2 L1B products used in this work. Centre Borelli is also with Université Paris Cité, SSA and INSERM.

References

  • [1] Jérémy Anger, Thibaud Ehret, Carlo de Franchis, and Gabriele Facciolo. Fast and accurate multi-frame super-resolution of satellite images. ISPRS Annals of Photogrammetry, Remote Sensing & Spatial Information Sciences, 5(1), 2020.
  • [2] Moab Arar, Yiftach Ginger, Dov Danon, Amit H Bermano, and Daniel Cohen-Or. Unsupervised multi-modal image registration via geometry preserving image-to-image translation. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 13410–13419, 2020.
  • [3] Md Rifat Arefin, Vincent Michalski, Pierre-Luc St-Charles, Alfredo Kalaitzis, Sookyung Kim, Samira E. Kahou, and Yoshua Bengio. Multi-image super-resolution for remote sensing using deep recurrent networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, pages 206–207, 2020.
  • [4] Simon Baker and Takeo Kanade. Limits on super-resolution and how to break them. IEEE Transactions on Pattern Analysis and Machine Intelligence, 24(9):1167–1183, 2002.
  • [5] Jianrui Cai, Hui Zeng, Hongwei Yong, Zisheng Cao, and Lei Zhang. Toward real-world single image super-resolution: A new benchmark and a new model. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 3086–3095, 2019.
  • [6] Si-Yuan Cao, Hui-Liang Shen, Shu-Jie Chen, and Chunguang Li. Boosting structure consistency for multispectral and multimodal image registration. IEEE Transactions on Image Processing, 29:5147–5162, 2020.
  • [7] Michel Deudon, Alfredo Kalaitzis, Israel Goytom, Md Rifat Arefin, Zhichao Lin, Kris Sankaran, Vincent Michalski, Samira E Kahou, Julien Cornebise, and Yoshua Bengio. Highres-net: Recursive fusion for multi-frame super-resolution of satellite imagery. arxiv 2020. arXiv preprint arXiv:2002.06460, 2020.
  • [8] Valéry Dewil, Jérémy Anger, Axel Davy, Thibaud Ehret, Gabriele Facciolo, and Pablo Arias. Self-supervised training for blind multi-frame video denoising. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, pages 2724–2734, 2021.
  • [9] Matthias Drusch, Umberto Del Bello, Sébastien Carlier, Olivier Colin, Veronica Fernandez, Ferran Gascon, Bianca Hoersch, Claudia Isola, Paolo Laberinti, Philippe Martimort, et al. Sentinel-2: Esa’s optical high-resolution mission for gmes operational services. Remote sensing of Environment, 120:25–36, 2012.
  • [10] Thibaud Ehret, Axel Davy, Pablo Arias, and Gabriele Facciolo. Joint demosaicing and denoising by overfitting of bursts of raw images. In The IEEE International Conference on Computer Vision (ICCV), 2019.
  • [11] Thibaud Ehret, Axel Davy, Jean-Michel Morel, Gabriele Facciolo, and Pablo Arias. Model-blind video denoising via frame-to-frame training. In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2019.
  • [12] Mikel Galar, Rubén Sesma, Christian Ayala, Lourdes Albizua, and Carlos Aranda. Super-resolution of sentinel-2 images using convolutional neural networks and real ground truth data. Remote Sensing, 12(18):2941, 2020.
  • [13] Massimiliano Gargiulo, Antonio Mazza, Raffaele Gaetano, Giuseppe Ruello, and Giuseppe Scarpa. Fast super-resolution of 20 m sentinel-2 bands using convolutional neural networks. Remote Sensing, 11(22):2635, 2019.
  • [14] Ferran Gascon, Catherine Bouzinac, Olivier Thépaut, Mathieu Jung, Benjamin Francesconi, Jérôme Louis, Vincent Lonjou, Bruno Lafrance, Stéphane Massera, Angélique Gaudel-Vacaresse, et al. Copernicus sentinel-2a calibration and products validation status. Remote Sensing, 9(6):584, 2017.
  • [15] Xavier Glorot and Yoshua Bengio. Understanding the difficulty of training deep feedforward neural networks. In Proceedings of the thirteenth international conference on artificial intelligence and statistics, pages 249–256. JMLR Workshop and Conference Proceedings, 2010.
  • [16] Jie Hu, Li Shen, and Gang Sun. Squeeze-and-excitation networks. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 7132–7141, 2018.
  • [17] Phillip Isola, Jun-Yan Zhu, Tinghui Zhou, and Alexei A Efros. Image-to-image translation with conditional adversarial networks. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 1125–1134, 2017.
  • [18] Xingyu Jiang, Jiayi Ma, Guobao Xiao, Zhenfeng Shao, and Xiaojie Guo. A review of multimodal image matching: Methods and applications. Information Fusion, 73:22–71, 2021.
  • [19] Gwantae Kim, Jaihyun Park, Kanghyu Lee, Junyeop Lee, Jeongki Min, Bokyeung Lee, David K Han, and Hanseok Ko. Unsupervised real-world super resolution with cycle generative adversarial network and domain discriminator. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, pages 456–457, 2020.
  • [20] Diederik P Kingma and Jimmy Ba. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980, 2014.
  • [21] Jamy Lafenetre, Ngoc Long Nguyen, Gabriele Facciolo, and Thomas Eboli. Handheld burst super-resolution meets multi-exposure satellite imagery. arXiv preprint arXiv:2303.05879, 2023.
  • [22] Charis Lanaras, José Bioucas-Dias, Silvano Galliani, Emmanuel Baltsavias, and Konrad Schindler. Super-resolution of Sentinel-2 images: Learning a globally applicable deep neural network. ISPRS Journal of Photogrammetry and Remote Sensing, 146:305–319, 2018.
  • [23] Daewon Lee, Matthias Hofmann, Florian Steinke, Yasemin Altun, Nathan D Cahill, and Bernhard Scholkopf. Learning similarity measure for multi-modal 3d image registration. In 2009 IEEE Conference on Computer Vision and Pattern Recognition, pages 186–193. IEEE, 2009.
  • [24] Jaakko Lehtinen, Jacob Munkberg, Jon Hasselgren, Samuli Laine, Tero Karras, Miika Aittala, and Timo Aila. Noise2Noise: Learning Image Restoration without Clean Data. In 35th International Conference on Machine Learning, ICML 2018, 2018.
  • [25] Xiangzeng Liu, Yunfeng Ai, Juli Zhang, and Zhuping Wang. A novel affine and contrast invariant descriptor for infrared and visible image registration. Remote Sensing, 10(4):658, 2018.
  • [26] Andreas Lugmayr, Martin Danelljan, and Radu Timofte. Unsupervised learning for real-world super-resolution. In 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW), pages 3408–3416. IEEE, 2019.
  • [27] Marcus Märtens, Dario Izzo, Andrej Krzic, and Daniël Cox. Super-resolution of proba-v images using convolutional neural networks. Astrodynamics, 3:387–402, 2019.
  • [28] Julien Michel, Juan Vinasco-Salinas, Jordi Inglada, and Olivier Hagolle. Sen2venμ\mus, a dataset for the training of sentinel-2 super-resolution algorithms. Data, 7(7):96, 2022.
  • [29] Ngoc Long Nguyen, Jérémy Anger, Axel Davy, Pablo Arias, and Gabriele Facciolo. Self-supervised multi-image super-resolution for push-frame satellite images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 1121–1131, 2021.
  • [30] Ngoc Long Nguyen, Jérémy Anger, Axel Davy, Pablo Arias, and Gabriele Facciolo. Self-supervised super-resolution for multi-exposure push-frame satellites. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 1858–1868, 2022.
  • [31] Ngoc Long Nguyen, Jérémy Anger, Lara Raad, Bruno Galerne, and Gabriele Facciolo. On the role of alias and band-shift for sentinel-2 super-resolution. arXiv preprint arXiv:2302.11494, 2023.
  • [32] Javier Sánchez Pérez, Enric Meinhardt-Llopis, and Gabriele Facciolo. Tv-l1 optical flow estimation. Image Processing On Line, 2013:137–150, 2013.
  • [33] Nicolas Pielawski, Elisabeth Wetzer, Johan Öfverstedt, Jiahao Lu, Carolina Wählby, Joakim Lindblad, and Natasa Sladoje. Comir: Contrastive multimodal image representation for registration. Advances in neural information processing systems, 33:18433–18444, 2020.
  • [34] Ferdinand Pineda, Victor Ayma, and César Beltran. A generative adversarial network approach for super-resolution of sentinel-2 satellite images. The International Archives of Photogrammetry, Remote Sensing and Spatial Information Sciences, 43:9–14, 2020.
  • [35] Olaf Ronneberger, Philipp Fischer, and Thomas Brox. U-net: Convolutional networks for biomedical image segmentation. In Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, October 5-9, 2015, Proceedings, Part III 18, pages 234–241. Springer, 2015.
  • [36] Luis Salgueiro Romero, Javier Marcello, and Verónica Vilaplana. Super-resolution of Sentinel-2 imagery using generative adversarial networks. Remote Sensing, 12(15):2424, 2020.
  • [37] Dev Yashpal Sheth, Sreyas Mohan, Joshua L. Vincent, Ramon Manzorro, Peter A. Crozier, Mitesh M. Khapra, Eero P. Simoncelli, and Carlos Fernandez-Granda. Unsupervised deep video denoising. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 1759–1768, 2021.
  • [38] Assaf Shocher, Nadav Cohen, and Michal Irani. “zero-shot” super-resolution using deep internal learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 3118–3126, 2018.
  • [39] Jae Woong Soh, Sunwoo Cho, and Nam Ik Cho. Meta-transfer learning for zero-shot super-resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 3516–3525, 2020.
  • [40] Celyn Walters, Oscar Mendez, Mark Johnson, and Richard Bowden. There and back again: Self-supervised multispectral correspondence estimation. In 2021 IEEE International Conference on Robotics and Automation (ICRA), pages 5147–5154. IEEE, 2021.
  • [41] Wei Wang, Haochen Zhang, Zehuan Yuan, and Changhu Wang. Unsupervised real-world super-resolution: A domain adaptation perspective. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 4318–4327, 2021.
  • [42] Zhihao Wang, Jian Chen, and Steven CH Hoi. Deep learning for image super-resolution: A survey. IEEE transactions on pattern analysis and machine intelligence, 43(10):3365–3387, 2020.
  • [43] Han Xu, Jiayi Ma, Jiteng Yuan, Zhuliang Le, and Wei Liu. Rfnet: Unsupervised network for mutually reinforcing multi-modal image registration and fusion. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 19679–19688, 2022.
  • [44] Yuanxin Ye, Jie Shan, Lorenzo Bruzzone, and Li Shen. Robust registration of multimodal remote sensing images based on structural similarity. IEEE Transactions on Geoscience and Remote Sensing, 55(5):2941–2958, 2017.
  • [45] Songhyun Yu, Bumjun Park, Junwoo Park, and Jechang Jeong. Joint learning of blind video denoising and optical flow estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, pages 500–501, 2020.
  • [46] Yuan Yuan, Siyuan Liu, Jiawei Zhang, Yongbing Zhang, Chao Dong, and Liang Lin. Unsupervised image super-resolution using cycle-in-cycle generative adversarial networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, pages 701–710, 2018.
  • [47] Maialen Zabalza and Angela Bernardini. Super-resolution of sentinel-2 images using a spectral attention mechanism. Remote Sensing, 14(12):2890, 2022.
  • [48] Junhao Zhang, Masoumeh Zareapoor, Xiangjian He, Donghao Shen, Deying Feng, and Jie Yang. Mutual information based multi-modal remote sensing image registration using adaptive feature weight. Remote Sensing Letters, 9(7):646–655, 2018.
  • [49] Yulun Zhang, Kunpeng Li, Kai Li, Lichen Wang, Bineng Zhong, and Yun Fu. Image super-resolution using very deep residual channel attention networks. In Proceedings of the European conference on computer vision (ECCV), pages 286–301, 2018.