In this work is presented a new adversarial training framework for deep learning neural networks for super-resolution of Sentinel 2 images, exploiting the data fusion techniques on 10 and 20 meters bands. The proposed scheme is fully convolutional and tries to answer the need for generalization in scale, producing realistic and detailed accurate images. Furthermore, the presence of a mathcal{L}-{1} loss limits the instability of GAN training, limiting possible problems of spectral dis-tortion. In our preliminary experiments, the GAN training scheme has shown comparable results in comparison with the baseline approach.

An Adversarial Training Framework for Sentinel-2 Image Super-Resolution

Mazza, A.;
2022

Abstract

In this work is presented a new adversarial training framework for deep learning neural networks for super-resolution of Sentinel 2 images, exploiting the data fusion techniques on 10 and 20 meters bands. The proposed scheme is fully convolutional and tries to answer the need for generalization in scale, producing realistic and detailed accurate images. Furthermore, the presence of a mathcal{L}-{1} loss limits the instability of GAN training, limiting possible problems of spectral dis-tortion. In our preliminary experiments, the GAN training scheme has shown comparable results in comparison with the baseline approach.
2022
Istituto di Metodologie per l'Analisi Ambientale - IMAA
978-1-6654-2792-0
978-1-6654-2791-3
978-1-6654-2793-7
Convo-lutional Neural Network
Data-Fusion
Deep Learning
Generative Adversarial Network
Sentinel-2
Super-Resolution
File in questo prodotto:
File Dimensione Formato  
An_Adversarial_Training_Framework_for_Sentinel-2_Image_Super-Resolution.pdf

solo utenti autorizzati

Licenza: NON PUBBLICO - Accesso privato/ristretto
Dimensione 692.98 kB
Formato Adobe PDF
692.98 kB Adobe PDF   Visualizza/Apri   Richiedi una copia

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/20.500.14243/516731
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 3
  • ???jsp.display-item.citation.isi??? 2
social impact