Sar2color: Learning Imaging Characteristics of SAR Images for SAR-to-Optical Transformation
Abstract
:1. Introduction
- We propose an end-to-end SAR-to-optical image transformation model called Sar2color, which takes into account the imaging characteristics of SAR images, reduces the adverse effects of SAR images brought to the generated optical images during the transformation process, improves the color quality of the generated optical images, and is conducive to assisting the interpretation of SAR images;
- In this paper, DCTRB and Light-ASPP modules are designed to reduce the negative effects of coherent speckle noise and geometric distortion characteristics in SAR images on the generation of optical images, and thus reduce the difficulty of SAR-to-optical image transformation task;
- A CCMB module is proposed to alleviate the problem of color deviation that occurs in generating optical images;
- This paper evaluates the results of the proposed method on the remote sensing image and optical image paired dataset SEN1-2 [33], while achieving the state-of-the-art effect on four different mainstream evaluation metrics, such as peak signal to noise ratio (PSNR) [34], structural similarity index metric (SSIM) [35], mean square error (MSE) [36], and learned perceptual image patch similarity (LPIPS) [37].
2. Methods
2.1. The Structure of Sar2color Model
2.1.1. Generator
- Firstly, the SAR image is processed by homomorphic transformation, and the transformed SAR image is obtained by the following formula: , and represent the transformed SAR image and SAR image under the spatial coordinates , respectively;
- Then, the transformed SAR image is divided into 8 blocks, DCT is applied in each sub-block, and then ZigZag [41] scanning is used to retain 10 low-frequency coefficients in the upper left corner of the coefficient matrix after DCT;
- For other medium and high-frequency coefficients, the average coefficient of each sub-block is set as the screening threshold, and the remaining 54 coefficients are compared with this threshold of each sub-block. If the coefficient is larger than the threshold, it is retained, and if less than 0, set it to 0. In this way, the medium and high-frequency coefficients with higher energy can be retained. Finally, the filtered coefficients of each sub-block are integrated.
2.1.2. Discriminator
2.2. Loss Functions
3. Results
3.1. Implementation Details
3.2. Dataset Setting
3.3. Metrics
3.4. Qualitative Evaluation
3.5. Quantitative Evaluation
3.6. Ablation Study
4. Discussion
5. Conclusions
Author Contributions
Funding
Data Availability Statement
Acknowledgments
Conflicts of Interest
Abbreviations
GAN | Generative adversarial network |
CGAN | Conditional generative adversarial network |
SAR | Synthetic aperture radar |
DCTRB | DCT residual block |
Light-ASPP | Light atrous spatial pyramid pooling |
CCMB | Correct color memory block |
PSNR | Peak signal-to-noise ratio |
SSIM | Structural similarity index metric |
MSE | Mean square error |
LPIPS | Learned perceptual image patch similarity |
References
- Scarpa, G.; Gargiulo, M.; Mazza, A.; Gaetano, R. A CNN-based fusion method for feature extraction from sentinel data. Remote Sens. 2018, 10, 236. [Google Scholar] [CrossRef] [Green Version]
- Lyu, H.; Lu, H.; Mou, L. Learning a transferable change rule from a recurrent neural network for land cover change detection. Remote Sens. 2016, 8, 506. [Google Scholar] [CrossRef] [Green Version]
- Balz, T.; Liao, M. Building-damage detection using post-seismic high-resolution SAR satellite data. Int. J. Remote Sens. 2010, 31, 3369–3391. [Google Scholar] [CrossRef]
- Singhroy, V.; Mattar, K.; Gray, A. Landslide characterisation in Canada using interferometric SAR and combined SAR and TM images. Adv. Space Res. 1998, 21, 465–476. [Google Scholar] [CrossRef]
- Santangelo, M.; Cardinali, M.; Bucci, F.; Fiorucci, F.; Mondini, A.C. Exploring event landslide mapping using Sentinel-1 SAR backscatter products. Geomorphology 2022, 397, 108021. [Google Scholar] [CrossRef]
- Zhang, T.; Zhang, X.; Shi, J.; Wei, S.; Wang, J.; Li, J.; Su, H.; Zhou, Y. Balance scene learning mechanism for offshore and inshore ship detection in SAR images. IEEE Geosci. Remote Sens. Lett. 2020, 19, 1–5. [Google Scholar] [CrossRef]
- Gao, J.; Yuan, Q.; Li, J.; Zhang, H.; Su, X. Cloud removal with fusion of high resolution optical and SAR images using generative adversarial networks. Remote Sens. 2020, 12, 191. [Google Scholar] [CrossRef] [Green Version]
- Cigna, F.; Bateson, L.B.; Jordan, C.J.; Dashwood, C. Simulating SAR geometric distortions and predicting Persistent Scatterer densities for ERS-1/2 and ENVISAT C-band SAR and InSAR applications: Nationwide feasibility assessment to monitor the landmass of Great Britain with SAR imagery. Remote Sens. Environ. 2014, 152, 441–466. [Google Scholar] [CrossRef] [Green Version]
- Maity, A.; Pattanaik, A.; Sagnika, S.; Pani, S. A comparative study on approaches to speckle noise reduction in images. In Proceedings of the 2015 International Conference on Computational Intelligence and Networks, Odisha, India, 12–13 January 2015; pp. 148–155. [Google Scholar]
- Zhang, Q.; Liu, X.; Liu, M.; Zou, X.; Zhu, L.; Ruan, X. Comparative analysis of edge information and polarization on sar-to-optical translation based on conditional generative adversarial networks. Remote Sens. 2021, 13, 128. [Google Scholar] [CrossRef]
- Guo, J.; He, C.; Zhang, M.; Li, Y.; Gao, X.; Song, B. Edge-Preserving Convolutional Generative Adversarial Networks for SAR-to-Optical Image Translation. Remote Sens. 2021, 13, 3575. [Google Scholar] [CrossRef]
- Kong, Y.; Hong, F.; Leung, H.; Peng, X. A Fusion Method of Optical Image and SAR Image Based on Dense-UGAN and Gram–Schmidt Transformation. Remote Sens. 2021, 13, 4274. [Google Scholar] [CrossRef]
- Chen, Y.; Bruzzone, L. Self-supervised sar-optical data fusion of sentinel-1/-2 images. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–11. [Google Scholar] [CrossRef]
- Goodfellow, I.J.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial networks. arXiv 2014, arXiv:1406.2661. [Google Scholar] [CrossRef]
- Mirza, M.; Osindero, S. Conditional generative adversarial nets. arXiv 2014, arXiv:1411.1784. [Google Scholar]
- Isola, P.; Zhu, J.Y.; Zhou, T.; Efros, A.A. Image-to-image translation with conditional adversarial networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1125–1134. [Google Scholar]
- Zhu, J.Y.; Park, T.; Isola, P.; Efros, A.A. Unpaired image-to-image translation using cycle-consistent adversarial networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2223–2232. [Google Scholar]
- Wang, T.C.; Liu, M.Y.; Zhu, J.Y.; Tao, A.; Kautz, J.; Catanzaro, B. High-resolution image synthesis and semantic manipulation with conditional gans. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8798–8807. [Google Scholar]
- Cho, W.; Choi, S.; Park, D.K.; Shin, I.; Choo, J. Image-to-image translation via group-wise deep whitening-and-coloring transformation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 10639–10647. [Google Scholar]
- Reed, S.; Akata, Z.; Yan, X.; Logeswaran, L.; Schiele, B.; Lee, H. Generative adversarial text to image synthesis. In Proceedings of the International Conference on Machine Learning (PMLR), New York, NY, USA, 19–24 June 2016; pp. 1060–1069. [Google Scholar]
- Zhang, H.; Xu, T.; Li, H.; Zhang, S.; Wang, X.; Huang, X.; Metaxas, D.N. Stackgan: Text to photo-realistic image synthesis with stacked generative adversarial networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 5907–5915. [Google Scholar]
- Bahng, H.; Yoo, S.; Cho, W.; Park, D.K.; Wu, Z.; Ma, X.; Choo, J. Coloring with words: Guiding image colorization through text-based palette generation. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018; pp. 431–447. [Google Scholar]
- Yoo, S.; Bahng, H.; Chung, S.; Lee, J.; Chang, J.; Choo, J. Coloring with limited data: Few-shot colorization via memory augmented networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 11283–11292. [Google Scholar]
- Choi, Y.; Choi, M.; Kim, M.; Ha, J.W.; Kim, S.; Choo, J. Stargan: Unified generative adversarial networks for multi-domain image-to-image translation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8789–8797. [Google Scholar]
- Pumarola, A.; Agudo, A.; Martinez, A.M.; Sanfeliu, A.; Moreno-Noguer, F. Ganimation: Anatomically-aware facial animation from a single image. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 818–833. [Google Scholar]
- Merkle, N.; Auer, S.; Müller, R.; Reinartz, P. Exploring the potential of conditional adversarial networks for optical and SAR image matching. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 1811–1820. [Google Scholar] [CrossRef]
- Doi, K.; Sakurada, K.; Onishi, M.; Iwasaki, A. GAN-Based SAR-to-Optical Image Translation with Region Information. In Proceedings of the IGARSS 2020—2020 IEEE International Geoscience and Remote Sensing Symposium, Waikoloa, HI, USA, 26 September–2 October 2020; pp. 2069–2072. [Google Scholar]
- Yu, T.; Zhang, J.; Zhou, J. Conditional GAN with Effective Attention for SAR-to-Optical Image Translation. In Proceedings of the 2021 3rd International Conference on Advances in Computer Technology, Information Science and Communication (CTISC), Shanghai, China, 23–25 April 2021; pp. 7–11. [Google Scholar]
- Darbaghshahi, F.N.; Mohammadi, M.R.; Soryani, M. Cloud removal in remote sensing images using generative adversarial networks and SAR-to-optical image translation. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–9. [Google Scholar] [CrossRef]
- Zuo, Z.; Li, Y. A SAR-to-Optical Image Translation Method Based on PIX2PIX. In Proceedings of the 2021 IEEE International Geoscience and Remote Sensing Symposium IGARSS, Brussels, Belgium, 11–16 July 2021; pp. 3026–3029. [Google Scholar]
- Turnes, J.N.; Castro, J.D.B.; Torres, D.L.; Vega, P.J.S.; Feitosa, R.Q.; Happ, P.N. Atrous cgan for sar to optical image translation. IEEE Geosci. Remote Sens. Lett. 2020, 19, 1–5. [Google Scholar] [CrossRef]
- Tan, D.; Liu, Y.; Li, G.; Yao, L.; Sun, S.; He, Y. Serial GANs: A Feature-Preserving Heterogeneous Remote Sensing Image Transformation Model. Remote Sens. 2021, 13, 3968. [Google Scholar] [CrossRef]
- Schmitt, M.; Hughes, L.H.; Zhu, X.X. The SEN1-2 dataset for deep learning in SAR-optical data fusion. arXiv 2018, arXiv:1807.01569. [Google Scholar] [CrossRef] [Green Version]
- Huynh-Thu, Q.; Ghanbari, M. Scope of validity of PSNR in image/video quality assessment. Electron. Lett. 2008, 44, 800–801. [Google Scholar] [CrossRef]
- Hore, A.; Ziou, D. Image quality metrics: PSNR vs. SSIM. In Proceedings of the 2010 20th International Conference on Pattern Recognition, Istanbul, Turkey, 23–26 August 2010; pp. 2366–2369. [Google Scholar]
- Chai, T.; Draxler, R.R. Root mean square error (RMSE) or mean absolute error (MAE)?–Arguments against avoiding RMSE in the literature. Geosci. Model Dev. 2014, 7, 1247–1250. [Google Scholar] [CrossRef] [Green Version]
- Zhang, R.; Isola, P.; Efros, A.A.; Shechtman, E.; Wang, O. The unreasonable effectiveness of deep features as a perceptual metric. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 586–595. [Google Scholar]
- Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention; Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
- Rubel, O.S.; Lukin, V.V.; De Medeiros, F.S. Prediction of Despeckling Efficiency of DCT-based filters Applied to SAR Images. In Proceedings of the 2015 International Conference on Distributed Computing in Sensor Systems, Fortaleza, Brazil, 10–12 June 2015; pp. 159–168. [Google Scholar]
- Meenakshi, K.; Swaraja, K.; Kora, P. A robust DCT-SVD based video watermarking using zigzag scanning. In Soft Computing and Signal Processing; Springer: Berlin/Heidelberg, Germany, 2019; pp. 477–485. [Google Scholar]
- Chen, L.C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-decoder with atrous separable convolution for semantic image segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 801–818. [Google Scholar]
- Guo, H.; Guo, Z.; Pan, Z.; Liu, X. Bilateral Res-Unet for Image Colorization with Limited Data via GANs. In Proceedings of the 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), Washington, DC, USA, 1–3 November 2021; pp. 729–735. [Google Scholar]
- Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25, 1097–1105. [Google Scholar] [CrossRef]
- Peters, A.F.; Peters, P. The Color Thief; Albert Whitman and Company: Park Ridge, IL, USA, 2015. [Google Scholar]
- Guo, G.; Wang, H.; Bell, D.; Bi, Y.; Greer, K. KNN model-based approach in classification. In Proceedings of the OTM Confederated International Conferences On the Move to Meaningful Internet Systems; Springer: Berlin/Heidelberg, Germany, 2003; pp. 986–996. [Google Scholar]
- Li, Y.; Chen, X.; Wu, F.; Zha, Z.J. Linestofacephoto: Face photo generation from lines with conditional self-attention generative adversarial networks. In Proceedings of the 27th ACM International Conference on Multimedia, Nice, France, 21–25 October 2019; pp. 2323–2331. [Google Scholar]
- Xian, W.; Sangkloy, P.; Agrawal, V.; Raj, A.; Lu, J.; Fang, C.; Yu, F.; Hays, J. Texturegan: Controlling deep image synthesis with texture patches. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8456–8465. [Google Scholar]
Season | Train | Test |
---|---|---|
Spring | 2000 | 500 |
Summer | 1600 | 400 |
Fall | 2000 | 500 |
Winter | 1600 | 400 |
Total | 7200 | 1800 |
PSNR↑ | SSIM↑ | MSE↓ | LPIPS↓ | |
---|---|---|---|---|
Pix2pix-HD [18] | 13.4253 | 0.2782 | 0.0487 | 0.3734 |
Faramarz [29] | 15.5617 | 0.3451 | 0.0374 | 0.2511 |
Serial GAN [32] | 16.3529 | 0.3876 | 0.0328 | 0.2649 |
Ours | 18.8145 | 0.4162 | 0.0315 | 0.1769 |
PSNR↑ | SSIM↑ | LPIPS↓ | |
---|---|---|---|
Base | 12.1417 | 0.2544 | 0.3659 |
Base + DCTRB | 16.1325 | 0.3715 | 0.2434 |
Base + Light-ASPP | 13.4263 | 0.2963 | 0.3187 |
Base + CCMB | 12.8766 | 0.2731 | 0.3428 |
Base + All of Above | 18.8145 | 0.4162 | 0.1769 |
PSNR↑ | SSIM↑ | LPIPS↓ | |
---|---|---|---|
Oral | 18.3427 | 0.4017 | 0.1936 |
Oral + | 18.5641 | 0.4058 | 0.1842 |
Oral + | 18.8145 | 0.4162 | 0.1769 |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Guo, Z.; Guo, H.; Liu, X.; Zhou, W.; Wang, Y.; Fan, Y. Sar2color: Learning Imaging Characteristics of SAR Images for SAR-to-Optical Transformation. Remote Sens. 2022, 14, 3740. https://doi.org/10.3390/rs14153740
Guo Z, Guo H, Liu X, Zhou W, Wang Y, Fan Y. Sar2color: Learning Imaging Characteristics of SAR Images for SAR-to-Optical Transformation. Remote Sensing. 2022; 14(15):3740. https://doi.org/10.3390/rs14153740
Chicago/Turabian StyleGuo, Zhe, Haojie Guo, Xuewen Liu, Weijie Zhou, Yi Wang, and Yangyu Fan. 2022. "Sar2color: Learning Imaging Characteristics of SAR Images for SAR-to-Optical Transformation" Remote Sensing 14, no. 15: 3740. https://doi.org/10.3390/rs14153740
APA StyleGuo, Z., Guo, H., Liu, X., Zhou, W., Wang, Y., & Fan, Y. (2022). Sar2color: Learning Imaging Characteristics of SAR Images for SAR-to-Optical Transformation. Remote Sensing, 14(15), 3740. https://doi.org/10.3390/rs14153740