Turn off MathJax
Article Contents
CHEN Yongkang, WANG Peng, ZHU Daiyin, et al. Research progress and prospects of synthetic aperture radar image colorization algorithms[J]. Journal of Radars, in press. doi: 10.12000/JR26035
Citation: CHEN Yongkang, WANG Peng, ZHU Daiyin, et al. Research progress and prospects of synthetic aperture radar image colorization algorithms[J]. Journal of Radars, in press. doi: 10.12000/JR26035

Research Progress and Prospects of Synthetic Aperture Radar Image Colorization Algorithms

DOI: 10.12000/JR26035 CSTR: 32380.14.JR26035
Funds:  The National Natural Science Foundation of China (92464204, 61801211, 62471221), Guangdong Basic and Applied Basic Research Foundation (2025A1515010258), Shenzhen Science and Technology Program (JCYJ20240813180005007, JCYJ20230807142000001), Open Project Funds for National Key Laboratory of Space Target Awareness (STA2025ZCC0403), Open Project Funds for National Key Experiment on Advanced Off Road System Technology (201NKL-2025-O-02-05), Shenzhen Longhua Science and Technology Innovation Special Funding Project (Industrial Sci-Tech Innovation Center of Low-Altitude Intelligent Networking)
More Information
  • Synthetic aperture radar (SAR) is a remote sensing technology that utilizes the principle of synthetic aperture to achieve high-resolution microwave imaging. SAR image colorization is a fundamental and crucial task in remote sensing. Unlike optical imaging, SAR imaging is unaffected by clouds and fog, enabling all-weather observation of the Earth. However, owing to its imaging principle, SAR images are grayscale images; hence, they lack spectral information and have extremely low visual clarity. Therefore, numerous studies have focused on enhancing the interpretability of SAR images by incorporating color information. This paper reviews existing SAR image colorization techniques and categorizes them into three types: traditional SAR image colorization techniques, deep-learning–based SAR-to-optical image colorization techniques, and SAR image colorization techniques based on radiometric property preservation. Finally, we summarize the application scenarios and future development directions.

     

  • loading
  • [1]
    吴谨. 关于合成孔径激光雷达成像研究[J]. 雷达学报, 2012, 1(4): 353–360. doi: 10.3724/SP.J.1300.2012.20076.

    WU Jin. On the development of synthetic aperture ladar imaging[J]. Journal of Radars, 2012, 1(4): 353–360. doi: 10.3724/SP.J.1300.2012.20076.
    [2]
    徐丰, 王海鹏, 金亚秋. 深度学习在SAR目标识别与地物分类中的应用[J]. 雷达学报, 2017, 6(2): 136–148. doi: 10.12000/JR16130.

    XU Feng, WANG Haipeng, and JIN Yaqiu. Deep learning as applied in SAR target recognition and terrain classification[J]. Journal of Radars, 2017, 6(2): 136–148. doi: 10.12000/JR16130.
    [3]
    ZHU Xiaoxiang, TUIA D, MOU Lichao, et al. Deep learning in remote sensing: A comprehensive review and list of resources[J]. IEEE Geoscience and Remote Sensing Magazine, 2017, 5(4): 8–36. doi: 10.1109/MGRS.2017.2762307.
    [4]
    ZHANG Yuanzhi, ZHANG Hongsheng, and LIN Hui. Improving the impervious surface estimation with combined use of optical and SAR remote sensing images[J]. Remote Sensing of Environment, 2014, 141: 155–167. doi: 10.1016/j.rse.2013.10.028.
    [5]
    KALECINSKI N I, SKAKUN S, TORBICK N, et al. Crop yield estimation at different growing stages using a synergy of SAR and optical remote sensing data[J]. Science of Remote Sensing, 2024, 10: 100153. doi: 10.1016/j.srs.2024.100153.
    [6]
    云烨, 吕孝雷, 付希凯, 等. 星载InSAR技术在地质灾害监测领域的应用[J]. 雷达学报, 2020, 9(1): 73–85. doi: 10.12000/JR20007.

    YUN Ye, LÜ Xiaolei, FU Xikai, et al. Application of spaceborne interferometric synthetic aperture radar to geohazard monitoring[J]. Journal of Radars, 2020, 9(1): 73–85. doi: 10.12000/JR20007.
    [7]
    LIU Chenfang, SUN Yulin, XU Yanjie, et al. A review of optical and SAR image deep feature fusion in semantic segmentation[J]. IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, 2024, 17: 12910–12930. doi: 10.1109/JSTARS.2024.3424831.
    [8]
    杜兰, 王兆成, 王燕, 等. 复杂场景下单通道SAR目标检测及鉴别研究进展综述[J]. 雷达学报, 2020, 9(1): 34–54. doi: 10.12000/JR19104.

    DU Lan, WANG Zhaocheng, WANG Yan, et al. Survey of research progress on target detection and discrimination of single-channel SAR images for complex scenes[J]. Journal of Radars, 2020, 9(1): 34–54. doi: 10.12000/JR19104.
    [9]
    张帆, 孟凡乐, 马飞, 等. 基于主散射时间熵的多时相极化SAR盐田区域分类[J/OL]. 雷达学报(中英文), 待出版. https://link.cnki.net/urlid/10.1030.TN.20250911.1302.002, 2025.

    ZHANG Fan, MENG Fanle, MA Fei, et al. Multi-temporal polarimetric synthetic aperture radar salt field regional classification based on dominant scattering temporal entropy[J/OL]. Journal of Radars, in press. https://link.cnki.net/urlid/10.1030.TN.20250911.1302.002, 2025.
    [10]
    ZHOU Xiaodong, ZHANG Chunhua, and LI Song. A perceptive uniform pseudo-color coding method of SAR images[C]. 2006 CIE International Conference on Radar, Shanghai, China, 2006: 1–4. doi: 10.1109/ICR.2006.343253.
    [11]
    FENG Zicheng, LIU Xiaolin, and PEI Bingzhi. Pseudo-color coding method for high-dynamic single-polarization SAR images[C]. 9th International Conference on Graphic and Image Processing (ICGIP 2017), Qingdao, China, 2018, 10615: 1222–1228. doi: 10.1117/12.2302684.
    [12]
    UHLMANN S, KIRANYAZ S, and GABBOUJ M. Polarimetric SAR classification using visual color features extracted over pseudo color images[C]. 2013 IEEE International Geoscience and Remote Sensing Symposium-IGARSS, Melbourne, Australia, 2013: 1999–2002. doi: 10.1109/IGARSS.2013.6723201.
    [13]
    UHLMANN S and KIRANYAZ S. Classification of dual- and single polarized SAR images by incorporating visual features[J]. ISPRS Journal of Photogrammetry and Remote Sensing, 2014, 90: 10–22. doi: 10.1016/j.isprsjprs.2014.01.005.
    [14]
    LI Zhijiang, LIU Jianping, and HUANG Jing. Dynamic range compression and pseudo-color presentation based on Retinex for SAR images[C]. 2008 Internation73al Conference on Computer Science and Software Engineering, Wuhan, China, 2008: 257–260. doi: 10.1109/CSSE.2008.1459.
    [15]
    UHLMANN S and KIRANYAZ S. Integrating color features in polarimetric SAR image classification[J]. IEEE Transactions on Geoscience and Remote Sensing, 2014, 52(4): 2197–2216. doi: 10.1109/TGRS.2013.2258675.
    [16]
    DENG Qiming, CHEN Yilun, ZHANG Weijie, et al. Colorization for polarimetric SAR image based on scattering mechanisms[C]. 2008 Congress on Image and Signal Processing, Sanya, China, 2008: 697–701. doi: 10.1109/CISP.2008.366.
    [17]
    LI Zewen, LIU Fan, YANG Wenjie, et al. A survey of convolutional neural networks: Analysis, applications, and prospects[J]. IEEE Transactions on Neural Networks and Learning Systems, 2022, 33(12): 6999–7019. doi: 10.1109/TNNLS.2021.3084827.
    [18]
    GOODFELLOW I, POUGET-ABADIE J, MIRZA M, et al. Generative adversarial networks[J]. Communications of the ACM, 2020, 63(11): 139–144. doi: 10.1145/3422622.
    [19]
    MIRZA M and OSINDERO S. Conditional generative adversarial nets[EB/OL]. https://arxiv.org/abs/1411.1784, 2014.
    [20]
    ZHU Junyan, PARK T, ISOLA P, et al. Unpaired image-to-image translation using cycle-consistent adversarial networks[C]. IEEE International Conference on Computer Vision, Venice, Italy, 2017: 2242–2251. doi: 10.1109/ICCV.2017.244.
    [21]
    HO J, JAIN A, and ABBEEL P. Denoising diffusion probabilistic models[C]. 34th International Conference on Neural Information Processing Systems, Vancouver, Canada, 2020: 574.
    [22]
    EVANS D L, FARR T G, VAN ZYL J J, et al. Radar polarimetry: Analysis tools and applications[J]. IEEE Transactions on Geoscience and Remote Sensing, 1988, 26(6): 774–789. doi: 10.1109/36.7709.
    [23]
    JORDAN R L, HUNEYCUTT B L, and WERNER M. The SIR-C/X-SAR synthetic aperture radar system[J]. Proceedings of the IEEE, 1991, 79(6): 827–838. doi: 10.1109/5.90161.
    [24]
    VAN ZYL J J. Unsupervised classification of scattering behavior using radar polarimetry data[J]. IEEE Transactions on Geoscience and Remote Sensing, 1989, 27(1): 36–45. doi: 10.1109/36.20273.
    [25]
    CLOUDE S R and POTTIER E. An entropy based classification scheme for land applications of polarimetric SAR[J]. IEEE Transactions on Geoscience and Remote Sensing, 1997, 35(1): 68–78. doi: 10.1109/36.551935.
    [26]
    POTTIER E, LEE J S, and FERRO-FAMIL L. Advanced concepts in polarimetry—Part 1 (Polarimetric target description, speckle filtering and decomposition theorems)[R]. RTO-EN-SET-081, 2005.
    [27]
    SWAIN M J and BALLARD D H. Color indexing[J]. International Journal of Computer Vision, 1991, 7(1): 11–32. doi: 10.1007/BF00130487.
    [28]
    MANJUNATH B S, OHM J R, VASUDEVAN V V, et al. Color and texture descriptors[J]. IEEE Transactions on Circuits and Systems for Video Technology, 2001, 11(6): 703–715. doi: 10.1109/76.927424.
    [29]
    LIU Hongying, WANG Shuang, HOU Biao, et al. Unsupervised classification of polarimetric SAR images integrating color features[C]. 2014 IEEE Geoscience and Remote Sensing Symposium, Quebec City, Canada, 2014: 2762–2765. doi: 10.1109/IGARSS.2014.6947048.
    [30]
    Liu P, Wang L, Ranjan R, et al. A survey on active deep learning: From model driven to data driven[J]. ACM Computing Surveys (CSUR), 2022, 54(10s): 1–34. doi: 10.1145/3510414.
    [31]
    LI Yu, FU Randi, MENG Xiangchao, et al. A SAR-to-optical image translation method based on conditional generation adversarial network (cGAN)[J]. IEEE Access, 2020, 8: 60338–60343. doi: 10.1109/ACCESS.2020.2977103.
    [32]
    ZUO Zongcheng and LI Yuanxiang. A SAR-to-optical image translation method based on PIX2PIX[C]. 2021 IEEE International Geoscience and Remote Sensing Symposium IGARSS, Brussels, Belgium, 2021: 3026–3029. doi: 10.1109/IGARSS47720.2021.9555111.
    [33]
    HU Pengcheng, WANG Yong, LIU Yifan, et al. Improved cGAN for SAR-to-optical image translation[C]. 2024 43rd Chinese Control Conference (CCC), Kunming, China, 2024: 7675–7680. doi: 10.23919/CCC63176.2024.10661422.
    [34]
    CHENG Fujian, KANG Yashu, CHEN Chunlei, et al. A strategy optimized Pix2pix approach for SAR-to-optical image translation task[EB/OL]. https://arxiv.org/abs/2206.13042, 2022.
    [35]
    YU Tianzhu, ZHANG Jiexin, and ZHOU Jianjiang. Conditional GAN with effective attention for SAR-to-optical image translation[C]. 2021 3rd International Conference on Advances in Computer Technology, Information Science and Communication (CTISC), Shanghai, China, 2021: 7–11. doi: 10.1109/CTISC52352.2021.00009.
    [36]
    TURNES J N, CASTRO J D B, TORRES D L, et al. Atrous cGAN for SAR to optical image translation[J]. IEEE Geoscience and Remote Sensing Letters, 2022, 19: 4003905. doi: 10.1109/IGARSS.2018.8518719.
    [37]
    CHEN L C, PAPANDREOU G, SCHROFF F, et al. Rethinking atrous convolution for semantic image segmentation[EB/OL].https://arxiv.org/abs/1706.05587, 2017.
    [38]
    LUO Jinjin, YANG Xuezhi, LIANG Hongbo, et al. Contextual inference feature extraction approach based on generative adversarial network for SAR-to-optical image translation[C]. IGARSS 2024–2024 IEEE International Geoscience and Remote Sensing Symposium, Athens, Greece, 2024: 7334–7337. doi: 10.1109/IGARSS53475.2024.10642046.
    [39]
    YANG Xi, ZHAO Jingyi, WEI Ziyu, et al. SAR-to-optical image translation based on improved CGAN[J]. Pattern Recognition, 2022, 121: 108208. doi: 10.1016/j.patcog.2021.108208.
    [40]
    WEI Juan, ZOU Huanxin, SUN Li, et al. CFRWD-GAN for SAR-to-optical image translation[J]. Remote Sensing, 2023, 15(10): 2547. doi: 10.3390/rs15102547.
    [41]
    ZHAN Tao, BIAN Jiarong, YANG Jing, et al. Improved conditional generative adversarial networks for SAR-to-optical image translation[C]. 6th Chinese Conference on Pattern Recognition and Computer Vision (PRCV), Xiamen, China, 2023: 279–291. doi: 10.1007/978-981-99-8462-6_23.
    [42]
    KONG Yingying, LIU Siyuan, and PENG Xiangyang. Multi-Scale translation method from SAR to optical remote sensing images based on conditional generative adversarial network[J]. International Journal of Remote Sensing, 2022, 43(8): 2837–2860. doi: 10.1080/01431161.2022.2072179.
    [43]
    GUO Zhe, GUO Haojie, LIU Xuewen, et al. Sar2color: Learning imaging characteristics of SAR images for SAR-to-optical transformation[J]. Remote Sensing, 2022, 14(15): 3740. doi: 10.3390/rs14153740.
    [44]
    WEI Juan, ZOU Huanxin, SUN Li, et al. Generative adversarial network for SAR-to-optical image translation with feature cross-fusion inference[C]. IGARSS 2022–2022 IEEE International Geoscience and Remote Sensing Symposium, Kuala Lumpur, Malaysia, 2022: 6025–6028. doi: 10.1109/IGARSS46834.2022.9884166.
    [45]
    LUO Qingli, LI Hong, CHEN Zhiyuan, et al. Add-UNet: An adjacent dual-decoder UNET for SAR-to-optical translation[J]. Remote Sensing, 2023, 15(12): 3125. doi: 10.3390/rs15123125.
    [46]
    YU Ning, MA Ailong, ZHONG Yanfei, et al. HFGAN: A heterogeneous fusion generative adversarial network for SAR-to-optical image translation[C]. IGARSS 2022–2022 IEEE International Geoscience and Remote Sensing Symposium, Kuala Lumpur, Malaysia, 2022: 2864–2867. doi: 10.1109/IGARSS46834.2022.9883519.
    [47]
    FENG Chenguo, LIU Yang, WANG Nan, et al. INR-ECGAN: An enhanced conditional GAN with Implicit neural representation for SAR-to-optical image translation[C]. 2024 China Automation Congress (CAC), Qingdao, China, 2024: 4358–4363. doi: 10.1109/CAC63892.2024.10864577.
    [48]
    CHOUHAN A, JINDAL N, SUR A, et al. EDCGAN: Encoder decoder based conditional GAN for SAR to optical image translation[C]. The 13th Indian Conference on Computer Vision, Graphics and Image Processing, Gandhinagar, India, 2022: 38. doi: 10.1145/3571600.3571639.
    [49]
    DOSOVITSKIY A, BEYER L, KOLESNIKOV A, et al. An image is worth 16x16 words: Transformers for image recognition at scale[J]. arXiv preprint arXiv:2010.11929, 2020. https://doi.org/10.48550/arXiv.2010.11929.
    [50]
    ZHAO Wenbo, JIANG Nana, LIAO Xiaoxin, et al. HVT-cGAN: Hybrid vision transformer cGAN for SAR-to-optical image translation[J]. IEEE Transactions on Geoscience and Remote Sensing, 2025, 63: 5202017. doi: 10.1109/TGRS.2024.3523040.
    [51]
    WANG Zhaobin, MA Yikun, and ZHANG Yaonan. Hybrid cGAN: Coupling global and local features for SAR-to-optical image translation[J]. IEEE Transactions on Geoscience and Remote Sensing, 2022, 60: 5236016. doi: 10.1109/TGRS.2022.3212208.
    [52]
    PARK S, LEE H, and LEE S. SAR-to-optical image translation using vision transformer-based CGAN[J]. IEEE Sensors Journal, 2025, 25(10): 18503–18514. doi: 10.1109/JSEN.2025.3555933.
    [53]
    DOI K, SAKURADA K, ONISHI M, et al. GAN-based SAR-to-optical image translation with region information[C]. IGARSS 2020–2020 IEEE International Geoscience and Remote Sensing Symposium, Waikoloa, USA, 2020: 2069–2072. doi: 10.1109/IGARSS39084.2020.9323085.
    [54]
    FU Xuanchao, KOUYAMA T, SEKI S, et al. Advanced SAR-to-optical image translation techniques using Jaxa's high-resolution land-use and land-cover map[C]. IGARSS 2024–2024 IEEE International Geoscience and Remote Sensing Symposium, Athens, Greece, 2024: 7367–7370. doi: 10.1109/IGARSS53475.2024.10641911.
    [55]
    LIANG Hongbo, YANG Xuezhi, YANG Xiangyu, et al. GFTT: Geographical feature tokenization transformer for SAR-to-optical image translation[J]. IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, 2025, 18: 2975–2989. doi: 10.1109/JSTARS.2024.3523274.
    [56]
    XIONG Quan, LI Guoqing, YAO Xiaochuang, et al. SAR-to-optical image translation and cloud removal based on conditional generative adversarial networks: Literature survey, taxonomy, evaluation indicators, limits and future directions[J]. Remote Sensing, 2023, 15(4): 1137. doi: 10.3390/rs15041137.
    [57]
    WANG Lei, XU Xin, YU Yue, et al. SAR-to-optical image translation using supervised cycle-consistent adversarial networks[J]. IEEE Access, 2019, 7: 129136–129149. doi: 10.1109/ACCESS.2019.2939649.
    [58]
    HWANG J, YU Chushi, and SHIN Y. SAR-to-optical image translation using SSIM and perceptual loss based cycle-consistent GAN[C]. 2020 International Conference on Information and Communication Technology Convergence (ICTC), Jeju, Korea (South), 2020: 191–194. doi: 10.1109/ICTC49870.2020.9289381.
    [59]
    SIDDIQUI A A, K S P, and AMBASTHA S. Converting SAR images to optical images using CycleGAN[C]. 2024 2nd International Conference on Emerging Trends in Information Technology and Engineering (ICETITE), Vellore, India, 2024: 1–5. doi: 10.1109/ic-ETITE58242.2024.10493784.
    [60]
    SHI Hao, ZHANG Bocheng, WANG Yupei, et al. SAR-to-optical image translating through generate-validate adversarial networks[J]. IEEE Geoscience and Remote Sensing Letters, 2022, 19: 4506905. doi: 10.1109/LGRS.2022.3168391.
    [61]
    CHEN Yongkang, ZHU Zuguo, HUANG Yan, et al. MSF: A multi-scale fusion generative adversarial network for SAR-to-optical image translation[C]. IGARSS 2024–2024 IEEE International Geoscience and Remote Sensing Symposium, Athens, Greece, 2024: 9058–9061. doi: 10.1109/IGARSS53475.2024.10640486.
    [62]
    FU Shilei, XU Feng, and JIN Yaqiu. Reciprocal translation between SAR and optical remote sensing images with cascaded-residual adversarial networks[J]. Science China Information Sciences, 2021, 64(2): 122301. doi: 10.1007/s11432-020-3077-5.
    [63]
    HWANG J and SHIN Y. SAR-to-optical image translation using SSIM loss based unpaired GAN[C]. 2022 13th International Conference on Information and Communication Technology Convergence (ICTC), Jeju Island, Korea, Republic of, 2022: 917–920. doi: 10.1109/ICTC55196.2022.9952649.
    [64]
    LI Huihui, GU Cang, WU Dongqing, et al. Multiscale generative adversarial network based on wavelet feature learning for SAR-to-optical image translation[J]. IEEE Transactions on Geoscience and Remote Sensing, 2022, 60: 5236115. doi: 10.1109/TGRS.2022.3211415.
    [65]
    YANG Xing, WANG Zihan, ZHAO Jingyi, et al. FG-GAN: A fine-grained generative adversarial network for unsupervised SAR-to-optical image translation[J]. IEEE Transactions on Geoscience and Remote Sensing, 2022, 60: 5621211. doi: 10.1109/TGRS.2022.3165371.
    [66]
    GUO Zhe, ZHANG Zhibo, CAI Qinglin, et al. MS-GAN: Learn to memorize scene for unpaired SAR-to-optical image translation[J]. IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, 2024, 17: 11467–11484. doi: 10.1109/JSTARS.2024.3411691.
    [67]
    LIU Yang, HAN Qingcen, YANG Hong, et al. High-resolution SAR-to-multispectral image translation based on S2MS-GAN[J]. Remote Sensing, 2024, 16(21): 4045. doi: 10.3390/rs16214045.
    [68]
    ZHANG Mingjin, ZHANG Peng, ZHANG Yuhan, et al. SAR-to-optical image translation via an interpretable network[J]. Remote Sensing, 2024, 16(2): 242. doi: 10.3390/rs16020242.
    [69]
    ZHANG Mingjin, HE Chengyu, ZHANG Jing, et al. SAR-to-optical image translation via neural partial differential equations[C]. The 31st International Joint Conference on Artificial Intelligence, Vienna, Austria, 2022: 1644–1650.
    [70]
    ZHANG Hannuo, LI Huihui, LIN Jiarui, et al. Seg-CycleGAN: SAR-to-optical image translation guided by a downstream task[J]. IEEE Geoscience and Remote Sensing Letters, 2025, 22: 4005205. doi: 10.1109/LGRS.2025.3538868.
    [71]
    GUO Jie, HE Chengyu, ZHANG Mingjin, et al. Edge-preserving convolutional generative adversarial networks for SAR-to-optical image translation[J]. Remote Sensing, 2021, 13(18): 3575. doi: 10.3390/rs13183575.
    [72]
    ZHANG Jiexin, ZHOU Jianjiang, and LU Xiwen. Feature-guided SAR-to-optical image translation[J]. IEEE Access, 2020, 8: 70925–70937. doi: 10.1109/ACCESS.2020.2987105.
    [73]
    LUO Yi and PI Dechang. SAR-to-optical image translation for quality enhancement[J]. Journal of Ambient Intelligence and Humanized Computing, 2023, 14(8): 9985–10000. doi: 10.1007/s12652-021-03665-0.
    [74]
    BRALET A, ATTO A M, CHANUSSOT J, et al. Deep learning of radiometrical and geometrical SAR distorsions for image modality translations[C]. 2022 IEEE International Conference on Image Processing (ICIP), Bordeaux, France, 2022: 1766–1770. doi: 10.1109/ICIP46576.2022.9897713.
    [75]
    YAN Weidong, YAN Pei, and CAO Li. Domain transfer net based on U-Net and transformer for synthetic aperture radar-to-optical image translation[J]. Journal of Applied Remote Sensing, 2024, 18(4): 046509. doi: 10.1117/1.JRS.18.046509.
    [76]
    WANG Haixia, ZHANG Zhigang, HU Zhanyi, et al. SAR-to-optical image translation with hierarchical latent features[J]. IEEE Transactions on Geoscience and Remote Sensing, 2022, 60: 5233812. doi: 10.1109/TGRS.2022.3200996.
    [77]
    PAN Yue, KHAN I A, and MENG Han. SAR-to-optical image translation using multi-stream deep ResCNN of information reconstruction[J]. Expert Systems with Applications, 2023, 224: 120040. doi: 10.1016/j.eswa.2023.120040.
    [78]
    HE Jingfei, MAO Yongfei, ZHAO Liangbo, et al. A dual network approach with enhanced feature guidance for SAR-to-optical image translation[C]. 2024 IEEE International Conference on Signal, Information and Data Processing (ICSIDP), Zhuhai, China, 2024: 1–6. doi: 10.1109/ICSIDP62679.2024.10868756.
    [79]
    ZHOU Hao, XIAO Xiao, and LU Yilong. Multi-modality translation from SAR image to optical image for remote sensing applications[C]. 2021 7th Asia-Pacific Conference on Synthetic Aperture Radar (APSAR), Bali, Indonesia, 2021: 1–4. doi: 10.1109/APSAR52370.2021.9688420.
    [80]
    DU Wenliang, ZHOU Yong, ZHU Hancheng, et al. A semi-supervised image-to-image translation framework for SAR-optical image matching[J]. IEEE Geoscience and Remote Sensing Letters, 2022, 19: 4516305. doi: 10.1109/LGRS.2022.3223353.
    [81]
    WANG Jinyu, YANG Haitao, LIU Zhengjun, et al. A semi-supervised method using cycle consistency and multi-perspective dilated for SAR-to-Optical translation[J]. iScience, 2025, 28(5): 112401. doi: 10.1016/j.isci.2025.112401.
    [82]
    GUO Zhe, LUO Rui, CAI Qinglin, et al. Scene-embedded generative adversarial networks for semi-supervised SAR-to-optical image translation[J]. IEEE Geoscience and Remote Sensing Letters, 2024, 21: 4018005. doi: 10.1109/LGRS.2024.3471553.
    [83]
    ZHAO Jiaqi, NI Wenxin, ZHOU Yong, et al. SAR-to-optical image translation by a variational generative adversarial network[J]. Remote Sensing Letters, 2022, 13(7): 672–682. doi: 10.1080/2150704X.2022.2068986.
    [84]
    BRALET A, NGO T N, TROUVÉ E, et al. DEM-assisted neural network for SAR-to-optical image translation[C]. IGARSS 2024–2024 IEEE International Geoscience and Remote Sensing Symposium, Athens, Greece, 2024: 7649–7652. doi: 10.1109/IGARSS53475.2024.10641788.
    [85]
    BAI Xinyu, PU Xinyang, and XU Feng. Conditional diffusion for SAR to optical image translation[J]. IEEE Geoscience and Remote Sensing Letters, 2024, 21: 4000605. doi: 10.1109/LGRS.2023.3337143.
    [86]
    SHI Hao, CUI Zihan, CHEN Liang, et al. A brain-inspired approach for SAR-to-optical image translation based on diffusion models[J]. Frontiers in Neuroscience, 2024, 18: 1352841. doi: 10.3389/fnins.2024.1352841.
    [87]
    BAI Xinyu and XU Feng. SAR to optical image translation with color supervised diffusion model[C]. IGARSS 2024–2024 IEEE International Geoscience and Remote Sensing Symposium, Athens, Greece, 2024: 963–966. doi: 10.1109/IGARSS53475.2024.10640647.
    [88]
    QIN Jiang, ZOU Bin, ZHANG Lamei, et al. SAR-to-optical image translation using conditional denoising diffusion probabilistic models[C]. 2024 International Radar Conference (RADAR), Rennes, France, 2024: 1–5. doi: 10.1109/RADAR58436.2024.10993599.
    [89]
    QIN Jiang, WANG Kai, ZOU Bin, et al. Conditional diffusion model with spatial-frequency refinement for SAR-to-optical image translation[J]. IEEE Transactions on Geoscience and Remote Sensing, 2024, 62: 5226914. doi: 10.1109/TGRS.2024.3491826.
    [90]
    YOU Ruixi, JIA Hecheng, and XU Feng. Unpaired object-level SAR-to-optical image translation for aircraft with keypoints-guided diffusion models[EB/OL]. https://arxiv.org/abs/2503.19798, 2025.
    [91]
    AYDIN K, HANNA J, and BORTH D. SAR-to-RGB translation with latent diffusion for earth observation[EB/OL]. https://arxiv.org/abs/2504.11154, 2025.
    [92]
    GOU Shuiping, WANG Xin, WANG Xinlin, et al. Interpretable matching of optical-SAR image via dynamically conditioned diffusion models[C]. The 32nd ACM International Conference on Multimedia, Melbourne, Australia, 2024: 4358–4367. doi: 10.1145/3664647.3681700.
    [93]
    BAI Xinyu and XU Feng. Accelerating diffusion for SAR-to-optical image translation via adversarial consistency distillation[EB/OL]. https://arxiv.org/abs/2407.06095, 2024.
    [94]
    QIN Jiang, ZOU Bin, LI Haolin, et al. Efficient end-to-end diffusion model for one-step SAR-to-optical translation[J]. IEEE Geoscience and Remote Sensing Letters, 2025, 22: 4000405. doi: 10.1109/LGRS.2024.3506566.
    [95]
    LI Bo, XUE Kaitao, LIU Bin, et al. BBDM: Image-to-image translation with Brownian bridge diffusion models[C]. IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, Canada, 2023: 1952–1961. doi: 10.1109/CVPR52729.2023.00194.
    [96]
    GUO Zhe, LIU Jiayi, CAI Qinglin, et al. Learning SAR-to-optical image translation via diffusion models with color memory[J]. IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, 2024, 17: 14454–14470. doi: 10.1109/JSTARS.2024.3439516.
    [97]
    KIM S H and CHUNG D. Conditional brownian bridge diffusion model for VHR SAR to optical image translation[J]. IEEE Geoscience and Remote Sensing Letters, 2025, 22: 4008305. doi: 10.1109/LGRS.2025.3562401.
    [98]
    王雪松, 陈思伟. 合成孔径雷达极化成像解译识别技术的进展与展望[J]. 雷达学报, 2020, 9(2): 259–276. doi: 10.12000/JR19109.

    WANG Xuesong and CHEN Siwei. Polarimetric synthetic aperture radar interpretation and recognition: Advances and perspectives[J]. Journal of Radars, 2020, 9(2): 259–276. doi: 10.12000/JR19109.
    [99]
    SCHMITT M, HUGHES L H, KÖRNER M, et al. Colorizing sentinel-1 SAR images using a variational autoencoder conditioned on sentinel-2 imagery[C]. The International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences, Riva del Garda, Italy, 2018: 1045–1051. doi: 10.5194/isprs-archives-XLII-2-1045-2018.
    [100]
    SHEN Kangqing, VIVONE G, YANG Xiaoyuan, et al. A benchmarking protocol for SAR colorization: From regression to deep learning approaches[J]. Neural Networks, 2024, 169: 698–712. doi: 10.1016/j.neunet.2023.10.058.
    [101]
    SHEN Kangqing, VIVONE G, LOLLI S, et al. IcGAN4CoLSAR: A novel multispectral conditional generative adversarial network approach for SAR image colorization[J]. IEEE Geoscience and Remote Sensing Letters, 2025, 22: 4004605. doi: 10.1109/LGRS.2025.3534795.
    [102]
    GAO Jianhao, YUAN Qiangqiang, LI Jie, et al. Cloud removal with fusion of high resolution optical and SAR images using generative adversarial networks[J]. Remote Sensing, 2020, 12(1): 191. doi: 10.3390/rs12010191.
    [103]
    XIANG Xuanyu, TAN Yihua, and YAN Longfei. Cloud-guided fusion with SAR-to-optical translation for thick cloud removal[J]. IEEE Transactions on Geoscience and Remote Sensing, 2024, 62: 5633715. doi: 10.1109/TGRS.2024.3431556.
    [104]
    WANG Peng, CHEN Yongkang, HUANG Bo, et al. MT_GAN: A SAR-to-optical image translation method for cloud removal[J]. ISPRS Journal of Photogrammetry and Remote Sensing, 2025, 225: 180–195. doi: 10.1016/j.isprsjprs.2025.04.011.
    [105]
    CHRISTOVAM L E, SHIMABUKURO M H, GALO M L B T, et al. Evaluation of SAR to optical image translation using conditional generative adversarial network for cloud removal in a crop dataset[C]. The International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences, Nice, France, 2021: 823–828. doi: 10.5194/isprs-archives-XLIII-B3-2021-823-2021.
    [106]
    BERMUDEZ J D, HAPP P N, OLIVEIRA D A B, et al. SAR to optical image synthesis for cloud removal with generative adversarial networks[C]. ISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences, Karlsruhe, Germany, 2018: 5–11. doi: 10.5194/isprs-annals-IV-1-5-2018.
    [107]
    LIU Rui, MENG Siqi, PENG Yini, et al. TransFusion-CR: Two-phase SAR-to-optical translation and deep feature fusion for cloud removal[J]. IEEE Transactions on Geoscience and Remote Sensing, 2024, 62: 5635911. doi: 10.1109/TGRS.2024.3439854.
    [108]
    DARBAGHSHAHI F N, MOHAMMADI M R, and SORYANI M. Cloud removal in remote sensing images using generative adversarial networks and SAR-to-optical image translation[J]. IEEE Transactions on Geoscience and Remote Sensing, 2022, 60: 4105309. doi: 10.1109/TGRS.2021.3131035.
    [109]
    HAN Zhen, LV Ning, SU Tao, et al. A structure consistency generative adversarial network for SAR-optical image registration[C]. 2024 IEEE International Conference on Smart Internet of Things (SmartIoT), Shenzhen, China, 2024: 197–203. doi: 10.1109/SmartIoT62235.2024.00037.
    [110]
    NIE Han, FU Zhitao, TANG Bohui, et al. A dual-generator translation network fusing texture and structure features for SAR and optical image matching[J]. Remote Sensing, 2022, 14(12): 2946. doi: 10.3390/rs14122946.
    [111]
    TORIYA H, DEWAN A, and KITAHARA I. SAR2OPT: Image alignment between multi-modal images using generative adversarial networks[C]. IGARSS 2019–2019 IEEE International Geoscience and Remote Sensing Symposium, Yokohama, Japan, 2019: 923–926. doi: 10.1109/IGARSS.2019.8898605.
    [112]
    JAMES L, NIDAMANURI R R, KRISHNAN S M, et al. A novel approach for SAR to optical image registration using deep learning[C]. 2023 International Conference on Machine Intelligence for GeoAnalytics and Remote Sensing (MIGARS), Hyderabad, India, 2023: 1–4. doi: 10.1109/MIGARS57353.2023.10064578.
    [113]
    HUANG Xuejun, WEN Liwu, and DING Jinshan. SAR and optical image registration method based on improved CycleGAN[C]. 2019 6th Asia-Pacific Conference on Synthetic Aperture Radar (APSAR), Xiamen, China, 2019: 1–6. doi: 10.1109/APSAR46974.2019.9048448.
    [114]
    HONG Yumeng, PAN Jun, XU Jiangong, et al. SMAF-net: Semantics-guided modality transfer and hierarchical feature fusion for optical-SAR image registration[J]. International Journal of Applied Earth Observation and Geoinformation, 2025, 143: 104827. doi: 10.1016/j.jag.2025.104827.
    [115]
    SONOBE R, TANI H, ZADI M, et al. Improving crop classification accuracy using Sentinel-1 C-SAR data and GAN-generated optical images[J]. Journal of Applied Remote Sensing, 2025, 19(2): 024505. doi: 10.1117/1.JRS.19.024505.
    [116]
    ZHANG Jiexin, ZHOU Jianjiang, LI Minglei, et al. Quality assessment of SAR-to-optical image translation[J]. Remote Sensing, 2020, 12(21): 3472. doi: 10.3390/rs12213472.
    [117]
    KWAK G H and PARK N W. Assessing the potential of multi-temporal conditional generative adversarial networks in SAR-to-optical image translation for early-stage crop monitoring[J]. Remote Sensing, 2024, 16(7): 1199. doi: 10.3390/rs16071199.
    [118]
    HU Xikun, ZHANG Puzhao, and BAN Yifang. Gan-based SAR to optical image translation in fire-disturbed regions[C]. IGARSS 2022–2022 IEEE International Geoscience and Remote Sensing Symposium, Kuala Lumpur, Malaysia, 2022: 1456–1459. doi: 10.1109/IGARSS46834.2022.9884234.
    [119]
    HU Xikun, ZHANG Puzhao, BAN Yifang, et al. GAN-based SAR and optical image translation for wildfire impact assessment using multi-source remote sensing data[J]. Remote Sensing of Environment, 2023, 289: 113522. doi: 10.1016/j.rse.2023.113522.
    [120]
    LEY A, DHONDT O, VALADE S, et al. Exploiting GAN-based SAR to optical image transcoding for improved classification via deep learning[C]. 12th European Conference on Synthetic Aperture Radar, Aachen, Germany, 2018: 1–6.
    [121]
    AMITRANO D. Multitemporal SAR-to-optical image translation using Pix2Pix with application to vegetation monitoring[J]. IEEE Access, 2024, 12: 124402–124413. doi: 10.1109/ACCESS.2024.3454513.
    [122]
    ZHANG Qian, LIU Xiangnan, LIU Meiling, et al. Comparative analysis of edge information and polarization on SAR-to-optical translation based on conditional generative adversarial networks[J]. Remote Sensing, 2021, 13(1): 128. doi: 10.3390/rs13010128.
    [123]
    BRUNE E and BAN Yifang. SAR-to-optical translation using conditional diffusion models for wildfire-burned area segmentation[C]. IGARSS 2024–2024 IEEE International Geoscience and Remote Sensing Symposium, Athens, Greece, 2024: 7011–7015. doi: 10.1109/IGARSS53475.2024.10641072.
    [124]
    LEE I H and PARK C G. SAR-to-virtual optical image translation for improving SAR automatic target recognition[J]. IEEE Geoscience and Remote Sensing Letters, 2023, 20: 4010105. doi: 10.1109/LGRS.2023.3312140.
    [125]
    SUN Yuchuang, YAN Kaijia, and LI Wangzhe. CycleGAN-based SAR-optical image fusion for target recognition[J]. Remote Sensing, 2023, 15(23): 5569. doi: 10.3390/rs15235569.
    [126]
    SUN Yuchuang, JIANG Wen, YANG Jiyao, et al. SAR target recognition using cGAN-based SAR-to-optical image translation[J]. Remote Sensing, 2022, 14(8): 1793. doi: 10.3390/rs14081793.
    [127]
    SCHMITT M, HUGHES L H, and ZHU X X. The SEN1-2 dataset for deep learning in SAR-optical data fusion[C]. ISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences, Karlsruhe, Germany, 2018: 141–146. doi: 10.5194/isprs-annals-IV-1-141-2018.
    [128]
    SUMBUL G, DE WALL A, KREUZIGER T, et al. BigEarthNet-MM: A large-scale, multimodal, multilabel benchmark archive for remote sensing image classification and retrieval [software and data sets][J]. IEEE Geoscience and Remote Sensing Magazine, 2021, 9(3): 174–180. doi: 10.1109/MGRS.2021.3089174.
    [129]
    LI Xue, ZHANG Guo, CUI Hao, et al. MCANet: A joint semantic segmentation framework of optical and SAR images for land use classification[J]. International Journal of Applied Earth Observation and Geoinformation, 2022, 106: 102638. doi: 10.1016/j.jag.2021.102638.
    [130]
    SCHMITT M, HUGHES L H, QIU Chunping, et al. SEN12MS--A curated dataset of georeferenced multi-spectral sentinel-1/2 imagery for deep learning and data fusion[EB/OL]. https://arxiv.org/abs/1906.07789, 2019.
    [131]
    HUANG Meiyu, XU Yao, QIAN Lixin, et al. The QXS-SAROPT dataset for deep learning in SAR-optical data fusion[EB/OL]. https://arxiv.org/abs/2103.08259, 2021.
    [132]
    ZHAO Yitao, CELIK T, LIU Nanqing, et al. A comparative analysis of GAN-based methods for SAR-to-optical image translation[J]. IEEE Geoscience and Remote Sensing Letters, 2022, 19: 3512605. doi: 10.1109/LGRS.2022.3177001.
    [133]
    SHERMEYER J, HOGAN D, BROWN J, et al. SpaceNet 6: Multi-sensor all weather mapping dataset[C]. IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, USA, 2020: 768–777. doi: 10.1109/CVPRW50498.2020.00106.
    [134]
    WANG Yuanyuan and ZHU Xiaoxiang. The SARptical dataset for joint analysis of SAR and optical image in dense urban area[C]. IGARSS 2018–2018 IEEE International Geoscience and Remote Sensing Symposium, Valencia, Spain, 2018: 6840–6843. doi: 10.1109/IGARSS.2018.8518298.
    [135]
    ENOMOTO K, SAKURADA K, WANG Weiming, et al. Image translation between SAR and optical imagery with generative adversarial nets[C]. IGARSS 2018–2018 IEEE International Geoscience and Remote Sensing Symposium, Valencia, Spain, 2018: 1752–1755.
    [136]
    KIRILLOV A, MINTUN E, RAVI N, et al. Segment anything[C]. IEEE/CVF International Conference on Computer Vision, Paris, France, 2023: 3992–4003. doi: 10.1109/ICCV51070.2023.00371.
    [137]
    RADFORD A, KIM J W, HALLACY C, et al. Learning transferable visual models from natural language supervision[C]. 38th International Conference on Machine Learning. PmLR, Virtual Event,2021: 8748–8763.
  • 加载中

Catalog

    通讯作者: 陈斌, bchen63@163.com
    • 1. 

      沈阳化工大学材料科学与工程学院 沈阳 110142

    1. 本站搜索
    2. 百度学术搜索
    3. 万方数据库搜索
    4. CNKI搜索
    Article views(75) PDF downloads(15) Cited by()
    Proportional views
    Related

    /

    DownLoad:  Full-Size Img  PowerPoint