Adversarial Optical‑Prior Transfer with Multiresolution Anisotropic Decoding for Coherence-Structured SAR Segmentation
🌟 DDSD-GANet: Adversarial Optical-Prior Transfer with Multiresolution Anisotropic Decoding for Coherence-Structured SAR Segmentation This project implements the Dual-Driven Spectral-Directional Cross-Modal Generative Adversarial Network (DDSD-GANet) proposed in the paper "Adversarial Optical-Prior Transfer with Multiresolution Anisotropic Decoding for Coherence-Structured SAR Segmentation".
DDSD-GANet is an advanced framework for Synthetic Aperture Radar (SAR) image semantic segmentation. It is designed to address the challenges inherent in SAR imagery, specifically multiplicative speckle noise, absence of optical-domain semantic priors (geometric semantic deficiency), and intrinsic scale inconsistency. By incorporating adversarial cross-modal learning and an innovative decoder structure , DDSD-GANet achieves significant performance improvements in the structured segmentation of SAR targets, such as buildings, vessels, and sea-land boundaries.
The Dual-Driven Spectral-Directional Cross-Modal Generative Adversarial Network is proposed to address the challenges of coherence-induced noise, scale inconsistency, and semantic deficiency in SAR imagery.Subfigures: (a) overall framework; (b) segmentation encoder; (c) segmentation decoder with MACU and CAG; (d) ResNet encoder; (e–f) SAR-to-optical GAN encoder/decoder; (g) Up Block for progressive upsampling; (h) OCF for cross-branch fusion and gating.
The proposed Multiresolution Anisotropic Coherence Unit couples AAPB with MRPB to capture directional structures while suppressing speckle noise.
The proposed Coherent-aware Gated Skip employs a dual-driven gating mechanism to suppress spurious patterns in shallow features and, under global semantic guidance, selectively pass true boundary information.
SARBuD:The dataset is available through \url{https://github.com/CAESAR-Radi/SARBuD}
HRSID:The dataset is available through \url{https://github.com/chaozhong2010/HRSID}
Sea-land Segmentation:The dataset is available through \url{https://radars.ac.cn/web/data/getData?dataType=sea-landsegmentation}
For fair comparison, all methods share the same data preprocessing and training settings. Input images are normalized and resized to a fixed resolution, then augmented with random cropping, intensity jitter, and simple affine transformations to increase data diversity. We train all networks for 100 epochs with a batch size of 2 using the AdamW optimizer with a learning rate of 1e-4. Experiments are implemented in PyTorch on a workstation equipped with NVIDIA RTX 3090 GPUs.
In evaluating network performance, a suite of distinct metrics was utilized, including the 95th Percentile Hausdorff Distance, Dice Coefficient, Intersection over Union, Cohen’s Kappa, and Matthews Correlation Coefficient.