Open Access
30 August 2023 Synthetic aperture radar and optical image registration using local and global feature learning by modality-shared attention network
Xin Hu, Yan Wu, Zhikang Li, Xiaoru Zhao, Xingyu Liu, Ming Li
Author Affiliations +
Abstract

The registration of synthetic aperture radar (SAR) and optical images is a meaningful but challenging multimodal task. Due to the large radiometric differences between SAR and optical images, it is difficult to obtain discriminative features only by mining local features in the traditional Siamese convolutional networks. We propose a modality-shared attention network (MSA-Net) that introduces nonlocal attention (NLA) to the partially shared two-stream network to jointly exploit local and global features. First, a modality-specific feature learning module is designed to efficiently extract shallow modality-specific features from SAR and optical images. Subsequently, a modality-shared feature learning (MShFL) module is designed to extract deep modality-shared features. The local feature extraction module and the NLA module in MShFL extract deep local and global features to enrich feature representations. Furthermore, a triplet loss function with a cross-modality similarity constraint is constructed to learn modality-shared feature representations, thereby reducing nonlinear radiometric differences between the two modalities. The MSA-Net is trained on a public SAR and optical dataset and tested on five pairs of SAR and optical images. In the registration results of five pairs of test SAR and optical images, the matching rate of the MSA-Net is 5% to 15% higher than that of other compared methods, and the matching errors of the matched inliers are on average reduced by about 0.28. Several ablation experiments verify the effectiveness of the partially shared network structure, the MShFL module, and the cross-modality similarity constraint.

1.

Introduction

Image registration is the process of matching two or more images of the same scene captured at different times, from different viewing angles, and using different sensors. This process achieves geometric alignment between the reference image and the image to be registered. Image registration is widely used in various fields, such as image fusion,13 image mosaic,4 image stitching,5,6 and multitemporal image change detection.7,8 With the rapid development of multisensor technology, many different modes of images can be obtained from the same scene. Synthetic aperture radar (SAR) and optical images are the two main ways of observing the earth. Because these two images have complementary information, they are widely used in multimodal remote image tasks.13,812 For these multimodal tasks to be accurate, the registration results of SAR and optical images are crucial.

However, nonlinear radiation differences between SAR and optical images and coherent speckles in SAR images make registration an unsolved and challenging research issue. Both traditional handcrafted methods1319 and deep learning-based ones have been used to address these issues.2028

The handcrafted image registration approaches mainly include intensity-based and feature-based methods. Due to the poor similarity of intensity-based descriptors, the former is less effective in multimodal image registration. The latter can deal with nonlinear radiation differences more effectively by mining the structural features of multimodal images to obtain descriptors. As a result, feature-based approaches are increasingly being used in image registration. The gradient-based descriptor algorithms, such as scale-invariant feature transform (SIFT),13 SAR-SIFT,15 and OS-SIFT,16 are applied to optical image registration, SAR image registration, and optical and SAR image registration, respectively. Nonetheless, gradient-based descriptors extract less valuable information from darker images. Compared with gradients, changes in light and radiation have no effect on phase congruency. The registration algorithms based on phase congruency, such as the histogram of phase congruency (HOPC),17 phase congruency structural descriptor (PCSD),18 and radiation-invariant feature transform (RIFT),19 have been successfully employed in SAR and optical image registration. Recently, Fan et al.29 proposed a new nonlinear diffusion-based Harris-Laplace detector and a new structural descriptor based on multiscale adaptive binning phase congruency that is more robust to geometric and radiometric differences, improving the number of matching points and reducing matching errors.

The handcrafted methods based on gradient and phase congruency mentioned above mine modality-shared features, i.e., structural features. The handcrafted methods use a fixed feature extraction pipeline to align a small number of images. They do not have the ability to handle the resolution, scale, and rotations that occur in the alignment of a large number of SAR and optical images. With its powerful data mining capabilities, deep learning has been widely used in various tasks involving remote sensing images. In recent years, many deep-learning registration methods have been proposed. For example, L2-Net,20 HardNet,21 and Siamese fully convolutional network (SFcNet)24 based on Siamese convolutional network have been proposed for single- or multimodal image matching or registration. MatchosNet26 and CNet27 based on pseudo-Siamese convolutional networks are proposed for SAR and optical image matching and registration.

The application of the above-mentioned Siamese or pseudo-Siamese convolutional networks to multimodal image registration has two problems. First, neither the Siamese nor pseudo-Siamese networks take into account both nonlinear radiation differences and mining modality-shared features. Since it is a two-stream network with completely shared parameters, the Siamese network ignores the radiation differences between SAR and optical images. The pseudo-Siamese network is a two-stream network that does not share parameters at all, which makes it unable to fully exploit the modality-shared features. Second, the local features extracted by convolutional networks are easily affected by radiation differences and noise, resulting in low similarity of extracted features.

For the first problem mentioned above, we propose a partially shared dual-stream convolutional neural network (CNN) to mine shared features that help improve registration accuracy. For the second problem mentioned above, we insert nonlocal attention (NLA) into the two-stream CNNs to extract both local and global features. Local features are limited due to receptive fields, which make it difficult to extract similar features from multimodal images. Global features have a larger range of receptive areas and are less affected by modal differences than local features. Therefore, considering global correlation can improve the robustness of features to a certain extent.

In this paper, we propose a modality-shared attention network (MSA-Net) to construct modality-shared feature descriptors for SAR and optical image registration. The main contributions of our method are summarized as follows.

  • (1) In MSA-Net, we built a partially shared network structure for mining the modality-specific and modality-shared features of SAR and optical images to overcome modal differences.

  • (2) In MSA-Net, a modality-specific feature learning (MSpFL) module is designed to extract shallow modality-specific features, and a modality-shared feature learning (MShFL) module is designed to extract deep modality-shared features. Specifically, the MShFL module consists of three local feature extraction (LFE) modules and three NLA modules, which can extract both local and global features.

  • (3) In MSA-Net, a triplet loss function with a cross-modality similarity constraint is designed to make the MSA-Net less susceptible to modal changes. The triplet loss can encourage the MSA-Net to learn shared features between SAR and optical images.

The reminder of this paper is organized as follows: Section 2 briefly introduces the related work of this paper. Section 3 elaborates on the proposed registration algorithm and the architecture of MSA-Net in detail. In Sec. 4, the experimental comparisons and analysis are performed on a public SAR and optical dataset. In addition, the comparisons between five test SAR and optical image pairs are carried out. Finally, Sec. 5 summarizes this paper.

2.

Related Work

2.1.

Deep Learning Registration Methods

Deep learning-based methods learn data-driven deep features through Siamese and pseudo-Siamese networks. L2Net20 and HardNet21 are methods applied to optical image matching. The similarity between the two is that they use a Siamese network with completely shared parameters to learn local feature descriptors. The difference is that L2-Net uses all positive and negative samples in the batch to optimize the network, whereas HardNet uses hard negative samples in a batch to complete the network optimization. Due to the successful application of HardNet in optical image matching, some papers22,24 used its improved Siamese convolutional network to extract deep features and realize multimodal image matching or registration. Bürgmann et al.22 input SAR and optical image patches of larger size into a modified Siamese network based on HardNet to obtain sparse feature descriptors and use the L2 distance of feature descriptors to obtain a similarity metric. Zhang et al.24 adopted the SFcNet to learn the similarity score between two different kinds of image patches.

However, the Siamese network ignores the radiometric differences between SAR and optical images and the scattered noise in SAR images. Hughes et al.23,24 explained that pseudo-Siamese networks are better suited for multimodal image matching. MatchosNet26 and CNet,27 based on pseudo-Siamese network structure, have achieved average results in SAR and optical image registration. MatchosNet is template matching, which can obtain better results only when there is no large translation or rotation transformation between the two images to be registered. However, the pseudo-Siamese network lacks the interaction between different modalities in the process of learning features, so it is more difficult to explore the shared features between different modalities. In addition, pseudo-Siamese networks increase the network parameters, which makes convergence more difficult. Therefore, we design a partially shared network to mine similar features in multimodal images while considering the differences between modalities.

2.2.

Nonlocal Attention

Due to the limited perceptual field of CNN, it is difficult to capture the global correlations of images. Thus it does not cope well with multiple transformations and nonlinear radiometric differences in multimodal image registration. Due to its sequence (image block) modeling structure, the transformer30 based on self-attention mechanism is particularly good at capturing global interactions between token embeddings. ViT31 has successfully applied the transformer30 from natural language processing to computer vision.3234 It demonstrates that transformer also has a strong ability to model spatial correlations of images. The main reason for the success of the transformer is its core component, the self-attention module that captures global information. The self-attention module computes the response at a position in a sequence (e.g., a sentence) by focusing on all positions and taking their weighted average in the embedding space. In the registration task, the two images to be registered may have scale, rotation, translation, and other transformations. Due to various transformations, the local information of an image changes. However, the relative position between two pixels in one image is constant, which is equivalent to the relative position of two words in a sentence in natural language. Mathematically, self-attention can be understood as a nonlocal averaging operation35 that captures long-distance dependencies. To fully exploit the features, we use a nonlocal operation similar to the self-attention module in the transformer to extract global features and CNN to extract local features.

3.

Proposed Method

The proposed algorithm framework in this paper is depicted in Fig. 1. First, we use the phase congruency maximum moments (MMPC)19 to extract keypoints from SAR and optical images. Then we cut patches centered on the keypoints to generate SAR and optical image patches. These patches are converted into 8-bit grayscale images and then fed into MSA-Net to learn deep features. The matching points are obtained by the nearest-neighbor matching strategy. Finally, the final registration results are obtained after removing the outliers by random sample consensus (RANSAC).

Fig. 1

Framework of the proposed registration algorithm.

JARS_17_3_036504_f001.png

As shown in Fig. 2, the proposed MSA-Net learns modality-shared features between the SAR and optical images to improve multimodal image registration performance. First, the MSpFL module is proposed to extract specific shallow features from SAR and optical images. Second, the MShFL module is used to extract deep local and global features shared between the two modalities. The MShFL module includes three LFE modules and three NLA modules. Third, a triplet loss function with a cross-modality similarity constraint is used to match multimodal features. The structure of MSA-Net is shown in Table 1.

Fig. 2

Framework of the proposed MSA-Net.

JARS_17_3_036504_f002.png

Table 1

Detailed structure of the proposed MSA-Net.

StructureDetailsOutput size
Input layerOptical image patch1×64×64
SAR image patch1×64×64
MSpFLSARConv(32×3×3)/BN/ReLU32×32×32
Conv(32×3×3)/BN32×32×32
MSpFLOPTConv(32×3×3)/BN/ReLU32×32×32
Conv(32×3×3)/BN32×32×32
MShFLLFE132×32×32
NLA132×32×32
LFE264×16×16
NLA264×16×16
LFE3128×8×8
NLA3128×8×8
Last layerDropout(0.1)/Conv(128×8×8)128×1

Given a set of SAR and optical patch pairs, pisps, piopo, i1,2,,N, where N is the number of the patches. When i=j, pis=pjo are the corresponding SAR and optical image patches. Thus pis=pjo are the noncorresponding SAR and optical image patches, when ij. The goal of the MSA-Net is to make the distances between the corresponding patch pairs closer and the distances between the noncorresponding patch pairs farther. First, {(pis,pio),i=1,2,,N} are separately fed into the MSpFLSAR and MSpFLOPT modules to obtain N pairs of modality-specific feature maps {(fis,fio)Rc×h×w,i=1,2,,N}, where h and w represent the spatial height and width of the feature map, and c is the channel dimension of each feature map. fis and fio represent two i’th shallow modality-specific features. Second, the MShFL module extracts deep modality-shared features. Each MShFL module includes three LFE modules and three NLA modules. The LFE module is used to extract local features, and the NLA module is used to extract global features. Third, the feature maps obtained by MShFL are fed into a kernel size with 8×8 convolution to obtain the 128-dimensional deep feature vectors {(eis,eio)R128×1,i=1,2,,N}. Finally, the triplet loss function with a cross-modality similarity constraint motivates the MSA-Net to learn more similar feature representations between corresponding SAR and optical image patches.

3.1.

Modality-Specific Feature Learning Module

Considering the nonlinear radiation differences between SAR and optical images, we design MSpFL modules that do not share parameters to extract modality-specific features, such as texture, in the shallow network. There are two reasons to propose MSpFL modules. On the one hand, since the difference between modalities is not considered, it is difficult for the Siamese network to extract effective features that are favorable for multimodal image registration. On the other hand, if a pseudo-Siamese network is used, it will increase the number of network parameters and the difficulty of convergence.

Based on the above analysis, this paper proposes a partially shared feature extraction network. The MSpFL module is used to extract the specific shallow features of SAR and optical images. Suppose that the SAR and optical image patch pairs are {(pis,pio),i=1,2,,N}, pis is the i’th SAR image patch, and pio is the corresponding optical image patch of pis. When they are fed into the MSpFL module, the modality-specific features learned from SAR and optical images can be expressed as follows:

Eq. (1)

fis=MSpFLSAR(pis),fio=MSpFLOPT(pio),
where MSpFLSAR and MSpFLOPT represent the branches that extract modality-specific features of SAR and optical images, resepctively. fis is the learned SAR feature of pis, and fio is the learned optical feature of pio.

MSpFLSAR and MSpFLOPT are two networks that do not share parameters but have the same structure. As shown in Table 1, MSpFLSAR and MSpFLOPT contain two convolutional layers with a kernel size of 3×3 and a number of convolutional filters of 32. The strides of the two convolutional layers are 2 and 1, respectively. The first convolutional layer is followed by a batch normalization (BN) layer and a rectified linear unit (ReLU) layer, and the second convolutional layer is followed by a BN layer.

3.2.

Modality-Shared Feature Learning Module

After extracting the shallow specific features from the two modalities, we need to further extract the shared deep features. Here we design the MShFL module to learn the shared features from SAR and optical images, which includes three LFE modules and three NLA modules:

Eq. (2)

eis=MShFL(fis),eio=MShFL(fio),
where eis is the shared feature of fis, eio is the shared feature of fio, and MShFL is the MShFL module.

3.2.1.

Local feature extract module

We design three LFE modules in the MShFL module to extract local features shared between SAR and optical images. Each LFE module includes a residual structure of two convolution layers with a kernel size of 3×3. The strides of the two convolution layers are 2 and 1, respectively. The channels of the two convolutional layers in the three modules are 32, 64, and 128. Residual connections are added to prevent the network from overfitting. The learned modality-shared local features can be represented as follows:

Eq. (3)

fi,k+1s=LFEk(fi,ks),fi,k+1o=LFEk(fi,ko),
where LFEk represents k’th shared LFE module, k=1, 2, 3. fi,ks and fi,ko represent the SAR and optical features input to LFEk, respectively. fi,k+1s and fi,k+1o represent the SAR and optical features output from LFEk, respectively.

3.2.2.

Nonlocal attention module

Although CNN has good perception abilities for local regions, it lacks the modeling of global information. For the registration task, the information of the local area is important, but the global information can better describe the long-distance correlation and has good robustness to changes in scale, rotation, etc. For example, in two images with the same scene, the content in the images remains the same even after a slight rotation and translation transformation. Local information cannot capture unchanged information due to its limited perceptual range, while global information can still capture relevant information due to its large perceptual region. The NLA module aims at strengthening the features of the current pixel position by aggregating the information from all other positions in the feature map.

As shown in Fig. 3, xRc×h×w is transformed into xRc×hw by the reshape operation R. Let x denotes the input feature map of the NLA module, where c is the number of channels, and hw is the number of positions in the feature map. We use three 1×1 convolutions that form three functions called α, g, and v, reducing the number of channels from c to c/r for the input x. In our experiments, we set the reduction factor r to 2. We use dot-product similarity30 to define the normalized pairwise relationship between positions i and j in a feature map:

Eq. (4)

wij=f(xi,xj)C(x)=Wαx,Wgxjhw,
where wijRhw×hw is a normalized similarity matrix, f(xi,xj) is the relationship between positions i and j, Wα and Wg are the weights of two convolution functions α and g. C(x) is a normalization factor, in this case C(x)=hw.

Fig. 3

NLA module extracts global features. “R” represents reshape operation, “” denotes matrix multiplication, and “” denotes element-wise sum.

JARS_17_3_036504_f003.png

Then the similarity matrix wij is multiplied by the value matrix Wv to obtain the output matrix:

Eq. (5)

yi=j=1hwwij(Wv·xj),
where yi is the attention value of position i, and Wv is the weight of the convolution function v.

The output of the NLA module is defined as

Eq. (6)

zi=Wzyi+xi,
where Wz is a 1×1 convolution with the channel number of c, “+” in the formula indicates residual connection, zi represents the output of the NLA module at position i. zRc×hw is transformed into zRc×h×w by the reshape operation R. The input feature dimension of the NLA module is the same as the output feature dimension, which allows us to insert it anywhere in the network.

The NLA module can be regarded as a global information construction module. The specific operation is to obtain the global contextual features of the current pixel by the weighted average of all positions.

3.3.

Triplet Loss with Similarity Constraint

In order to learn modality-shared features between the SAR and the corresponding optical patch pairs, the triple loss function is used to minimize the distances of the features of corresponding patch pairs and maximize the distances of the features of noncorresponding patch pairs. The margin m is a hyperparameter that aims to keep a certain distance between the corresponding and noncorresponding features. The triple loss Lt is calculated as follows:

Eq. (7)

Lt=1Nti=1Ntmax(d(eis,eio)d(eis,ejo)+m,0),
where d(eis,eio)=22eiseio represents the L2 pairwise distance of the corresponding features, d(eis,ejo)=22eisejo, ij, represents the L2 pairwise distance of the noncorresponding features, and Nt represents the number of samples in a batch, m=1.

Negative sampling is usually required when calculating the triple loss function, which usually includes random sampling in all samples20 and hard negative sample sampling in batches.21 We adopt the latter to improve the optimization performance of the MSA-Net. Under hard negative sample mining, the triple loss is rewritten as

Eq. (8)

Lt=1Nti=1Ntmax(d(eis,eio)d(eis,ehardo)+m,0),
where ehardo is the hard negative sample in the minibatch of eis.

In order to reduce the modal differences between corresponding patch pairs of SAR and optical images, it is necessary to add similarity constraints between them. In this way, the network can learn more compact feature representations between corresponding patches. The cross-modality similarity constraint loss function Ls is defined as

Eq. (9)

Ls=1Nti=1Ntlog(1+exp(d(eis,eio))),
where log and exp represent a logarithmic and exponential function, respectively.

Therefore, the overall loss function is expressed as

Eq. (10)

L=Lt+λLs,
where λ is the weighting coefficient of the loss function.

4.

Experiments and Discussion

To validate the performance of the proposed MSA-Net, the experiments are performed on a public SAR and optical image dataset SEN1-236 and five pairs of SAR and optical images. We compare our method with three handcrafted and three deep learning methods, including OS-SIFT,16 PCSD,18 RIFT,19 HardNet,21 MatchosNet,26 and CNet.27 Section 4.1 describes a data description. Section 4.2 introduces implementation details for the training stage, testing stage, and matching stage. Section 4.3 describes the evaluation criteria used to evaluate algorithm performance. Section 4.4 presents the results and analysis of the experiments. Section 4.4.1 analyzes the influence of three weight-sharing strategies on the model’s performance. In Secs. 4.4.2 and 4.4.3, we verify the effectiveness of the NLA module and the cross-modality similarity constraint. In Sec. 4.4.4, we conduct a parameter sensitivity analysis to analyze the influence of parameter λ on the matching accuracy. In Sec. 4.4.5, we analyze the influence of fine tuning on registration results. In Sec. 4.4.6, we compare the registration results of seven algorithms on five pairs of SAR and optical images, including qualitative analysis and quantitative evaluation. The computational times of seven methods are shown in Sec. 4.4.7.

4.1.

Data Description

The proposed MSA-Net is trained on SEN1-2,36 which is a public SAR and optical image dataset. The SEN1-2 includes 282,384 co-registered SAR and optical images, each with a size of 256×256 and a resolution of 10 m. The SEN1-2 dataset contains four folders: ROIs1158_spring, ROIs1868_summer, ROIs1970_fall, and ROIs2017_winter. In order to ensure that the training samples and test samples do not overlap, we select the data in ROIs1868_summer as the training samples and the data in ROIs1970_fall as the test samples and filter out fuzzy images without obvious ground objects, as shown in Fig. 4. We uniformly sample the images selected from SEN1-2 with a stride of 32 to obtain a patch of size 64×64. After rotation and zooming, we obtain 70,193 training sample pairs and 11,147 validation sample pairs. The information about five pairs of SAR and optical images used for testing is shown in Table 2. The “*” in Table 2 represents unknown information. The SAR and optical images in pairs 1 to 3 are from TerraSAR-X and Google Earth, and the SAR and optical images in pairs 4 to 5 are from GaoFen-3 and Google Earth. In this paper, all experiments are conducted on a desktop computer with Windows 10, an RTX 3060 GPU, and 24 GB RAM. The MSA-Net proposed in this paper is built using the PyTorch37 deep learning framework.

Fig. 4

Examples of SAR and optical images with no obvious objects filtered out from SEN1-2. A pair of corresponding (a), (c) SAR and (b), (d) optical images.

JARS_17_3_036504_f004.png

Table 2

Information about five pairs of SAR and optical images used for testing.

PairModalitySensorSize (m)Resolution (m)
1SARTerraSAR-X550×512*
OpticalGoogle Earth550×512*
1SARTerraSAR-X550×512*
OpticalGoogle Earth550×512*
1SARTerraSAR-X550×512*
OpticalGoogle Earth550×512*
4SARGaoFen-3512×5121
OpticalGoogle Earth512×5121
4SARGaoFen-3512×5121
OpticalGoogle Earth512×5121

4.2.

Implementation Details

4.2.1.

Training stage

We first train our network from scratch on 70,193 pairs of training samples in SEN1-2 and validate it on 11,147 pairs of validation samples. The batch size is 300. The SGD optimizer with an initial learning rate of 0.1 is adopted. The training epoch is 20. Next, the network is fine-tuned with the five pairs of images before testing. Specifically, we manually select four points on each SAR and optical image and use HOPC17 to get registered images. According to the stride of 16, we cut the registered image into patches with a size of 64×64. Some blank patches without content are removed. The reason for using a stride of 16 instead of 32 for the test image is that the number of test data is small, and reducing the stride can generate more patches. Meanwhile, reducing the stride results in more overlap between adjacent patches, leading to increased similarity between their contents. This forces the network to learn more discriminative features, so that the feature distance between different patches is as large as possible. These patches are amplified using the same data enhancement techniques as SEN1-2. The purpose of using overlapping blocks instead of nonoverlapping blocks is to make MSA-Net learn more discriminative features from certain similar samples. When one pair of images is tested, the remaining pairs of images are used to fine-tune the network. The parameters m and λ in the proposed loss function are 1 and 0.1, respectively.

4.2.2.

Testing stage

The feature points are extracted using the MMPC.19 Next, a local patch with a size of 64×64 is cropped around each feature point. Then the local patch is fed into the network to obtain the 128-dimensional feature vector, that is, the proposed modality-shared feature descriptor.

4.2.3.

Matching stage

We match the modality-shared feature descriptors corresponding to each feature point through nearest-neighbor matching to obtain the initial matching points. The affine transformation parameters from the image to be registered (optical image) to the reference image (SAR image) are calculated by using the initial matching points. Finally, outliers are deleted by RANSAC to obtain the final registration result.

4.3.

Evaluation Criteria

We use two metrics to evaluate the matching performance, including FPR95 and accuracy. The FPR95 refers to the false positive rate in which the true positive rate equals 95%, the accuracy refers to the probability that the network correctly predicts the matching label of two patches. Therefore, the smaller the FPR95 is, the higher the accuracy is, and the better the matching performance will be.

We quantitatively evaluate different registration methods using the number of correct matches (NCM), the ratio of correct matches (RCM), and the root-mean-square error (RMSE) as evaluation metrics:

Eq. (11)

RCM=NCMNM,

Eq. (12)

RMSE=1NCMi=1NCM[(mimi)2+(nini)2],
where NM is the total of matches, (mi,ni) is the coordinate of the image to be registered after the transformation matrix, and (mi,ni) is the coordinate of the reference image.

4.4.

Experimental Results and Analysis

4.4.1.

Influence of three weight-sharing strategies

As the analysis in Sec. 2.1 shows, although the Siamese network with completely shared weights can mine shared features, it may ignore modal differences. The pseudo-Siamese network with completely unshared weights can extract modality-shared features, but it does not take modality-shared features into account. Therefore, this paper discusses three weight-sharing strategies, including completely unshared, completely shared, and partially shared weights. The experiment is carried out on the SEN1-2 dataset, and the experimental results are shown in Table 3. The “Dist_np” in Table 3 represents the L2 pairwise distance between positive samples and negative samples. The larger the distance is, the better the network’s ability to distinguish between positive and negative samples will be.

Table 3

Matching results of different weight-sharing strategies.

Weight-sharing strategyAccuracyFPR95Dist_np
Completely unshared0.95390.23560.2035
Completely shared0.93460.24970.1943
Partially shared0.96690.17690.2299

From Table 3, we can see that the network structure with completely shared weights achieves the lowest matching accuracy. The FPR95 is the highest and Dist_np is the lowest among the three weight-sharing strategies, which indicates that the network with completely shared weights does not cope well with differences of multimodal images. The third and fourth rows in Table 3 are the experimental results of the completely unshared weight network and the partially shared weight network, respectively. The latter achieves a higher matching accuracy, about 1.1% higher than the former, and the FPR95 of the latter is 0.1769, about 0.06 lower than the former. The Dist_np of the partially shared weight network is 0.2299, about 0.03 higher than the completely unshared weight network. Therefore, we propose a dual-stream network with partially shared weights can better mine the modality-shared features of SAR and optical images and improve matching accuracy.

4.4.2.

Effectiveness of the NLA module

Figures 5(a) and 5(b) show the changing trend of FPR95 and accuracy under three network structures, which are HardNet, MSA-Net without NLA, and the proposed MSA-Net. MSA-Net without NLA represents the network structure after removing the NLA module from MSA-Net. It can be seen that the matching accuracy of MSA-Net is the highest and the FPR95 of MSA-Net is the lowest among the three models. The NLA module mines global features with a larger receptive field. This is so that MSA-Net can cope with radiation differences to a certain extent, thus improving matching performance. We can also see that MSA-Net without NLA achieves higher matching accuracy and lower FPR95 in fewer epochs compared to HardNet. The difference between the two models lies in whether the weights of the feature network are completely shared and whether residual connections are used. HardNet learns features from SAR and optical images using a network with completely shared weights and no residual connections, both of which make HardNet less effective than MSA-Net without NLA in handling multimodal data.

Fig. 5

Changing trend of (a) FPR95 and (b) accuracy with epoch under different models. Changing trend of (c) FPR95 and (d) accuracy with epoch under different loss functions.

JARS_17_3_036504_f005.png

4.4.3.

Influence of loss function

Figures 5(c) and 5(d) show the FPR95 and accuracy change curves of different loss functions. After adding cross-modality similarity constraints to the loss function, the matching accuracy is further improved and the FPR95 is further reduced. It proves that the combination of triplet loss and cross-modality similarity constraint loss allows the MSA-Net to further explore the similarity between SAR and optical images.

4.4.4.

Parameter sensitivity analysis

In this section, we analyze the influence of the weight coefficient λ on matching accuracy. To observe the changing trend of accuracy in the test set of SEN1-2, we gradually increase the value of λ from 0 to 1 with an interval of 0.1. Figure 6 reports the average of three experiments. When the value of λ is between 0 and 1, the value of accuracy is between 0.978 and 0.982. It indicates that adding cross-modality similarity constraints to the loss function will not affect the stability of the network. When λ is 0.1, accuracy reaches its maximum value. It can be seen that a small weight for Ls can make the network learn well. When λ increases to 1, accuracy gradually decreases. We conclude that giving a larger λ to cross-modal similarity constraints is not conducive to learning features of multimodal images. We should not let the network pay too much attention to learning similar features without considering that the multimodal data itself is different. So it is necessary to obtain an appropriate weight coefficient through experiments.

Fig. 6

Changing trend of accuracy with λ.

JARS_17_3_036504_f006.png

4.4.5.

Influence of fine-tuning on registration results

In Fig. 7, we show the influence of fine-tuning on the registration results of five test images, including NCM and RMSE. Figure 7(a) shows that the NCM increases significantly after fine-tuning, especially in pairs 1, 2, 4, and 5. As shown in Fig. 7(b), the RMSEs of pairs 1 and 2 decrease after fine-tuning. The RMSE of pair 3 is higher after fine-tuning. The RMSEs of pairs 4 and 5 do not change significantly before and after fine-tuning. It can be seen that registration results can be improved to some extent through fine-tuning. Even without fine-tuning, relatively good results can be obtained, indicating that the model trained with SEN1-2 data has transfer ability. Considering that different data are distributed differently, fine-tuning can improve the adaptability of the network and obtain better registration results.

Fig. 7

Influence of fine-tuning on the registration results of five test image pairs. (a) NCM obtained by MSA-Net in five pairs of test images with or without fine-tuning. (b) RMSE obtained by MSA-Net in five pairs of test images with or without fine-tuning.

JARS_17_3_036504_f007.png

4.4.6.

Comparison with the other registration methods

We select five pairs of SAR and optical images for evaluating seven registration methods. In pairs 1 to 3, there are not only radiation differences between SAR and optical images but also rotation and translation transformations. Pairs 4 and 5 have obvious radiation differences, noise, slight translation transformations, and no rotation transformation. Figures 8(a)8(g)12(a)12(g) show the matching points obtained by seven methods in five pairs of SAR and optical images. Figures 8(h)12(h) show the registration results of MSA-Net. We superimpose the outlines (as shown by the yellow curve) of transformed optical images on the original SAR images and use several red rectangles to highlight the registration results for certain regions.

Fig. 8

Registration results of seven methods in image pair 1. (a)–(g) Matches found using OS-SIFT, PCSD, RIFT, HardNet, MatchosNet, CNet, and MSA-Net, respectively. (h) Registration result of our proposed MSA-Net.

JARS_17_3_036504_f008.png

From Figs. 8(a)8(c)12(a)12(c), we can see that the three handcrafted algorithms have achieved relatively good registration results for the five pairs of images. RIFT obtains more matching points than OS-SIFT and PCSD thanks to its use of radiation-invariant phase congruency to extract keypoints and construct descriptors. OS-SIFT has the fewest matching points among the three handcrafted algorithms because the descriptor based on gradient is worse than the one based on phase congruency in dealing with radiation differences. PCSD matches descriptors in the local window, resulting in fewer matching points obtained.

Among the four deep learning methods, MatchosNet obtains few matching points with obvious errors in pairs 1 to 3 [as shown in Figs. 8(e)10(e)], but a certain number of correct matching points can be found in pairs 4 and 5 [as shown in Figs. 11(e) and 12(e)]. The main reason is that MatchosNet is a position-matching algorithm, which needs to know the approximate offset of the two images to be registered in advance. Therefore, it is often difficult for MatchosNet to find the correct number of matching points in a pair of images with rotation and translation transformations. HardNet, CNet, and MSA-Net are patch-based feature extraction networks. The difference between the three is the weight-sharing strategy of the networks. Specifically, HardNet is a Siamese network with weight sharing, CNet is a pseudo-Siamese network without weight sharing completely, and MSA-Net is a network structure with partial weight sharing. HardNet does not consider the modal differences, and CNet does not take the mining of the modality-shared features into account, so the matching points obtained by both [as shown in Figs. 8(d) and 8(f)12(d) and 12(f)] are not as much as those obtained by MSA-Net. As shown in Figs. 8(g)12(g), our proposed MSA-Net achieves the highest number of matching points among the seven registration methods. MSA-Net combines two-stream CNNs and NLA to extract both local and global features and obtains features with different perceptual ranges to cope with radiation differences between SAR and optical images. In Figs. 8(h)12(h), after carefully observing the registration results, we find that the outlines of the transformed optical image and the corresponding SAR images almost overlap each other. Meanwhile, the regions circled by the red rectangles are well-matched. In summary, compared with the other six methods, the proposed MSA-Net achieves the best registration results and the most matches on the five pairs of images, which can prove the effectiveness of the proposed method.

Fig. 9

Registration results of seven methods in image pair 2. (a)–(g) Matches found using OS-SIFT, PCSD, RIFT, HardNet, MatchosNet, CNet, and MSA-Net, respectively. (h) Registration result of our proposed MSA-Net.

JARS_17_3_036504_f009.png

Fig. 10

Registration results of seven methods in image pair 3. (a)–(g) Matches found using OS-SIFT, PCSD, RIFT, HardNet, MatchosNet, CNet, and MSA-Net, respectively. (h) Registration result of our proposed MSA-Net.

JARS_17_3_036504_f010.png

Fig. 11

Registration results of seven methods in image pair 4. (a)–(g) Matches found using OS-SIFT, PCSD, RIFT, HardNet, MatchosNet, CNet, and MSA-Net, respectively. (h) Registration result of our proposed MSA-Net.

JARS_17_3_036504_f011.png

Fig. 12

Registration results of seven methods in image pair 5. (a)–(g) Matches found using OS-SIFT, PCSD, RIFT, HardNet, MatchosNet, CNet, and MSA-Net, respectively. (h) Registration result of our proposed MSA-Net.

JARS_17_3_036504_f012.png

Table 4 shows the quantitative comparison of the seven methods on five pairs of test images, including NCM, RCM, and RMSE. The higher the NCM and RCM and the lower the RMSE, the better the performance of the registration algorithm. The “*” in Table 4 indicates that the method is invalid for this pair of images. Even though the algorithm finds some matching points, the error of these matching points is relatively large.

Table 4

Quantitative comparison of the proposed MSA-Net with OS-SIFT, PCSD, RIFT, HardNet, MatchosNet, and CNet. The bold values represent the best results among all compared methods.

PairMetricOS-SIFTPCSDRIFTHardNetMatchosNetCNetMSA-Net
1NCM50738511013101119
RCM (%)33.3348.6756.6773.33*67.3379.33
RMSE1.35161.20181.05230.9816*1.12430.9520
2NCM40546567175778
RCM (%)26.673643.3344.66*3852
RMSE1.04251.14430.97950.9589*1.17900.8790
3NCM21232526162129
RCM (%)42465042*4258
RMSE1.36591.10391.03010.9659*1.14140.8863
4NCM10164716223469
RCM (%)*5.9318.576.318.6913.4329.99
RMSE*1.77551.66571.55701.50851.64141.4346
5NCM18254813271762
RCM (%)8.8710.3424.136.4013.308.3740.78
RMSE2.04261.80891.65241.51981.64841.46341.4201

Our proposed MSA-Net obtains the highest NCM and RCM and the lowest RMSE among the seven algorithms. This is mainly due to the fact that our algorithm not only considers the radiation differences between SAR and optical images but also mines the shared features between the two modalities to improve the similarity of features. In order to achieve the above, MSA-Net constructs two MSpFL modules and three MShFL modules to extract local and global features and designs a loss function that considers similarity constraints. PCSD, RIFT, HardNet, and CNet all find some correct matching points in the five pairs of test images. PCSD and RIFT construct descriptors based on phase congruency, which can cope with radiation differences. HardNet and CNet are patch-based deep feature extraction networks, and the network can mine similar features of the two modalities through data learning. However, as the radiation difference increases, the matching points will decrease accordingly. PCSD, RIFT, HardNet, and CNet obtain more matching points in pairs 1 and 2 than in pairs 3 to 5. OS-SIFT and MatchosNet fail on some image pairs. OS-SIFT is a gradient-based descriptor that is susceptible to radiation differences. MatchosNet is an algorithm based on position matching, which fails in image pairs with rotation and translation transformations.

4.4.7.

Computational time of the seven methods

The computational time of the seven algorithms on the five pairs of images is shown in Table 5. Among the seven algorithms, RIFT has the fastest computational efficiency. This is because it uses the FAST detector to extract key points, the phase consistency algorithm to extract descriptors, and the ratio threshold for matching. The calculation time of OS-SIFT is much longer than that of RIFT. Because OS-SIFT extracts keypoints and constructs descriptors in the multiscale space, and it takes a certain amount of time to construct a multiscale space. PCSD takes the longest time, mainly because it uses template matching. Each keypoint calculates its similarity to all keypoints in the local window. The larger the window is, the more the keypoints are, and the longer the time will be. The computational time of the four deep learning algorithms is relatively close. HardNet takes the least time because it is a Siamese network with the least network parameters. MatchosNet and CNet are pseudo-Siamese networks with more parameters than HardNet. MSA-Net has a partially shared network structure and three NLA modules. The computational complexity of the NLA module is the square of the number of samples. So the computation time of MSA-Net is longer than that of MatchosNet and CNet.

Table 5

Computational time of the seven methods.

PairOS-SIFT (s)PCSD (s)RIFT (s)HardNet (s)MatchosNet (s)CNet (s)MSA-Net (s)
131.3244.4512.7313.7917.1616.0519.33
225.3241.269.5711.8614.2817.8914.34
324.9245.267.718.8210.6513.1413.34
419.7342.248.639.6510.6516.1418.47
520.9243.107.719.7810.4315.4216.87

5.

Conclusion

This paper proposes a deep learning algorithm, called MSA-Net, for accomplishing SAR and optical image registration. The algorithm combines a dual-stream network with NLA to construct modality-shared features descriptors for SAR and optical images. In MSA-Net, the MSpFL module extracts shallow features from SAR and optical images to obtain modality-specific features, whereas the MShFL module combines the LFE module and NLA module to obtain richer deep modality-shared features, including both local and global features. A loss function that combines triplet loss with cross-modality similarity constraint is also proposed to further improve matching accuracy by constraining the network to be immune to radiometric differences. The experiments are conducted on a public dataset and five pairs of test SAR and optical images. Specifically, the validation experiments verify that the partially shared weight strategy, the NLA, and cross-modality similarity constraints in MSA-Net help to improve the matching accuracy of SAR and optical images. In addition, the comparative experimental results demonstrate that our proposed method achieves the best registration results among the seven registration methods. In future work, we would like to study more diverse attention mechanisms and feature fusion techniques to further enhance the capabilities of the network for SAR and optical image registration.

Acknowledgements

The authors would like to thank all the anonymous reviewers for their valuable comments that helped to improve the quality of this paper. No potential conflicts of interest were reported by the author(s). This work was supported by the National Natural Science Foundation of China (Grant No. 62172321) and the Civil Space Thirteen Five Years Pre-Research Project (Grant No. D040114).

References

1. 

Y. Wu et al., “Fusion of synthetic aperture radar and visible images based on variational multiscale image decomposition,” J. Appl. Remote Sens, 11 (2), 025006 https://doi.org/10.1117/1.JRS.11.025006 (2017). Google Scholar

2. 

A. Shakya, M. Biswas and M. Pal, “CNN-based fusion and classification of SAR and Optical data,” Int. J. Remote Sens., 41 (22), 8839 –8861 https://doi.org/10.1080/01431161.2020.1783713 IJSEDK 0143-1161 (2020). Google Scholar

3. 

X. Jiang et al., “Building damage detection via superpixel-based belief fusion of space-borne SAR and optical images,” IEEE Sens. J., 20 (4), 2008 –2022 https://doi.org/10.1109/JSEN.2019.2948582 ISJEAZ 1530-437X (2020). Google Scholar

4. 

Z. Zhang et al., “Endoscope image mosaic based on pyramid ORB,” Biomed. Signal Process. Control, 71 103261 https://doi.org/10.1016/j.bspc.2021.103261 (2022). Google Scholar

5. 

S. Eken et al., “Resource-and content-aware, scalable stitching framework for remote sensing images,” Arab. J. Geosci., 12 1 –13 https://doi.org/10.1007/s12517-019-4388-y (2019). Google Scholar

6. 

S. Eken and A. Sayar, “A MapReduce-based distributed and scalable framework for stitching of satellite mosaic images,” Arab. J. Geosci., 14 (18), 1 –16 https://doi.org/10.1007/s12517-021-07500-w (2021). Google Scholar

7. 

G. N. Vivekananda, R. Swathi and A. Sujith, “Multi-temporal image analysis for LULC classification and change detection,” Eur. J. Remote Sens., 54 (suppl 2), 189 –199 https://doi.org/10.1080/22797254.2020.1771215 (2021). Google Scholar

8. 

Y. Wu et al., “Commonality autoencoder: learning common features for change detection from heterogeneous images,” IEEE Trans. Neural Networks Learn. Syst., 33 4257 –4270 https://doi.org/10.1109/TNNLS.2021.3056238 (2021). Google Scholar

9. 

X. Li et al., “Collaborative attention-based heterogeneous gated fusion network for land cover classification,” IEEE Trans. Geosci. Remote Sens., 59 (5), 3829 –3845 https://doi.org/10.1109/TGRS.2020.3015389 IGRSD2 0196-2892 (2020). Google Scholar

10. 

W. Kang et al., “CFNet: a cross fusion network for joint land cover classification using optical and SAR images,” IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens., 15 1562 –1574 https://doi.org/10.1109/JSTARS.2022.3144587 (2022). Google Scholar

11. 

X. Li et al., “Dense adaptive grouping distillation network for multimodal land cover classification with privileged modality,” IEEE Trans. Geosci. Remote Sens., 60 1 –14 https://doi.org/10.1109/TGRS.2022.3176936. IGRSD2 0196-2892 (2022). Google Scholar

12. 

X. Li, L. Lei and G. Kuang, “Locality-constrained bilinear network for land cover classification using heterogeneous images,” IEEE Geosci. Remote Sens. Lett., 19 1 –5 https://doi.org/10.1109/LGRS.2021.3086592 (2022). Google Scholar

13. 

D. G. Lowe, “Distinctive image features from scale-invariant keypoints,” Int. J. Comput. Vision, 60 (2), 91 –110 https://doi.org/10.1023/B:VISI.0000029664.99615.94 IJCVEQ 0920-5691 (2004). Google Scholar

14. 

J. Fan et al., “SAR image registration using phase congruency and nonlinear diffusion-based SIFT,” IEEE Geosci. Remote Sens. Lett., 12 (3), 562 –566 https://doi.org/10.1109/LGRS.2014.2351396 (2015). Google Scholar

15. 

F. Dellinger et al., “SAR-SIFT: a SIFT-like algorithm for SAR images,” IEEE Trans. Geosci. Remote Sens., 53 (1), 453 –466 https://doi.org/10.1109/TGRS.2014.2323552 IGRSD2 0196-2892 (2015). Google Scholar

16. 

Y. Xiang, F. Wang and H. You, “OS-SIFT: a robust SIFT-like algorithm for high-resolution Optical-to-SAR image registration in suburban areas,” IEEE Trans. Geosci. Remote Sens., 56 (6), 3078 –3090 https://doi.org/10.1109/TGRS.2018.2790483 IGRSD2 0196-2892 (2018). Google Scholar

17. 

Y. Ye et al., “Robust registration of multimodal remote sensing images based on structural similarity,” IEEE Trans. Geosci. Remote Sens., 55 (5), 2941 –2958 https://doi.org/10.1109/TGRS.2017.2656380 IGRSD2 0196-2892 (2017). Google Scholar

18. 

J. Fan et al., “SAR and optical image registration using nonlinear diffusion and phase congruency structural descriptor,” IEEE Trans. Geosci. Remote Sens., 56 (9), 5368 –5379 https://doi.org/10.1109/TGRS.2018.2815523 IGRSD2 0196-2892 (2018). Google Scholar

19. 

J. Li, Q. Hu and M. Ai, “RIFT: multi-modal image matching based on radiation-variation insensitive feature transform,” IEEE Trans. Image Process., 29 3296 –3310 https://doi.org/10.1109/TIP.2019.2959244 IIPRE4 1057-7149 (2020). Google Scholar

20. 

Y. Tian, B. Fan and F. Wu, “L2-Net: deep learning of discriminative patch descriptor in Euclidean space,” in Proc. IEEE Conf. Comput. Vision and Pattern Recognit., 661 –669 (2017). https://doi.org/10.1109/CVPR.2017.649 Google Scholar

21. 

A. Mishchuk et al., “Working hard to know your neighbor’s margins: local descriptor learning loss,” in Adv. Neural Inf. Process. Syst., (2017). Google Scholar

22. 

T. Bürgmann, W. Koppe and M. Schmitt, “Matching of TerraSAR-X derived ground control points to optical image patches using deep learning,” ISPRS J. Photogramm. Remote Sens., 158 241 –248 https://doi.org/10.1016/j.isprsjprs.2019.09.010 IRSEE9 0924-2716 (2019). Google Scholar

23. 

L. H. Hughes et al., “Identifying corresponding patches in SAR and optical images with a pseudo-Siamese CNN,” IEEE Geosci. Remote Sens. Lett., 15 (5), 784 –788 https://doi.org/10.1109/LGRS.2018.2799232 (2018). Google Scholar

24. 

H. Zhang et al., “Registration of multimodal remote sensing image based on deep fully convolutional neural network,” IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens., 12 (8), 3028 –3042 https://doi.org/10.1109/JSTARS.2019.2916560 (2019). Google Scholar

25. 

L. H. Hughes et al., “A deep learning framework for matching of SAR and optical imagery,” ISPRS J. Photogramm. Remote Sens., 169 166 –179 https://doi.org/10.1016/j.isprsjprs.2020.09.012 IRSEE9 0924-2716 (2020). Google Scholar

26. 

Y. Liao et al., “Feature matching and position matching between optical and SAR with local deep feature descriptor,” IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens., 15 448 –462 https://doi.org/10.1109/JSTARS.2021.3134676 (2021). Google Scholar

27. 

D. Quan et al., “Deep feature correlation learning for multi-modal remote sensing image registration,” IEEE Trans. Geosci. Remote Sens., 60 1 –16 https://doi.org/10.1109/TGRS.2022.3187015 IGRSD2 0196-2892 (2022). Google Scholar

28. 

H. Han, C. Li and X. Qiu, “Multi-modal remote sensing image matching method based on deep learning technology,” J. Phys. Conf. Ser., 2083 (3), 032093 https://doi.org/10.1088/1742-6596/2083/3/032093 JPCSDZ 1742-6588 (2021). Google Scholar

29. 

J. Fan et al., “A novel multiscale adaptive binning phase congruency feature for SAR and optical image registration,” IEEE Trans. Geosci. Remote Sens., 60 1 –16 https://doi.org/10.1109/TGRS.2022.3206804 IGRSD2 0196-2892 (2022). Google Scholar

30. 

A. Dosovitskiy et al., “An image is worth 16×16 words: transformers for image recognition at scale,” (2020). https://doi.org/10.48550/arXiv.2010.11929 Google Scholar

31. 

A. Vaswani et al., “Attention is all you need,” (2017). https://doi.org/10.48550/arXiv.1706.03762 Google Scholar

32. 

H. Dong, L. Zhang and B. Zou, “Exploring vision transformers for polarimetric SAR image classification,” IEEE Trans. Geosci. Remote Sens., 60 1 –15 https://doi.org/10.1109/TGRS.2021.3137383 IGRSD2 0196-2892 (2022). Google Scholar

33. 

H. Chen, Z. Qi and Z. Shi, “Remote sensing image change detection with transformers,” IEEE Trans. Geosci. Remote Sens., 60 1 –14 https://doi.org/10.1109/TGRS.2021.3095166 IGRSD2 0196-2892 (2022). Google Scholar

34. 

X. Liu et al., “High resolution SAR image classification using global-local network structure based on vision transformer and CNN,” IEEE Geosci. Remote Sens. Lett., 19 1 –5 https://doi.org/10.1109/LGRS.2022.3151353 (2022). Google Scholar

35. 

X. Wang et al., “Non-local neural networks,” in Proc. IEEE Conf. Comput. Vision and Pattern Recognit., 7794 –7803 (2018). https://doi.org/10.1109/CVPR.2018.00813 Google Scholar

36. 

M. Schmitt, L. H. Hughes and X. X. Zhu, “The SEN1-2 dataset for deep learning in SAR-optical data fusion,” in ISPRS Ann. Photogramm. Remote Sens. Spatial Inf. Sci., IV-1, 141 –146 (2018). Google Scholar

37. 

A. Paszke et al., “Automatic differentiation in PyTorch,” in NIPS Workshops, (2017). Google Scholar

Biography

Xin Hu received her BE degree from Xi’an University of Technology in 2019. She is currently pursuing her PhD with the Remote Sensing Image Processing and Fusion Group of the School of Electronic Engineering at Xidian University. Her main research direction is remote sensing image registration.

Yan Wu received her BS degree in information processing and her MA and PhD degrees in signal and information processing from Xidian University, Xi’an, China, in 1987, 1998, and 2003, respectively. Since 2005, she has been a professor in the Department of Electronic Engineering, Xidian University. Her broad research interests include remote sensing image analysis and interpretation, data fusion of multisensory images, synthetic aperture radar autotarget recognition, and statistical learning theory and application.

Zhikang Li received his BS degree from the Xidian University, Xi’an, China, in 2021. He is currently pursuing his PhD with the Remote Sensing Image Processing and Fusion Group, School of Electronic Engineering, Xidian University, Xi’an, China. His research interests include synthetic aperture radar image analysis and feature extraction.

Xiaoru Zhao received his BS degree from the Taiyuan University of Technology, Taiyuan, China, in 2020. He is currently pursuing his MA degree with the Remote Sensing Image Processing and Fusion Group, School of Electronic Engineering, Xidian University, Xi’an, China. His research interests include synthetic aperture radar image analysis and feature extraction.

Xingyu Liu received her BS degree in engineering from Xidian University, Shaanxi, China, in 2020, majoring in measurement and control technology and instrumentation program. Currently, she continues to pursue her MS degree with the Remote Sensing Image Processing and Fusion Group, Xidian University. Her main research direction is synthetic aperture radar image feature extraction and classification.

Ming Li received his BS degree in electrical engineering and his MS and PhD degrees in signal processing from Xidian University, Xi’an, China, in 1987, 1990, and 2007, respectively. In 1987, he joined the Department of Electronic Engineering, Xidian University, where he is currently a professor with the National Key Laboratory of Radar Signal Processing. His research interests include adaptive signal processing, detection theory, ultrawideband, and synthetic aperture radar image processing.

CC BY: © The Authors. Published by SPIE under a Creative Commons Attribution 4.0 International License. Distribution or reproduction of this work in whole or in part requires full attribution of the original publication, including its DOI.
Xin Hu, Yan Wu, Zhikang Li, Xiaoru Zhao, Xingyu Liu, and Ming Li "Synthetic aperture radar and optical image registration using local and global feature learning by modality-shared attention network," Journal of Applied Remote Sensing 17(3), 036504 (30 August 2023). https://doi.org/10.1117/1.JRS.17.036504
Received: 5 October 2022; Accepted: 8 August 2023; Published: 30 August 2023
Advertisement
Advertisement
KEYWORDS
Synthetic aperture radar

Image registration

Feature extraction

Radio optics

Mining

Education and training

Image fusion

Back to Top