[go: up one dir, main page]

Next Article in Journal
AIPI: Network Status Identification on Multi-Protocol Wireless Sensor Networks
Previous Article in Journal
Securing IoT Networks Against DDoS Attacks: A Hybrid Deep Learning Approach
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Method Combining Discrete Cosine Transform with Attention for Multi-Temporal Remote Sensing Image Matching

1
Key Laboratory of Opto-Electronic Information Processing, Chinese Academy of Sciences, Shenyang 110016, China
2
Shenyang Institute of Automation, Chinese Academy of Sciences, Shenyang 110016, China
3
University of Chinese Academy of Sciences, Beijing 100049, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(5), 1345; https://doi.org/10.3390/s25051345 (registering DOI)
Submission received: 11 January 2025 / Revised: 13 February 2025 / Accepted: 17 February 2025 / Published: 22 February 2025
(This article belongs to the Section Remote Sensors)
Figure 1
<p>Overall framework of the proposed method.</p> ">
Figure 2
<p>The details in the feature extraction module. (<b>a</b>) The structure of the blocks. BN is short for Batch Normalization. (<b>b</b>) The structure of the enhanced frequency channel attention (eFCA).</p> ">
Figure 3
<p>The flowchart of DCT-guided sparse attention. For self-attention, <math display="inline"><semantics> <mi mathvariant="bold-italic">Q</mi> </semantics></math>, <math display="inline"><semantics> <mi mathvariant="bold-italic">K</mi> </semantics></math>, and <math display="inline"><semantics> <mi mathvariant="bold-italic">V</mi> </semantics></math> all come from the feature map to be updated. For cross-attention, <math display="inline"><semantics> <mi mathvariant="bold-italic">Q</mi> </semantics></math> comes from the feature map to be updated, while <math display="inline"><semantics> <mi mathvariant="bold-italic">K</mi> </semantics></math> and <math display="inline"><semantics> <mi mathvariant="bold-italic">V</mi> </semantics></math> come from the other one.</p> ">
Figure 4
<p>Examples of the datasets [<a href="#B51-sensors-25-01345" class="html-bibr">51</a>,<a href="#B52-sensors-25-01345" class="html-bibr">52</a>,<a href="#B53-sensors-25-01345" class="html-bibr">53</a>].</p> ">
Figure 5
<p>Box plots of ablation experiment results. The dots represent outliers of the results, and the orange lines indicate the median. (<b>a</b>) Box plot of PCK results. (<b>b</b>) Box plot of ACE results.</p> ">
Figure 6
<p>The line charts show the proportion of results that met different PCK thresholds. The x-axis represents the set PCK thresholds, and the y-axis represents the proportion of images in the dataset that satisfied each threshold. (<b>a</b>) Results on the DSIFN dataset. (<b>b</b>) Results on the LEVIR-CD dataset.</p> ">
Figure 7
<p>Visualization of some results from the DSIFN dataset and the LEVIR-CD dataset. The two groups on the left were from the DSIFN dataset, while the two groups on the right were from the LEVIR-CD dataset. The green dots indicate correctly matched keypoints, and the red dots indicate incorrectly matched keypoints.</p> ">
Figure 8
<p>This is a visualization of qualitative comparison, with correct matching keypoints represented in green and incorrect matching keypoints represented in red.</p> ">
Versions Notes

Abstract

:
Multi-temporal remote sensing image matching is crucial for tasks such as drone positioning under satellite-denial conditions, natural disaster monitoring, and land-cover-change detection. However, the significant differences between multi-temporal images often lead to the reduced accuracy or even failure of most image matching methods in these scenarios. To address this challenge, this paper introduces a Discrete Cosine Transform (DCT) for frequency analysis tailored to the characteristics of remote sensing images, and proposes a network that combines the DCT with attention mechanisms for multi-scale feature matching. First, DCT-enhanced channel attention is embedded in the multi-scale feature extraction module to capture richer ground object information. Second, in coarse-scale feature matching, DCT-guided sparse attention is proposed for feature enhancement, which suppresses the impact of temporal differences on matching while making the amount of computation controllable. The coarse-scale matching results are further refined in the fine-scale feature map to obtain the final matches. Our method achieved correct keypoint percentages of 81.92% and 88.48%, with average corner errors of 4.27 and 2.98 pixels on the DSIFN dataset and LEVIR-CD dataset, respectively, while maintaining a high inference speed. The experimental results demonstrate that our method outperformed the state-of-art methods in terms of both robustness and efficiency in the multi-temporal scenarios.

1. Introduction

Multi-temporal remote sensing image matching is a critical task in geospatial analysis, facilitating applications such as environmental monitoring, urban development tracking, and disaster assessment. The accurate alignment of images captured at different times is essential for detecting changes and ensuring the integrity of subsequent analyses. Traditional image matching methods often rely on feature detection and description [1,2]. However, maintaining robust correspondences becomes challenging in the presence of significant temporal variations, such as different sensors, seasonal effects, and changes in urban structures.
Recent advances in deep learning have introduced sophisticated image matching frameworks, notably SuperGlue [3] and LoFTR [4]. SuperGlue introduces an innovative application of self-attention and cross-attention mechanisms for feature matching. LoFTR builds on this application of attention mechanisms and further extends this approach by combining a pixel-wise dense matching strategy and a coarse-to-fine matching method. Although these methods have demonstrated impressive performances in general image matching tasks, their effectiveness in multi-temporal remote sensing applications is limited due to the significant appearance variations inherent in such data.
To address these limitations, this paper presents an enhanced image matching method based on the LoFTR framework that was specifically designed for multi-temporal remote sensing images. Our approach integrates Discrete Cosine Transform (DCT)-based attention mechanisms to leverage frequency domain information, which is rich in the structural and textural patterns crucial for distinguishing different land types and monitoring temporal changes. By incorporating the DCT into both the feature extraction and matching stages, our method enhances feature discrimination and reduces computational complexity, ultimately improving the matching accuracy and efficiency.
The proposed methodology introduces two key innovations. On the one hand, DCT-based channel attention modules are integrated into the CNN backbone of LoFTR, enabling the network to leverage frequency information to emphasize salient features. Different land covers in remote sensing images often have unique frequency distributions [5]. We exploited this character to innovatively introduce frequency information into the feature extraction process. By utilizing frequency information to generate channel attention, we enhanced the robustness and discrimination of feature representations, especially when dealing with significant appearance changes in multi-temporal remote sensing images. On the other hand, a DCT-guided sparse attention mechanism was introduced in the coarse matching stage that was inspired by DCT’s application in image compression. By innovatively integrating DCT with a sparse attention mechanism, we extracted frequency information to generate attention guides. This approach allowed us to perform self-attention or cross-attention calculations on the most relevant windows, minimizing interference from temporal difference regions in the attention mechanism and enhancing the computational efficiency. Through extensive experiments on diverse multi-temporal remote sensing datasets, we demonstrated that integrating DCT-based attention mechanisms into the LoFTR framework significantly enhanced both the matching accuracy and computational efficiency compared with existing image matching methods.
In summary, the main contributions of this paper are as follows:
1.
We introduce a novel method for integrating DCT-based frequency channel attention into the CNN backbone, enhancing the feature robustness and discrimination in multi-temporal remote sensing scenarios.
2.
We propose a frequency-guided sparse attention mechanism to enhance coarse-scale features. By narrowing the attention scope, this module minimizes noise introduced by the temporal difference region and simultaneously reduces the computational complexity.
3.
Through comprehensive experiments, we validated that our DCT-integrated approach outperformed existing image matching frameworks in terms of the robustness and efficiency on multi-temporal remote sensing datasets.

2. Related Works

2.1. Image Matching Methods for Multi-Temporal Remote Sensing Images

Remote sensing image matching methods can be broadly categorized into area-based methods and feature-based methods [6,7,8,9]. Feature-based methods extract local invariant features for matching, which is more resistant to temporal changes and is therefore more suitable for multi-temporal scenarios. Traditional feature-based matching methods typically involve first detecting features, followed by a feature description [2,10,11,12]. Rasmy et al. [13] proposed a method combining Fourier phase correlation and the Harris detector in order to achieve a sub-pixel registration accuracy. Aiming to improve the temporal robustness, OFM [14] uses multioriented filters to convolute images and extract the orientation indices to construct the feature maps. However, this type of manual design method relies on prior knowledge and is only applicable to multi-temporal scenes with small changes in ground objects. The introduction of deep learning can bring significant performance improvements, especially in scene robustness. Yang et al. [15] used a pre-trained VGG network to generate robust multi-scale feature descriptors. Liu et al. [16] used spatial multi-scale convolutional layers to construct a Siamese network to align multi-temporal features. These methods still follow the framework of feature detection followed by feature description. However, the complex temporal variations make it challenging to reliably detect keypoints. Recently, detector-free image matching methods [17,18,19], designed for general tasks, have emerged, offering new solutions to this problem. NBR-Net [20] used the bidirectional estimation of pixel flow to enhance the multi-temporal image matching consistency, but its performance remained constrained by the limited receptive field of CNNs. SC-Net [21] used an attention mechanism to fuse local features and global features of a single image to improve the recognition of repeated scenes, but it still lacks the ability to resist obvious temporal differences.

2.2. Deep Learning-Based Image Matching Methods for General Tasks

With the rise of deep learning, researchers have increasingly utilized deep learning models to autonomously extract image features, replacing traditional low-level image information with high-level features learned. Early image matching methods mainly employed Siamese networks, where two sub-networks with shared weights were used to extract meaningful features and then describe them [22,23,24,25]. Innovatively, SuperPoint [26] proposed a dual-branch network framework for self-supervised training, enabling simultaneous keypoint detection and descriptor generation. D2-Net [27] further developed this idea, transforming convolutional networks into dense feature descriptors and detectors. NCNet [17] adopted a detector-free approach to directly learn dense correspondences in an end-to-end manner. In recent years, Transformer [28] was introduced into computer vision [29]. To handle large viewpoint variations, SuperGlue [3] used a Transformer to expand the receptive field of keypoints. LoFTR [4] combined the Transformer with the detector-free strategy to solve the matching problem in low-texture scenes. The framework of LoFTR has inspired numerous subsequent works [30,31,32]. However, in multi-temporal scenarios, the challenge is not the lack of texture or the large change in viewpoint, but rather the interference caused by temporal variations that disrupt the texture consistency. These variations introduce significant noise and distortions, which are often difficult to differentiate from the invariant features, presenting a unique challenge compared with general tasks.

2.3. DCT in Deep Learning

Common methods for space–frequency domain conversion are the Discrete Fourier Transform (DFT) and the DCT. The DFT uses complex numbers to represent frequency domain information, which introduces additional computational complexity. In contrast, the DCT uses real numbers. It is computationally efficient and has good energy concentration, so it is often used for image denoising [33,34,35] and image compression [36,37,38]. In recent years, the DCT was introduced into deep learning as an efficient feature extraction tool [39,40,41]. Ghosh et al. [42] believed that the combination of DCT and convolutional networks can generate sparser weight matrices. FCANet [43] first used the DCT to generate a channel attention mechanism that achieved excellent performance in multiple tasks. Chaudhury et al. [44] performed a DCT transformation on the input and found that the network trained in the cosine domain is more resistant to noise than that trained in the spatial domain. These works inspired us to integrate the DCT into our method.

3. Method

Our method can be divided into three stages: DCT-enhanced multi-scale feature extraction, DCT-guided coarse-scale feature matching, and fine-scale feature matching. The overall structure is shown in Figure 1.

3.1. DCT-Enhanced Multi-Scale Feature Extraction

Given a pair of images A and B to be matched, they are input into a multi-scale feature extraction module that integrates a feature pyramid network (FPN) [45] with enhanced frequency channel attention (eFCA). This module extracts deep features from the images and ultimately outputs both coarse-scale feature maps F ¯ A and F ¯ B and fine-scale feature maps F ˜ A and F ˜ B . The overall framework of the module is shown in Figure 1 and more details are in Figure 2.
Remote sensing images usually contain a wealth of information, such as the topography, vegetation cover, and water distribution. An FPN enhances the model’s ability to handle scenes with large scale differences in remote sensing images through multi-scale feature fusion. However, the downsampling operation in an FPN reduces the image resolution, leading to a loss of some fine details that may be essential [46]. This presents a significant challenge for extracting abundant features from remote sensing images. Frequency domain analysis can capture information that is difficult to distinguish in the spatial domain by separating different frequency components within the image. For example, the high-frequency components typically contain detailed features, such as edges and lines, which have high stability and recognizability. Additionally, frequency domain analysis aids in distinguishing various land-cover types, thereby enhancing the ability to perceive changes in land objects. Therefore, before downsampling the feature map, we used frequency domain information to generate the eFCA to assist the network in adjusting the saliency of different channels adaptively. This module enhances the representation of key information in the feature map and suppresses irrelevant interference, ensuring that the downsampled feature map retains as many critical details as possible.
For the feature extraction stage, we focused on the structural and textural features of the image, which are mainly reflected in the amplitude of the image frequency domain representation. To ensure computational efficiency and maintain the ability to perceive image information, we chose the DCT rather than the DFT.
Assuming the image f ( x , y ) size is M × N , Two-Dimensional (2D) DCT can be expressed as
F ( u , v ) = α ( u ) α ( v ) x = 0 M 1 y = 0 N 1 f ( x , y ) cos π ( 2 x + 1 ) u 2 M cos π ( 2 y + 1 ) v 2 N
where u and v are the 2D frequency components corresponding to f ( x , y ) , while α ( u ) and α ( v ) are normalizing constants. α ( u ) and α ( v ) have the following values:
α ( u ) = 1 M , u = 0 2 M , u 0 , α ( v ) = 1 N , v = 0 2 N , v 0
Inspired by FCANet [43], we designed an enhanced frequency channel attention (eFCA) module, as shown in Figure 2. In FCANet, the features are divided into several parts evenly along the channel dimension. Each part of the features uses only one specific set of frequency components to obtain the channel attention. This strategy of uniformly grouping features and considering only a single frequency component limits the integrity of the information. To address this limitation, we improved the approach by using multiple sets of frequency components to compute feature maps for all channels. We then integrated different frequency components through linear mapping to generate the channel attention. Thus, more comprehensive frequency information is introduced in the eFCA, enabling the channel attention to better capture the diversity and complexity of the feature details. The experimental comparison of the channel attention is shown in detail in Section 4.4.
Further considering the computational complexity, it is impractical to calculate all the M × N frequency components for each feature map. To ensure the stability of the results during training and inference, it is necessary to maintain consistent relative frequency positions across the DCTs of varying sizes. Therefore, in the implementation, a predefined list of n f groups of u and v values is established for a 7 × 7 DCT size. For DCT transforms of different sizes, corresponding scaling is performed to ensure the consistency of relative frequency positions. This can effectively incorporate important frequency information into the decision-making process of the model while avoiding too many redundant calculations or too much noise. The model is able to adaptively adjust the weights of these frequency components through learning according to different image features, giving priority to those frequency components that are conducive to matching.
Suppose F R C × H × W is the feature map in the networks, and u i , v i ( i [ 0 , n f 1 ] , i Z ) is one set of frequency components from the u and v lists. The DCT coefficients of these frequencies corresponding to the j-th ( j [ 0 , C 1 ] , j Z ) channel feature can be calculated using the above Equation (1). Therefore, the DCT coefficient of F at [ u i , v i ] can be expressed as
Freq i = DCT [ u i , v i ] ( F )
After obtaining Freq i R 1 × C , F is calculated at all the selected n f groups of frequency components to obtain Freq R n f × C . The channel attention Attn ch R C is obtained through linear mapping, which can be written as
Attn ch = Sigmoid ( FC ( ReLU ( FC ( Freq ) ) ) )
Each channel of F is scaled by the corresponding attention value:
F = Attn ch F
Through the above process, the eFCA integrates the rich frequency domain information from the feature map and assists the FPN in adjusting the saliency of different channels through learning, thereby preserving important features during convolutional downsampling.

3.2. DCT-Guided Coarse-Scale Matching

Before matching, we used sine and cosine functions of different frequencies to perform 2D position encoding [47] on the feature maps F ¯ A and F ¯ B obtained from the backbone. We then applied DCT-guided sparse attention for self-attention and cross-attention to further enhance both of them.
The advantages of attention-mechanism-enhanced features were demonstrated in numerous image matching works [48,49,50]. The input vectors of the attention layer are typically called the query ( Q ), key ( K ), and value ( V ). The attention operation computes the similarity between Q and each K to assign weights to the corresponding V , with the final output being the weighted sum of V . However, remote sensing images often have a high resolution and large size. Even at a coarse scale, the computational burden imposed by the vanilla attention [28] is still unbearable. Moreover, attention interactions involve many redundant calculations. Not all features are helpful for updating the current features. For example, interactions with regions with large temporal differences exhibiting large temporal differences can introduce unnecessary noise, hindering the effective update of current features.
To address these issues, this paper proposes a DCT-guided sparse attention (DSA) mechanism, which aims to use the frequency information of the feature map to measure the relevance and similarity between the current region and the other regions, generating attention guides that focus the attention on the most relevant features, thus reducing the scope of attention interaction and alleviating the computational burden. The reason DCT is chosen again here is that it concentrates the image’s energy in the low-frequency components, which are sufficient to represent the general information of the image. In this stage, this advantage enables DCT to effectively represent the window content, as it captures the approximate features of the window by analyzing only the low-frequency components so as to effectively assess the similarity. The flowchart of the DSA is shown as Figure 3. We took self-attention as an example to explain the DSA in detail.
Given a feature map X R H × W × C , it is first partitioned into n non-overlapping windows, each of size w c × w c , where
n = H w c × W w c
The query, key, and value representations for attention are obtained through linear projections:
Q = X W Q , K = X W K , V = X W V
where W Q , W K , W V R C × d are learnable projection matrices.
As shown in Figure 3, we first applied the 2D DCT to each window to compress it spatially:
Q ^ i = DCT m ( Q i ) ,
K ^ i = DCT m ( K i ) for i = 1 , , n
where DCT m ( · ) denotes the operation that retains m frequency components after the 2D DCT transformation [43]. This compression results in Q ^ i , K ^ i R m × d for the i-th window.
The compressed representations are then projected using a learnable projection matrix:
Q ˜ = Q ^ W P , K ˜ = K ^ W P
where W P R m × 1 is the projection matrix, resulting in Q ˜ , K ˜ R n × d .
The similarity matrix S R n × n between windows is computed using
S = Q ˜ K ˜ d
For each query window i, the top-k most similar key and value windows are selected:
Ω i = top k S i , :
For each pixel p in the i-th query window, its attention is computed with all pixels in the selected k key and value windows:
DSA i p ( Q , K , V ) = j Ω i q P softmax Q i p ( K j q ) d V j q
where p , q P denote pixel positions within a window, Q i p R 1 × d is the query feature at position p in the i-th window, and K j q , V j q R 1 × d are the key and value features at position q in the j-th window. This results in a sparse attention pattern where each query window (comprising w c × w c pixels) attends to k selected key and value windows (a total of k × w c × w c pixels).
Using DCT guidance, the attention is made sparse, which has a significant effect on reducing the amount of computation. The specific comparison of the computational complexities of vanilla attention and DSA are presented as follows:
1.
Vanilla attention:
  • Each query pixel attends to all H × W pixels.
  • Feature dimension: d.
  • Computational complexity: O ( H W ) 2 d .
2.
Our DCT-guided sparse attention (DSA):
  • Each query pixel attends to k windows, each containing w × w pixels.
  • Total number of windows: n = H W w 2 .
  • Selected windows per query: k n .
  • Pixels per window: w 2 .
  • Feature dimension: d.
  • Number of DCT basis (constant): m.
  • Computational complexity:
    DCT compression: O m H W d .
    Window similarity calculation: O n 2 d .
    Attention calculation:
    O n w 2 · k w 2 d = O H W w 2 · w 2 · k w 2 d = O H W k w 2 d .
    Total:
    O m H W d + n 2 d + H W k w 2 d = O ( n 2 + H W ( m + k w 2 ) ) d .
Therefore, the computational reduction ratio between the vanilla attention and the DSA can be calculated by
Reduction Ratio = ( H W ) 2 d ( n 2 + H W ( m + k w 2 ) ) d = ( H W ) 2 n 2 + H W ( m + k w 2 )
For a typical setting of our method:
  • Image size: H = W = 64 .
  • Window size: w = 8 .
  • Selected windows: k = 8 .
  • DCT basis: m = 4 .
The computational reduction ratio between the vanilla attention and the DSA can be calculated by
Reduction Ratio = ( 64 × 64 ) 2 64 2 + 64 × 64 × ( 8 × 8 2 + 4 ) 8 .
We employed the DSA introduced above to perform alternating self-attention and cross-attention on F ¯ A and F ¯ B for N C times to enhance the features. Self-attention expands the receptive field, enabling the integration of the most relevant information for current feature updates, thereby making the key features more distinctive. Cross-attention allows the model to perceive similarities and differences between the two images, reducing noise from temporal-difference areas and obtaining a more robust feature representation. The coarse-scale feature maps after enhancement are denoted as F ¯ A and F ¯ B . Assuming F ¯ A , F ¯ B R C ¯ × H ¯ × W ¯ , in order to measure the similarity, they are flattened and the similarity matrix S C is constructed using the dot product:
S C ( i , j ) = < F ¯ A ( i ) , F ¯ B ( j ) > , i , j ( 0 , H ¯ × W ¯ ]
in which F ¯ A ( i ) represents the i-th feature after F ¯ A is flattened, and F ¯ B ( j ) represents the j-th feature in F ¯ B . < · , · > denotes the dot product operation. The softmax function is applied to both the rows and columns of S C to derive the matching probability matrix, which can be expressed as follows:
P C ( i , j ) = Softmax r o w = i ( S C ) · Softmax c o l = j ( S C )
Matches with probabilities below the threshold θ C are filtered and further refined using mutual nearest neighbor (MNN) criteria. The coarse-scale matches set M C is obtained.

3.3. Fine-Scale Matching

As shown in Figure 1, F ˜ A and F ˜ B are first updated by
F ˜ A ˙ = FC ( [ F ˜ A | Upsampling ( FC ( F ¯ A ) ) ] ) ,
F ˜ B ˙ = FC ( [ F ˜ B | Upsampling ( FC ( F ¯ B ) ) ] )
in which [ · | · ] denotes the concentration and F ˜ A ˙ , F ˜ B ˙ R C ˜ × H ˜ × W ˜ . This update implicitly embeds the position encoding from the coarse scale into the fine scale, eliminating additional position encoding at this stage.
For every coarse match [ i ¯ c , j ¯ c ] in M c , the corresponding feature patches F ˜ c r o p A , F ˜ c r o p B R C ˜ × w f × w f are cropped with a window size of w f × w f from F ˜ A ˙ and F ˜ B ˙ . After obtaining a set of feature map patch pairs, the vanilla attention mechanism is employed to perform self-attention and cross-attention feature enhancement:
F ˜ c r o p A = CA ( SA ( F ˜ c r o p A ) ) , F ˜ c r o p B = CA ( SA ( F ˜ c r o p B ) )
in which CA ( · ) denotes the vanilla cross-attention and SA ( · ) denotes the vanilla self-attention. Since the feature maps are cropped at this time, the computational complexity of the vanilla attention is within an acceptable range.
Then, F ˜ c r o p A and F ˜ c r o p B are flattened, and denote that F ˜ c r o p A is centered on the i ˜ -th vector. To locate the matching keypoints precisely, F ˜ c r o p A ( i ˜ ) is taken as the query and the similarity with the dot product of all vectors in F ˜ c r o p B is calculated:
S F ( j ) = < F ˜ c r o p A ( i ˜ ) , F ˜ c r o p B ( j ) > , j ( 0 , w f × w f ]
A probability distribution can be obtained after applying softmax to S F , which indicates the possibility that each point in F ¯ c r o p B is the match point of F ˜ c r o p A ( i ˜ ) :
P F ( j ) = Softmax ( S F ( j ) )
By calculating the expectation of P F , the final match j ˜ in F ¯ c r o p B with sub-pixel accuracy on B can be obtained.

4. Experimental Results and Discussion

4.1. Experimental Setup

The experiments in this study were conducted on the Ubuntu 22.04 operating system using two NVIDIA GeForce RTX 3090 GPUs, each with 24 GB of memory. The Adam optimizer was used for the training, with an initial learning rate of 0.001. The learning rate was warmed up over the first three epochs, and a weight decay factor of 0.01 was applied. The momentum parameter was set to [ β 1 , β 2 ] = [ 0.9 , 0.999 ] , and the batch size was 4. The training proceeded for 50 epochs. The loss function followed the settings used in LoFTR, which comprised a coarse-scale cross-entropy loss and a fine-scale Euclidean distance loss.
In the feature extraction stage, we tended to select n f = 16 groups of frequency components that were verified to be the most beneficial to feature extraction accuracy [43]. In the coarse-scale matching, we selected m = 4 groups of the lowest frequency components, with the aim to balance the matching accuracy and computational efficiency. The window size was set to w c = 8 and the number of selected windows was k = 8 . We set N C = 2 . w c was selected based on the input size during the model training. For different input sizes, we preferred to keep the number of windows consistent rather than the window size. The feature dimension remained C ¯ = 256 during this stage. In the fine-scale matching, the window size was w f = 5 and the feature dimension remained C ˜ = 128 .

4.2. Datasets

The three datasets used in this study were the LEVIR-CD dataset [51], the DSIFN dataset [52], and the WHU building change detection dataset [53] (Figure 4). The LEVIR-CD dataset contained 637 bitemporal image pairs with a time span of 5 to 14 years, of which 445 image pairs were used for training, 64 image pairs were used for validation, and 128 image pairs were used for testing. The image pair size was 1024 × 1024. The DSIFN dataset contained high-resolution bitemporal image pairs of six cities in China collected from Google Earth. Five cities were cropped into 394 sub-image pairs as training sets, and one city was cropped into 48 sub-image pairs as test sets. The WHU dataset consisted of two high-resolution (32,507 × 15,354) TIFF optical remote sensing images. The WHU dataset was only used for training purposes and not for testing. Specifically, we cropped the pair of images vertically with a width of 512 pixels to create the validation set, and used the remaining portion of the image for training. For the validation set, we applied non-overlapping 512 × 512 pixel windows to generate the validation patches. In contrast, for the training set, we cropped overlapping 512 × 512 pixel patches from the remaining part of the image, which helped create a more diverse and comprehensive set of training examples. The whole process ensured that both the original images were cropped at the same locations, and thus, formed the corresponding image pairs. We cropped all the images to size 512 × 512 and expanded them by mirror flipping and rotating them by 90°, 180°, and 270° to form the training datasets.
The specific steps to generate a single training instance were as follows: First, a pair of images I A and I B were cropped from the same position of two images from the initial dataset, with I A as image A . Then, I B was randomly translated, scaled, and rotated within a certain range to form image B , and the corresponding homography transformation matrix H was calculated. This resulted in A and B as a pair of cross-temporal remote sensing images, with H as the label. Additionally, random changes, such as adjustments in brightness and contrast, were applied to the images. In each round of training, the image pairs input into the model were randomly generated, which effectively prevented the model from overfitting during training.

4.3. Matching Experiment Evaluation Metrics

This study used the following metrics to evaluate the matching experiments: success rate (SR), percentage of correct keypoints (PCK), average corner error (ACE), number of trainable parameters (Params), and running time for one pair of images (RT). The images to be matched are denoted as A and B. The matching set of the image pair is M = { ( k A , k B ) | k A A , k B B } . H is the homography transformation matrix obtained by applying the RANSAC [54] algorithm from A to B. The details of some metrics are as follows:
1.
The success rate is calculated by
S R = N s u c c e s s N t o t a l ,
in which N t o t a l denotes the total number of image pairs in the dataset and N s u c c e s s denotes the number of successfully matched image pairs. We considered image pairs with more than 10 matching point pairs and an ACE of less than 100 pixels as successful matches.
2.
The percentage of correct keypoints is calculated by
P C K = n c o r r e c t n t o t a l ,
in which n t o t a l represents the total number of point pairs in M and n c o r r e c t denote the number of correctly matched point pairs in M. The coordinates of the actual corresponding point of each k A in B are marked as k B g t . The Euclidean distance between k B g t and k B is calculated. If this distance is less than the threshold ε (set to 3 pixels), the keypoint pair is considered correctly matched. PCK can be calculated for each pair of images, and the average PCK across all pairs is used for comparison.
3.
The average corner error is calculated by
A C E = 1 N i = 1 N c i c i g t 2 ,
in which N denotes the number of corner points. We take the four vertices of the image as corner points, which means N = 4 . c i represents the position of the i-th corner point after the H transformation, and c i g t represents the true position of the i-th corner point after transformation. We compared the ACE average values of the successfully matched image pairs.

4.4. Ablation Study and Discussion

In order to prove the effectiveness of the eFCA in the backbone and the DSA in the coarse-scale matching module, we conducted ablation experiments. All models were trained on the same dataset, the optimizer settings were kept consistent, and the experimental results were measured on our validation set. The network without the eFCA and DSA was regarded as the baseline. Two attention mechanisms were added to the baseline and compared with our network (Experiment 4). The results are shown in Table 1, and more details are shown as box plots in Figure 5.
As shown in the table, the matching performance of the model was significantly influenced by these modules. The box plots show that the result distribution became notably more concentrated after incorporating attention mechanisms, indicating enhanced robustness and improved accuracy. It is evident that using frequency domain information to generate channel attention in multi-temporal remote sensing image matching contributed to obtaining richer detailed features. These two mechanisms effectively improved the accuracy. Overall, the DSA module had a greater impact on the matching effect than the eFCA module. The eFCA refined the importance of different feature channels by emphasizing relevant frequency components and suppressing irrelevant ones, which laid a solid foundation for the DSA to generate frequency-based attention guidance. The DSA used this frequency guidance to apply window attention, which enhanced the quality of the feature representation. In summary, the eFCA enhanced the quality of the feature extraction and overall feature representation, while the DSA further refined the feature representation by focusing on the most relevant regions of the image. Together, these modules improved both the global feature representation and local feature accuracy, ultimately boosting the model’s performance in image matching. Therefore, we believe that although the eFCA and DSA work independently on different stages, their combined use creates a synergistic effect.
In addition, we also compared the performance of the eFCA module with the original FCA module from the FCANet [43] article. The experimental results are presented in Table 2. As shown in this table, our method with the eFCA module significantly outperformed the one with original FCA module, particularly in terms of the ACE. Specifically, Experiment 4 achieved a reduction of 1.5 pixels compared with Experiment 5. This improvement can be attributed to the integration of richer frequency information by the eFCA. This enhancement allowed the model to better focus on key features, and thus, significantly improved the overall performance.
Furthermore, we also conducted comparative experiments between the DSA and the vanilla attention, and the results are shown in Table 3. Adding attention to the coarse-scale matching module significantly increased the trainable parameters. Compared with Experiment 2, the trainable parameters of Experiment 6 were almost doubled, but the attention brought about a greater performance improvement. The SR of Experiment 6 was 0.77 pixels higher than that of Experiment 4, but the comparisons of PCK and ACE were 1.88% and 0.43 pixels lower than Experiment 4, with a significantly longer inference time. The distribution of the results of Experiment 6, as shown in Figure 5, was more dispersed compared with Experiment 4. The experimental results verified that the global interaction of the features of the vanilla attention mechanism could improve the robustness of the model in complex scenes, but also introduced a large amount of redundant calculations. This led to an increased inference time and increased interference in the temporal difference areas during the attention allocation process, which negatively impacted the final accuracy of the model. In contrast, the DSA used frequency information to perceive the overall situation in advance and focused on regions with high feature similarity for the attention interaction, which greatly reduced the computational complexity and improved both the efficiency and practicality of the model.

4.5. Comparative Experiments and Discussion

In this section, we compare our methods with two categories of existing methods. The first category included matching methods based on manually designed features. We selected two prominent algorithms proposed in recent years: RIFT [12] and OFM [14]. The second category consisted of an end-to-end matching method based on deep learning. Super (SuperPoint [26] + SuperGlue [3]), LoFTR [4], and DeDoDe [55] were selected here. All methods were evaluated using the publicly available code and optimal parameter settings provided by the respective authors. The dataset and training settings were consistent with those used in this study.

4.5.1. Quantitative Comparison

Table 4 presents the quantitative evaluation results of the six algorithms on the DSIFN dataset. Figure 6a shows the proportion of images in the dataset that met different thresholds of the PCK. In the evaluation on the DSIFN dataset, the method proposed in this paper outperformed all the other algorithms in terms of the SR and matching accuracy. In comparison, the SR of the manually designed methods was lower than that of the deep learning-based methods, indicating that the scene robustness of the manually designed method was limited. This limitation arose because these methods rely on human prior knowledge and rules. When the changes in the image exceeded expected ranges, the manual methods often failed to match or were inaccurate. However, in their applicable scenarios, they could achieve impressive matching accuracies. The line chart demonstrates that in some images, the accuracy of the detected keypoints approached 1, a result that was challenging for the deep learning-based methods to achieve. Among the deep learning methods, the method proposed in this paper was the best. When all the images in the dataset were successfully matched, the average PCK was 81.92%, which was 5.19% ahead of the second-best approach. The average ACE after the homography was only 4.27 pixels, which was 2.46 pixels better than the second-best approach. In addition, it is worth noting that our method had the fewest trainable parameters and the fastest running speed.
Table 5 presents the quantitative evaluation results of the six algorithms on the LEVIR-CD dataset, while Figure 6b illustrates the detailed results of the PCK. From the perspective of the SR, all algorithms performed better on the LEVIR-CD dataset than the DSIFN dataset. This was because most image pairs in the LEVIR-CD dataset had prominent and invariant geometric structures (e.g., roads) that facilitated matching, whereas the DSIFN dataset had a larger element size and contained more significant variations in the ground objects, which made matching more challenging. In the comparison on the LEVIR-CD dataset, our method achieved the highest matching success rate. Although its PCK was 1.93% lower than LoFTR, and the ACE was 0.94 pixels higher, our method outperformed LoFTR in terms of the inference speed. The slight differences in the PCK and ACE could be attributed to the trade-off between the accuracy and computational efficiency. Our method was designed for scenes with temporal differences and adopted a DCT-guided sparse attention mechanism to enhance the computational efficiency while reducing the interference in temporal regions. The greater the difference in the ground objects, the more pronounced the advantage of our method, as confirmed by the comparison on the DSIFN dataset. In scenes with fewer temporal changes, the amount of interference information naturally decreased, and more feature interactions tended to lead to a higher accuracy. However, the limited number of attention windows in our method may constrain its performance in such cases. In contrast, LoFTR had a more complex network structure, particularly its global attention mechanism, which allowed it to capture more information. As a result, LoFTR achieved a slightly higher accuracy in scenes with fewer temporal changes between images. Nevertheless, the advantages of our method in scene adaptability and faster inference speed make it a more practical solution for applications that require high efficiency and scalability. In terms of the running time, the proposed method was 25.96 ms slower than Super. This was because our method is a dense matching method, and the running time is influenced by the regional similarity. While the computational efficiency was less prominent than that in DSIFN, our method remained the second fastest algorithm among all the compared methods on LEVIR-CD.
Figure 7 visualizes some of the quantitative results. It can be observed that the keypoints detected by the manually designed methods were significantly clustered. This clustering could lead to overfitting in these regions when calculating the homography transformation, which resulted in poor adaptability to other regions. This explains why the ACE of the manually designed method was low. In contrast, the keypoints detected by the deep learning-based methods were more dispersed, which enhanced their ability to handle complex and changing scenes, which made it more robust. This also explains why the SRs of the deep learning-based methods were higher than those of the manually designed methods. In addition, the keypoints detected by our method could effectively avoid the temporal difference area. Considering the results across the two datasets, our method achieved the highest SR, indicating superior robustness in multi-temporal scenarios. The high PCK and low ACE further demonstrated that our method provided accurate matching while maintaining a high reliability in successful matching. Our method was also lightweight and fast to infer.

4.5.2. Qualitative Comparison

To provide a comprehensive evaluation, we selected several representative image pairs from different datasets and conducted a qualitative analysis of all the algorithms. The multi-temporal scene changes were categorized into four types: seasonal changes, building changes, water changes, and natural disaster changes. The matching results are visualized in Figure 8. As shown in the figure, the performances of the RIFT and OFM methods were largely similar. Both algorithms achieved a high accuracy in scenes with minor changes, but they exhibited significant deviations in scenarios with large-scale building changes, water changes, or natural disasters (floods). Super could detect a certain number of matching points across all scenes, but it also produced more mismatches. LoFTR generally performed well in matching, though it struggled with large-scale building changes and water changes, which led to more mismatched points. DeDoDe, on the other hand, showed substantial errors and could only successfully match seasonal changes and natural disasters (forest fires). In contrast, our method effectively avoided regions with significant changes during the keypoint detection across all four scenes, which resulted in a higher accuracy. This qualitative analysis aligned with the quantitative results and validated the strong robustness and high precision of our proposed method.

5. Conclusions

To address the challenges posed by surface differences in multi-temporal remote sensing image matching, this paper proposes a remote sensing image matching network that combines a Discrete Cosine Transform (DCT) and attention mechanisms for multi-scale feature matching. The aim was to improve the robustness and accuracy of image matching in multi-temporal scenarios. The innovation of this network lies primarily in the enhanced frequency channel attention mechanism and the frequency-guided sparse attention mechanism. The former is used for multi-scale feature extraction, where the DCT is employed to integrate the frequency domain information of the feature map, forming weights that enhance the expression of key information in the feature map. The latter is used for coarse-scale matching, where the DCT transforms the feature map to generate attention guides, suppressing noise caused by temporal variations while reducing the computational burden. The experimental results show that our method outperformed the existing methods in terms of the inference speed and accuracy, and was more suitable for matching multi-temporal remote sensing images. This study mainly focused on the significant differences in ground objects between multi-temporal remote sensing images. However, the differences in multi-temporal scenarios may arise not only from changes in ground objects but also from day–night variations and sensor differences. In future work, we will explore the impacts of these additional factors to gain a deeper understanding of the task, enrich the dataset, and further improve the model structure to better address these challenges.

Author Contributions

Conceptualization, Q.Z. and B.H.; methodology, Q.Z. and B.H.; software, Q.Z.; validation, Q.Z.; formal analysis, Q.Z. and B.H.; investigation, Q.Z.; resources, B.H., Z.L., Z.X. and M.H.; data curation, Q.Z.; writing—original draft preparation, Q.Z.; writing—review and editing, Q.Z. and B.H.; visualization, Q.Z.; supervision, B.H.; funding acquisition, Z.L., Z.X. and M.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Defense Pre-Research Foundation of China [2023-JCJQ-JJ-0163].

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data that support the findings of this study are available from the corresponding author upon reasonable request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Harris, C.; Stephens, M. A combined corner and edge detector. In Proceedings of the Alvey Vision Conference, Citeseer, Manchester, UK, 31 August – 2 September 1988; Volume 15, pp. 10–5244. [Google Scholar]
  2. Lowe, D.G. Distinctive image features from scale-invariant keypoints. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
  3. Sarlin, P.E.; DeTone, D.; Malisiewicz, T.; Rabinovich, A. Superglue: Learning feature matching with graph neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 4938–4947. [Google Scholar]
  4. Sun, J.; Shen, Z.; Wang, Y.; Bao, H.; Zhou, X. LoFTR: Detector-free local feature matching with transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 8922–8931. [Google Scholar]
  5. Turkar, V.; Deo, R.; Rao, Y.S.; Mohan, S.; Das, A. Classification Accuracy of Multi-Frequency and Multi-Polarization SAR Images for Various Land Covers. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2012, 5, 936–941. [Google Scholar] [CrossRef]
  6. Tondewad, M.P.S.; Dale, M.M.P. Remote Sensing Image Registration Methodology: Review and Discussion. Procedia Comput. Sci. 2020, 171, 2390–2399. [Google Scholar] [CrossRef]
  7. Xinghua, L.; Wenhao, A.; Ruitao, F.; Shaojie, L. Survey of remote sensing image registration based on deep learning. Natl. Remote Sens. Bull. 2023, 27, 267–284. [Google Scholar]
  8. Zhu, B.; Zhou, L.; Pu, S.; Fan, J.; Ye, Y. Advances and challenges in multimodal remote sensing image registration. IEEE J. Miniaturization Air Space Syst. 2023, 4, 165–174. [Google Scholar] [CrossRef]
  9. Fu, Z.; Zhang, J.; Tang, B.H. Multi-Temporal Snow-Covered Remote Sensing Image Matching via Image Transformation and Multi-Level Feature Extraction. Optics 2024, 5, 392–405. [Google Scholar] [CrossRef]
  10. Bay, H.; Tuytelaars, T.; Van Gool, L. Surf: Speeded up robust features. In Proceedings of the 2006 European Conference on Computer Vision (ECCV), Graz, Austria, 7–13 May 2006; pp. 404–417. [Google Scholar]
  11. Rublee, E.; Rabaud, V.; Konolige, K.; Bradski, G. ORB: An efficient alternative to SIFT or SURF. In Proceedings of the 2011 International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011; pp. 2564–2571. [Google Scholar]
  12. Li, J.; Hu, Q.; Ai, M. RIFT: Multi-modal image matching based on radiation-variation insensitive feature transform. IEEE Trans. Image Process. 2019, 29, 3296–3310. [Google Scholar] [CrossRef] [PubMed]
  13. Rasmy, L.; Sebari, I.; Ettarid, M. Automatic sub-pixel co-registration of remote sensing images using phase correlation and Harris detector. Remote Sens. 2021, 13, 2314. [Google Scholar] [CrossRef]
  14. Fan, Z.; Wang, M.; Pi, Y.; Liu, Y.; Jiang, H. A robust oriented filter-based matching method for multisource, multitemporal remote sensing images. IEEE Trans. Geosci. Remote Sens. 2023, 61, 4703316. [Google Scholar] [CrossRef]
  15. Yang, Z.; Dan, T.; Yang, Y. Multi-temporal remote sensing image registration using deep convolutional features. IEEE Access 2018, 6, 38544–38555. [Google Scholar] [CrossRef]
  16. Liu, J.; Li, Y.; Chen, Y. Multi-temporal remote sensing image registration based on siamese network. In Proceedings of the 2021 International Conference on Computer Engineering and Application (ICCEA), Kunming, China, 25–27 June 2021; pp. 333–337. [Google Scholar]
  17. Rocco, I.; Cimpoi, M.; Arandjelović, R.; Torii, A.; Pajdla, T.; Sivic, J. Neighbourhood consensus networks. Adv. Neural Inf. Process. Syst. 2018, 31, 1658–1669. [Google Scholar]
  18. Li, X.; Han, K.; Li, S.; Prisacariu, V. Dual-resolution correspondence networks. Adv. Neural Inf. Process. Syst. 2020, 33, 17346–17357. [Google Scholar]
  19. Zhou, Q.; Sattler, T.; Leal-Taixe, L. Patch2pix: Epipolar-guided pixel-level correspondences. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 4669–4678. [Google Scholar]
  20. Xu, Y.; Li, J.; Du, C.; Chen, H. Nbr-net: A nonrigid bidirectional registration network for multitemporal remote sensing images. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5620715. [Google Scholar] [CrossRef]
  21. Zhang, J.; Zhao, S.; Li, B. Selective Context Network With Neighbourhood Consensus for Aerial Image Registration. In Proceedings of the 2022 6th International Conference on Computer Science and Artificial Intelligence, Beijing, China, 9–11 December 2022; pp. 258–264. [Google Scholar]
  22. Han, X.; Leung, T.; Jia, Y.; Sukthankar, R.; Berg, A.C. Matchnet: Unifying feature and metric learning for patch-based matching. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3279–3286. [Google Scholar]
  23. Yi, K.M.; Trulls, E.; Lepetit, V.; Fua, P. Lift: Learned invariant feature transform. In Proceedings of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Proceedings, Part VI 14. Springer: Berlin/Heidelberg, Germany, 2016; pp. 467–483. [Google Scholar]
  24. Ono, Y.; Trulls, E.; Fua, P.; Yi, K.M. LF-Net: Learning local features from images. Adv. Neural Inf. Process. Syst. 2018, 31, 6237–6247. [Google Scholar]
  25. Georgakis, G.; Karanam, S.; Wu, Z.; Ernst, J.; Košecká, J. End-to-end learning of keypoint detector and descriptor for pose invariant 3D matching. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 1965–1973. [Google Scholar]
  26. DeTone, D.; Malisiewicz, T.; Rabinovich, A. SuperPoint: Self-Supervised Interest Point Detection and Description. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Salt Lake City, UT, USA, 18–22 June 2018; pp. 337–33712. [Google Scholar] [CrossRef]
  27. Dusmanu, M.; Rocco, I.; Pajdla, T.; Pollefeys, M.; Sivic, J.; Torii, A.; Sattler, T. D2-net: A trainable cnn for joint description and detection of local features. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 8092–8101. [Google Scholar]
  28. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30, 5998–6008. [Google Scholar]
  29. Dosovitskiy, A. An image is worth 16 × 16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  30. Chen, H.; Luo, Z.; Zhou, L.; Tian, Y.; Zhen, M.; Fang, T.; Mckinnon, D.; Tsin, Y.; Quan, L. Aspanformer: Detector-free image matching with adaptive span transformer. In Proceedings of the European Conference on Computer Vision (ECCV), Tel Aviv, Israel, 23–27 October 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 20–36. [Google Scholar]
  31. Tan, D.; Liu, J.J.; Chen, X.; Chen, C.; Zhang, R.; Shen, Y.; Ding, S.; Ji, R. Eco-tr: Efficient correspondences finding via coarse-to-fine refinement. In Proceedings of the European Conference on Computer Vision (ECCV), Tel Aviv, Israel, 23–27 October 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 317–334. [Google Scholar]
  32. Wang, Q.; Zhang, J.; Yang, K.; Peng, K.; Stiefelhagen, R. Matchformer: Interleaving attention in transformers for feature matching. In Proceedings of the Asian Conference on Computer Vision, Macau SAR, China, 4–8 December 2022; pp. 2746–2762. [Google Scholar]
  33. Rubel, O.; Rubel, A.; Lukin, V.; Egiazarian, K. Blind DCT-based prediction of image denoising efficiency using neural networks. In Proceedings of the 2018 7th European Workshop on Visual Information Processing (EUVIP), Tampere, Finland, 26–28 November 2018; pp. 1–6. [Google Scholar]
  34. Herbreteau, S.; Kervrann, C. DCT2net: An interpretable shallow CNN for image denoising. IEEE Trans. Image Process. 2022, 31, 4292–4305. [Google Scholar] [CrossRef]
  35. Karaoğlu, H.H.; Ekşioğlu, E.M. Revisiting DCT in Deep Learning Era: An Initial Denoising Application. In Proceedings of the 2024 32nd Signal Processing and Communications Applications Conference (SIU), Mersin, Turkiye, 15–18 May 2024; pp. 1–4. [Google Scholar]
  36. Raid, A.; Khedr, W.; El-Dosuky, M.A.; Ahmed, W. Jpeg image compression using discrete cosine transform—A survey. arXiv 2014, arXiv:1405.6147. [Google Scholar]
  37. Xue, J.; Yin, L.; Lan, Z.; Long, M.; Li, G.; Wang, Z.; Xie, X. 3D DCT based image compression method for the medical endoscopic application. Sensors 2021, 21, 1817. [Google Scholar] [CrossRef] [PubMed]
  38. Peng, Y.; Fu, C.; Cao, G.; Song, W.; Chen, J.; Sham, C.W. JPEG-compatible Joint Image Compression and Encryption Algorithm & nbsp;with File Size Preservation. ACM Trans. Multimed. Comput. Commun. Appl. 2024, 20, 1–20. [Google Scholar] [CrossRef]
  39. Duan, C.; Hu, B.; Liu, W.; Ma, T.; Ma, Q.; Wang, H. Infrared small target detection method based on frequency domain clutter suppression and spatial feature extraction. IEEE Access 2023, 11, 85549–85560. [Google Scholar] [CrossRef]
  40. Xu, Y.; Nakayama, H. Dct-based fast spectral convolution for deep convolutional neural networks. In Proceedings of the 2021 International Joint Conference on Neural Networks (IJCNN), Shenzhen, China, 18–22 July 2021; pp. 1–8. [Google Scholar]
  41. Xu, R.; Kang, X.; Li, C.; Chen, H.; Ming, A. DCT-FANet: DCT based frequency attention network for single image super-resolution. Displays 2022, 74, 102220. [Google Scholar] [CrossRef]
  42. Ghosh, A.; Chellappa, R. Deep feature extraction in the DCT domain. In Proceedings of the 2016 23rd International Conference on Pattern Recognition (ICPR), Cancun, Mexico, 4–8 December 2016; pp. 3536–3541. [Google Scholar]
  43. Qin, Z.; Zhang, P.; Wu, F.; Li, X. Fcanet: Frequency channel attention networks. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 783–792. [Google Scholar]
  44. Chaudhury, S.; Yamasaki, T. Adversarial Robustness of Convolutional Models Learned in the Frequency Domain. In Proceedings of the ICASSP 2024—2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Seoul, Republic of Korea, 14–19 April 2024; pp. 7455–7459. [Google Scholar]
  45. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  46. Minaee, S.; Boykov, Y.; Porikli, F.; Plaza, A.; Kehtarnavaz, N.; Terzopoulos, D. Image Segmentation Using Deep Learning: A Survey. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 3523–3542. [Google Scholar] [CrossRef]
  47. Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End-to-End Object Detection with Transformers. In Proceedings of the European Conference on Computer Vision (ECCV), Glasgow, UK, 23–28 August 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 213–229. [Google Scholar] [CrossRef]
  48. Hu, Y.; Liu, Y.; Hui, B. Combining OpenStreetMap with Satellite Imagery to Enhance Cross-View Geo-Localization. Sensors 2025, 25, 44. [Google Scholar] [CrossRef] [PubMed]
  49. Quan, D.; Wang, S.; Li, Y.; Yang, B.; Huyan, N.; Chanussot, J.; Hou, B.; Jiao, L. Multi-relation attention network for image patch matching. IEEE Trans. Image Process. 2021, 30, 7127–7142. [Google Scholar] [CrossRef] [PubMed]
  50. Liu, M.; Zhou, G.; Ma, L.; Li, L.; Mei, Q. SIFNet: A self-attention interaction fusion network for multisource satellite imagery template matching. Int. J. Appl. Earth Obs. Geoinf. 2023, 118, 103247. [Google Scholar] [CrossRef]
  51. Chen, H.; Shi, Z. A Spatial-Temporal Attention-Based Method and a New Dataset for Remote Sensing Image Change Detection. Remote Sens. 2020, 12, 1662. [Google Scholar] [CrossRef]
  52. Zhang, C.; Yue, P.; Tapete, D.; Jiang, L.; Shangguan, B.; Huang, L.; Liu, G. A deeply supervised image fusion network for change detection in high resolution bi-temporal remote sensing images. ISPRS J. Photogramm. Remote Sens. 2020, 166, 183–200. [Google Scholar] [CrossRef]
  53. Ji, S.; Wei, S.; Lu, M. Fully Convolutional Networks for Multisource Building Extraction From an Open Aerial and Satellite Imagery Data Set. IEEE Trans. Geosci. Remote Sens. 2019, 57, 574–586. [Google Scholar] [CrossRef]
  54. Fischler, M.A.; Bolles, R.C. Random sample consensus: A paradigm for model fitting with applications to image analysis and automated cartography. Commun. ACM 1981, 24, 381–395. [Google Scholar] [CrossRef]
  55. Edstedt, J.; Bökman, G.; Wadenbäck, M.; Felsberg, M. DeDoDe: Detect, Don’t Describe—Describe, Don’t Detect for Local Feature Matching. In Proceedings of the 2024 International Conference on 3D Vision (3DV), Davos, Switzerland, 18–21 March 2024. [Google Scholar]
Figure 1. Overall framework of the proposed method.
Figure 1. Overall framework of the proposed method.
Sensors 25 01345 g001
Figure 2. The details in the feature extraction module. (a) The structure of the blocks. BN is short for Batch Normalization. (b) The structure of the enhanced frequency channel attention (eFCA).
Figure 2. The details in the feature extraction module. (a) The structure of the blocks. BN is short for Batch Normalization. (b) The structure of the enhanced frequency channel attention (eFCA).
Sensors 25 01345 g002
Figure 3. The flowchart of DCT-guided sparse attention. For self-attention, Q , K , and V all come from the feature map to be updated. For cross-attention, Q comes from the feature map to be updated, while K and V come from the other one.
Figure 3. The flowchart of DCT-guided sparse attention. For self-attention, Q , K , and V all come from the feature map to be updated. For cross-attention, Q comes from the feature map to be updated, while K and V come from the other one.
Sensors 25 01345 g003
Figure 4. Examples of the datasets [51,52,53].
Figure 4. Examples of the datasets [51,52,53].
Sensors 25 01345 g004
Figure 5. Box plots of ablation experiment results. The dots represent outliers of the results, and the orange lines indicate the median. (a) Box plot of PCK results. (b) Box plot of ACE results.
Figure 5. Box plots of ablation experiment results. The dots represent outliers of the results, and the orange lines indicate the median. (a) Box plot of PCK results. (b) Box plot of ACE results.
Sensors 25 01345 g005
Figure 6. The line charts show the proportion of results that met different PCK thresholds. The x-axis represents the set PCK thresholds, and the y-axis represents the proportion of images in the dataset that satisfied each threshold. (a) Results on the DSIFN dataset. (b) Results on the LEVIR-CD dataset.
Figure 6. The line charts show the proportion of results that met different PCK thresholds. The x-axis represents the set PCK thresholds, and the y-axis represents the proportion of images in the dataset that satisfied each threshold. (a) Results on the DSIFN dataset. (b) Results on the LEVIR-CD dataset.
Sensors 25 01345 g006
Figure 7. Visualization of some results from the DSIFN dataset and the LEVIR-CD dataset. The two groups on the left were from the DSIFN dataset, while the two groups on the right were from the LEVIR-CD dataset. The green dots indicate correctly matched keypoints, and the red dots indicate incorrectly matched keypoints.
Figure 7. Visualization of some results from the DSIFN dataset and the LEVIR-CD dataset. The two groups on the left were from the DSIFN dataset, while the two groups on the right were from the LEVIR-CD dataset. The green dots indicate correctly matched keypoints, and the red dots indicate incorrectly matched keypoints.
Sensors 25 01345 g007
Figure 8. This is a visualization of qualitative comparison, with correct matching keypoints represented in green and incorrect matching keypoints represented in red.
Figure 8. This is a visualization of qualitative comparison, with correct matching keypoints represented in green and incorrect matching keypoints represented in red.
Sensors 25 01345 g008
Table 1. Ablation experiment evaluation results.
Table 1. Ablation experiment evaluation results.
NoApproachParamsSR (%) ↑PCK > 80 (%) ↑ACE (pix) ↓
1Baseline4.09 M89.0683.336.92
2Baseline + eFCA4.10 M93.7589.945.25
3Baseline + DSA5.69 M96.0985.764.08
4Baseline + eFCA + DSA (ours)5.70 M98.4494.442.23
↑: A higher metric value indicates better performance. ↓: A lower metric value indicates better performance.
Table 2. Comparison of different channel attention effects.
Table 2. Comparison of different channel attention effects.
NoApproachParamsSR (%) ↑PCK > 80 (%) ↑ACE (pix) ↓
4Baseline + eFCA + DSA (ours)5.70 M98.4494.442.23
5Baseline + FCA + DSA5.70 M96.8690.583.73
↑: A higher metric value indicates better performance. ↓: A lower metric value indicates better performance.
Table 3. Comparison of different attentions in coarse-scale matching effects.
Table 3. Comparison of different attentions in coarse-scale matching effects.
NoApproachParamsSR (%) ↑PCK > 80 (%) ↑ACE (pix) ↓RT (ms) ↓
4Baseline + eFCA + DSA (ours)5.70 M98.4494.442.2366.39
6Baseline + eFCA + VanillaAttn8.05 M99.2192.562.66108.24
↑: A higher metric value indicates better performance. ↓: A lower metric value indicates better performance.
Table 4. Results of comparisons on DSIFN datasets.
Table 4. Results of comparisons on DSIFN datasets.
ApproachParamsSR (%) ↑PCK (%) ↑ACE (pix) ↓RT (ms) ↓
RIFT-87.5072.0313.41-
OFM-70.8362.0310.62-
Super13.32 M89.5858.0511.1491.43
LoFTR11.56 M97.9276.736.73193.05
DeDoDe13.52 M85.4264.8710.67155.72
Ours5.70 M10081.924.2770.90
↑: A higher metric value indicates better performance. ↓: A lower metric value indicates better performance.
Table 5. Results of comparisons on LEVIR-CD dataset.
Table 5. Results of comparisons on LEVIR-CD dataset.
ApproachParamsSR (%) ↑PCK (%) ↑ACE (pix) ↓RT (ms) ↓
RIFT-90.6370.4710.48-
OFM-92.1883.725.92-
Super13.32 M98.4475.955.78172.43
LoFTR11.56 M98.4490.412.04656.35
DeDoDe13.52 M89.0672.368.27406.30
Ours5.70 M99.2288.482.98198.39
↑: A higher metric value indicates better performance. ↓: A lower metric value indicates better performance.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zeng, Q.; Hui, B.; Liu, Z.; Xu, Z.; He, M. A Method Combining Discrete Cosine Transform with Attention for Multi-Temporal Remote Sensing Image Matching. Sensors 2025, 25, 1345. https://doi.org/10.3390/s25051345

AMA Style

Zeng Q, Hui B, Liu Z, Xu Z, He M. A Method Combining Discrete Cosine Transform with Attention for Multi-Temporal Remote Sensing Image Matching. Sensors. 2025; 25(5):1345. https://doi.org/10.3390/s25051345

Chicago/Turabian Style

Zeng, Qinyan, Bin Hui, Zhaoji Liu, Zheng Xu, and Miao He. 2025. "A Method Combining Discrete Cosine Transform with Attention for Multi-Temporal Remote Sensing Image Matching" Sensors 25, no. 5: 1345. https://doi.org/10.3390/s25051345

APA Style

Zeng, Q., Hui, B., Liu, Z., Xu, Z., & He, M. (2025). A Method Combining Discrete Cosine Transform with Attention for Multi-Temporal Remote Sensing Image Matching. Sensors, 25(5), 1345. https://doi.org/10.3390/s25051345

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop