[go: up one dir, main page]

Next Article in Journal
Unleashing the Power of Contrastive Learning for Zero-Shot Video Summarization
Previous Article in Journal
Leveraging Perspective Transformation for Enhanced Pothole Detection in Autonomous Vehicles
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Hybrid Approach for Image Acquisition Methods Based on Feature-Based Image Registration

1
Department of Computer Science Engineering and Application, Sambalpur University Institute of Information Technology (SUIIT), Burla, Sambalpur 768018, India
2
School of Computer Sciences, Veer Surendra Sai University of Technology (VSSUT), Burla, Sambalpur 768018, India
3
Department of Digital Systems, University of Thessaly, 41500 Larissa, Greece
4
Department of Informatics, Ionian University, 49100 Corfu, Greece
5
Department of Computer Engineering-AI, Marwadi University, Rajkot 360003, India
6
Department of Planning and Regional Development, University of Thessaly, 38334 Volos, Greece
*
Authors to whom correspondence should be addressed.
J. Imaging 2024, 10(9), 228; https://doi.org/10.3390/jimaging10090228
Submission received: 6 June 2024 / Revised: 6 September 2024 / Accepted: 11 September 2024 / Published: 14 September 2024
(This article belongs to the Section Image and Video Processing)
Figure 1
<p>Flow diagram of the proposed methodology.</p> ">
Figure 2
<p>Diagonal approach for hybrid feature-detection method.</p> ">
Figure 3
<p>Sampled color images from AID database [<a href="#B46-jimaging-10-00228" class="html-bibr">46</a>].</p> ">
Figure 4
<p>Grayscale conversion of sampled color images from AID database [<a href="#B46-jimaging-10-00228" class="html-bibr">46</a>].</p> ">
Figure 5
<p>Various rotation angles applied on park and railway station grayscale aerial images.</p> ">
Figure 6
<p>Scaling transformations applied to VSSUT gate and Hirakud dam images. (<b>a</b>) 0.7 scaling factor on VSSUT gate. (<b>b</b>) 0.7 scaling factor on Hirakud dam. (<b>c</b>) 2.0 scaling factor on VSSUT gate. (<b>d</b>) 2.0 scaling factor on Hirakud dam.</p> ">
Figure 7
<p>Detection of feature keypoints in the park image under <math display="inline"><semantics> <msup> <mn>150</mn> <mo>∘</mo> </msup> </semantics></math> rotation, showcasing the performance of different detectors. Green markers highlight the keypoints detected, with each subfigure corresponding to the output using a different feature-detection method.</p> ">
Figure 8
<p>Detection of feature keypoints in the railway station image under <math display="inline"><semantics> <msup> <mn>150</mn> <mo>∘</mo> </msup> </semantics></math> rotation, showcasing the performance of different detectors. Green markers indicate the keypoints, and each subfigure corresponds to the output using a different feature-detection method.</p> ">
Figure 9
<p>Extraction of feature keypoints from the park image under <math display="inline"><semantics> <msup> <mn>150</mn> <mo>∘</mo> </msup> </semantics></math> rotation. Green markers demonstrate the keypoints extracted, emphasizing the nuances of each algorithm with each subfigure showing results using a different feature extraction method.</p> ">
Figure 10
<p>Extraction of feature keypoints from the railway station image under <math display="inline"><semantics> <msup> <mn>150</mn> <mo>∘</mo> </msup> </semantics></math> rotation. Each subfigure demonstrates the results using a different feature extraction method, with green markers used to emphasize keypoint locations and algorithmic nuances.</p> ">
Figure 11
<p>Matching of feature keypoints in the park image across different rotational views under <math display="inline"><semantics> <msup> <mn>150</mn> <mo>∘</mo> </msup> </semantics></math> rotation. Subfigures (<b>a</b>–<b>f</b>) display the matched keypoints separately to illustrate individual detector performance clearly. Subfigure (<b>g</b>) shows an overlaid result of the hybrid detector to demonstrate the integration of multiple detection outcomes, providing a comprehensive view of the keypoints matched by the proposed method. Each image aims to highlight the effectiveness of each feature detector in achieving consistent matching across transformations.</p> ">
Figure 12
<p>Matching of feature keypoints in the railway station image across different rotational views under <math display="inline"><semantics> <msup> <mn>150</mn> <mo>∘</mo> </msup> </semantics></math> rotation. Each subfigure highlights the effectiveness of each feature detector in achieving consistent matching.</p> ">
Figure 13
<p>Sequential presentation of detection, extraction, and matching phases for various feature detectors on two sets of airport aerial images. Each row represents a different detector and showcases the process from detection to matching.</p> ">
Figure 14
<p>Sequential presentation of detection, extraction, and matching phases for various feature detectors on two sets of bridge aerial images. Each row represents a different detector and showcases the process from detection to matching.</p> ">
Figure 15
<p>Comparison of feature-detection performance using MSER, BRISK, and Hybrid detectors on two different images under scaling transformations. Each row demonstrates the response of the detectors at scaling factors of the original, 0.7, and 2.0, highlighting the adaptability of these algorithms to changes in image scale. (<b>a</b>) MSER: VSSUT Gate Image. (<b>b</b>) BRISK: VSSUT Gate Image. (<b>c</b>) MSER: HD Image. (<b>d</b>) BRISK: HD Image. (<b>e</b>) Hybrid: VSSUT Gate Image. (<b>f</b>) Hybrid: HD Image. (<b>g</b>) MSER: VSSUT Gate Image, Scale 0.7. (<b>h</b>) BRISK: VSSUT Gate Image, Scale 0.7. (<b>i</b>) MSER: HD Image, Scale 0.7. (<b>j</b>) BRISK: HD Image, Scale 0.7. (<b>k</b>) Hybrid: VSSUT Gate Image, Scale 0.7. (<b>l</b>) Hybrid: HD Image, Scale 0.7. (<b>m</b>) MSER: VSSUT Gate Image, Scale 2.0. (<b>n</b>) BRISK: VSSUT Gate Image, Scale 2.0. (<b>o</b>) MSER: HD Image, Scale 2.0. (<b>p</b>) BRISK: HD Image, Scale 2.0. (<b>q</b>) Hybrid: VSSUT Gate Image, Scale 2.0. (<b>r</b>) Hybrid: HD Image, Scale 2.0.</p> ">
Figure 16
<p>Extraction of feature keypoints using various extractors based on scaling factors of <math display="inline"><semantics> <mrow> <mn>0.7</mn> </mrow> </semantics></math> and <math display="inline"><semantics> <mrow> <mn>2.0</mn> </mrow> </semantics></math>. Each row demonstrates the impact of scaling on the effectiveness of feature extraction across different images and detectors. (<b>a</b>) MSER: VSSUT Gate Image. (<b>b</b>) BRISK: VSSUT Gate Image. (<b>c</b>) MSER: HD Image. (<b>d</b>) BRISK: HD Image. (<b>e</b>) Hybrid: VSSUT Gate Image. (<b>f</b>) Hybrid: HD Image. (<b>g</b>) MSER: VSSUT Gate Image, Scale 0.7. (<b>h</b>) BRISK: VSSUT Gate Image, Scale 0.7. (<b>i</b>) MSER: HD Image, Scale 0.7. (<b>j</b>) BRISK: HD Image, Scale 0.7. (<b>k</b>) Hybrid: VSSUT Gate Image, Scale 0.7. (<b>l</b>) Hybrid: HD Image, Scale 0.7. (<b>m</b>) MSER: VSSUT Gate Image, Scale 2.0. (<b>n</b>) BRISK: VSSUT Gate Image, Scale 2.0. (<b>o</b>) MSER: HD Image, Scale 2.0. (<b>p</b>) BRISK: HD Image, Scale 2.0. (<b>q</b>) Hybrid: VSSUT Gate Image, Scale 2.0. (<b>r</b>) Hybrid: HD Image, Scale 2.0.</p> ">
Figure 17
<p>Matching of feature keypoints using various detectors on VSSUT gate and Hirakud dam images under two scaling factors, 0.7 and 2.0. Each image series demonstrates the effect of scaling on feature matching performance. (<b>a</b>) MSER: VSSUT Gate Image, Scale 0.7. (<b>b</b>) BRISK: VSSUT Gate Image, Scale 0.7. (<b>c</b>) MSER: Hirakud Dam Image, Scale 0.7. (<b>d</b>) BRISK: Hirakud Dam Image, Scale 0.7. (<b>e</b>) Hybrid: VSSUT Gate Image, Scale 0.7. (<b>f</b>) Hybrid: Hirakud Dam Image, Scale 0.7. (<b>g</b>) MSER: VSSUT Gate Image, Scale 2.0. (<b>h</b>) BRISK: VSSUT Gate Image, Scale 2.0. (<b>i</b>) MSER: Hirakud Dam Image, Scale 2.0. (<b>j</b>) BRISK: Hirakud Dam Image, Scale 2.0. (<b>k</b>) Hybrid: VSSUT Gate Image, Scale 2.0. (<b>l</b>) Hybrid: Hirakud Dam Image, Scale 2.0.</p> ">
Figure 18
<p>Registered images of different scenes using the Hybrid feature detector. Each subfigure shows a different aerial or scene image, highlighting the detailed synthesis achieved through the registration process.</p> ">
Figure 19
<p>Performance comparison of various feature detectors on park scene images. Each subplot visually demonstrates how each feature detector identifies keypoints within the same environmental setting. This provides insights into the adaptability and precision of each method under similar conditions, highlighting their strengths and limitations in detecting significant image features effectively.</p> ">
Versions Notes

Abstract

:
This paper presents a novel hybrid approach to feature detection designed specifically for enhancing Feature-Based Image Registration (FBIR). Through an extensive evaluation involving state-of-the-art feature detectors such as BRISK, FAST, ORB, Harris, MinEigen, and MSER, the proposed hybrid detector demonstrates superior performance in terms of keypoint detection accuracy and computational efficiency. Three image acquisition methods (i.e., rotation, scene-to-model, and scaling transformations) are considered in the comparison. Applied across a diverse set of remote-sensing images, the proposed hybrid approach has shown marked improvements in match points and match rates, proving its effectiveness in handling varied and complex imaging conditions typical in satellite and aerial imagery. The experimental results have consistently indicated that the hybrid detector outperforms conventional methods, establishing it as a valuable tool for advanced image registration tasks.

1. Introduction

Image registration is the process of aligning multiple scene images into a single, integrated image. This technique addresses common issues such as image rotation, scale, and skew, which often arise when overlaying multiple images. The primary goal of image registration is to automatically establish correspondence between different images, a crucial step for further processing in various applications [1]. These images may be acquired at different times, from different devices, or may produce different types of information. Image-registration methods can be classified into two main types: area-based approaches and feature-based approaches. Area-based methods compare intensity patterns in images through correlation matrices, while feature-based methods establish correspondence between different image features such as lines, points, and contours. Feature-based methods are generally more reliable than area-based methods but require complex calculations to establish correspondence between the source image and the target image.
The process of image registration has extensive applications in diverse fields such as medical imaging, remote sensing, and computer vision. In medical imaging, for example, image registration is critical for combining data from different imaging modalities (such as CT and MRI) to provide comprehensive information about a patient’s anatomy. In remote sensing, it enables the integration of images taken from different sensors and at different times to monitor environmental changes. In computer vision, it is essential for tasks such as object recognition and 3D reconstruction. The robustness and accuracy of the registration process directly impact the effectiveness of these applications, making the development of efficient registration algorithms a significant research focus [2].
Feature-Based Image Registration (FBIR) consists of four steps: feature detection, feature matching, transform model estimation, and image resampling and transformation. In the first step, features are detected in both the source and target images. These features can include regions, contours, edges, and corners [3]. In the feature-matching step, it is determined whether pixels from the source image correspond to pixels from the target image. If correspondence is established, matching is performed. In the transform model estimation step, a mapping function is built, with types and parameters estimated. Finally, in the image resampling and transformation step, the source image is transformed using the transform model.
The accuracy and efficiency of the feature-detection step are crucial as they directly influence the subsequent steps in the registration process. Several feature-detection algorithms have been developed, each with its own strengths and weaknesses. For example, Speeded Up Robust Features (SURF) and Scale-Invariant Feature Transform (SIFT) are known for their robustness but can be computationally expensive. Features from Accelerated Segment Test (FAST) is computationally efficient but lacks invariance to scale and rotation. Binary Robust Invariant Scalable Keypoints (BRISK) offers robustness to both scale and rotation, while Oriented FAST and Rotated BRIEF (ORB) provides a good balance between computational efficiency and robustness. Other algorithms like MinEigen and Maximally Stable Extremal Regions (MSER) focus on specific aspects such as stability and sensitivity to feature regions [4].
Here, we briefly define several pivotal terms to aid understanding of feature-based image registration: Robustness refers to the ability of a feature-detection algorithm to deliver consistent results under varying conditions such as noise, illumination changes, and occlusion. Invariance to scale and rotation describes the capability of an algorithm to identify features correctly regardless of image scaling and rotation. Stability indicates the consistency of feature detection across different images or different instances of the same scene. Sensitivity to feature region measures the degree to which an algorithm can detect subtle changes or small features within an image. These definitions provide a foundation for discussing the strengths and limitations of various algorithms used in the field.
In the realm of digital image processing, image registration is a critical task that involves aligning two or more images—often from different sensors, times, or viewpoints—into a single, cohesive framework. This process is fundamental in applications ranging from satellite imagery analysis to medical imaging and automated surveillance systems. The primary challenges in image registration include ensuring high accuracy in matching diverse images, reducing the computational time required to process these images, and addressing failures in registration due to complex image transformations. This study focuses on innovating feature-detection techniques that enhance the precision and efficiency of image registration, particularly tackling the computational demands and robustness against image variations.
To overcome the limitations of existing algorithms, we propose a novel hybrid feature-detection algorithm that combines the strengths of BRISK and FAST [5,6]. This hybrid algorithm aims to reduce the time required for feature detection while maintaining robustness to scale and rotation. By leveraging the complementary strengths of BRISK and FAST, the hybrid algorithm can provide more reliable and efficient feature detection, which is essential for accurate image registration.
This study introduces a novel hybrid feature detector that integrates the robustness of BRISK with the speed of FAST, addressing common limitations found in traditional methods. Unlike conventional detectors, our hybrid approach is designed to provide high accuracy and efficiency under diverse operational conditions, offering significant improvements in both scale and rotation invariance.
The remainder of this paper is structured as follows: Section 2 reviews related work, providing a critical examination of previous studies and developments in feature-detection techniques and their applications in image registration. Section 3 details the methodology, including the development and implementation of the hybrid feature-detection algorithm, and describes the image acquisition methods used. In Section 4, we present the simulation setup and discuss the results obtained from testing the proposed algorithm against established detectors, demonstrating its effectiveness through various performance metrics. Finally, Section 5 summarizes the findings and contributions of this study and outlines potential avenues for future research, highlighting opportunities for further enhancements and applications of the proposed hybrid feature detector.

2. Related Work

Image registration is a fundamental and crucial task in image processing, utilized to match two or more images acquired at different times, from different sensors, or from different viewpoints [7,8]. It plays an important role in integrating and analyzing images from various sources. Both classic and recent image-registration methods have been extensively reviewed, highlighting the advantages, drawbacks, and future research directions [9].
A comparative study of well-known feature detectors and descriptors, including SIFT, MSER with SIFT, SURF with SURF, BRISK with BRISK, FAST with BRIEF, and ORB with ORB, has been conducted [10]. Additionally, the performance of feature descriptors extracted by the Harris–Affine detector has been compared [11].
These methods often face challenges such as computational complexity, sensitivity to changes in lighting and scale, and robustness against noise and occlusion. Robustness refers to a method’s ability to deliver consistent performance under varying conditions, while scale-invariance denotes the capability of a method to handle images of different sizes and orientations effectively. Stability, another crucial term, indicates the consistency of a method in detecting features across similar or varying scenes.
Furthermore, the existing algorithms often exhibit limitations in terms of their adaptability to different application contexts. For instance, while algorithms like SIFT and SURF provide excellent feature detection and matching under uniform lighting conditions, they may underperform in scenarios with variable lighting or when capturing images from rapidly moving objects. This variability demands a nuanced understanding of each algorithm’s operational environment to optimize performance effectively. Moreover, the trade-off between computational demand and accuracy is a critical consideration, particularly in real-time applications where processing speed is paramount.
In response to the limitations of existing feature detectors, our research introduces a novel hybrid algorithm that combines the strengths of several well-established methods. This hybrid approach aims to mitigate individual weaknesses and enhance overall performance, particularly in challenging environments typical of remote sensing and automated surveillance. By leveraging composite techniques, the proposed method not only improves detection accuracy and speed but also enhances robustness against variations in lighting, scale, and movement, making it highly suitable for modern image registration demands.
A novel algorithm for multispectral facial recognition, incorporating both visible and infrared (IR) images using various feature detectors, has been proposed [12]. Multispectral image registration with scale-invariant feature transform (SIFT) and random sample consensus (RANSAC) were described in [13]. In another study, a feature-based image registration (FBIR) method using HOG (Histogram of Oriented Gradients) for keypoint matching and a six-parametric offline transformation model was introduced [14]. Registration methods and their challenges have been reviewed, with a performance evaluation based on registration accuracy [15].
Traditional and advanced methods for multimodal remote-sensing (MMRS) image registration algorithms have been discussed [16]. An enhanced affine transformation (EAT) algorithm for non-rigid IR and visible (VIS) image registration has been presented [17]. A novel spatially invariant feature-matching scheme with higher performance using similarity matrices based on normalized eigenvector correlation and signal directional differences has been proposed [18].
Various algorithms for feature detection and description have been investigated [19]. A new feature descriptor, histogram of angle, and maximal edge orientation distribution (HAED), has been developed to address multi-source image-matching problems [20]. A comprehensive survey on different feature-based image-matching procedures and methods has been conducted [21].
An image-matching algorithm known as Dominant Orientation of Gradient (DOG) has been found to be robust to nonlinear intensity variations [22]. A novel local statistics-based image registration scheme, robust to contrast changes and geometric transformations, has been introduced [23]. An intelligent framework using a hybrid structural feature extraction technique for estimating transformation parameters using ground truth images has been proposed [24]. Authors in [25] focused on real-time image registration for an accurate geographic position using a UAV aerial view of images.
A hybrid feature extraction technique for medical images has been presented [26], along with a robust coarse-to-fine registration (CCFR) algorithm [27]. A feature matching algorithm combining FAST feature points and SURF descriptors has been proposed [28]. To address the difficulty of accurately registering low-texture images, a high-precision image registration algorithm based on line segment features has been developed [29].
An optimization algorithm called the normal vibration distribution search-based differential evolution algorithm (NVSA) has been introduced for SAR and optical image registration [30,31]. The Efficient Attention Pyramid Transformer (EAPT) has been proposed to address the problem of patch detection, using deformable attention, encode-decode communication modules, and position encoding for patches of any dimension [32].
A combination of traditional machine learning algorithms and deep neural networks for feature extraction in color and space aspects has been utilized to develop a one-stop deep portrait photographing guidance system [33]. A temporally broad learning system (TBLS) has been proposed to maintain temporal consistency between frames, consisting of original frames and corresponding frames in temporally inconsistent videos [34].
A novel deep convolutional neural network utilizing the multimodal cascaded method for detecting and classifying domestic waste has been proposed, along with a smart trash bin (STB) as the front-end carrier for waste disposal [35]. A Generative Parking Spot Detection (GPSD) algorithm using corner points to recover parking spots has been developed, featuring a layered analytical illumination balance method and a fast micro-target detection network [36]. Lastly, a broad attentive graph fusion network (BaGFN) has been designed to strengthen high-order feature representation under graph structures and refine high-order feature interactions at a bitwise level [37].
In the realm of image processing, the selection of an appropriate feature-detection and registration method is pivotal for achieving high accuracy and efficiency. Table 1 provides a succinct overview of various well-established and novel algorithms discussed in the literature. These methods are characterized by their unique attributes and are applied across a diverse array of scenarios.

3. Methodology

This paper proposes a novel algorithm resulting from a combination of the BRISK and FAST feature-detection algorithms. The proposed methodology is outlined in the flow diagram shown in Figure 1. Detailed explanations of all the feature detectors and descriptors used in this study are provided in the following section.
The experiments conducted in this study focus on three distinct types of image registrations based on the manner of image acquisition: (i) image registration based on different viewpoints with varying rotation angles, (ii) scene-to-model registration using two different scenes with some common portions, and (iii) image registration based on different viewpoints with scaling transformations. These specific registrations were chosen to methodically assess and demonstrate the capabilities of our hybrid algorithm under controlled variations.
By separating these transformations in the initial experiments, we aimed to isolate the effects of each type of manipulation on the detection performance, providing a clear understanding of how each adjustment affects the overall efficiency and accuracy of our proposed hybrid detector. This approach allows for a more granular analysis of performance under specific conditions, which is essential for developing a robust feature-detection system.
The registrations are performed on a diverse set of remote-sensing and scene images, encompassing a broad range of real-world scenarios to evaluate the performance of the proposed hybrid algorithm comprehensively. Future work may include combining these transformations to simulate more complex real-world scenarios, further testing the adaptability and robustness of the hybrid detector.
In our manuscript, we focus on empirical performance metrics such as matching rate, authentication rate, and computation time. These metrics are directly applicable to the practical deployment of image registration techniques and provide clear, measurable outcomes that can be compared across different methods; while theoretical frameworks involving cost function minimization are valuable for certain analytical or optimization-focused studies, our approach prioritizes direct evaluation of the methods in terms of their operational effectiveness in real-world scenarios. This choice is aligned with the needs of applications that require fast and accurate feature detection and registration, such as remote sensing and automated surveillance.

3.1. Feature Detectors and Descriptors

A novel feature detector for keypoint detection, description, and matching known as BRISK (Binary Robust Invariant Scalable Keypoints) has been proposed [40]. This method is recognized for its robustness to scale and rotation, making it suitable for various image-registration tasks. BRISK (Binary Robust Invariant Scalable Keypoints) provides a fast and robust solution for keypoint detection and description, ideal for real-time applications. It creates scale–space pyramids to achieve scale invariance and uses a pattern-based descriptor for robustness against rotation. However, it may produce higher false positive rates compared to slower, more complex methods like SIFT.
FAST (Features from Accelerated Segment Test) employs machine learning algorithms to enhance the efficiency of feature detection [39]. A subsequent work presents a heuristic approach for feature detection, further improved by machine learning techniques [41], significantly enhancing the speed and performance of the algorithm. FAST is renowned for its computational speed by using a decision tree to quickly assess pixel intensities. Although highly efficient, FAST lacks rotational invariance and is not inherently scale-invariant, which may limit its application in environments where orientation and scale vary significantly. Extensions like ORB have been developed to incorporate scale and rotation invariance into FAST.
ORB (Oriented FAST and Rotated BRIEF) builds on FAST by adding a pyramid scheme for scale invariance and a learning-based orientation mechanism. ORB is partially scale-invariant and more robust to rotation variations compared to FAST alone, making it a versatile choice for multi-scale, orientation-varied feature-detection tasks. However, it may still struggle with high levels of image noise and significant scale changes.
In a comprehensive comparison of feature detectors such as SIFT, ORB, AKAZE, BRISK, MinEigen, and SURF, SIFT and BRISK have been identified as more accurate, while ORB and BRISK demonstrate higher efficiency [42,43]. Additionally, SURF, BRISK, and SIFT are noted for their scale-invariance, with ORB being less so. This evaluation underscores the varying capabilities and suitability of these algorithms for different image-registration tasks, emphasizing the need for a hybrid approach to combine their strengths.
The MSER (Maximally Stable Extremal Region) detector, which is known for its robustness and efficiency, utilizes the component tree as an efficient data structure, allowing quasi-linear time calculation of MSERs [38]. This feature detector is particularly useful for tracking applications due to its stability. MSER detects regions in an image that are stable and distinctive. Stability ensures that regions remain consistent under slight image perturbations such as noise or geometric transformations, while distinctiveness means regions visually stand out from their surroundings. This makes MSER particularly effective for identifying text or other significant structures within varied imaging conditions. It analyzes intensity variations, identifying connected areas with similar intensity that maintain their shape and intensity over transformations, making it suitable for applications like text recognition and object segmentation. Despite its robustness, MSER’s performance may vary depending on image contrast and noise, sometimes requiring combination with other techniques for optimal results.
The Harris corner detector examines the eigenvalues of the autocorrelation matrix to detect corners and interest points in images [44]. This method is well regarded for its precision in identifying distinct image features.
By integrating the strengths of BRISK and FAST, the proposed hybrid algorithm aims to achieve efficient and robust feature detection suitable for a wide range of image-registration tasks. The proposed hybrid detector combines the high-speed processing of FAST with the scale and rotation invariance of BRISK, enhancing both the accuracy and the efficiency of keypoint detection. This improvement is particularly beneficial in complex image transformations and diverse operational scenarios, as detailed in the subsequent sections on methodology and simulation results.
Moreover, the application of the hybrid detector in challenging environments, such as remote sensing and automated surveillance, has demonstrated superior performance in terms of both detection accuracy and computational efficiency. The hybrid approach effectively addresses the limitations of individual feature detectors by dynamically adjusting to the characteristics of the input images and the specific requirements of the task. This adaptability is crucial in environments where image quality and scene complexity can vary significantly, thus requiring a more robust and flexible approach to feature detection.

Hybrid Feature-Detection Technique

The proposed hybrid approach combines the strengths of both the BRISK and FAST algorithms to address their individual limitations and enhance the overall performance of feature detection. This innovative technique leverages a ‘diagonal strategy’ within a circle of sixteen pixels for efficient corner detection, focusing on only four key pixels. This method not only speeds up the detection process but also maintains high accuracy by categorizing pixels into brighter, darker, and similar sections based on their intensity.
By integrating the detailed, scale-invariant detection capabilities of BRISK with the high-speed, efficient processing of FAST, the hybrid method significantly reduces both time and computational complexity. This dual approach ensures that the hybrid detector can quickly and accurately process images, making it especially suitable for complex image transformations and diverse operational scenarios found in remote sensing and automated surveillance.
Furthermore, the adaptability of the hybrid detector to varying imaging conditions—owing to its combined algorithmic structure—provides enhanced robustness against changes in lighting, scale, and motion. This robust performance is crucial for applications requiring reliable and precise feature detection in dynamic environments.

3.2. Feature-Based Image Registration (FBIR)

The problem of image registration, particularly FBIR, is a crucial issue in the image processing and computer vision domains. FBIR is mainly applicable in scenarios where it is necessary to integrate and analyze information from different sources, which may include different sensors, multiple photographs, various times, depths, or viewpoints. The diverse application areas of FBIR include image fusion, change detection, and multi-channel image restoration.

3.2.1. Feature Detection and Extraction Using Proposed Hybrid Feature Detector

The first step of FBIR involves feature detection and extraction. An automatic machine-based method is developed to extract structures and features from images. Features can be regions like forests, lakes, and fields, or points such as region corners and line intersections.
An automatic hybrid approach for feature detection and description is developed to address the limitations of the BRISK and FAST algorithms [5]. The proposed hybrid detector and descriptor aim to reduce both time complexity and computational complexity, and while the BRISK algorithm is robust to rotation and scale, it is computationally intensive. Conversely, the FAST algorithm is faster but not scale-invariant and depends on a threshold value. By combining the advantages of both BRISK and FAST, the proposed approach overcomes their respective drawbacks.
In the proposed hybrid feature detector and descriptor, a corner pixel is constructed by considering a circle of sixteen pixels. These pixels are divided into three groups: brighter, darker, and similar portions, identified by their threshold and intensity values. Suppose the threshold value is represented by t and the intensity of the pixel by i.
Intensity of a pixel = t + i if brighter t i if darker t i i t + i if similar
For the brighter section, i is added to t; for the darker section, i is subtracted from t; and for the similar section, i lies between t i and t + i . Using these three types of pixels, a diagonal approach is employed for the input sixteen pixels.
As shown in Figure 2, instead of testing all sixteen pixels, which is time consuming, the proposed hybrid algorithm tests only four pixels. This diagonal approach uses two slanting lines: one from pixel 3 to pixel 11 (right to left) and another from pixel 15 to pixel 7 (left to right). The intersection of these two diagonal lines identifies the center corner pixel. This method accelerates the process by reducing the number of comparisons. Finally, a local gradient method is applied to the center corner pixel to determine if the pixel’s distance is below or above the threshold value. The main advantage of the proposed algorithm is the reduction in both time and computational complexity compared to BRISK, FAST, MSER, MinEigen, ORB, and Harris algorithms. Additionally, the proposed algorithm outperforms existing algorithms in terms of match points and match rate.
Algorithm 1 outlines the steps employed in the Feature-Based Image Registration (FBIR) using the proposed hybrid feature detector and descriptor. The algorithm aims to optimize the feature-detection and matching process by applying various transformations and utilizing a hybrid approach that combines the strengths of BRISK and FAST algorithms. The result is an improved FBIR system that enhances the efficiency and accuracy of image registration, crucial for applications in areas like remote sensing, medical imaging, and computer vision.
Algorithm 1 FBIR with Proposed Hybrid Algorithm.
Require: Original image
Ensure: Registered image using FBIR with hybrid feature detector and descriptor
  1:
Obtain the required input image from the database and convert it to its grayscale equivalent.
  2:
Apply transformations on the grayscale image:
  • For rotation of an image using different angles:
    R ( x , y ) = x cos θ y sin θ x sin θ + y cos θ
    where R is the rotated resultant image, x and y are the rotated pixel coordinates, and  θ is the angle of rotation.
  • For scaling transformation of an image:
    S ( x , y ) = x · s x y · s y
    where S is the scaled image, x and y are the new scale coordinates, and  s x and s y are the scaling factors.
  • Perform scene-to-model description and detection using two different input images with some common portions.
  3:
Detect feature keypoints from both the reference image and the sensed image using various detectors like BRISK, FAST, MSER, ORB, MinEigen, Harris, and the proposed hybrid feature detector.
  4:
Extract features from both detected images using the transformations described in previous steps with various detectors and the proposed hybrid detector.
  5:
Match the key feature points extracted from both the reference image and the sensed image using affine transformation with bicubic/bilinear interpolation.
  6:
Estimate the time for rotation, scaling, and scene-to-model registration for all detectors, including the proposed hybrid detector.
  7:
If the matched points from both images are successfully extracted, obtain the registered image.
To further elucidate the operation of our proposed Hybrid Feature Detector, we utilize a strategic subset of pixels within a predefined circular pattern. Initially, all 16 pixels within this pattern are assessed based on their intensity values to classify them into three categories: brighter, darker, and similar. This categorization allows the algorithm to focus computational efforts on four pivotal pixels, which are determined through a diagonal evaluation approach. This method not only streamlines the feature-detection process by reducing unnecessary computations but also maintains high detection accuracy by focusing on the most informative pixels.
The hybrid detector is designed to integrate the robust detection capabilities of BRISK, which is adept at handling scale and rotation variations, with the computational efficiency of FAST. This integration addresses the time-consuming keypoint detection in BRISK and the scale limitations of FAST, providing a balanced solution that is both fast and scalable.
To aid in the understanding of this integration, the following Algorithm 2 illustrates the process of selecting key pixels and combining the strengths of BRISK and FAST in our hybrid detection approach:
Algorithm 2 Hybrid Feature-Detection Process.
Require: Image
Ensure: Keypoints
  1:
Start with a circle of 16 pixels around each candidate pixel.
  2:
Classify 16 pixels into three categories based on intensity:
  • Brighter
  • Darker
  • Similar
  3:
Apply a diagonal strategy to select 4 key pixels:
  • Draw two diagonal lines across the circle.
  • Select the intersection points as key pixels.
  4:
Use BRISK for scale and rotation invariant detection on selected pixels.
  5:
Apply FAST for quick detection on the reduced pixel set.
  6:
Combine results to obtain final keypoints.
This comprehensive approach ensures that our hybrid detector not only optimizes computational resources but also adapts dynamically to various imaging conditions, thereby enhancing both the performance and applicability of the feature-detection process in real-world scenarios.

3.2.2. Feature Matching Using a Hybrid Algorithm

Feature matching, the second step of FBIR, plays a crucial role in establishing the mapping between two images of the same fields acquired from different sources. Keypoints are identified in both the reference image and the sensed image to perform matching. The goal is to find a better correspondence between these images by comparing each feature keypoint of the sensed image to those of the reference image and measuring the distance between these points using BRISK, FAST, and the hybrid feature descriptors. The Euclidean distance between two keypoints ( a 1 , b 1 ) and ( a 2 , b 2 ) is calculated using the following formula:
distance = ( a 2 a 1 ) 2 + ( b 2 b 1 ) 2
The distance is calculated for various feature detectors, including BRISK, FAST, MSER, MinEigen, ORB, Harris, and the hybrid algorithm. To compare the execution time of each algorithm, three time-based parameters are used: elapsed time, CPU time, and PMT time. Experimental results show that the proposed hybrid feature matching algorithm takes less time compared to existing feature detectors and descriptor algorithms.

3.2.3. Feature-Based Transform Model Estimation

The third step of FBIR involves estimating both the type and parameters of the transformation needed to align the source image with the target image. The goal is to find an accurate transform model that appropriately transforms the source image. Three methods for parameter estimation in transform model estimation are discussed, with affine transformation being used in this paper.
Similarity transformation is a shape-preserving mapping transformation that preserves angles and curvatures, while affine transformation is a linear mapping method that preserves points, straight lines, and planes. Affine transformation is applied to correct geometric deformations that occur due to non-ideal camera angles, and it is a particular case of projective transformation.

3.2.4. Image Resampling and Transformation

The final step of FBIR is image resampling, which involves changing the pixel dimensions of an image, effectively altering its resolution. The registered image obtained from the previous step is convolved with an interpolation kernel. Interpolation techniques reduce the bandwidth of the signal by employing a low-pass filter on the discrete signal.
Three interpolation techniques are compared: Nearest Neighbor, Bilinear, and Bicubic. These techniques are evaluated based on image quality parameters such as Mean Squared Error (MSE), Root Mean Squared Error (RMSE), Signal-to-Noise Ratio (SNR), and Peak Signal-to-Noise Ratio (PSNR), as shown in Table 2.
From Table 2, it can be observed that the MSE for Bicubic interpolation is 0.00214, 0.00219, and 0.00221 for affine, similarity, and projective transformations, respectively. These values indicate that Bicubic interpolation has the lowest error content compared to Bilinear and Nearest Neighbor interpolation schemes. Similar observations can be made for RMSE, SNR, and PSNR values, where Bicubic interpolation outperforms the other methods. Lower values of MSE and RMSE imply less error content, while higher values of SNR and PSNR indicate lower noise content in the image. Hence, Bicubic interpolation, which produces a smoother interpolation surface, is used in the image resampling and transformation step of image registration in this paper.

4. Simulation and Results

This section validates the proposed hybrid algorithm using a series of experiments conducted on eight aerial images. The validation considers three primary types of transformations: rotation at various angles, scene-to-model transformation using different instances of the same image where some parts share common features, and scaling transformations at varying scales.
The results of these experiments, presented in both numerical and visual formats, demonstrate that the proposed hybrid feature-detection algorithm outperforms existing detectors such as BRISK [40], FAST [41], ORB [43], Harris [44], MSER [38], and MinEigen [42]. Our method not only reduces the computational time required for feature detection but also improves the accuracy of keypoint matching, making it a valuable tool for a wide range of image registration applications.
To facilitate transparency and allow for in-depth validation by the research community, the complete source code used in our experiments is available on GitHub [45]. This includes the Matlab 2019a implementation of our algorithm to assist with setup and replication of our results.

4.1. Experimental Setup and Image Data

The experiments were conducted using eight different images from the AID (Aerial Image Dataset) database, representing a variety of scenes including parks, railway stations, airports, bridges, a university gate (VSSUT gate), and a large dam (Hirakud dam) [46]. These images were chosen to cover a broad spectrum of typical scenarios in remote sensing and feature-detection tasks.
These images are subjected to three types of transformations to test the robustness and effectiveness of the hybrid feature-detection algorithm:
  • Rotation: Images are rotated at angles of 30 , 70 , 90 , 120 , 150 , and 180 .
  • Scene-to-Model Transformation: This involves using two different instances of the same scene (e.g., different views of an airport and a bridge) where parts of these images share common features.
  • Scaling: Images are scaled by factors of 0.7 and 2.0 to evaluate the algorithm’s performance under size variations.
Figure 3 presents the original color aerial images from the AID database. Figure 4 shows these images converted to grayscale. Figure 5 illustrates the effects of various rotational angles applied to the park and railway station images. Figure 6 displays the scaling transformations applied to the VSSUT gate and Hirakud dam images.

4.1.1. Time Measurement Definitions

In our experimental analysis, we utilize three primary metrics to evaluate the computational efficiency of the feature-detection algorithms:
  • Elapsed Time: total time from the initiation to the completion of the feature-detection process.
  • CPU Time: the amount of processing time the CPU spends to execute the feature-detection tasks, excluding any idle time.
  • PMT (Performance Measuring Time): this metric assesses the performance efficiency of the algorithm, focusing on the active processing time.
These metrics help in understanding the computational demand and efficiency of the proposed methods under different operational conditions.

4.1.2. Validation of Detected Keypoints

To ensure the accuracy of detected keypoints, our analysis relies on the established performance metrics such as precision and matching rate, which have been detailed in previous sections. These metrics serve as indicators of the correctness of the keypoint identification:
  • Precision assesses the proportion of detected keypoints that are true positives, helping to confirm that the keypoints are genuine features of the images rather than noise or errors.
  • Matching Rate evaluates how well the keypoints from different transformations of the same image correlate with each other. A high matching rate indicates a successful identification of consistent and reliable keypoints across different versions of the images.
This analytical approach allows us to validate the effectiveness of the keypoint detection algorithm without the need for additional experimental validation. The high performance metrics reported reflect the robustness of our feature-detection algorithm, underscoring its reliability in identifying correct keypoints even under challenging conditions such as rotation and scaling.

4.2. Rotation with Different Angles

This subsection meticulously evaluates the performance of various feature-detection algorithms under rotation transformations, focusing on their robustness and effectiveness. Two distinct aerial images—a serene park and a bustling railway station—serve as test subjects. These images were methodically rotated at six pivotal angles: 30 , 70 , 90 , 120 , 150 , and 180 . This setup aims to rigorously test the resilience of the feature-detection methods, including an innovative hybrid algorithm developed as part of this study.
Several renowned feature detectors were employed in these experiments: BRISK [40], FAST [41], ORB [43], Harris [44], MinEigen [42], and MSER [38], alongside the newly proposed hybrid detector. Each detector’s capability to consistently identify and track feature keypoints across various rotation angles was analyzed.
The results of these experiments are illustrated in Figure 7 and Figure 8. These figures not only depict the detection of feature keypoints but also highlight the comparative performance and distinctive traits of each detector under rotational stress. Such detailed visualization aids in understanding the practical impacts of rotational transformations on feature-detection reliability. The feature keypoints, indicated in green, represent significant aspects of the image such as edges, corners, or other specific patterns that are crucial for alignment tasks in image registration. This detailed visualization helps to better understand how different detectors perform under the challenge of rotation, providing insight into their robustness and effectiveness.
Subsequent Figure 9 and Figure 10 extend this analysis by detailing the extraction processes of these keypoints for both the park and railway images. The robust extraction capabilities of each feature detector are crucial for accurate feature matching in applications such as image stitching and object recognition in computer vision. The green markers in these figures specifically illustrate the keypoints that each detector has identified as crucial for successful image analysis and manipulation. This step-by-step visualization showcases the effectiveness of each algorithm in maintaining keypoint integrity even through complex transformations, ensuring accurate subsequent image registration.
In our evaluation of the feature detectors, we carefully analyze how each method performs under various transformations. Figure 11 and Figure 12 are designed to provide a clear comparison between the detectors. To make the differences more discernible, the images are presented separately for individual detectors to allow for a focused analysis of each method’s capabilities in isolation. For the hybrid detector, we present an overlaid image to demonstrate the synergistic effect of combining multiple detection techniques, showcasing our proposed method’s comprehensive matching capability. This format aids in visualizing the distinct performance traits and alignment precision of each detector.
In addition to visual assessments, the performance of the feature-detection algorithms was quantitatively evaluated, as depicted in Table 3 and Table 4. These tables enumerate the detected keypoints, extracted keypoints, and matched keypoints across six rotational angles for both the park and railway station images, employing a variety of feature detectors including BRISK, FAST, ORB, Harris, MinEigen, MSER, and the newly proposed hybrid detector.
The hybrid algorithm, notably, demonstrated superior performance in terms of match rate percentages, significantly outperforming traditional detectors. For example, at a 120 rotation, the hybrid detector achieved a match rate of 28.84 % , which is the highest among the detectors tested. Moreover, in scenarios with 90 and 180 rotations, the hybrid and ORB detectors achieved a perfect match rate of 100 % , indicating robust performance in standard upright and inverted orientations.
Table 3 includes key statistical metrics—mean, variance, and standard deviation—across various rotation angles to provide a clearer, summarized view of the performance of different feature detectors. These statistical summaries help discern the general trends and variability in the performance of each method without the clutter of individual data points. For instance, at a 90 rotation, the mean matching rate is notably high, reflecting the robustness of the detectors under orthogonal rotations. The standard deviation at this angle is lower compared to other angles, indicating more consistent performance across different detectors. This statistical approach not only simplifies the comparative analysis but also enhances the readability and interpretability of the results, supporting a stronger, more justified conclusion about the superiority of specific methods, such as the Hybrid detector which consistently shows high efficiency and accuracy.
The efficiency of the algorithms was also gauged through performance metrics such as elapsed time, CPU time, and PMT (Processor Memory Time). The hybrid algorithm consistently showed the lowest time consumption across these metrics, suggesting its suitability for real-time applications. For the park image at 150 rotation, the hybrid algorithm required only 3.7125 s of CPU time, which is significantly less than the other detectors like BRISK (5.1875 s) and ORB (11.2813 s).
This numerical analysis confirms the effectiveness of the hybrid algorithm not only in maintaining high accuracy in feature matching across varied rotations but also in ensuring computational efficiency. Such attributes make the hybrid algorithm particularly advantageous for applications in remote sensing and medical imaging where rapid and reliable feature detection is critical.

4.3. Scene-to-Model Registration

Scene-to-model registration involves comparing different instances of the same image, typically referred to as the reference image and the sensed image, which share some common features. This process is crucial for applications such as satellite image analysis, where changes over time within the same geographic area need to be identified accurately. In this experiment, instances of aerial images from an airport and a bridge were selected to demonstrate the effectiveness of various feature-detection and matching algorithms.
Figure 13 and Figure 14 illustrate the process of feature point detection, extraction, and matching for these aerial images. Each figure sequences through the stages of detecting features in individual images, extracting those features, and then matching them between two images of the same scene. This sequence is demonstrated using different feature detectors and descriptors, highlighting how each algorithm performs under the same conditions.
The quantitative analysis of these experiments is presented in Table 5, which details the performance of each detection method across various metrics such as the number of keypoints detected, extracted, and successfully matched, as well as the efficiency metrics including match rate percentage and execution times (elapsed, CPU, and PMT). For instance, when comparing the two instances of the airport images, the hybrid algorithm significantly outperformed other methods with a matching rate of 73.90%, which is the highest among all the detectors. This high performance is consistent across the different images, underscoring the hybrid algorithm’s robustness and efficiency, particularly noted by its minimal processing time.
The results for the bridge images show similar trends, where the hybrid algorithm again demonstrates superior performance, especially in terms of execution time, making it an ideal candidate for real-time applications in remote sensing and aerial reconnaissance.
Overall, the scene-to-model registration experiments validate the efficacy of the proposed hybrid algorithm, not only in achieving high match rates but also in maintaining lower computational costs, making it suitable for real-time image analysis applications.

4.4. Scaling Transformations with Differet Scale Vectors

This subsection explores the effects of scaling transformations on feature detection and matching using three different feature detectors: BRISK, MSER, and Hybrid. The choice of BRISK and MSER stems from their known scale–space invariant properties, making them ideal for studying scaling impacts. The Hybrid detector is introduced to assess potential improvements in scaling performance.

4.4.1. Comparative Analysis of Feature Detectors

The analysis is supported by three detailed tables that compare the performance of these detectors under different scaling vectors on two distinct images—the VSSUT entrance and the Hirakud dam. These tables (Table 6, Table 7 and Table 8) quantify the effects of scaling on detection accuracy and computational efficiency, essential for understanding the scalability of each detector.
The comprehensive performance comparison facilitated by Table 6, Table 7 and Table 8 provides a robust basis to evaluate the effectiveness of the BRISK, MSER, and Hybrid feature detectors under various scaling transformations. These tables incorporate crucial image quality metrics such as PSNR (Peak Signal-to-Noise Ratio) and MSE (Mean Squared Error) along with essential time-based metrics including elapsed time, CPU time, and PMT (Preprocessing and Matching Time). This extensive dataset not only validates the efficiency of the feature-detection algorithms but also highlights the computational demands associated with each method.
A consistent pattern is observed in the data, which underscores the superior performance of the Hybrid detector in maintaining high image quality metrics and managing computational time effectively across different scaling scenarios. For example, under a scaling vector of 2.0, the Hybrid detector consistently demonstrates higher PSNR values and lower MSE, indicating better image reconstruction with fewer errors compared to the BRISK and MSER methods. This efficiency is also reflected in the processing times, where the Hybrid detector often equals or surpasses the speed of the other detectors while delivering more accurate results.
This evaluation emphasizes the importance of selecting an appropriate feature-detection strategy based on specific application requirements, especially in environments that involve significant variations in image scale. The Hybrid detector, with its robust performance across various scales, emerges as an exceptionally effective tool in scenarios where both precision and efficiency are crucial. It proves to be an excellent choice for applications such as aerial imaging, surveillance, and other forms of remote sensing where images may undergo various transformations and require high fidelity and rapid processing for timely decision-making.

4.4.2. Impact of Scaling on Feature Detection

Figure 15, Figure 16 and Figure 17 demonstrate the efficacy of MSER, BRISK, and Hybrid feature detectors under varying scaling conditions, highlighting their capabilities and limitations in handling images from two distinctive scenarios: the VSSUT entrance and the Hirakud dam. Each image undergoes transformations using scaling vectors of 0.7 and 2.0, simulating conditions of both under-scaling and over-scaling, which are common challenges in practical applications.
In the initial set of images, MSER shows robustness in detecting features on the VSSUT gate under normal conditions, but its performance slightly degrades when the image is downscaled (0.7 scaling factor), suggesting a drop in sensitivity to smaller scale features. Conversely, BRISK maintains a consistent detection rate across scales, likely due to its design that balances scale invariance and feature stability. The Hybrid detector, expectedly, outperforms the individual MSER and BRISK in both scenarios by integrating their strengths and mitigating their weaknesses, particularly in the over-scaled (2.0 scaling factor) images where finer details are magnified.
Moving to the feature extraction figures, the pattern is somewhat consistent with the initial detection tests. The MSER detector captures denser clusters of keypoints in high-detail areas, which becomes sparse in scaled-down images. BRISK again shows uniform performance, proving to be less sensitive to scale changes compared to MSER. Hybrid’s advanced algorithm synergizes the detection process, ensuring that the quality and quantity of keypoints remain relatively stable across different scales. This adaptability makes it particularly suitable for applications requiring high precision in feature extraction across varied imaging conditions.
These observations underscore the importance of choosing the right feature-detection and extraction methods based on the specific requirements of the application, especially when dealing with images subjected to significant scale transformations. The Hybrid detector emerges as a strong candidate for scenarios requiring robust performance under diverse scaling conditions, providing a balanced solution that leverages the strengths of both MSER and BRISK.

4.4.3. Advanced Analysis Using Registered Images

Figure 18 displays the results of image registration using the Hybrid feature detector. This set of images showcases enhanced informational content through the integration of various scene details, achieving a more comprehensive representation of the original scenes. These registered images, processed through the Hybrid detector, exemplify the effectiveness of the method in synthesizing high-quality composite views from multiple aerial and terrestrial photographs.
In Figure 19, each subplot visually demonstrates the keypoints detected by different feature-detection algorithms when applied to a park scene. These visual representations allow us to assess the density and distribution of keypoints identified by each algorithm, which are critical factors in evaluating their effectiveness and reliability for practical applications such as image registration and object recognition.

4.4.4. Analysis of Feature-Detection Metrics

The performance of feature-detection algorithms is evaluated based on several key metrics, each offering insights into different aspects of the algorithm’s effectiveness. The Number of Keypoints detected is indicative of the algorithm’s ability to identify features across the image, which is crucial for comprehensive analysis and accurate matching. However, a larger number of keypoints does not necessarily equate to better performance, as the relevance and accuracy of these points are paramount.
Precision is a critical measure that assesses the accuracy and relevance of the detected keypoints. It measures the proportion of true positive keypoints among all detected keypoints, reflecting the accuracy of the detection process. The Matching Rate further complements these metrics by examining how effectively the keypoints from different images correlate, which is essential for applications like image stitching and 3D reconstruction. High matching rates suggest that the keypoints are not only accurately detected but are also meaningful in the context of aligning multiple images.
By evaluating these metrics together, we can form a comprehensive view of a detector’s performance, considering both the quantity and quality of the detected features. This approach ensures that the feature detectors are not only prolific in terms of keypoint generation but also precise and practical for real-world applications.

4.5. Discussion

This study’s comprehensive evaluation highlights the enhanced capabilities of our novel hybrid feature-detection algorithm within the context of Feature-Based Image Registration (FBIR). By integrating the strengths of established detectors such as BRISK, FAST, ORB, Harris, MinEigen, and MSER, the hybrid detector excels in both accuracy and efficiency, especially notable in complex image transformations like rotation, scaling, and scene-to-model changes. Such enhancements are crucial for applications in remote sensing and automated surveillance, where precise and reliable feature matching is paramount.
The experimental results demonstrate a significant reduction in time complexity alongside improvements in the detection and matching of keypoints. These improvements are quantified through detailed performance metrics, underlining the hybrid detector’s robustness across varied operational scenarios. This robustness ensures that the hybrid approach is well-suited to the dynamic and often unpredictable environments typical of remote-sensing applications, which demand high levels of adaptability and precision.
Furthermore, the comprehensive statistical analysis and the enhanced performance metrics clearly justify the efficacy of the hybrid feature-detection method. This approach not only meets but exceeds the capabilities of traditional detectors, particularly in handling complex image transformations.
Despite the promising outcomes, this study also acknowledges inherent limitations linked to the hybrid detector’s complexity. The integration of multiple detection methods into a cohesive algorithm introduces challenges in balancing computational efficiency with detection efficacy. This balance is particularly delicate when considering the diverse and often conflicting characteristics of the individual detectors involved. For instance, while some detectors may excel in speed, others might offer greater accuracy, necessitating careful calibration and tuning to harness their collective strengths effectively.
Moreover, the performance of our hybrid detector, while superior, still depends on the quality and diversity of the input data. This dependency suggests that the hybrid system’s adaptability might be constrained by less variable datasets, potentially limiting its effectiveness in less-controlled environments. Addressing these challenges involves not only refining the algorithm’s architecture but also ensuring that it remains flexible and responsive to the evolving landscape of image registration technologies and methodologies.

5. Conclusions and Future Work

This study introduced a combined approach using a novel feature detector and descriptor, enhancing all four steps of Feature-Based Image Retrieval (FBIR). Our evaluations, focusing on metrics such as elapsed time, CPU time, and performance measurement time, demonstrate that the proposed hybrid detector surpasses existing state-of-the-art detectors in terms of both efficiency and accuracy.
The hybrid detector not only improves the accuracy of detecting feature keypoints but also significantly reduces time complexity compared to conventional methods. This makes it particularly valuable for real-time image processing applications where speed and accuracy are crucial.
Throughout the testing phase, which included three types of image transformations—rotation, scene-to-model, and scaling—the proposed detector consistently outperformed other detectors, delivering superior visual and numerical results while also reducing execution times. These findings confirm that the proposed feature detector is more efficient compared to existing feature detectors, making it a robust and effective tool for image-analysis tasks.
In future studies, we aim to extend the reach and effectiveness of our proposed hybrid feature detector. A primary focus will be on optimizing the algorithm to enhance its adaptability and performance under a broader range of image conditions and transformations. Additionally, integrating deep learning techniques may offer substantial improvements in feature-detection capabilities, particularly for complex image scenarios, thereby expanding the utility and accuracy of our approach. Testing the hybrid detector across various platforms and media types will also be crucial to thoroughly validate its effectiveness and robustness in diverse operational contexts.
Applying the detector in real-world scenarios such as surveillance, autonomous driving, and medical imaging is essential to assess its practical utility and operational efficiency. Moreover, exploring the performance of the hybrid detector on different hardware configurations could lead to optimizations for energy efficiency and processing speed, making it well-suited for use in embedded systems and mobile devices. Through these initiatives, we hope to refine the capabilities of the hybrid detector further and expand its applicability to meet the evolving challenges in digital image processing and analysis [8].

Author Contributions

Conceptualization, A.K. (Anchal Kumawat), S.P., V.C.G., A.K. (Andreas Kanavos), B.A. and S.M.; Methodology, A.K. (Anchal Kumawat), S.P., V.C.G., A.K. (Andreas Kanavos), B.A. and S.M.; Software, A.K. (Anchal Kumawat), S.P., V.C.G., A.K. (Andreas Kanavos), B.A. and S.M.; Validation, A.K. (Anchal Kumawat), S.P., V.C.G., A.K. (Andreas Kanavos), B.A. and S.M.; Data curation, A.K. (Anchal Kumawat), S.P., V.C.G., A.K. (Andreas Kanavos), B.A. and S.M.; Writing—original draft, A.K. (Anchal Kumawat), S.P., V.C.G., A.K. (Andreas Kanavos), B.A. and S.M.; Writing—review & editing, A.K. (Anchal Kumawat), S.P., V.C.G., A.K. (Andreas Kanavos), B.A. and S.M.; Project administration, A.K. (Anchal Kumawat), S.P., V.C.G., A.K. (Andreas Kanavos), B.A. and S.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Experimental aerial images were obtained from the following repository: https://captain-whu.github.io/AID/, (accessed on 29 August 2024).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Yuan, W.; Poosa, S.R.P.; Dirks, R.F. Comparative Analysis of Color Space and Channel, Detector, and Descriptor for Feature-Based Image Registration. J. Imaging 2024, 10, 105. [Google Scholar] [CrossRef] [PubMed]
  2. Dai, X.; Khorram, S. A Feature-based Image Registration Algorithm using Improved Chain-Code Representation Combined with Invariant Moments. IEEE Trans. Geosci. Remote Sens. 1999, 37, 2351–2362. [Google Scholar]
  3. Guan, S.Y.; Wang, T.M.; Meng, C.; Wang, J.C. A Review of Point Feature Based Medical Image Registration. Chin. J. Mech. Eng. 2018, 31, 76. [Google Scholar] [CrossRef]
  4. Kuppala, K.; Banda, S.; Barige, T.R. An Overview of Deep Learning Methods for Image Registration with Focus on Feature-based Approaches. Int. J. Image Data Fusion 2020, 11, 113–135. [Google Scholar] [CrossRef]
  5. Kumawat, A.; Panda, S. Feature Detection and Description in Remote Sensing Images Using a Hybrid Feature Detector. Procedia Comput. Sci. 2018, 132, 277–287. [Google Scholar] [CrossRef]
  6. Kumawat, A.; Panda, S. Feature Extraction and Matching of River Dam Images in Odisha Using a Novel Feature Detector. In Proceedings of the Computational Intelligence in Pattern Recognition (CIPR); Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2020; pp. 703–713. [Google Scholar]
  7. Pratt, W.K. Digital Image Processing: PIKS Scientific Inside; Wiley Online Library: Hoboken, NJ, USA, 2007; Volume 4. [Google Scholar]
  8. Sridhar. Digital Image Processing; Oxford University Press, Inc.: Oxford, UK, 2011. [Google Scholar]
  9. Zitova, B.; Flusser, J.; Sroubek, F. Image Registration: A Survey and Recent Advances. In Proceedings of the International Conference on Image Processing, Genoa, Italy, 11–14 September 2005. [Google Scholar]
  10. Işık, Ş. A Comparative Evaluation of Well-Known Feature Detectors and Descriptors. Int. J. Appl. Math. Electron. Comput. 2014, 3, 1–6. [Google Scholar] [CrossRef]
  11. Mikolajczyk, K.; Schmid, C. A Performance Evaluation of Local Descriptors. IEEE Trans. Pattern Anal. Mach. Intell. (TPAMI) 2005, 27, 1615–1630. [Google Scholar] [CrossRef]
  12. Mamadou, D.; Gouton, P.; Adou, K.J. A Comparative Study of Descriptors and Detectors in Multispectral Face Recognition. In Proceedings of the 12th International Conference on Signal-Image Technology & Internet-Based Systems (SITIS), Naples, Italy, 28 November–1 December 2016; pp. 209–214. [Google Scholar]
  13. Rana, S.; Gerbino, S.; Crimaldi, M.; Cirillo, V.; Carillo, P.; Sarghini, F.; Maggio, A. Comprehensive Evaluation of Multispectral Image Registration Strategies in Heterogenous Agriculture Environment. J. Imaging 2024, 10, 61. [Google Scholar] [CrossRef]
  14. Abraham, E.; Mishra, S.; Tripathi, N.; Sukumaran, G. HOG Descriptor Based Registration (A New Image Registration Technique). In Proceedings of the 15th International Conference on Advanced Computing Technologies (ICACT), Rajampet, India, 21–22 September 2013; pp. 1–4. [Google Scholar]
  15. Tondewad, M.P.S.; Dale, M.M.P. Remote Sensing Image Registration Methodology: Review and Discussion. Procedia Comput. Sci. 2020, 171, 2390–2399. [Google Scholar] [CrossRef]
  16. Zhang, X.; Leng, C.; Hong, Y.; Pei, Z.; Cheng, I.; Basu, A. Multimodal Remote Sensing Image Registration Methods and Advancements: A Survey. Remote Sens. 2021, 13, 5128. [Google Scholar] [CrossRef]
  17. Min, C.; Gu, Y.; Li, Y.; Yang, F. Non-Rigid Infrared and Visible Image Registration by Enhanced Affine Transformation. Pattern Recognit. 2020, 106, 107377. [Google Scholar] [CrossRef]
  18. Kahaki, S.M.M.; Nordin, M.J.; Ashtari, A.H.; Zahra, S.J. Invariant Feature Matching for Image Registration Application Based on New Dissimilarity of Spatial Features. PLoS ONE 2016, 11, e0149710. [Google Scholar]
  19. Salahat, E.; Qasaimeh, M. Recent Advances in Features Extraction and Description Algorithms: A Comprehensive Survey. In Proceedings of the International Conference on Industrial Technology (ICIT), Toronto, ON, Canada, 22–25 March 2017; pp. 1059–1063. [Google Scholar]
  20. Xu, G.; Wu, Q.; Cheng, Y.; Yan, F.; Li, Z.; Yu, Q. A Robust Deformed Image Matching Method for Multi-Source Image Matching. Infrared Phys. Technol. 2021, 115, 103691. [Google Scholar] [CrossRef]
  21. Ma, J.; Jiang, X.; Fan, A.; Jiang, J.; Yan, J. Image Matching from Handcrafted to Deep Features: A Survey. Int. J. Comput. Vis. 2021, 129, 23–79. [Google Scholar] [CrossRef]
  22. Liang, D.; Ding, J.; Zhang, Y. Efficient Multisource Remote Sensing Image Matching Using Dominant Orientation of Gradient. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 2194–2205. [Google Scholar] [CrossRef]
  23. Liu, X.; Xue, J.; Xu, X.; Lu, Z.; Liu, R.; Zhao, B.; Li, Y.; Miao, Q. Robust Multimodal Remote Sensing Image Registration Based on Local Statistical Frequency Information. Remote Sens. 2022, 14, 1051. [Google Scholar] [CrossRef]
  24. Hazra, J.; Chowdhury, A.R.; Dasgupta, K.; Dutta, P. A Hybrid Structural Feature Extraction-Based Intelligent Predictive Approach for Image Registration. In Innovations in Systems and Software Engineering; Springer: Berlin/Heidelberg, Germany, 2022; pp. 1–9. [Google Scholar]
  25. Liu, Z.; Xu, G.; Xiao, J.; Yang, J.; Wang, Z.; Cheng, S. A Real-Time Registration Algorithm of UAV Aerial Images Based on Feature Matching. J. Imaging 2023, 9, 67. [Google Scholar] [CrossRef]
  26. Madhu; Kumar, R. A Hybrid Feature Extraction Technique for Content Based Medical Image Retrieval Using Segmentation and Clustering Techniques. Multimed. Tools Appl. 2022, 81, 8871–8904. [Google Scholar] [CrossRef]
  27. Zhang, P.; Luo, X.; Ma, Y.; Wang, C.; Wang, W.; Qian, X. Coarse-to-Fine Image Registration for Multi-Temporal High Resolution Remote Sensing Based on a Low-Rank Constraint. Remote Sens. 2022, 14, 573. [Google Scholar] [CrossRef]
  28. Karim, S.; Zhang, Y.; Brohi, A.A.; Asif, M.R. Feature Matching Improvement Through Merging Features for Remote Sensing Imagery. 3D Res. 2018, 9, 52. [Google Scholar] [CrossRef]
  29. Zhang, D.; Wei, H.; Huang, X.; Ni, H. Research on High Precision Image Registration Method Based on Line Segment Feature and ICP Algorithm. In Proceedings of the International Conference on Optics and Machine Vision (ICOMV), Changsha, China, 6–8 January 2023; Volume 12634, pp. 90–96. [Google Scholar]
  30. Gui, P.; He, F.; Ling, B.W.K.; Zhang, D.; Ge, Z. Normal Vibration Distribution Search-Based Differential Evolution Algorithm for Multimodal Biomedical Image Registration. Neural Comput. Appl. 2023, 35, 16223–16245. [Google Scholar] [CrossRef] [PubMed]
  31. Zhang, W.; Zhao, Y. SAR and Optical Image Registration Based on Uniform Feature Points Extraction and Consistency Gradient Calculation. Appl. Sci. 2023, 13, 1238. [Google Scholar] [CrossRef]
  32. Lin, X.; Sun, S.; Huang, W.; Sheng, B.; Li, P.; Feng, D.D. EAPT: Efficient Attention Pyramid Transformer for Image Processing. IEEE Trans. Multimed. 2023, 25, 50–61. [Google Scholar] [CrossRef]
  33. Jiang, N.; Sheng, B.; Li, P.; Lee, T.Y. PhotoHelper: Portrait Photographing Guidance Via Deep Feature Retrieval and Fusion. IEEE Trans. Multimed. 2023, 25, 2226–2238. [Google Scholar] [CrossRef]
  34. Sheng, B.; Li, P.; Ali, R.; Chen, C.L.P. Improving Video Temporal Consistency via Broad Learning System. IEEE Trans. Cybern. 2022, 52, 6662–6675. [Google Scholar] [CrossRef]
  35. Li, J.; Chen, J.; Sheng, B.; Li, P.; Yang, P.; Feng, D.D.; Qi, J. Automatic Detection and Classification System of Domestic Waste via Multimodel Cascaded Convolutional Neural Network. IEEE Trans. Ind. Inform. 2022, 18, 163–173. [Google Scholar] [CrossRef]
  36. Chen, Z.; Qiu, J.; Sheng, B.; Li, P.; Wu, E. GPSD: Generative Parking Spot Detection Using Multi-Clue Recovery Model. Vis. Comput. 2021, 37, 2657–2669. [Google Scholar] [CrossRef]
  37. Xie, Z.; Zhang, W.; Sheng, B.; Li, P.; Chen, C.L.P. BaGFN: Broad Attentive Graph Fusion Network for High-Order Feature Interactions. IEEE Trans. Neural Netw. Learn. Syst. 2023, 34, 4499–4513. [Google Scholar] [CrossRef]
  38. Donoser, M.; Bischof, H. Efficient Maximally Stable Extremal Region (MSER) Tracking. In Proceedings of the Computer Society Conference on Computer Vision and Pattern Recognition (CVPR), New York, NY, USA, 17–22 June 2006; pp. 553–560. [Google Scholar]
  39. Rosten, E.; Drummond, T. Machine Learning for High-Speed Corner Detection. In Proceedings of the 9th European Conference on Computer Vision (ECCV), Graz, Austria, 7–13 May 2006; Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2006; Volume 3951, pp. 430–443. [Google Scholar]
  40. Leutenegger, S.; Chli, M.; Siegwart, R. BRISK: Binary Robust Invariant Scalable Keypoints. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Barcelona, Spain, 6–13 November 2011; pp. 2548–2555. [Google Scholar]
  41. Rosten, E.; Porter, R.B.; Drummond, T. Faster and Better: A Machine Learning Approach to Corner Detection. IEEE Trans. Pattern Anal. Mach. Intell. (TPAMI) 2010, 32, 105–119. [Google Scholar] [CrossRef]
  42. Jeyapal, A.; Ganesan, J.; Savarimuthu, S.R.; Perumal, I.; Eswaran, P.M.; Subramanian, L.; Anbalagan, N. A Comparative Study of Feature Detection Techniques for Navigation of Visually Impaired Person in an Indoor Environment. J. Comput. Theor. Nanosci. 2020, 17, 21–26. [Google Scholar] [CrossRef]
  43. Tareen, S.A.K.; Saleem, Z. A Comparative Analysis of SIFT, SURF, KAZE, AKAZE, ORB, and BRISK. In Proceedings of the International Conference on Computing, Mathematics and Engineering Technologies (iCoMET), Sukkur, Pakistan, 3–4 March 2018; pp. 1–10. [Google Scholar]
  44. Sánchez, J.; Monzón, N.; Salgado, A. An Analysis and Implementation of the Harris Corner Detector. Image Process. Line 2018, 8, 305–328. [Google Scholar] [CrossRef]
  45. Hybrid Approach for FBIR. Available online: https://github.com/Anchal2016/Hybrid-approach-for-FBIR (accessed on 29 August 2024).
  46. Xia, G.S.; Hu, J.; Hu, F.; Shi, B.; Bai, X.; Zhong, Y.; Zhang, L.; Lu, X. AID: A Benchmark Data Set for Performance Evaluation of Aerial Scene Classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 3965–3981. [Google Scholar] [CrossRef]
Figure 1. Flow diagram of the proposed methodology.
Figure 1. Flow diagram of the proposed methodology.
Jimaging 10 00228 g001
Figure 2. Diagonal approach for hybrid feature-detection method.
Figure 2. Diagonal approach for hybrid feature-detection method.
Jimaging 10 00228 g002
Figure 3. Sampled color images from AID database [46].
Figure 3. Sampled color images from AID database [46].
Jimaging 10 00228 g003
Figure 4. Grayscale conversion of sampled color images from AID database [46].
Figure 4. Grayscale conversion of sampled color images from AID database [46].
Jimaging 10 00228 g004
Figure 5. Various rotation angles applied on park and railway station grayscale aerial images.
Figure 5. Various rotation angles applied on park and railway station grayscale aerial images.
Jimaging 10 00228 g005
Figure 6. Scaling transformations applied to VSSUT gate and Hirakud dam images. (a) 0.7 scaling factor on VSSUT gate. (b) 0.7 scaling factor on Hirakud dam. (c) 2.0 scaling factor on VSSUT gate. (d) 2.0 scaling factor on Hirakud dam.
Figure 6. Scaling transformations applied to VSSUT gate and Hirakud dam images. (a) 0.7 scaling factor on VSSUT gate. (b) 0.7 scaling factor on Hirakud dam. (c) 2.0 scaling factor on VSSUT gate. (d) 2.0 scaling factor on Hirakud dam.
Jimaging 10 00228 g006
Figure 7. Detection of feature keypoints in the park image under 150 rotation, showcasing the performance of different detectors. Green markers highlight the keypoints detected, with each subfigure corresponding to the output using a different feature-detection method.
Figure 7. Detection of feature keypoints in the park image under 150 rotation, showcasing the performance of different detectors. Green markers highlight the keypoints detected, with each subfigure corresponding to the output using a different feature-detection method.
Jimaging 10 00228 g007
Figure 8. Detection of feature keypoints in the railway station image under 150 rotation, showcasing the performance of different detectors. Green markers indicate the keypoints, and each subfigure corresponds to the output using a different feature-detection method.
Figure 8. Detection of feature keypoints in the railway station image under 150 rotation, showcasing the performance of different detectors. Green markers indicate the keypoints, and each subfigure corresponds to the output using a different feature-detection method.
Jimaging 10 00228 g008
Figure 9. Extraction of feature keypoints from the park image under 150 rotation. Green markers demonstrate the keypoints extracted, emphasizing the nuances of each algorithm with each subfigure showing results using a different feature extraction method.
Figure 9. Extraction of feature keypoints from the park image under 150 rotation. Green markers demonstrate the keypoints extracted, emphasizing the nuances of each algorithm with each subfigure showing results using a different feature extraction method.
Jimaging 10 00228 g009
Figure 10. Extraction of feature keypoints from the railway station image under 150 rotation. Each subfigure demonstrates the results using a different feature extraction method, with green markers used to emphasize keypoint locations and algorithmic nuances.
Figure 10. Extraction of feature keypoints from the railway station image under 150 rotation. Each subfigure demonstrates the results using a different feature extraction method, with green markers used to emphasize keypoint locations and algorithmic nuances.
Jimaging 10 00228 g010
Figure 11. Matching of feature keypoints in the park image across different rotational views under 150 rotation. Subfigures (af) display the matched keypoints separately to illustrate individual detector performance clearly. Subfigure (g) shows an overlaid result of the hybrid detector to demonstrate the integration of multiple detection outcomes, providing a comprehensive view of the keypoints matched by the proposed method. Each image aims to highlight the effectiveness of each feature detector in achieving consistent matching across transformations.
Figure 11. Matching of feature keypoints in the park image across different rotational views under 150 rotation. Subfigures (af) display the matched keypoints separately to illustrate individual detector performance clearly. Subfigure (g) shows an overlaid result of the hybrid detector to demonstrate the integration of multiple detection outcomes, providing a comprehensive view of the keypoints matched by the proposed method. Each image aims to highlight the effectiveness of each feature detector in achieving consistent matching across transformations.
Jimaging 10 00228 g011
Figure 12. Matching of feature keypoints in the railway station image across different rotational views under 150 rotation. Each subfigure highlights the effectiveness of each feature detector in achieving consistent matching.
Figure 12. Matching of feature keypoints in the railway station image across different rotational views under 150 rotation. Each subfigure highlights the effectiveness of each feature detector in achieving consistent matching.
Jimaging 10 00228 g012
Figure 13. Sequential presentation of detection, extraction, and matching phases for various feature detectors on two sets of airport aerial images. Each row represents a different detector and showcases the process from detection to matching.
Figure 13. Sequential presentation of detection, extraction, and matching phases for various feature detectors on two sets of airport aerial images. Each row represents a different detector and showcases the process from detection to matching.
Jimaging 10 00228 g013
Figure 14. Sequential presentation of detection, extraction, and matching phases for various feature detectors on two sets of bridge aerial images. Each row represents a different detector and showcases the process from detection to matching.
Figure 14. Sequential presentation of detection, extraction, and matching phases for various feature detectors on two sets of bridge aerial images. Each row represents a different detector and showcases the process from detection to matching.
Jimaging 10 00228 g014
Figure 15. Comparison of feature-detection performance using MSER, BRISK, and Hybrid detectors on two different images under scaling transformations. Each row demonstrates the response of the detectors at scaling factors of the original, 0.7, and 2.0, highlighting the adaptability of these algorithms to changes in image scale. (a) MSER: VSSUT Gate Image. (b) BRISK: VSSUT Gate Image. (c) MSER: HD Image. (d) BRISK: HD Image. (e) Hybrid: VSSUT Gate Image. (f) Hybrid: HD Image. (g) MSER: VSSUT Gate Image, Scale 0.7. (h) BRISK: VSSUT Gate Image, Scale 0.7. (i) MSER: HD Image, Scale 0.7. (j) BRISK: HD Image, Scale 0.7. (k) Hybrid: VSSUT Gate Image, Scale 0.7. (l) Hybrid: HD Image, Scale 0.7. (m) MSER: VSSUT Gate Image, Scale 2.0. (n) BRISK: VSSUT Gate Image, Scale 2.0. (o) MSER: HD Image, Scale 2.0. (p) BRISK: HD Image, Scale 2.0. (q) Hybrid: VSSUT Gate Image, Scale 2.0. (r) Hybrid: HD Image, Scale 2.0.
Figure 15. Comparison of feature-detection performance using MSER, BRISK, and Hybrid detectors on two different images under scaling transformations. Each row demonstrates the response of the detectors at scaling factors of the original, 0.7, and 2.0, highlighting the adaptability of these algorithms to changes in image scale. (a) MSER: VSSUT Gate Image. (b) BRISK: VSSUT Gate Image. (c) MSER: HD Image. (d) BRISK: HD Image. (e) Hybrid: VSSUT Gate Image. (f) Hybrid: HD Image. (g) MSER: VSSUT Gate Image, Scale 0.7. (h) BRISK: VSSUT Gate Image, Scale 0.7. (i) MSER: HD Image, Scale 0.7. (j) BRISK: HD Image, Scale 0.7. (k) Hybrid: VSSUT Gate Image, Scale 0.7. (l) Hybrid: HD Image, Scale 0.7. (m) MSER: VSSUT Gate Image, Scale 2.0. (n) BRISK: VSSUT Gate Image, Scale 2.0. (o) MSER: HD Image, Scale 2.0. (p) BRISK: HD Image, Scale 2.0. (q) Hybrid: VSSUT Gate Image, Scale 2.0. (r) Hybrid: HD Image, Scale 2.0.
Jimaging 10 00228 g015
Figure 16. Extraction of feature keypoints using various extractors based on scaling factors of 0.7 and 2.0 . Each row demonstrates the impact of scaling on the effectiveness of feature extraction across different images and detectors. (a) MSER: VSSUT Gate Image. (b) BRISK: VSSUT Gate Image. (c) MSER: HD Image. (d) BRISK: HD Image. (e) Hybrid: VSSUT Gate Image. (f) Hybrid: HD Image. (g) MSER: VSSUT Gate Image, Scale 0.7. (h) BRISK: VSSUT Gate Image, Scale 0.7. (i) MSER: HD Image, Scale 0.7. (j) BRISK: HD Image, Scale 0.7. (k) Hybrid: VSSUT Gate Image, Scale 0.7. (l) Hybrid: HD Image, Scale 0.7. (m) MSER: VSSUT Gate Image, Scale 2.0. (n) BRISK: VSSUT Gate Image, Scale 2.0. (o) MSER: HD Image, Scale 2.0. (p) BRISK: HD Image, Scale 2.0. (q) Hybrid: VSSUT Gate Image, Scale 2.0. (r) Hybrid: HD Image, Scale 2.0.
Figure 16. Extraction of feature keypoints using various extractors based on scaling factors of 0.7 and 2.0 . Each row demonstrates the impact of scaling on the effectiveness of feature extraction across different images and detectors. (a) MSER: VSSUT Gate Image. (b) BRISK: VSSUT Gate Image. (c) MSER: HD Image. (d) BRISK: HD Image. (e) Hybrid: VSSUT Gate Image. (f) Hybrid: HD Image. (g) MSER: VSSUT Gate Image, Scale 0.7. (h) BRISK: VSSUT Gate Image, Scale 0.7. (i) MSER: HD Image, Scale 0.7. (j) BRISK: HD Image, Scale 0.7. (k) Hybrid: VSSUT Gate Image, Scale 0.7. (l) Hybrid: HD Image, Scale 0.7. (m) MSER: VSSUT Gate Image, Scale 2.0. (n) BRISK: VSSUT Gate Image, Scale 2.0. (o) MSER: HD Image, Scale 2.0. (p) BRISK: HD Image, Scale 2.0. (q) Hybrid: VSSUT Gate Image, Scale 2.0. (r) Hybrid: HD Image, Scale 2.0.
Jimaging 10 00228 g016
Figure 17. Matching of feature keypoints using various detectors on VSSUT gate and Hirakud dam images under two scaling factors, 0.7 and 2.0. Each image series demonstrates the effect of scaling on feature matching performance. (a) MSER: VSSUT Gate Image, Scale 0.7. (b) BRISK: VSSUT Gate Image, Scale 0.7. (c) MSER: Hirakud Dam Image, Scale 0.7. (d) BRISK: Hirakud Dam Image, Scale 0.7. (e) Hybrid: VSSUT Gate Image, Scale 0.7. (f) Hybrid: Hirakud Dam Image, Scale 0.7. (g) MSER: VSSUT Gate Image, Scale 2.0. (h) BRISK: VSSUT Gate Image, Scale 2.0. (i) MSER: Hirakud Dam Image, Scale 2.0. (j) BRISK: Hirakud Dam Image, Scale 2.0. (k) Hybrid: VSSUT Gate Image, Scale 2.0. (l) Hybrid: Hirakud Dam Image, Scale 2.0.
Figure 17. Matching of feature keypoints using various detectors on VSSUT gate and Hirakud dam images under two scaling factors, 0.7 and 2.0. Each image series demonstrates the effect of scaling on feature matching performance. (a) MSER: VSSUT Gate Image, Scale 0.7. (b) BRISK: VSSUT Gate Image, Scale 0.7. (c) MSER: Hirakud Dam Image, Scale 0.7. (d) BRISK: Hirakud Dam Image, Scale 0.7. (e) Hybrid: VSSUT Gate Image, Scale 0.7. (f) Hybrid: Hirakud Dam Image, Scale 0.7. (g) MSER: VSSUT Gate Image, Scale 2.0. (h) BRISK: VSSUT Gate Image, Scale 2.0. (i) MSER: Hirakud Dam Image, Scale 2.0. (j) BRISK: Hirakud Dam Image, Scale 2.0. (k) Hybrid: VSSUT Gate Image, Scale 2.0. (l) Hybrid: Hirakud Dam Image, Scale 2.0.
Jimaging 10 00228 g017
Figure 18. Registered images of different scenes using the Hybrid feature detector. Each subfigure shows a different aerial or scene image, highlighting the detailed synthesis achieved through the registration process.
Figure 18. Registered images of different scenes using the Hybrid feature detector. Each subfigure shows a different aerial or scene image, highlighting the detailed synthesis achieved through the registration process.
Jimaging 10 00228 g018
Figure 19. Performance comparison of various feature detectors on park scene images. Each subplot visually demonstrates how each feature detector identifies keypoints within the same environmental setting. This provides insights into the adaptability and precision of each method under similar conditions, highlighting their strengths and limitations in detecting significant image features effectively.
Figure 19. Performance comparison of various feature detectors on park scene images. Each subplot visually demonstrates how each feature detector identifies keypoints within the same environmental setting. This provides insights into the adaptability and precision of each method under similar conditions, highlighting their strengths and limitations in detecting significant image features effectively.
Jimaging 10 00228 g019
Table 1. Summary of feature-detection and image-registration methods.
Table 1. Summary of feature-detection and image-registration methods.
Method/AlgorithmCharacteristicsApplications
SIFT [13]Scale-invariant, robust to rotationMultispectral image registration
MSER [38]Stable regions, distinctive featuresText detection, multi-source matching
SURF [13]Fast, robust to scale and rotationMultispectral matching
BRISK [10]Fast, scale and rotation invariantGeneric image registration
FAST [39]Very fast, lacks rotational invarianceHigh-speed feature detection
ORB [10]Combines FAST and BRIEF, with rotation invarianceCost-effective real-time applications
Harris–Affine [11]High precision in detecting corners, not scale-invariantCorner detection in images
Multispectral Facial Recognition [12]Incorporates visible and IR images using various detectorsFacial recognition across spectrums
HOG [14]Histogram of Oriented Gradients for keypoint matchingOffline transformation models
Table 2. Comparison of different interpolation (INR) techniques.
Table 2. Comparison of different interpolation (INR) techniques.
Image Quality ParametersINR TechniquesTransformation Types
AffineSimilarityProjective
MSENearest Neighbor0.004380.004450.00431
Bilinear0.002850.002930.00286
Bicubic0.002140.002190.00221
RMSENearest Neighbor0.066200.066740.06565
Bilinear0.053350.054110.05348
Bicubic0.046260.046780.04704
SNRNearest Neighbor18.2835218.2114618.35780
Bilinear20.1610120.0356120.18970
Bicubic21.3979721.2989721.24642
PSNRNearest Neighbor23.5826223.5122923.65465
Bilinear25.4577225.3352125.48598
Bicubic26.6954926.5982626.79086
Table 3. Quantitative evaluation of feature-detection performance at various rotation angles for the railway station image. The table displays keypoint detection, extraction, and matching statistics for each detector.
Table 3. Quantitative evaluation of feature-detection performance at various rotation angles for the railway station image. The table displays keypoint detection, extraction, and matching statistics for each detector.
DetectorDet.
Kpts1
Det.
Kpts2
Ext.
Kpts1
Ext.
Kpts2
Matched
Kpts
Match
Rate (%)
Elapsed
Time (s)
CPU
Time (s)
PMT
Time (s)
Rotation Angle: 30 , Sum: 146.7800, Mean: 20.9685, Variance: 64.7572, Std. Dev.: 8.0471
BRISK5727364306807210.5911.4212.8411.42
FAST2342912072885418.754.714.194.71
MSER6785916785917813.206.798.566.79
ORB6753993667539936303730.574.925.034.93
Harris6655255885049719.254.805.274.80
MinEigen414037853573374884722.603.583.803.59
Hybrid56974656974823831.823.343.223.34
Rotation Angle: 70 , Sum: 149.5300, Mean: 21.3614, Variance: 48.5798, Std. Dev.: 6.9699
BRISK5727304306777410.9311.6913.8311.70
FAST2342632072566826.564.474.094.45
MSER67858667858614224.237.467.307.46
ORB6753953567539535307332.235.014.705.02
Harris6654795884507516.673.483.113.49
MinEigen414036403573359370119.513.263.033.27
Hybrid56973256973214219.402.862.442.85
Rotation Angle: 90 , Sum: 637.2900, Mean: 91.0414, Variance: 186.4632, Std. Dev.: 13.6551
BRISK57256943042626862.913.993.784.00
FAST23423420720720599.033.953.703.95
MSER678678678678678100.005.895.445.89
ORB67536753675367536753100.004.183.894.19
Harris66566558858951887.953.573.253.57
MinEigen4140414035733572312287.402.912.342.91
Hybrid569569569569569100.002.712.412.71
Rotation Angle: 120 , Sum: 148.2700, Mean: 21.1814, Variance: 56.2275, Std. Dev.: 7.4985
BRISK5727164306639213.883.493.643.49
FAST2342912072884917.014.684.474.68
MSER6735916785917813.207.738.457.73
ORB6753993667539936303730.575.175.425.19
Harris66552558850410120.044.074.134.07
MinEigen414037353573374781521.753.152.723.15
Hybrid56974856974823831.822.982.612.99
Rotation Angle: 150 , Sum: 150.1200, Mean: 21.4457, Variance: 45.5085, Std. Dev.: 6.7460
BRISK5727224306739514.126.497.866.48
FAST2342952072894214.533.893.303.88
MSER67858067858010718.457.577.867.57
ORB675310,381675310,381296428.555.104.915.08
Harris6654715884518418.633.423.303.42
MinEigen414034243573338883824.733.483.193.48
Hybrid56973656973622931.112.883.192.88
Rotation Angle: 180 , Sum: 682.85, Mean: 97.5500, Variance: 33.1407, Std. Dev.: 5.7567
BRISK57256843042636084.513.602.443.59
FAST234234207207207100.003.983.033.98
MSER67867867867867499.416.237.116.23
ORB67536753675367536753100.003.753.753.75
Harris66566558858858599.493.643.273.63
MinEigen4140414035733568354899.443.042.883.04
Hybrid569569569569569100.002.842.582.84
Table 4. Quantitative evaluation of feature-detection performance at various rotation angles for the park image. The table displays keypoint detection, extraction, and matching statistics for each detector.
Table 4. Quantitative evaluation of feature-detection performance at various rotation angles for the park image. The table displays keypoint detection, extraction, and matching statistics for each detector.
DetectorDetected
Kpts1
Detected
Kpts2
Extracted
Kpts1
Extracted
Kpts2
Matched
Kpts
Matched
Rate (%)
Elapsed
Time (s)
CPU
Time (s)
PMT
Time (s)
Rotation Angle: 30
BRISK16341973149919511738.8613.9216.0813.93
FAST8941128859112517915.914.846.724.84
MSER76777976777912616.717.517.787.50
ORB13,70418,52113,70418,521517727.957.4510.487.44
Harris117610811119104916215.444.614.134.60
MinEigen521345904608455064514.174.353.894.34
Hybrid9761009976100929028.743.633.593.64
Rotation Angle: 70
BRISK16341906149918721779.4513.2913.2513.29
FAST89495185994417018.004.885.114.88
MSER76773976773917924.226.747.886.74
ORB13,70417,71313,70417,713539330.448.9110.458.92
Harris117612701119123614211.484.843.924.85
MinEigen521346234608456849510.833.834.753.84
Hybrid9761063976106332230.293.443.863.44
Rotation Angle: 90
BRISK163416481499151293862.034.744.454.75
FAST89489485985979792.783.423.133.43
MSER76776776776775598.4316.6923.3016.68
ORB13,70413,70413,70413,70413,704100.005.677.225.66
Harris117611761119111993383.374.233.614.23
MinEigen5213521346084613360078.043.643.723.64
Hybrid976976976976976100.002.552.772.55
Rotation Angle: 120
BRISK16341972149919481568.004.314.724.31
FAST8941128859112518516.443.614.413.61
MSER76777976777912616.176.587.916.59
ORB13,70418,52113,70418,521517727.958.3211.368.33
Harris117610811119104915815.064.915.734.89
MinEigen521345904608455062013.623.373.773.37
Hybrid9761009976100929128.843.334.443.33
Rotation Angle: 150
BRISK16341932149918991799.424.555.194.55
FAST8941144859113716314.334.554.774.56
MSER76772676772616322.456.496.646.49
ORB13,70418,28213,70418,282513228.077.5711.287.57
Harris117612101119118214912.604.014.114.02
MinEigen521346324608459255912.173.733.773.73
Hybrid9761022976102226726.123.613.713.61
Rotation Angle: 180
BRISK1634163414991501132588.272.952.752.95
FAST89489485986185999.764.063.004.06
MSER76776776776775498.306.786.336.79
ORB13,70413,70413,70413,70413,704100.006.567.886.56
Harris1176117611191121111899.734.243.224.24
MinEigen5213521346084615459099.453.512.733.50
Hybrid976976976976976100.002.812.942.81
Table 5. Scene-to-model registration, i.e., different images of the same scene applied on two sets of aerial images: airport and bridge.
Table 5. Scene-to-model registration, i.e., different images of the same scene applied on two sets of aerial images: airport and bridge.
Detection
Method
Detected
Kpts1
Detected
Kpts2
Extracted
Kpts1
Extracted
Kpts2
Matched
Kpts
Matched
Rate (%)
Elapsed
Time
CPU
Time
PMT
Time
Airport Aerial Images
BRISK2787311956042419.854.934.734.93
FAST2014641504042834.656.115.256.12
MSER1732701732703412.596.265.306.25
ORB125337591253375912917.155.514.565.51
Harris1533421172892136.305.524.835.52
MinEigen9552176697168910029.605.594.095.59
Hybrid89257892573873.904.483.864.47
Bridge Aerial Images
BRISK83057764441278.455.694.695.68
FAST475294397239918.804.534.144.53
MSER55838555838579.056.806.986.80
ORB380535733805357312617.605.085.025.08
Harris4353823503291218.205.114.645.13
MinEigen3664346531012897488.254.944.954.94
Hybrid3672823672821424.804.454.544.45
Table 6. Different sizes, i.e., scaling vectors applied on VSSUT entrance and Hirakud dam images using the BRISK detection method.
Table 6. Different sizes, i.e., scaling vectors applied on VSSUT entrance and Hirakud dam images using the BRISK detection method.
Image NameScaling VectorScaled SizeIQABicubicBilinearNearest
VSSUT0.7717 × 538PSNR30.3129.5226.74
1024 × 768 134 KB 65.7 KBMSE0.000930.001120.00212
Hirakud dam0.7385 × 289PSNR31.7529.3326.70
550 × 412 34.7 KB 15.4 KBMSE0.000670.001170.00214
VSSUT2.02048 × 1536PSNR26.6025.9424.38
1024 × 768 134 KB 330 KBMSE0.002190.002490.00364
Hirakud dam2.01100 × 824PSNR31.3130.2028.81
550 × 412 34.7 KB 73.2 KBMSE0.000740.000950.00131
Table 7. Different sizes, i.e., scaling vectors applied on VSSUT entrance and Hirakud dam images using the MSER detection method.
Table 7. Different sizes, i.e., scaling vectors applied on VSSUT entrance and Hirakud dam images using the MSER detection method.
Image NameScaling VectorScaled SizeIQABicubicBilinearNearest
VSSUT0.7717 × 538PSNR30.6630.3126.59
1024 × 768 134 KB 65.7 KBMSE0.000860.000930.00219
Hirakud dam0.7385 × 289PSNR29.8329.1425.68
550 × 412 34.7 KB 15.4 KBMSE0.001040.001220.00270
VSSUT2.02048 × 1536PSNR26.8725.9224.21
1024 × 768 134 KB 330 KBMSE0.002060.002560.00379
Hirakud dam2.01100 × 824PSNR30.5728.0325.47
550 × 412 34.7 KB 73.2 KBMSE0.000880.001570.00283
Table 8. Different sizes, i.e., scaling vectors applied on VSSUT entrance and Hirakud dam images using the Hybrid detection method.
Table 8. Different sizes, i.e., scaling vectors applied on VSSUT entrance and Hirakud dam images using the Hybrid detection method.
Image NameScaling VectorScaled SizeIQABicubicBilinearNearest
VSSUT0.7717 × 538PSNR31.4730.3427.02
1024 × 768 134 KB 65.7 KBMSE0.000710.000930.00198
Hirakud dam0.7385 × 289PSNR34.1131.6626.78
550 × 412 34.7 KB 15.4 KBMSE0.000390.000680.00210
VSSUT2.02048 × 1536PSNR26.8926.0424.38
1024 × 768 134 KB 330 KBMSE0.002050.002490.00364
Hirakud dam2.01100 × 824PSNR31.3129.7525.93
550 × 412 34.7 KB 73.2 KBMSE0.000740.001060.00255
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kumawat, A.; Panda, S.; Gerogiannis, V.C.; Kanavos, A.; Acharya, B.; Manika, S. A Hybrid Approach for Image Acquisition Methods Based on Feature-Based Image Registration. J. Imaging 2024, 10, 228. https://doi.org/10.3390/jimaging10090228

AMA Style

Kumawat A, Panda S, Gerogiannis VC, Kanavos A, Acharya B, Manika S. A Hybrid Approach for Image Acquisition Methods Based on Feature-Based Image Registration. Journal of Imaging. 2024; 10(9):228. https://doi.org/10.3390/jimaging10090228

Chicago/Turabian Style

Kumawat, Anchal, Sucheta Panda, Vassilis C. Gerogiannis, Andreas Kanavos, Biswaranjan Acharya, and Stella Manika. 2024. "A Hybrid Approach for Image Acquisition Methods Based on Feature-Based Image Registration" Journal of Imaging 10, no. 9: 228. https://doi.org/10.3390/jimaging10090228

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop