[go: up one dir, main page]

 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (129)

Search Parameters:
Keywords = bilateral filter

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
18 pages, 3952 KiB  
Article
WGCAMNet: Wasserstein Generative Adversarial Network Augmented and Custom Attention Mechanism Based Deep Neural Network for Enhanced Brain Tumor Detection and Classification
by Fatema Binte Alam, Tahasin Ahmed Fahim, Md Asef, Md Azad Hossain and M. Ali Akber Dewan
Information 2024, 15(9), 560; https://doi.org/10.3390/info15090560 - 11 Sep 2024
Viewed by 237
Abstract
Brain tumor detection and categorization of its subtypes are essential for early diagnosis and improving patient outcomes. This research presents a cutting-edge approach that employs advanced data augmentation and deep learning methodologies for brain tumor classification. For this work, a dataset of 6982 [...] Read more.
Brain tumor detection and categorization of its subtypes are essential for early diagnosis and improving patient outcomes. This research presents a cutting-edge approach that employs advanced data augmentation and deep learning methodologies for brain tumor classification. For this work, a dataset of 6982 MRI images from the IEEE Data Port was considered, in which a total of 5712 images of four classes (1321 glioma, 1339 meningioma, 1595 no tumor, and 1457 pituitary) were used in the training set and a total of 1270 images of the same four classes were used in the testing set. A Wasserstein Generative Adversarial Network was implemented to generate synthetic images to address class imbalance, resulting in a balanced and consistent dataset. A comparison was conducted between various data augmentation metholodogies demonstrating that Wasserstein Generative Adversarial Network-augmented results perform excellently over traditional augmentation (such as rotation, shift, zoom, etc.) and no augmentation. Additionally, a Gaussian filter and normalization were applied during preprocessing to reduce noise, highlighting its superior accuracy and edge preservation by comparing its performance to Median and Bilateral filters. The classifier model combines parallel feature extraction from modified InceptionV3 and VGG19 followed by custom attention mechanisms for effectively capturing the characteristics of each tumor type. The model was trained for 64 epochs using model checkpoints to save the best-performing model based on validation accuracy and learning rate adjustments. The model achieved a 99.61% accuracy rate on the testing set, with precision, recall, AUC, and loss of 0.9960, 0.9960, 0.0153, and 0.9999, respectively. The proposed architecture’s explainability has been enhanced by t-SNE plots, which show unique tumor clusters, and Grad-CAM representations, which highlight crucial areas in MRI scans. This research showcases an explainable and robust approach for correctly classifying four brain tumor types, combining WGAN-augmented data with advanced deep learning models in feature extraction. The framework effectively manages class imbalance and integrates a custom attention mechanism, outperforming other models, thereby improving diagnostic accuracy and reliability in clinical settings. Full article
(This article belongs to the Special Issue Applications of Deep Learning in Bioinformatics and Image Processing)
Show Figures

Figure 1

Figure 1
<p>Detailed Workflow Diagram of the Brain Tumor MRI Classification Process.</p>
Full article ">Figure 2
<p>The sample collection includes axial, coronal, and sagittal images of four brain tumor types: (<b>a</b>) gliomas, (<b>b</b>) meningiomas, (<b>c</b>) no tumor, and (<b>d</b>) pituitary tumors.</p>
Full article ">Figure 3
<p>Effects of Different Filtering Methods on Sample Brain Tumor Image.</p>
Full article ">Figure 4
<p>Distribution of samples per class before and after WGAN augmentation.</p>
Full article ">Figure 5
<p>Sample WGAN genarated images.</p>
Full article ">Figure 6
<p>Block Diagram of the Proposed Model for Brain Tumor Classification.</p>
Full article ">Figure 7
<p>Performance metrics of the proposed architecture over 64 epochs, including (<b>a</b>) accuracy, (<b>b</b>) loss, (<b>c</b>) precision, (<b>d</b>) recall, and (<b>e</b>) AUC.</p>
Full article ">Figure 7 Cont.
<p>Performance metrics of the proposed architecture over 64 epochs, including (<b>a</b>) accuracy, (<b>b</b>) loss, (<b>c</b>) precision, (<b>d</b>) recall, and (<b>e</b>) AUC.</p>
Full article ">Figure 7 Cont.
<p>Performance metrics of the proposed architecture over 64 epochs, including (<b>a</b>) accuracy, (<b>b</b>) loss, (<b>c</b>) precision, (<b>d</b>) recall, and (<b>e</b>) AUC.</p>
Full article ">Figure 8
<p>Confusion matrix showing the classification accuracy across different brain tumor types.</p>
Full article ">Figure 9
<p>t-SNE projection illustrating the separation of different brain tumor classes.</p>
Full article ">Figure 10
<p>Grad-CAM visualization highlighting important regions in the MRI images for model predictions.</p>
Full article ">
19 pages, 7835 KiB  
Article
Auxiliary Diagnosis of Dental Calculus Based on Deep Learning and Image Enhancement by Bitewing Radiographs
by Tai-Jung Lin, Yen-Ting Lin, Yuan-Jin Lin, Ai-Yun Tseng, Chien-Yu Lin, Li-Ting Lo, Tsung-Yi Chen, Shih-Lun Chen, Chiung-An Chen, Kuo-Chen Li and Patricia Angela R. Abu
Bioengineering 2024, 11(7), 675; https://doi.org/10.3390/bioengineering11070675 - 2 Jul 2024
Cited by 2 | Viewed by 1075
Abstract
In the field of dentistry, the presence of dental calculus is a commonly encountered issue. If not addressed promptly, it has the potential to lead to gum inflammation and eventual tooth loss. Bitewing (BW) images play a crucial role by providing a comprehensive [...] Read more.
In the field of dentistry, the presence of dental calculus is a commonly encountered issue. If not addressed promptly, it has the potential to lead to gum inflammation and eventual tooth loss. Bitewing (BW) images play a crucial role by providing a comprehensive visual representation of the tooth structure, allowing dentists to examine hard-to-reach areas with precision during clinical assessments. This visual aid significantly aids in the early detection of calculus, facilitating timely interventions and improving overall outcomes for patients. This study introduces a system designed for the detection of dental calculus in BW images, leveraging the power of YOLOv8 to identify individual teeth accurately. This system boasts an impressive precision rate of 97.48%, a recall (sensitivity) of 96.81%, and a specificity rate of 98.25%. Furthermore, this study introduces a novel approach to enhancing interdental edges through an advanced image-enhancement algorithm. This algorithm combines the use of a median filter and bilateral filter to refine the accuracy of convolutional neural networks in classifying dental calculus. Before image enhancement, the accuracy achieved using GoogLeNet stands at 75.00%, which significantly improves to 96.11% post-enhancement. These results hold the potential for streamlining dental consultations, enhancing the overall efficiency of dental services. Full article
Show Figures

Figure 1

Figure 1
<p>Dental calculus symptoms on a BW image: (<b>a</b>) dental calculus symptoms; (<b>b</b>) absence of dental calculus symptoms.</p>
Full article ">Figure 2
<p>The flowchart used in this study.</p>
Full article ">Figure 3
<p>Image-annotation step preserves edges on both sides of the tooth: (<b>a</b>) less of the tooth edge is retained; (<b>b</b>) more of the tooth edge is retained.</p>
Full article ">Figure 4
<p>The flowchart used in single-tooth image segmentation.</p>
Full article ">Figure 5
<p>The BW image-preprocessing results: (<b>a</b>) mean filter; (<b>b</b>) binarization.</p>
Full article ">Figure 6
<p>The results of pixel projection: (<b>a</b>) horizontal pixel projection; (<b>b</b>) horizontal pixel projection-coordinate graph; (<b>c</b>)vertical pixel projection; (<b>d</b>) vertical pixel projection-coordinate graph.</p>
Full article ">Figure 7
<p>Image-enhancement flowchart.</p>
Full article ">Figure 8
<p>Image-enhancement results: (<b>a</b>) binarization; (<b>b</b>) mathematical morphology; (<b>c</b>) added green line represents canny; (<b>d</b>) overlap onto the original image.</p>
Full article ">Figure 9
<p>Data augmentation results. (<b>a</b>) Dental calculus. (<b>b</b>) Without dental calculus.</p>
Full article ">Figure 10
<p>YOLO validation results: (<b>a</b>) YOLOv8 detect results; (<b>b</b>) YOLOv8 validation PR curve.</p>
Full article ">Figure 11
<p>The results of tooth extraction based on YOLOv8.</p>
Full article ">Figure 12
<p>CNN training process.</p>
Full article ">Figure 13
<p>GoogLeNet loss process.</p>
Full article ">
21 pages, 1647 KiB  
Article
Artificial Intelligence Approach for Classifying Images of Upper-Atmospheric Transient Luminous Events
by Axi Aguilera and Vidya Manian
Sensors 2024, 24(10), 3208; https://doi.org/10.3390/s24103208 - 18 May 2024
Cited by 1 | Viewed by 633
Abstract
Transient Luminous Events (TLEs) are short-lived, upper-atmospheric optical phenomena associated with thunderstorms. Their rapid and random occurrence makes manual classification laborious and time-consuming. This study presents an effective approach to automating the classification of TLEs using state-of-the-art Convolutional Neural Networks (CNNs) and a [...] Read more.
Transient Luminous Events (TLEs) are short-lived, upper-atmospheric optical phenomena associated with thunderstorms. Their rapid and random occurrence makes manual classification laborious and time-consuming. This study presents an effective approach to automating the classification of TLEs using state-of-the-art Convolutional Neural Networks (CNNs) and a Vision Transformer (ViT). The ViT architecture and four different CNN architectures, namely, ResNet50, ResNet18, GoogLeNet, and SqueezeNet, are employed and their performance is evaluated based on their accuracy and execution time. The models are trained on a dataset that was augmented using rotation, translation, and flipping techniques to increase its size and diversity. Additionally, the images are preprocessed using bilateral filtering to enhance their quality. The results show high classification accuracy across all models, with ResNet50 achieving the highest accuracy. However, a trade-off is observed between accuracy and execution time, which should be considered based on the specific requirements of the task. This study demonstrates the feasibility and effectiveness of using transfer learning and pre-trained CNNs for the automated classification of TLEs. Full article
(This article belongs to the Special Issue Applications of Video Processing and Computer Vision Sensor II)
Show Figures

Figure 1

Figure 1
<p>CNN architecture for TLE classification.</p>
Full article ">Figure 2
<p>Architecture of the ViT for TLE classification based on [<a href="#B27-sensors-24-03208" class="html-bibr">27</a>].</p>
Full article ">Figure 3
<p>Sample of images in the collected data. (<b>a</b>) Blue jet; (<b>b</b>) elve; (<b>c</b>) gigantic jet; (<b>d</b>) halo.</p>
Full article ">Figure 4
<p>Sample of images of sprites in the collected data. (<b>a</b>) sprite; (<b>b</b>) sprite–halo; (<b>c</b>) sprite–jellyfish.</p>
Full article ">Figure 5
<p>Loss and accuracy across different models using the ADAM optimizer.</p>
Full article ">Figure 5 Cont.
<p>Loss and accuracy across different models using the ADAM optimizer.</p>
Full article ">Figure 6
<p>Loss and accuracy across different models using the SGD optimizer.</p>
Full article ">Figure 7
<p>Confusion matrix assessment values for ResNet50. (<b>a</b>) ADAM optimizer; (<b>b</b>) SGD optimizer.</p>
Full article ">Figure 8
<p>Confusion matrix assessment values for the Vision Transformer. (<b>a</b>) ADAM optimizer; (<b>b</b>) SGD optimizer.</p>
Full article ">Figure 9
<p>Loss and accuracy across different pre-trained CNN models using the ADAM optimizer.</p>
Full article ">Figure 10
<p>Loss and accuracy across different pre-trained CNN models using the SGD optimizer.</p>
Full article ">Figure 11
<p>Confusion matrix assessment values for ResNet50 using transfer learning. (<b>a</b>) ADAM optimizer; (<b>b</b>) SGD optimizer.</p>
Full article ">
14 pages, 3868 KiB  
Article
Research on Tire Surface Damage Detection Method Based on Image Processing
by Jiaqi Chen, Aijuan Li, Fei Zheng, Shanshan Chen, Weikai He and Guangping Zhang
Sensors 2024, 24(9), 2778; https://doi.org/10.3390/s24092778 - 26 Apr 2024
Viewed by 675
Abstract
The performance of the tire has a very important impact on the safe driving of the car, and in the actual use of the tire, due to complex road conditions or use conditions, it will inevitably cause immeasurable wear, scratches and other damage. [...] Read more.
The performance of the tire has a very important impact on the safe driving of the car, and in the actual use of the tire, due to complex road conditions or use conditions, it will inevitably cause immeasurable wear, scratches and other damage. In order to effectively detect the damage existing in the key parts of the tire, a tire surface damage detection method based on image processing was proposed. In this method, the image of tire side is captured by camera first. Then, the collected images are preprocessed by optimizing the multi-scale bilateral filtering algorithm to enhance the detailed information of the damaged area, and the optimization effect is obvious. Thirdly, the image segmentation based on clustering algorithm is carried out. Finally, the Harris corner detection method is used to capture the “salt and pepper” corner of the target region, and the segmsegmed binary image is screened and matched based on histogram correlation, and the target region is finally obtained. The experimental results show that the similarity detection is accurate, and the damage area can meet the requirements of accurate identification. Full article
Show Figures

Figure 1

Figure 1
<p>Sample demos. (<b>a</b>) The fifth sample of the total sample; (<b>b</b>) The 11th sample of the total sample; (<b>c</b>) The 13th sample of the total sample.</p>
Full article ">Figure 2
<p>Flow chart of the proposed algorithm.</p>
Full article ">Figure 3
<p>Schematic diagram of bilateral filtering principle. (<b>a</b>) Image noise distribution before bilateral filtering; (<b>b</b>) Edge protection and noise reduction; (<b>c</b>) Image noise distribution after bilateral filtering.</p>
Full article ">Figure 4
<p>Multi-scale bilateral filter flow chart.</p>
Full article ">Figure 5
<p>Comparison of blur results. (<b>a</b>) Original image; (<b>b</b>) Grayscale histogram of the original image; (<b>c</b>) After optimizing bilateral filtering; (<b>d</b>) Grayscale histogram of the Optimize bilateral filter.</p>
Full article ">Figure 6
<p>Experimental results of clustering image segmentation.</p>
Full article ">Figure 7
<p>Corner characteristic diagram.</p>
Full article ">Figure 8
<p>(<b>a</b>) Binarized image after capturing corner points; (<b>b</b>) Binarized image after capturing optimized.</p>
Full article ">Figure 9
<p>(<b>a</b>) Filter the matched target area; (<b>b</b>) The final experimental results.</p>
Full article ">Figure 10
<p>(<b>a</b>) Diagram of experimental anomaly results; (<b>b</b>) Pic. 13 Binary image; (<b>c</b>) Pic. 13 Corner point binarization image.</p>
Full article ">
16 pages, 7993 KiB  
Article
A New Method for Extracting Refined Sketches of Ancient Murals
by Zhiji Yu, Shuqiang Lyu, Miaole Hou, Yutong Sun and Lihong Li
Sensors 2024, 24(7), 2213; https://doi.org/10.3390/s24072213 - 29 Mar 2024
Viewed by 757
Abstract
Mural paintings, as the main components of painted cultural relics, have essential research value and historical significance. Due to their age, murals are easily damaged. Obtaining intact sketches is the first step in the conservation and restoration of murals. However, sketch extraction often [...] Read more.
Mural paintings, as the main components of painted cultural relics, have essential research value and historical significance. Due to their age, murals are easily damaged. Obtaining intact sketches is the first step in the conservation and restoration of murals. However, sketch extraction often suffers from problems such as loss of details, too thick lines, or noise interference. To overcome these problems, a mural sketch extraction method based on image enhancement and edge detection is proposed. The experiments utilize Contrast Limited Adaptive Histogram Equalization (CLAHE) and bilateral filtering to enhance the mural images. This can enhance the edge features while suppressing the noise generated by over-enhancement. Finally, we extract the refined sketch of the mural using the Laplacian Edge with fine noise remover (FNR). The experimental results show that this method is superior to other methods in terms of visual effect and related indexes, and it can extract the complex line regions of the mural. Full article
Show Figures

Figure 1

Figure 1
<p>Simulated mural and its corresponding original sketch. (<b>a</b>,<b>b</b>) The original sketch and image of “SHUIYUE”, (<b>c</b>,<b>d</b>) the original sketch and image of “DAFO”, respectively.</p>
Full article ">Figure 2
<p>The proposed approach.</p>
Full article ">Figure 3
<p>The process of CLAHE on murals.</p>
Full article ">Figure 4
<p>Comparison of results for different filters. (<b>a</b>) The original image (CLAHE), (<b>b</b>) the result of (<b>a</b>) using Bilateral filter, and (<b>c</b>) the result of (<b>a</b>) using Gaussian filter. The figures below (<b>a1</b>–<b>c1</b>) are the corresponding partially enlarged details.</p>
Full article ">Figure 5
<p>The left shows the detection results of the original Laplacian method, and the right shows the detection results of the Laplacian Edge used in this paper.</p>
Full article ">Figure 6
<p>Schematic diagram of fine noise remover, where <span class="html-italic">n</span> and <span class="html-italic">m</span> are settable terms.</p>
Full article ">Figure 7
<p>(<b>a</b>–<b>d</b>) Images enhanced by CLAHE. (<b>a</b>–<b>d</b>) The images obtained by setting the threshold of CLAHE to 1, 2, 3, and 4, respectively; (<b>e</b>) the result after directly extracting the original image, (<b>f</b>) the result after extracting the enhanced image, and (<b>e1</b>,<b>f1</b>) the corresponding local zoom-in details.</p>
Full article ">Figure 8
<p>(<b>a</b>) The image after being enhanced by CLAHE, (<b>b</b>) the result of bilateral filtering of (<b>a</b>); (<b>a1</b>,<b>b1</b>) the corresponding sketch extraction results, respectively.</p>
Full article ">Figure 9
<p>(<b>a</b>) The original painted cultural relics images; (<b>b</b>) ground truth; (<b>c</b>) the result after processing of the original image using the Laplacian Edge; (<b>d</b>) the result after processing of the enhanced image using the Laplacian Edge; (<b>e</b>) the result after processing of the enhanced image using the Laplacian Edge with FNR.</p>
Full article ">Figure 10
<p>Ablation experiment: (<b>a</b>) ground truth; (<b>b</b>) the result of CLAHE-enhanced; (<b>c</b>) the result of Bilateral filtering-enhanced; (<b>d</b>) ours.</p>
Full article ">Figure 11
<p>Comparison with other methods: (<b>a</b>) The original painted cultural relics images; (<b>b</b>) ground truth; (<b>c</b>) Canny; (<b>d</b>) HED; (<b>e</b>) LDC; (<b>f</b>) PiDiNet; (<b>g</b>) ours. The figures below (<b>a1</b>–<b>g1</b>) are the corresponding partially enlarged details.</p>
Full article ">Figure 12
<p>Comparison with other methods: (<b>a</b>) actual cultural relics images; (<b>b</b>) Canny; (<b>c</b>) HED; (<b>d</b>) LDC; (<b>e</b>) PiDiNet; (<b>f</b>) ours. The figures below (<b>a1</b>–<b>f1</b>) are the corresponding partially enlarged details.</p>
Full article ">
16 pages, 21787 KiB  
Article
Expanding Sparse Radar Depth Based on Joint Bilateral Filter for Radar-Guided Monocular Depth Estimation
by Chen-Chou Lo and Patrick Vandewalle
Sensors 2024, 24(6), 1864; https://doi.org/10.3390/s24061864 - 14 Mar 2024
Viewed by 734
Abstract
Radar data can provide additional depth information for monocular depth estimation. It provides a cost-effective solution and is robust in various weather conditions, particularly when compared with lidar. Given the sparse and limited vertical field of view of radar signals, existing methods employ [...] Read more.
Radar data can provide additional depth information for monocular depth estimation. It provides a cost-effective solution and is robust in various weather conditions, particularly when compared with lidar. Given the sparse and limited vertical field of view of radar signals, existing methods employ either a vertical extension of radar points or the training of a preprocessing neural network to extend sparse radar points under lidar supervision. In this work, we present a novel radar expansion technique inspired by the joint bilateral filter, tailored for radar-guided monocular depth estimation. Our approach is motivated by the synergy of spatial and range kernels within the joint bilateral filter. Unlike traditional methods that assign a weighted average of nearby pixels to the current pixel, we expand sparse radar points by calculating a confidence score based on the values of spatial and range kernels. Additionally, we propose the use of a range-aware window size for radar expansion instead of a fixed window size in the image plane. Our proposed method effectively increases the number of radar points from an average of 39 points in a raw radar frame to an average of 100 K points. Notably, the expanded radar exhibits fewer intrinsic errors when compared with raw radar and previous methodologies. To validate our approach, we assess our proposed depth estimation model on the nuScenes dataset. Comparative evaluations with existing radar-guided depth estimation models demonstrate its state-of-the-art performance. Full article
(This article belongs to the Special Issue Sensing and Processing for 3D Computer Vision: 3rd Edition)
Show Figures

Figure 1

Figure 1
<p>Samples from nuScenes [<a href="#B17-sensors-24-01864" class="html-bibr">17</a>] with lidar and different radar formats: (<b>a</b>) an image with 1 sweep of sparse lidar projection, (<b>b</b>) 5 sweeps of raw sparse radar projection, (<b>c</b>) height-extended radar [<a href="#B14-sensors-24-01864" class="html-bibr">14</a>], (<b>d</b>) <math display="inline"><semantics> <msup> <mi>S</mi> <mn>3</mn> </msup> </semantics></math> radar (ad hoc) [<a href="#B18-sensors-24-01864" class="html-bibr">18</a>], (<b>e</b>) MER with RC-PDA <math display="inline"><semantics> <mrow> <mo>≥</mo> <mn>0.5</mn> </mrow> </semantics></math> [<a href="#B15-sensors-24-01864" class="html-bibr">15</a>], (<b>f</b>) proposed joint bilateral filter expansion. All the point sizes are dilated for better visualization. The color of lidar and radar data indicates the distance, ranging from 0 m (blue) to 80 m (dark red).</p>
Full article ">Figure 2
<p>Illustration of the proposed joint bilateral filter expansion process. The expansion window for each radar point is initially determined by a predefined width and height, alongside the distance of the radar point under consideration, highlighted with red frames. Subsequently, both spatial and range kernels are employed to determine the expansion confidence score for every point within the window. The final radar expansion is determined by considering the bilateral confidence alongside a predefined threshold. The details of the proposed joint bilateral expansion method are summarized in Algorithm 1. The color of radar data indicates the distance, ranging from 0 m (blue) to 80 m (dark red).</p>
Full article ">Figure 3
<p>Schematic diagram illustrating the proposed radar expansion method. The sparse radar depth and color intensity from camera images are given features. Following the computation of the expansion window for each sparse radar point, range and spatial confidence maps are calculated based on color and distance differences. The JBF confidence map is obtained by multiplying the range and spatial confidence maps, and the expansion map is generated after applying a threshold on the JBF confidence map. Finally, the expanded radar depth is obtained by combining the raw sparse radar depth with the expansion map.</p>
Full article ">Figure 4
<p>Samples of the proposed radar expansion. <b>Top</b> row: RGB image with the 5-frame raw radar. <b>Bottom</b> row: RGB image with the proposed JBF radar with <math display="inline"><semantics> <mrow> <msub> <mi>σ</mi> <mi>s</mi> </msub> <mo>=</mo> <mn>25</mn> </mrow> </semantics></math> and <math display="inline"><semantics> <mrow> <msub> <mi>σ</mi> <mi>r</mi> </msub> <mo>=</mo> <mn>10</mn> </mrow> </semantics></math>. All the point sizes are dilated for better visualization and better viewing in color. The color of expanded radar indicates the distance, ranging from 0 m (blue) to 80 m (dark red).</p>
Full article ">Figure 5
<p>Qualitative comparison of results for radar-guided depth estimation experiments. From top to bottom: input monocular image, DORN<sub><span class="html-italic">radar</span></sub> [<a href="#B14-sensors-24-01864" class="html-bibr">14</a>], RC-PDA [<a href="#B15-sensors-24-01864" class="html-bibr">15</a>], Lin [<a href="#B13-sensors-24-01864" class="html-bibr">13</a>], RCDPT [<a href="#B34-sensors-24-01864" class="html-bibr">34</a>], our proposed radar with RCDPT. The color of the estimated depth indicates the distance, ranging from 0 m (blue) to 80 m (dark red).</p>
Full article ">Figure 6
<p>Samples of the proposed radar expansion with different <math display="inline"><semantics> <msub> <mi>σ</mi> <mi>s</mi> </msub> </semantics></math> and <math display="inline"><semantics> <msub> <mi>σ</mi> <mi>r</mi> </msub> </semantics></math> in spatial and range kernels. The columns from left to right show the RGB image with 5-frame raw radar. The proposed JBF radar with <math display="inline"><semantics> <mrow> <msub> <mi>σ</mi> <mi>s</mi> </msub> <mo>=</mo> <mn>10</mn> </mrow> </semantics></math> and <math display="inline"><semantics> <mrow> <msub> <mi>σ</mi> <mi>r</mi> </msub> <mo>=</mo> <mn>5</mn> </mrow> </semantics></math>; <math display="inline"><semantics> <mrow> <msub> <mi>σ</mi> <mi>s</mi> </msub> <mo>=</mo> <mn>25</mn> </mrow> </semantics></math> and <math display="inline"><semantics> <mrow> <msub> <mi>σ</mi> <mi>r</mi> </msub> <mo>=</mo> <mn>10</mn> </mrow> </semantics></math>; <math display="inline"><semantics> <mrow> <msub> <mi>σ</mi> <mi>s</mi> </msub> <mo>=</mo> <mn>50</mn> </mrow> </semantics></math> and <math display="inline"><semantics> <mrow> <msub> <mi>σ</mi> <mi>r</mi> </msub> <mo>=</mo> <mn>20</mn> </mrow> </semantics></math>. All the point sizes are dilated for better visualization and better viewing in color. The color of radar data indicates the distance, ranging from 0 m (blue) to 80 m (dark red).</p>
Full article ">Figure 7
<p>Samples of expanded radar by either a single kernel or both kernels. The columns from left to right show the RGB image with 5-frame raw radar, the proposed JBF radar with <math display="inline"><semantics> <mrow> <msub> <mi>σ</mi> <mi>s</mi> </msub> <mo>=</mo> <mn>10</mn> </mrow> </semantics></math> and <math display="inline"><semantics> <mrow> <msub> <mi>σ</mi> <mi>r</mi> </msub> <mo>=</mo> <mn>5</mn> </mrow> </semantics></math>, range kernel only with <math display="inline"><semantics> <mrow> <msub> <mi>σ</mi> <mi>r</mi> </msub> <mo>=</mo> <mn>10</mn> </mrow> </semantics></math>, and spatial kernel only with <math display="inline"><semantics> <mrow> <msub> <mi>σ</mi> <mi>s</mi> </msub> <mo>=</mo> <mn>25</mn> </mrow> </semantics></math>. All the point sizes are dilated for better visualization and better viewing in color. The color of radar data indicates the distance, ranging from 0 m (blue) to 80 m (dark red).</p>
Full article ">
23 pages, 9387 KiB  
Article
Cloud–Aerosol Classification Based on the U-Net Model and Automatic Denoising CALIOP Data
by Xingzhao Zhou, Bin Chen, Qia Ye, Lin Zhao, Zhihao Song, Yixuan Wang, Jiashun Hu and Ruming Chen
Remote Sens. 2024, 16(5), 904; https://doi.org/10.3390/rs16050904 - 4 Mar 2024
Cited by 1 | Viewed by 1256
Abstract
Precise cloud and aerosol identification hold paramount importance for a thorough comprehension of atmospheric processes, enhancement of meteorological forecasts, and mitigation of climate change. This study devised an automatic denoising cloud–aerosol classification deep learning algorithm, successfully achieving cloud–aerosol identification in atmospheric vertical profiles [...] Read more.
Precise cloud and aerosol identification hold paramount importance for a thorough comprehension of atmospheric processes, enhancement of meteorological forecasts, and mitigation of climate change. This study devised an automatic denoising cloud–aerosol classification deep learning algorithm, successfully achieving cloud–aerosol identification in atmospheric vertical profiles utilizing CALIPSO L1 data. The algorithm primarily consists of two components: denoising and classification. The denoising task integrates an automatic denoising module that comprehensively assesses various methods, such as Gaussian filtering and bilateral filtering, automatically selecting the optimal denoising approach. The results indicated that bilateral filtering is more suitable for CALIPSO L1 data, yielding SNR, RMSE, and SSIM values of 4.229, 0.031, and 0.995, respectively. The classification task involves constructing the U-Net model, incorporating self-attention mechanisms, residual connections, and pyramid-pooling modules to enhance the model’s expressiveness and applicability. In comparison with various machine learning models, the U-Net model exhibited the best performance, with an accuracy of 0.95. Moreover, it demonstrated outstanding generalization capabilities, evaluated using the harmonic mean F1 value, which accounts for both precision and recall. It achieved F1 values of 0.90 and 0.97 for cloud and aerosol samples from the lidar profiles during the spring of 2019. The study endeavored to predict low-quality data in CALIPSO VFM using the U-Net model, revealing significant differences with a consistency of 0.23 for clouds and 0.28 for aerosols. Utilizing U-Net confidence and a 532 nm attenuated backscatter coefficient to validate medium- and low-quality predictions in two cases from 8 February 2019, the U-Net model was found to align more closely with the CALIPSO observational data and exhibited high confidence. Statistical comparisons of the predicted geographical distribution revealed specific patterns and regional characteristics in the distribution of clouds and aerosols, showcasing the U-Net model’s proficiency in identifying aerosols within cloud layers. Full article
Show Figures

Figure 1

Figure 1
<p>The architecture of the proposed algorithm. The overall research process is illustrated in (<b>a</b>), while (<b>b</b>) depicts the automated-denoising classification model. The model is primarily divided into two modules: denoising and classification. The denoising module can automatically select the most suitable method from among four denoising techniques: Gaussian filtering, three-point smoothing, bilateral filtering, and median filtering. The classification module is mainly built on the U-Net neural network and incorporates pyramid pooling, a self-attention mechanism, and residual connections.</p>
Full article ">Figure 2
<p>(<b>a</b>) Feature importance score chart. Feature importance scores are obtained based on the extreme tree model. (<b>b</b>) Feature correlation chart. Darker colors indicate higher correlations. The scale axis in (<b>a</b>) is reversed, with higher values closer to the center. Note: 532 nm TAB = 532 nm total attenuated backscatter, 532 nm PAB = 532 nm perpendicular attenuated backscatter, 1064 nm AB = 1064 nm attenuated backscatter.</p>
Full article ">Figure 3
<p>Cluster analysis was conducted using sample data collected from the geographical coordinates of 80–90°E and 40–45°N. The cluster analysis results are presented in (<b>a</b>) for 532 nm total attenuated backscatter and 1064 nm attenuated backscatter, (<b>b</b>) for 532 nm total attenuated backscatter and 523 nm perpendicular attenuated backscatter, and (<b>c</b>) for 1064 nm attenuated backscatter and 523 nm perpendicular attenuated backscatter. Histogram distributions of 523 nm total attenuated backscatter, 1064 nm attenuated backscatter, and 532 nm perpendicular attenuated backscatter are depicted in (<b>d</b>), (<b>e</b>), and (<b>f</b>), respectively.</p>
Full article ">Figure 4
<p>Visualization of automatic denoising effects using 532 nm attenuated backscatter coefficient data as an example. CALIPSO Vertical Feature Mask: 0 = invalid (bad or missing data), 1 = clear air, 2 = cloud, 3 = tropospheric aerosol, 4 = stratospheric aerosol, 5 = surface, 6 = subsurface, 7 = no signal. The figure displays the signal-to-noise ratio (SNR), root-mean-square error (RMSE), and structural-similarity index (SSIM) obtained by four denoising methods in this case.</p>
Full article ">Figure 5
<p>Composition comparison of CALIPSO VFM product and U-Net model results. The left disk illustrates the proportion of various data categories in the VFM product, with background data encompassing all atmospheric features and surface characteristics except for clouds and aerosols. The right histogram depicts the proportion of clouds and aerosols in both the VFM product and the predictions from the U-Net model.</p>
Full article ">Figure 6
<p>Correlation chart showing the confidence level of U-Net model predictions and accuracy (using high-quality data).</p>
Full article ">Figure 7
<p>Comparison between U-Net model predictions and CALIPSO VFM product results; 17:58 UTC, 8 February 2019. (<b>a</b>–<b>c</b>) show the model’s predictions for high-quality data; (<b>d</b>,<b>e</b>) show the model’s predictions for medium- and low-quality data. (<b>a</b>,<b>d</b>) CALIPSO VFM Product; (<b>b</b>,<b>e</b>) U-Net model identification results; (<b>c</b>,<b>f</b>) Identification discrepancies. Note: Green indicates cloud layers in the VFM product identified as aerosol layers by the model, while red indicates aerosol layers in the VFM product identified as cloud layers by the model.</p>
Full article ">Figure 8
<p>(<b>a</b>) CALIPSO L1 532 nm attenuated backscatter coefficient; (<b>b</b>) Cloud-layer confidence; (<b>c</b>) Aerosol-layer confidence. The confidence levels for cloud and aerosol layers are directly obtained from the U-Net model output, with data in the yellow-circled region indicating medium- and low-quality. Due to the fact that the majority of regions without clouds and aerosols fall within the low-confidence interval, in order to better visualize the confidence differences recognized by the model, image elements with confidence levels below 10% in (<b>b</b>,<b>c</b>) have been set to blank.</p>
Full article ">Figure 9
<p>Comparison between U-Net model predictions and CALIPSO VFM product results for data from 19:37 UTC, 8 February 2019. (<b>a</b>–<b>c</b>) show the model’s predictions for high-quality data; (<b>d</b>,<b>e</b>) show the model’s predictions for medium- and low-quality data. (<b>a</b>,<b>d</b>) CALIPSO VFM Product; (<b>b</b>,<b>e</b>) U-Net model identification results; (<b>c</b>,<b>f</b>) Identification discrepancies. Note: Similar to <a href="#remotesensing-16-00904-f007" class="html-fig">Figure 7</a>, green indicates cloud layers in the VFM product identified as aerosol layers by the model, while red indicates aerosol layers in the VFM product identified as cloud layers by the model.</p>
Full article ">Figure 10
<p>The composition of this figure is similar to that of <a href="#remotesensing-16-00904-f008" class="html-fig">Figure 8</a>. (<b>a</b>) CALIPSO L1 532 nm attenuated backscatter coefficient; (<b>b</b>) Cloud-layer confidence; (<b>c</b>) Aerosol-layer confidence. The confidence levels for cloud and aerosol layers are directly obtained from the U-Net model output, with data in the yellow-circled region indicating medium- and low-quality. Image elements with confidence levels below 10% in (<b>b</b>,<b>c</b>) are set to blank.</p>
Full article ">Figure 11
<p>Spatial distribution of cloud–aerosol samples from the spring of 2019. (<b>a</b>) U-Net model predictions for the high-quality data; (<b>b</b>) VFM product for the high-quality data; (<b>c</b>) U-Net model predictions for the medium- and low-quality data; (<b>d</b>) VFM product for the medium- and low-quality data. The study area has been gridded, and the occurrence frequency of cloud and aerosol samples at each grid point has been systematically recorded.</p>
Full article ">Figure 12
<p>Vertical distribution of cloud–aerosol samples from the spring of 2019. This figure primarily illustrates the comparison between the results of the U-Net model and CALIPSO products in the vertical atmospheric profiles. (<b>a</b>) Cloud frequency comparisons for high-quality data; (<b>b</b>) Aerosol frequency comparisons for high-quality data; (<b>c</b>) Cloud frequency comparisons for medium- and low-quality data; (<b>d</b>) Aerosol frequency comparisons for medium- and low-quality data. In practical applications, although the proportions of atmospheric features and land cover were considered, they are not explicitly shown in the figure as they do not impact the cloud–aerosol frequency comparisons.</p>
Full article ">
12 pages, 4523 KiB  
Article
Dual-Band Image Fusion Approach Using Regional Weight Analysis Combined with a Multi-Level Smoothing Filter
by Jia Yi, Huilin Jiang, Xiaoyong Wang and Yong Tan
Optics 2024, 5(1), 76-87; https://doi.org/10.3390/opt5010006 - 21 Feb 2024
Viewed by 965
Abstract
Image fusion is an effective and efficient way to express the feature information of an infrared image and abundant detailed information of a visible image in a single fused image. However, obtaining a fused result with good visual effect, while preserving and inheriting [...] Read more.
Image fusion is an effective and efficient way to express the feature information of an infrared image and abundant detailed information of a visible image in a single fused image. However, obtaining a fused result with good visual effect, while preserving and inheriting those characteristic details, seems a challenging problem. In this paper, by combining a multi-level smoothing filter and regional weight analysis, a dual-band image fusion approach is proposed. Firstly, a series of dual-band image layers with different details are obtained using smoothing results. With different parameters in a bilateral filter, different smoothed results are achieved at different levels. Secondly, regional weight maps are generated for each image layer, and then we fuse the dual-band image layers with their corresponding regional weight map. Finally, by imposing proper weights, those fused image layers are synthetized. Through comparison with seven excellent fusion methods, both subjective and objective evaluations for the experimental results indicate that the proposed approach can produce the best fused image, which has the best visual effect with good contrast, and those small details are preserved and highlighted, too. In particular, for the image pairs with a size of 640 × 480, the algorithm could provide a good visual effect result within 2.86 s, and the result has almost the best objective metrics. Full article
(This article belongs to the Section Engineering Optics)
Show Figures

Figure 1

Figure 1
<p>Smoothed result <math display="inline"><semantics> <mrow> <msub> <mi>g</mi> <mi>p</mi> </msub> </mrow> </semantics></math> with different <math display="inline"><semantics> <mrow> <msub> <mi>σ</mi> <mi>s</mi> </msub> </mrow> </semantics></math> and fixed <math display="inline"><semantics> <mrow> <msub> <mi>σ</mi> <mi>r</mi> </msub> </mrow> </semantics></math> = 0.03. (<b>a</b>) Original image, (<b>b</b>) result with <math display="inline"><semantics> <mrow> <msub> <mi>σ</mi> <mi>s</mi> </msub> </mrow> </semantics></math> = 5, (<b>c</b>) result with <math display="inline"><semantics> <mrow> <msub> <mi>σ</mi> <mi>s</mi> </msub> </mrow> </semantics></math> = 11, and (<b>d</b>) result with <math display="inline"><semantics> <mrow> <msub> <mi>σ</mi> <mi>s</mi> </msub> </mrow> </semantics></math> = 17.</p>
Full article ">Figure 2
<p>Smoothed result with different <math display="inline"><semantics> <mrow> <msub> <mi>σ</mi> <mi>r</mi> </msub> </mrow> </semantics></math> and fixed <math display="inline"><semantics> <mrow> <msub> <mi>σ</mi> <mi>s</mi> </msub> </mrow> </semantics></math> = 11. (<b>a</b>) Original image, (<b>b</b>) result with <math display="inline"><semantics> <mrow> <msub> <mi>σ</mi> <mi>r</mi> </msub> </mrow> </semantics></math> = 0.03, (<b>c</b>) result with <math display="inline"><semantics> <mrow> <msub> <mi>σ</mi> <mi>r</mi> </msub> </mrow> </semantics></math> = 0.13, and (<b>d</b>) result with <math display="inline"><semantics> <mrow> <msub> <mi>σ</mi> <mi>r</mi> </msub> </mrow> </semantics></math> = 0.23.</p>
Full article ">Figure 3
<p>Examples of regional weight map <math display="inline"><semantics> <mrow> <mi>R</mi> <msub> <mi>W</mi> <mrow> <mi>map</mi> </mrow> </msub> </mrow> </semantics></math>. (<b>a</b>) Original image 1, (<b>b</b>) <math display="inline"><semantics> <mrow> <mi>R</mi> <msub> <mi>W</mi> <mrow> <mi>map</mi> </mrow> </msub> </mrow> </semantics></math> of (<b>a</b>), (<b>c</b>) original image 2, and (<b>d</b>) <math display="inline"><semantics> <mrow> <mi>R</mi> <msub> <mi>W</mi> <mrow> <mi>map</mi> </mrow> </msub> </mrow> </semantics></math> of (<b>c</b>).</p>
Full article ">Figure 4
<p>Flowchart of the proposed approach.</p>
Full article ">Figure 5
<p>Source images: (<b>a</b>,<b>b</b>) are VI and IR images (434 × 340), mainly including road, cars, and people. (<b>c</b>,<b>d</b>) are VI and IR images (640 × 480), mainly including trees, buildings, and a person.</p>
Full article ">Figure 6
<p>Fused results for <a href="#optics-05-00006-f005" class="html-fig">Figure 5</a>a,b: (<b>a</b>–<b>h</b>) are the results based on the eight methods, respectively. (<b>a</b>) WLS, (<b>b</b>) BF, (<b>c</b>) SVD (<b>d</b>) FG, (<b>e</b>) MST, (<b>f</b>) CSTH, (<b>g</b>) MBSE, and (<b>h</b>) ours.</p>
Full article ">Figure 7
<p>Fused results for <a href="#optics-05-00006-f005" class="html-fig">Figure 5</a>c,d: (<b>a</b>–<b>h</b>) are the results based on the eight methods, respectively. (<b>a</b>) WLS, (<b>b</b>) BF, (<b>c</b>) SVD (<b>d</b>) FG, (<b>e</b>) MST, (<b>f</b>) CSTH, (<b>g</b>) MBSE, and (<b>h</b>) ours.</p>
Full article ">
25 pages, 5734 KiB  
Article
Deep Learning-Based 6-DoF Object Pose Estimation Considering Synthetic Dataset
by Tianyu Zheng, Chunyan Zhang, Shengwen Zhang and Yanyan Wang
Sensors 2023, 23(24), 9854; https://doi.org/10.3390/s23249854 - 15 Dec 2023
Viewed by 2006
Abstract
Due to the difficulty in generating a 6-Degree-of-Freedom (6-DoF) object pose estimation dataset, and the existence of domain gaps between synthetic and real data, existing pose estimation methods face challenges in improving accuracy and generalization. This paper proposes a methodology that employs higher [...] Read more.
Due to the difficulty in generating a 6-Degree-of-Freedom (6-DoF) object pose estimation dataset, and the existence of domain gaps between synthetic and real data, existing pose estimation methods face challenges in improving accuracy and generalization. This paper proposes a methodology that employs higher quality datasets and deep learning-based methods to reduce the problem of domain gaps between synthetic and real data and enhance the accuracy of pose estimation. The high-quality dataset is obtained from Blenderproc and it is innovatively processed using bilateral filtering to reduce the gap. A novel attention-based mask region-based convolutional neural network (R-CNN) is proposed to reduce the computation cost and improve the model detection accuracy. Meanwhile, an improved feature pyramidal network (iFPN) is achieved by adding a layer of bottom-up paths to extract the internalization of features of the underlying layer. Consequently, a novel convolutional block attention module–convolutional denoising autoencoder (CBAM–CDAE) network is proposed by presenting channel attention and spatial attention mechanisms to improve the ability of AE to extract images’ features. Finally, an accurate 6-DoF object pose is obtained through pose refinement. The proposed approach is compared to other models using the T-LESS and LineMOD datasets. Comparison results demonstrate the proposed approach outperforms the other estimation models. Full article
(This article belongs to the Section Sensing and Imaging)
Show Figures

Figure 1

Figure 1
<p>Framework of the methodology proposed in this paper.</p>
Full article ">Figure 2
<p>BlenderProc-based synthesized data production process.</p>
Full article ">Figure 3
<p>Results of bilateral filtering.</p>
Full article ">Figure 4
<p>M-ST network structure.</p>
Full article ">Figure 5
<p>Patch merging layer schematic.</p>
Full article ">Figure 6
<p>Swin Transformer block.</p>
Full article ">Figure 7
<p>Isometric sampling based on ortho-20 facets.</p>
Full article ">Figure 8
<p>CBAM–CDAE network structure.</p>
Full article ">Figure 9
<p>CBAM network structure.</p>
Full article ">Figure 10
<p>The histogram of rotation error for the 5th object, one view-dependent symmetry.</p>
Full article ">Figure 11
<p>The histogram of rotation error for the 29th object, two view-dependent symmetry.</p>
Full article ">Figure 12
<p>The 6-DoF pose visualization.</p>
Full article ">Figure 13
<p>Schematic diagram of the LineMOD dataset.</p>
Full article ">Figure 14
<p>Schematic diagram of the T-LESS dataset.</p>
Full article ">
23 pages, 8734 KiB  
Article
Motorcycle Detection and Collision Warning Using Monocular Images from a Vehicle
by Zahra Badamchi Shabestari, Ali Hosseininaveh and Fabio Remondino
Remote Sens. 2023, 15(23), 5548; https://doi.org/10.3390/rs15235548 - 28 Nov 2023
Cited by 2 | Viewed by 2229
Abstract
Motorcycle detection and collision warning are essential features in advanced driver assistance systems (ADAS) to ensure road safety, especially in emergency situations. However, detecting motorcycles from videos captured from a car is challenging due to the varying shapes and appearances of motorcycles. In [...] Read more.
Motorcycle detection and collision warning are essential features in advanced driver assistance systems (ADAS) to ensure road safety, especially in emergency situations. However, detecting motorcycles from videos captured from a car is challenging due to the varying shapes and appearances of motorcycles. In this paper, we propose an integrated and innovative remote sensing and artificial intelligence (AI) methodology for motorcycle detection and distance estimation based on visual data from a single camera installed in the back of a vehicle. Firstly, MD-TinyYOLOv4 is used for detecting motorcycles, refining the neural network through SPP (spatial pyramid pooling) feature extraction, Mish activation function, data augmentation techniques, and optimized anchor boxes for training. The proposed algorithm outperforms eight existing YOLO versions, achieving a precision of 81% at a speed of 240 fps. Secondly, a refined disparity map of each motorcycle’s bounding box is estimated by training a Monodepth2 with a bilateral filter for distance estimation. The proposed fusion model (motorcycle’s detection and distance from vehicle) is evaluated with depth stereo camera measurements, and the results show that 89% of warning scenes are correctly detected, with an alarm notification time of 0.022 s for each image. Outcomes indicate that the proposed integrated methodology provides an effective solution for ADAS, with promising results for real-world applications, and can be suitable for running on mobility services or embedded computing boards instead of the super expensive and powerful systems used in some high-tech unmanned vehicles. Full article
(This article belongs to the Special Issue Photogrammetry Meets AI)
Show Figures

Figure 1

Figure 1
<p>An example of a motorcycle covered with a black windshield.</p>
Full article ">Figure 2
<p>An outline of the three steps of the proposed methodology for motorcycle range estimation: a fusion is performed between object detection and depth estimation tasks using a single camera installed in the back of a vehicle.</p>
Full article ">Figure 3
<p>The MD-TinyYOLOv4 architecture, with its flowchart and proposed refinements.</p>
Full article ">Figure 4
<p>Mish activation function [<a href="#B68-remotesensing-15-05548" class="html-bibr">68</a>].</p>
Full article ">Figure 5
<p>Sample images for data augmentation.</p>
Full article ">Figure 6
<p>The histogram diagram of the depth values of a detected motorcycle bounding box.</p>
Full article ">Figure 7
<p>Samples of a rectified pair of images taken by the Mynt-Eye camera.</p>
Full article ">Figure 8
<p>K-means++ clustering results for the captured dataset.</p>
Full article ">Figure 9
<p>Predicted dimensions of the anchor boxes based on K = 10.</p>
Full article ">Figure 10
<p>Motorcycle detection with different distances and street congestion using the proposed MD-TinyYOLOv4 model.</p>
Full article ">Figure 11
<p>Samples of images from Image-Net, KITTI, and the dataset used in this paper.</p>
Full article ">Figure 12
<p>Comparing the colored disparity map produced by the fine-tuned Monodepth1 (<b>left</b>) and Monodepth2 (<b>right</b>) models.</p>
Full article ">Figure 13
<p>Example of disparity maps, before (<b>a</b>) and after (<b>b</b>) refining. The darker the color, the larger the depth and camera-to-object distance.</p>
Full article ">Figure 14
<p>Monodepth2 model’s RMSE of the computed distances at different camera-to-motorcycle distances.</p>
Full article ">Figure 15
<p>Motorcycle detection and range estimation results for the proposed MD-TinyYOLOv4 model which (<b>a</b>,<b>d</b>,<b>e</b>) are shown as a dangerous situation, and (<b>b</b>,<b>c</b>,<b>f</b>) are illustrated as a safe situation. Metric distances are also provided for each box.</p>
Full article ">
18 pages, 7221 KiB  
Article
Fast Local Laplacian Filter Based on Modified Laplacian through Bilateral Filter for Coronary Angiography Medical Imaging Enhancement
by Sarwar Shah Khan, Muzammil Khan and Yasser Alharbi
Algorithms 2023, 16(12), 531; https://doi.org/10.3390/a16120531 - 21 Nov 2023
Cited by 1 | Viewed by 1745
Abstract
Contrast enhancement techniques serve the purpose of diminishing image noise and increasing the contrast of relevant structures. In the context of medical images, where the differentiation between normal and abnormal tissues can be quite subtle, precise interpretation might become challenging when noise levels [...] Read more.
Contrast enhancement techniques serve the purpose of diminishing image noise and increasing the contrast of relevant structures. In the context of medical images, where the differentiation between normal and abnormal tissues can be quite subtle, precise interpretation might become challenging when noise levels are relatively elevated. The Fast Local Laplacian Filter (FLLF) is proposed to deliver a more precise interpretation and present a clearer image to the observer; this is achieved through the reduction of noise levels. In this study, the FLLF strengthened images through its unique contrast enhancement capabilities while preserving important image details. It achieved this by adapting to the image’s characteristics and selectively enhancing areas with low contrast, thereby improving the overall visual quality. Additionally, the FLLF excels in edge preservation, ensuring that fine details are retained and that edges remain sharp. Several performance metrics were employed to assess the effectiveness of the proposed technique. These metrics included Peak Signal-to-Noise Ratio (PSNR), Mean Squared Error (MSE), Root Mean Squared Error (RMSE), Normalization Coefficient (NC), and Correlation Coefficient. The results indicated that the proposed technique achieved a PSNR of 40.12, an MSE of 8.6982, an RMSE of 2.9492, an NC of 1.0893, and a Correlation Coefficient of 0.9999. The analysis highlights the superior performance of the proposed method when contrast enhancement is applied, especially when compared to existing techniques. This approach results in high-quality images with minimal information loss, ultimately aiding medical experts in making more accurate diagnoses. Full article
(This article belongs to the Section Algorithms for Multidisciplinary Applications)
Show Figures

Figure 1

Figure 1
<p>The schematic framework of the proposed technique.</p>
Full article ">Figure 2
<p>Input image and its corresponding contrast enhancement results of Test Image 1: (<b>a</b>) input image, (<b>b</b>) Retinax, (<b>c</b>) CS, (<b>d</b>) GC, (<b>e</b>) HE, (<b>f</b>) LBC, (<b>g</b>) LTHE, (<b>h</b>) OMC, (<b>i</b>) PLT, (<b>j</b>) Sigmoid, (<b>k</b>) AHE, (<b>l</b>) BHE, (<b>m</b>) BBHE, (<b>n</b>) CLAHE, (<b>o</b>) DSIHE, (<b>p</b>) LT, (<b>q</b>) GTHE, (<b>r</b>) MHE, (<b>s</b>) MSRCR, and (<b>t</b>) FLLF.</p>
Full article ">Figure 2 Cont.
<p>Input image and its corresponding contrast enhancement results of Test Image 1: (<b>a</b>) input image, (<b>b</b>) Retinax, (<b>c</b>) CS, (<b>d</b>) GC, (<b>e</b>) HE, (<b>f</b>) LBC, (<b>g</b>) LTHE, (<b>h</b>) OMC, (<b>i</b>) PLT, (<b>j</b>) Sigmoid, (<b>k</b>) AHE, (<b>l</b>) BHE, (<b>m</b>) BBHE, (<b>n</b>) CLAHE, (<b>o</b>) DSIHE, (<b>p</b>) LT, (<b>q</b>) GTHE, (<b>r</b>) MHE, (<b>s</b>) MSRCR, and (<b>t</b>) FLLF.</p>
Full article ">Figure 3
<p>Input image and its corresponding contrast enhancement results of Test Image 1: (<b>a</b>) input image, (<b>b</b>) Retinax, (<b>c</b>) CS, (<b>d</b>) GC, (<b>e</b>) HE, (<b>f</b>) LBC, (<b>g</b>) LTHE, (<b>h</b>) OMC, (<b>i</b>) PLT, (<b>j</b>) Sigmoid, (<b>k</b>) AHE, (<b>l</b>) BHE, (<b>m</b>) BBHE, (<b>n</b>) CLAHE, (<b>o</b>) DSIHE, (<b>p</b>) LT, (<b>q</b>) GTHE, (<b>r</b>) MHE, (<b>s</b>) MSRCR, and (<b>t</b>) FLLF.</p>
Full article ">Figure 3 Cont.
<p>Input image and its corresponding contrast enhancement results of Test Image 1: (<b>a</b>) input image, (<b>b</b>) Retinax, (<b>c</b>) CS, (<b>d</b>) GC, (<b>e</b>) HE, (<b>f</b>) LBC, (<b>g</b>) LTHE, (<b>h</b>) OMC, (<b>i</b>) PLT, (<b>j</b>) Sigmoid, (<b>k</b>) AHE, (<b>l</b>) BHE, (<b>m</b>) BBHE, (<b>n</b>) CLAHE, (<b>o</b>) DSIHE, (<b>p</b>) LT, (<b>q</b>) GTHE, (<b>r</b>) MHE, (<b>s</b>) MSRCR, and (<b>t</b>) FLLF.</p>
Full article ">Figure 4
<p>Comparing the top three noise removal techniques using five different measures on test image 1.</p>
Full article ">Figure 5
<p>Comparing the top three noise removal techniques using five different measures on test image 2.</p>
Full article ">Figure 6
<p>Illustrates the time complexity (s) of the proposed technique in comparison with the existing state-of-the-art methods on test image 1.</p>
Full article ">Figure 7
<p>Illustrates the time complexity (s) of the proposed technique in comparison with the existing state-of-the-art methods on test image 2.</p>
Full article ">
20 pages, 8718 KiB  
Article
Apple Surface Defect Detection Based on Gray Level Co-Occurrence Matrix and Retinex Image Enhancement
by Lei Yang, Dexu Mu, Zhen Xu and Kaiyu Huang
Appl. Sci. 2023, 13(22), 12481; https://doi.org/10.3390/app132212481 - 18 Nov 2023
Cited by 1 | Viewed by 1151
Abstract
Aiming at the problems of uneven light reflectivity on the spherical surface and high similarity between the stems/calyxes and scars that exist in the detection of surface defects in apples, this paper proposed a defect detection method based on image segmentation and stem/calyx [...] Read more.
Aiming at the problems of uneven light reflectivity on the spherical surface and high similarity between the stems/calyxes and scars that exist in the detection of surface defects in apples, this paper proposed a defect detection method based on image segmentation and stem/calyx recognition to realize the detection and recognition of surface defects in apples. Preliminary defect segmentation results were obtained by eliminating the interference of light reflection inhomogeneity through adaptive bilateral filtering-based single-scale Retinex (SSR) luminance correction and using adaptive gamma correction to enhance the Retinex reflective layer, and later segmenting the Retinex reflective layer by using a region-growing algorithm. The texture features of apple surface defects under different image processing methods were analyzed based on the gray level co-occurrence matrix, and a support vector machine was introduced for binary classification to differentiate between stems/calyxes and scars. Deploying the proposed defect detection method into the embedded device OpenMV4H7Plus, the accuracy of stem/calyx recognition reached 93.7%, and the accuracy of scar detection reached 94.2%. It has conclusively been shown that the proposed defect detection method can effectively detect apple surface defects in the presence of uneven light reflectivity and stem/calyx interference. Full article
(This article belongs to the Topic New Advances in Food Analysis and Detection)
Show Figures

Figure 1

Figure 1
<p>Defect type. (<b>a</b>) Stem; (<b>b</b>) calyx; (<b>c</b>) lightly browned scar; (<b>d</b>) heavily browned scar.</p>
Full article ">Figure 2
<p>Detection device sketch.</p>
Full article ">Figure 3
<p>Schematic of the background segmentation process.</p>
Full article ">Figure 4
<p>The algorithmic framework of this paper.</p>
Full article ">Figure 5
<p>Schematic diagram of the Retinex principle.</p>
Full article ">Figure 6
<p>Halation phenomenon. (<b>a</b>) Apple image; (<b>b</b>) MSR reflection component image.</p>
Full article ">Figure 7
<p>Support vector machine model.</p>
Full article ">Figure 8
<p>RGB color space decomposition results. (<b>a</b>) Original picture; (<b>b</b>) R-component picture; (<b>c</b>) G-component picture; (<b>d</b>) B-component picture.</p>
Full article ">Figure 9
<p>Luminance decomposition effect. (<b>a</b>) SSR result; (<b>b</b>) MSR result; (<b>c</b>) improved SSR decomposition result.</p>
Full article ">Figure 10
<p>Adaptive gamma correction effect. (<b>a</b>) Reflected component figure; (<b>b</b>) adaptive gamma correction result; (<b>c</b>) reflected component gray intensity surface figure; (<b>d</b>) adaptive gamma correction resulting in gray intensity surface figure.</p>
Full article ">Figure 11
<p>Effectiveness of each method for detection of apple stem images. (<b>a</b>) Original figure; (<b>b</b>) genetic algorithm segmentation result; (<b>c</b>) edge detection algorithm result; (<b>d</b>) segmentation effect of the proposed image segmentation method.</p>
Full article ">Figure 12
<p>Detection effect of each method on calyx images. (<b>a</b>) Original figure; (<b>b</b>) genetic algorithm segmentation result; (<b>c</b>) edge detection algorithm result; (<b>d</b>) segmentation effect of the proposed image segmentation method.</p>
Full article ">Figure 13
<p>Detection effectiveness of each method on complex images with light reflectance. (<b>a</b>) Original figure; (<b>b</b>) genetic algorithm segmentation result; (<b>c</b>) edge detection algorithm result; (<b>d</b>) segmentation effect of the proposed image segmentation method.</p>
Full article ">Figure 14
<p>Detection effectiveness of each method on complex images with edge information. (<b>a</b>) original figure; (<b>b</b>) genetic algorithm segmentation result; (<b>c</b>) edge detection algorithm result; (<b>d</b>) segmentation effect of the proposed image segmentation method.</p>
Full article ">Figure 15
<p>Some SSR image enhancement results. (<b>a</b>) Scars; (<b>b</b>) stems/calyxes.</p>
Full article ">Figure 16
<p>Texture characterization data for each type of defect sample. (<b>a</b>) R-component map stem/calyx texture data; (<b>b</b>) R-component map scar texture data; (<b>c</b>) SSR image enhanced stem/calyx texture data; (<b>d</b>) SSR image enhanced scar texture data; (<b>e</b>) stem/calyx texture data corrected by SSR and gamma; (<b>f</b>) scar texture data corrected by SSR and gamma.</p>
Full article ">Figure 17
<p>Detection effect of the proposed defect detection method. (<b>a</b>) Apple samples; (<b>b</b>) detection results of the proposed defect detection method.</p>
Full article ">
15 pages, 3942 KiB  
Article
Analytical Method for Bridge Damage Using Deep Learning-Based Image Analysis Technology
by Kukjin Jang, Taegeon Song, Dasran Kim, Jinsick Kim, Byeongsoo Koo, Moonju Nam, Kyungil Kwak, Jooyeoun Lee and Myoungsug Chung
Appl. Sci. 2023, 13(21), 11800; https://doi.org/10.3390/app132111800 - 28 Oct 2023
Viewed by 1060
Abstract
Bridge inspection methods using unmanned vehicles have been attracting attention. In this study, we devised an efficient and reliable method for visually inspecting bridges using unmanned vehicles. For this purpose, we developed the BIRD U-Net algorithm, which is an evolution of the U-Net [...] Read more.
Bridge inspection methods using unmanned vehicles have been attracting attention. In this study, we devised an efficient and reliable method for visually inspecting bridges using unmanned vehicles. For this purpose, we developed the BIRD U-Net algorithm, which is an evolution of the U-Net algorithm that utilizes images taken by unmanned vehicles. Unlike the U-Net algorithm, however, this algorithm identifies the optimal function by setting the epoch to 120 and uses the Adam optimization algorithm. In addition, a bilateral filter was applied to highlight the damaged areas of the bridge, and a different color was used for each of the five types of abnormalities detected, such as cracks. Next, we trained and tested 135,696 images of exterior bridge damage, including concrete delamination, water leakage, and exposed rebar. Through the analysis, we confirmed an analysis method that yields an average inspection reproduction rate of more than 95%. In addition, we compared and analyzed the inspection reproduction rate of the method with that of BIRD U-Net after using the same method and images for training as the existing U-Net and ResNet algorithms for validation. In addition, the algorithm developed in this study is expected to yield objective results through automatic damage analysis. It can be applied to regular inspections that involve unmanned mobile vehicles in the field of bridge maintenance, thereby reducing the associated time and cost. Full article
(This article belongs to the Special Issue Advances in Big Data Analysis and Visualization)
Show Figures

Figure 1

Figure 1
<p>Example photos of collected data.</p>
Full article ">Figure 2
<p>A&amp;R-topped drone.</p>
Full article ">Figure 3
<p>U-Net architecture.</p>
Full article ">Figure 4
<p>Algorithm process.</p>
Full article ">Figure 5
<p>BIRD U-Net characteristics.</p>
Full article ">Figure 6
<p>ResNet process.</p>
Full article ">Figure 7
<p>U-Net process.</p>
Full article ">Figure 8
<p>Verification process.</p>
Full article ">Figure 9
<p>Confusion matrix.</p>
Full article ">Figure 10
<p>Detection rate comparison.</p>
Full article ">
21 pages, 6303 KiB  
Article
Impact of Traditional and Embedded Image Denoising on CNN-Based Deep Learning
by Roopdeep Kaur, Gour Karmakar and Muhammad Imran
Appl. Sci. 2023, 13(20), 11560; https://doi.org/10.3390/app132011560 - 22 Oct 2023
Cited by 3 | Viewed by 2133
Abstract
In digital image processing, filtering noise is an important step for reconstructing a high-quality image for further processing such as object segmentation, object detection, and object recognition. Various image-denoising approaches, including median, Gaussian, and bilateral filters, are available in the literature. Since convolutional [...] Read more.
In digital image processing, filtering noise is an important step for reconstructing a high-quality image for further processing such as object segmentation, object detection, and object recognition. Various image-denoising approaches, including median, Gaussian, and bilateral filters, are available in the literature. Since convolutional neural networks (CNN) are able to directly learn complex patterns and features from data, they have become a popular choice for image-denoising tasks. As a result of their ability to learn and adapt to various denoising scenarios, CNNs are powerful tools for image denoising. Some deep learning techniques such as CNN incorporate denoising strategies directly into the CNN model layers. A primary limitation of these methods is their necessity to resize images to a consistent size. This resizing can result in a loss of vital image details, which might compromise CNN’s effectiveness. Because of this issue, we utilize a traditional denoising method as a preliminary step for noise reduction before applying CNN. To our knowledge, a comparative performance study of CNN using traditional and embedded denoising against a baseline approach (without denoising) is yet to be performed. To analyze the impact of denoising on the CNN performance, in this paper, firstly, we filter the noise from the images using traditional means of denoising method before their use in the CNN model. Secondly, we embed a denoising layer in the CNN model. To validate the performance of image denoising, we performed extensive experiments for both traffic sign and object recognition datasets. To decide whether denoising will be adopted and to decide on the type of filter to be used, we also present an approach exploiting the peak-signal-to-noise-ratio (PSNRs) distribution of images. Both CNN accuracy and PSNRs distribution are used to evaluate the effectiveness of the denoising approaches. As expected, the results vary with the type of filter, impact, and dataset used in both traditional and embedded denoising approaches. However, traditional denoising shows better accuracy, while embedded denoising shows lower computational time for most of the cases. Overall, this comparative study gives insights into whether denoising will be adopted in various CNN-based image analyses, including autonomous driving, animal detection, and facial recognition. Full article
(This article belongs to the Special Issue IoT in Smart Cities and Homes)
Show Figures

Figure 1

Figure 1
<p>Overview of the comparative study.</p>
Full article ">Figure 2
<p>Different backgrounds in CURE-OR dataset [<a href="#B19-applsci-13-11560" class="html-bibr">19</a>].</p>
Full article ">Figure 3
<p>Images of the shadow without denoising and denoising with median and Gaussian filter for CURE-TSR dataset. (<b>a</b>) Without filter. (<b>b</b>) After median filter. (<b>c</b>) After Gaussian filter.</p>
Full article ">Figure 4
<p>Images of lens blur without denoising and denoising with median and Gaussian filter for CURE-OR dataset. (<b>a</b>) Without filter. (<b>b</b>) After median filter. (<b>c</b>) After Gaussian filter.</p>
Full article ">Figure 5
<p>Images of lens dirty without denoising and denoising with median and Gaussian filter for CURE-OR dataset. (<b>a</b>) Without filter. (<b>b</b>) After median filter. (<b>c</b>) After Gaussian filter.</p>
Full article ">Figure 6
<p>Overview of the embedded denoising approach.</p>
Full article ">Figure 7
<p>Epochs vs. L=loss for CURE-OR dataset.</p>
Full article ">Figure 8
<p>Mean PSNR vs. recognition accuracy in CURE-TSR dataset. Here, LBWD: lens blur without denoising; LBM: lens blur with median filtering; LBG: lens blur with Gaussian filtering; LDWD: lens dirty without denoising; LDM: lens dirty with median filtering; LDG: lens dirty with Gaussian filtering; RWD: rain without denoising; RM: rain with median filtering; RG: rain with Gaussian filtering; SWD: shadow without denoising; SM: shadow with median filtering; SG: shadow with Gaussian filtering; DWD: darkness without denoising; DM: darkness with median filtering; and DG: darkness with Gaussian filtering.</p>
Full article ">Figure 9
<p>Comparative histogram of PSNR of underexposure without denoising and with Gaussian and median filtering for CURE-OR dataset. (<b>a</b>) Before and after Gaussian filtering. (<b>b</b>) Before and after median filtering.</p>
Full article ">Figure 10
<p>Comparative histogram of PSNR of lens blur without denoising and with Gaussian and median filtering for CURE-OR dataset. (<b>a</b>) Before and after Gaussian filtering. (<b>b</b>) Before and after median filtering.</p>
Full article ">Figure 11
<p>Comparative histogram of PSNR of the shadow without denoising and with Gaussian and median filtering for CURE-TSR dataset. (<b>a</b>) Before and after Gaussian filtering. (<b>b</b>) Before and after median filtering.</p>
Full article ">Figure 12
<p>Comparative histogram of PSNR of darkness without denoising and with Gaussian and median filtering for CURE-TSR dataset. (<b>a</b>) Before and after Gaussian filtering. (<b>b</b>) Before and after median filtering.</p>
Full article ">
18 pages, 5568 KiB  
Article
A Method for Extracting Contours of Building Facade Hollowing Defects Using Polarization Thermal Images Based on Improved Canny Algorithm
by Darong Zhu, Jianguo Li, Fangbin Wang, Xue Gong, Wanlin Cong, Ping Wang and Yanli Liu
Buildings 2023, 13(10), 2563; https://doi.org/10.3390/buildings13102563 - 10 Oct 2023
Cited by 3 | Viewed by 1300
Abstract
During the service process of high-rise buildings, hollowing defects may be produced in the decorative layer, which not only affect the appearance, but also create a safety hazard of wall covering and shattered plaster peeling. Numerous studies have shown that hollowing can be [...] Read more.
During the service process of high-rise buildings, hollowing defects may be produced in the decorative layer, which not only affect the appearance, but also create a safety hazard of wall covering and shattered plaster peeling. Numerous studies have shown that hollowing can be detected using infrared thermal imagery under normal conditions. However, it is difficult to detect the edge and calculate the area of the hollowing on an exterior facade accurately because of the low contrast and fuzzy boundaries of the obtained infrared thermal images. To address these problems, a method for extracting the contours of building facade hollowing defects using polarization thermal images based on an improved Canny algorithm has been proposed in this paper. Firstly, the principle of thermal polarization imaging was introduced for hollowing detection. Secondly, considering the shortcomings of the Canny edge detection algorithm and the features of polarization thermal images, an improved Canny edge detection algorithm is proposed, including adaptive bilateral filtering to improve noise reduction ability while ensuring defect edges are not virtualized, Laplacian sharpening and histogram equalization to achieve contour sharpening and contrast enhancement, and eight-direction gradient templates for calculating image gradients, which make interpolation with non-maximum suppression more accurate, and the Tsallis entropy threshold segmentation algorithm based on the OTSU algorithm verification makes the image contour information more complete and accurate. Finally, a long-wave infrared polarization thermal imaging experimental platform was established and validation experiments were conducted. The experimental results demonstrate that the distinct, smooth, and precise location edges of the hollowing polarization infrared thermal images can be obtained, and the average error of the detected hollowing area is about 10% using the algorithm proposed in this paper. Full article
(This article belongs to the Section Construction Management, and Computers & Digitization)
Show Figures

Figure 1

Figure 1
<p>Schematic diagram of heat transfer process of experimental wall and the structural diagram of actual wall.</p>
Full article ">Figure 2
<p>Improved Canny algorithm operation flowchart.</p>
Full article ">Figure 3
<p>Improvement direction division diagram.</p>
Full article ">Figure 4
<p>Experimental wall: (<b>a</b>) Position distribution diagram of preset hollowing defects; (<b>b</b>) Size of preset hollowing defects; (<b>c</b>) Visual image; (<b>d</b>) IR image; (<b>e</b>) Polarization thermal image.</p>
Full article ">Figure 5
<p>Schematic diagram of experimental setup.</p>
Full article ">Figure 6
<p>Experimental equipment diagram. (1) A CCD long-wave infrared cooling camera; (2) A high-precision turntable; (3) An infrared metal grating polarizer.</p>
Full article ">Figure 7
<p>Comparison of image filtering processing effects: (<b>a</b>–<b>f</b>) The polarization thermal original and cropped images processed using Gaussian filtering; median filtering; mean filtering; bilateral filtering; and improved bilateral filtering.</p>
Full article ">Figure 8
<p>The polarization thermal image and image enhancement processing result.</p>
Full article ">Figure 9
<p>The schematic diagram of infrared thermal images and polarized thermal images.</p>
Full article ">Figure 10
<p>The results of hollowing defect images processed using the traditional Canny algorithm and morphological methods: (<b>a1</b>) Output edge of cropped infrared image; (<b>a2</b>) The result of (<b>a1</b>) after application of morphological methods; (<b>b1</b>) Output edge of cropped polarization image; (<b>b2</b>) The result of (<b>b1</b>) after application of morphological methods.</p>
Full article ">Figure 11
<p>Edge contour detection results of hollowing defects: (<b>a</b>–<b>f</b>) Roberts, Sobel, Prewitt, Log, Canny, and improved Canny algorithm.</p>
Full article ">Figure 12
<p>Morphological processing results: (<b>a1</b>–<b>f1</b>) Roberts, Sobel, Prewitt, Log, Canny, and improved Canny algorithm.</p>
Full article ">Figure 13
<p>The error rate of different algorithms in the size of hollowing defects.</p>
Full article ">
Back to TopTop