[go: up one dir, main page]

Next Issue
Volume 8, January
Previous Issue
Volume 7, November
 
 

J. Imaging, Volume 7, Issue 12 (December 2021) – 33 articles

Cover Story (view full-size image): In this work, we address the problem of estimating a 3D body from single images of people wearing loose clothes. To this aim, we make use of the SMPL parametric body model and observe that shape parameters encoding the body shape should not change regardless of whether the subject is wearing clothes or not. To improve shape estimation under clothing, we train a deep network to regress the shape parameters from a single image. To increase robustness to clothing, we build our training dataset by associating the shape parameters of a “minimally clothed” person to other samples of the same person wearing looser clothes.View this paper
  • Issues are regarded as officially published after their release is announced to the table of contents alert mailing list.
  • You may sign up for e-mail alerts to receive table of contents of newly released issues.
  • PDF is the official format for papers published in both, html and pdf forms. To view the papers in pdf format, click on the "PDF Full-text" link, and use the free Adobe Reader to open them.
Order results
Result details
Section
Select all
Export citation of selected articles as:
13 pages, 3398 KiB  
Article
On the Relationship between Corneal Biomechanics, Macrostructure, and Optical Properties
by Francisco J. Ávila, Maria Concepción Marcellán and Laura Remón
J. Imaging 2021, 7(12), 280; https://doi.org/10.3390/jimaging7120280 - 18 Dec 2021
Cited by 4 | Viewed by 2920
Abstract
Optical properties of the cornea are responsible for correct vision; the ultrastructure allows optical transparency, and the biomechanical properties govern the shape, elasticity, or stiffness of the cornea, affecting ocular integrity and intraocular pressure. Therefore, the optical aberrations, corneal transparency, structure, and biomechanics [...] Read more.
Optical properties of the cornea are responsible for correct vision; the ultrastructure allows optical transparency, and the biomechanical properties govern the shape, elasticity, or stiffness of the cornea, affecting ocular integrity and intraocular pressure. Therefore, the optical aberrations, corneal transparency, structure, and biomechanics play a fundamental role in the optical quality of human vision, ocular health, and refractive surgery outcomes. However, the inter-relationships of those properties are not yet reported at a macroscopic scale within the hierarchical structure of the cornea. This work explores the relationships between the biomechanics, structure, and optical properties (corneal aberrations and optical density) at a macro-structural level of the cornea through dual Placido–Scheimpflug imaging and air-puff tonometry systems in a healthy young adult population. Results showed correlation between optical transparency, corneal macrostructure, and biomechanics, whereas corneal aberrations and in particular spherical terms remained independent. A compensation mechanism for the spherical aberration is proposed through corneal shape and biomechanics. Full article
(This article belongs to the Special Issue Biomechanical Techniques for Biomedical Imaging)
Show Figures

Figure 1

Figure 1
<p>Total corneal wavefront aberration map (<b>left</b>) and anterior segment Scheimpflug image (<b>right</b>) from a volunteer of the study.</p>
Full article ">Figure 2
<p>Optical density measurement at the posterior corneal location and horizontal viewing of the Scheimpflug camera.</p>
Full article ">Figure 3
<p>ORA measurement from a participant in our study. P1, P2, Max P, and CH correspond to the first and second applanation, maximum pressure, and corneal hysteresis, respectively.</p>
Full article ">Figure 4
<p>Standard deviation of OTI values as a function of the number of subjects per cluster.</p>
Full article ">Figure 5
<p>Mean clustered OTI values as function of total corneal astigmatism (TCA) (<b>a</b>) and posterior eccentricity (PE) (<b>b</b>) for all subjects. Standard deviation of the clustered data, equations, confident, and prediction bands of the regression analysis are included.</p>
Full article ">Figure 6
<p>Dynamical representation of corneal applanation as a function of time from a volunteer of the study. Air pulse pressure is scaled in arbitrary units and shown in the bottom right corner legend.</p>
Full article ">Figure 7
<p>Mean clustered OTI values as function of first (<b>a</b>) and second applanation pressures (<b>b</b>) at the ORA device for all subjects. Standard deviation of the clustered data, equations, confidence, and prediction bands of the regression analysis are included.</p>
Full article ">Figure 8
<p>Mean clustered posterior eccentricity versus total corneal astigmatism values (<b>a</b>) and mean clustered applanation pressures at ORA versus total corneal astigmatism values (<b>b</b>). Linear regression fits are included.</p>
Full article ">
20 pages, 1564 KiB  
Article
A Robust Tensor-Based Submodule Clustering for Imaging Data Using l12 Regularization and Simultaneous Noise Recovery via Sparse and Low Rank Decomposition Approach
by Jobin Francis, Baburaj Madathil, Sudhish N. George and Sony George
J. Imaging 2021, 7(12), 279; https://doi.org/10.3390/jimaging7120279 - 17 Dec 2021
Cited by 4 | Viewed by 3257
Abstract
The massive generation of data, which includes images and videos, has made data management, analysis, information extraction difficult in recent years. To gather relevant information, this large amount of data needs to be grouped. Real-life data may be noise corrupted during data collection [...] Read more.
The massive generation of data, which includes images and videos, has made data management, analysis, information extraction difficult in recent years. To gather relevant information, this large amount of data needs to be grouped. Real-life data may be noise corrupted during data collection or transmission, and the majority of them are unlabeled, allowing for the use of robust unsupervised clustering techniques. Traditional clustering techniques, which vectorize the images, are unable to keep the geometrical structure of the images. Hence, a robust tensor-based submodule clustering method based on l12 regularization with improved clustering capability is formulated. The l12 induced tensor nuclear norm (TNN), integrated into the proposed method, offers better low rankness while retaining the self-expressiveness property of submodules. Unlike existing methods, the proposed method employs a simultaneous noise removal technique by twisting the lateral image slices of the input data tensor into frontal slices and eliminates the noise content in each image, using the principles of the sparse and low rank decomposition technique. Experiments are carried out over three datasets with varying amounts of sparse, Gaussian and salt and pepper noise. The experimental results demonstrate the superior performance of the proposed method over the existing state-of-the-art methods. Full article
Show Figures

Figure 1

Figure 1
<p>Self-expressiveness property of free submodules. Red fibers represent non-zero fibers and greyish fibers represent zero value fibers. Non-zero fibers represent coefficients from intra-cluster. Zero fibers denote coefficients from inter-clusters.</p>
Full article ">Figure 2
<p>Unit ball representation of (<b>a</b>) <math display="inline"><semantics> <msub> <mi>l</mi> <mo>∞</mo> </msub> </semantics></math> norm (<b>b</b>) <math display="inline"><semantics> <msub> <mi>l</mi> <mn>2</mn> </msub> </semantics></math> norm (<b>c</b>) <math display="inline"><semantics> <msub> <mi>l</mi> <mn>1</mn> </msub> </semantics></math> norm (<b>d</b>) <math display="inline"><semantics> <msub> <mi>l</mi> <mfrac> <mn>1</mn> <mn>2</mn> </mfrac> </msub> </semantics></math> norm and (<b>e</b>) <math display="inline"><semantics> <msub> <mi>l</mi> <mn>0</mn> </msub> </semantics></math> norm, in the three dimensional space <math display="inline"><semantics> <msup> <mi mathvariant="double-struck">R</mi> <mn>3</mn> </msup> </semantics></math>.</p>
Full article ">Figure 3
<p>Illustration of noise removal of a single image from Coil20 dataset. (<b>a</b>) Input image (<b>b</b>) Image with <math display="inline"><semantics> <mrow> <mn>20</mn> <mo>%</mo> </mrow> </semantics></math> sparse noise (<b>c</b>) Sparse noise content (<b>d</b>) Noise removed image.</p>
Full article ">Figure 4
<p>Illustrated noise-removed images achieved using proposed method from UCSD dataset (<b>a</b>–<b>g</b>) and Coil20 dataset (<b>h</b>–<b>k</b>) for various levels of sparse noise. The sparse noise levels are indicated on the top of each input image. First row: original input image, second row: images that have been corrupted by various levels of sparse noise, third row: eliminated sparse noise content from each image, fourth row: noise-removed images.</p>
Full article ">Figure 5
<p>Quantitative comparison of purity, NMI and ARI metrics of the proposed method and state-of-the art algorithms under various levels of sparse noise using Coil20 and UCSD datasets.</p>
Full article ">Figure 6
<p>Noise-removed images achieved using proposed method from UCSD and Coil20 datasets for various levels of salt and pepper and Gaussian noise. The noise levels of salt and pepper noise and Gaussian noise are indicated on the top of each input image. First row: original input image, second row: images that have been corrupted by various levels of salt and pepper and Gaussian noise, third row: noise-removed images.</p>
Full article ">Figure 7
<p>Quantitative Comparison (<b>a</b>,<b>b</b>): purity metric and ARI metric for UCSD dataset for various levels of salt and pepper noise. (<b>c</b>,<b>d</b>): purity metric and ARI metric for Coil20 dataset for various levels of Gaussian noise. of purity and ARI metrics of the proposed method and state-of-the art algorithms under salt and pepper noise (<span class="html-italic">d</span>) and Gaussian noise (<math display="inline"><semantics> <msubsup> <mo>σ</mo> <mi>n</mi> <mn>2</mn> </msubsup> </semantics></math>).</p>
Full article ">Figure 8
<p>Sensitivity analysis of the proposed method with the evaluation metrics NMI and ARI. (<b>a</b>) Sensitivity analysis of <math display="inline"><semantics> <msub> <mo>λ</mo> <mn>1</mn> </msub> </semantics></math> and <math display="inline"><semantics> <msub> <mo>λ</mo> <mn>2</mn> </msub> </semantics></math> with ARI metric using Coil20 dataset, (<b>b</b>) Sensitivity analysis of <math display="inline"><semantics> <msub> <mo>λ</mo> <mn>3</mn> </msub> </semantics></math> and <math display="inline"><semantics> <msub> <mo>λ</mo> <mn>4</mn> </msub> </semantics></math> with NMI metric using Coil20 dataset. (<b>c</b>) Sensitivity analysis of <math display="inline"><semantics> <msub> <mo>λ</mo> <mn>3</mn> </msub> </semantics></math> and <math display="inline"><semantics> <msub> <mo>λ</mo> <mn>4</mn> </msub> </semantics></math> with ARI metric using UCSD dataset.</p>
Full article ">Figure 9
<p>Convergence analysis of the proposed method. (<b>a</b>) Convergence analysis of the proposed method with NMI metric using UCSD dataset, (<b>b</b>) Convergence analysis of the proposed method with MCR (<span class="html-italic">m</span>) metric using Coil20 dataset.</p>
Full article ">
13 pages, 6982 KiB  
Article
Word Spotting as a Service: An Unsupervised and Segmentation-Free Framework for Handwritten Documents
by Konstantinos Zagoris, Angelos Amanatiadis and Ioannis Pratikakis
J. Imaging 2021, 7(12), 278; https://doi.org/10.3390/jimaging7120278 - 17 Dec 2021
Cited by 3 | Viewed by 3179
Abstract
Word spotting strategies employed in historical handwritten documents face many challenges due to variation in the writing style and intense degradation. In this paper, a new method that permits efficient and effective word spotting in handwritten documents is presented that relies upon document-oriented [...] Read more.
Word spotting strategies employed in historical handwritten documents face many challenges due to variation in the writing style and intense degradation. In this paper, a new method that permits efficient and effective word spotting in handwritten documents is presented that relies upon document-oriented local features that take into account information around representative keypoints and a matching process that incorporates a spatial context in a local proximity search without using any training data. The method relies on a document-oriented keypoint and feature extraction, along with a fast feature matching method. This enables the corresponding methodological pipeline to be both effectively and efficiently employed in the cloud so that word spotting can be realised as a service in modern mobile devices. The effectiveness and efficiency of the proposed method in terms of its matching accuracy, along with its fast retrieval time, respectively, are shown after a consistent evaluation of several historical handwritten datasets. Full article
Show Figures

Figure 1

Figure 1
<p>The proposed word spotting client-server pipeline. Blue and green arrows denote the indexing and the query processing sequence, respectively.</p>
Full article ">Figure 2
<p>The proposed word spotting architecture. The OFFLINE procedure is only performed once to create the appropriate data structures. The ONLINE process is visible to the user and executes when locating the word.</p>
Full article ">Figure 3
<p>The proposed word spotting indexing and matching pipeline for the Greek handwritten word of ’Aristotle’. (<b>a</b>) Query image; (<b>b</b>) filtered gradient image <math display="inline"><semantics> <msub> <mi>I</mi> <mi>x</mi> </msub> </semantics></math>; (<b>c</b>) filtered gradient image <math display="inline"><semantics> <msub> <mi>I</mi> <mi>y</mi> </msub> </semantics></math>; (<b>d</b>) quantization of the gradient orientation; (<b>e</b>) keypoints; (<b>f</b>) query keypoint (yellow) and the ellipse area enclosed keypoints (blue) that should be matched during the matching process.</p>
Full article ">Figure 4
<p>Presentation of the Dolf’s location for various handwritten words: (<b>a</b>) ‘adjuncto’; (<b>b</b>) ‘adjunctur’; (<b>c</b>,<b>d</b>) ‘andere’; (<b>e</b>,<b>f</b>) ‘however’.</p>
Full article ">Figure 5
<p>Feature extraction: (<b>a</b>) scale invariant window size definition; (<b>b</b>) features calculation; (<b>c</b>) descriptor structure.</p>
Full article ">Figure 6
<p>The overall architecture for the quantization of the DoLF descriptors using multiple Bag of Visual Words.</p>
Full article ">Figure 7
<p>Memory and storage structures and their relationships.</p>
Full article ">Figure 8
<p>(<b>a</b>) The visual representation for the descriptor quantization using four different Bag of Visual Words; (<b>b</b>) the descriptor histogram, before and after applying quantization.</p>
Full article ">Figure 9
<p>Representative document images from (<b>a</b>) English dataset; (<b>b</b>) German dataset; (<b>c</b>) Finnish dataset; (<b>d</b>) Greek dataset.</p>
Full article ">
10 pages, 3143 KiB  
Article
Liouville Integrability in a Four-Dimensional Model of the Visual Cortex
by Ivan Galyaev and Alexey Mashtakov
J. Imaging 2021, 7(12), 277; https://doi.org/10.3390/jimaging7120277 - 17 Dec 2021
Cited by 1 | Viewed by 2088
Abstract
We consider a natural extension of the Petitot–Citti–Sarti model of the primary visual cortex. In the extended model, the curvature of contours is taken into account. The occluded contours are completed via sub-Riemannian geodesics in the four-dimensional space M of positions, orientations, and [...] Read more.
We consider a natural extension of the Petitot–Citti–Sarti model of the primary visual cortex. In the extended model, the curvature of contours is taken into account. The occluded contours are completed via sub-Riemannian geodesics in the four-dimensional space M of positions, orientations, and curvatures. Here, M=R2×SO(2)×R models the configuration space of neurons of the visual cortex. We study the problem of sub-Riemannian geodesics on M via methods of geometric control theory. We prove complete controllability of the system and the existence of optimal controls. By application of the Pontryagin maximum principle, we derive a Hamiltonian system that describes the geodesics. We obtain the explicit parametrization of abnormal extremals. In the normal case, we provide three functionally independent first integrals. Numerical simulations indicate the existence of one more first integral that results in Liouville integrability of the system. Full article
Show Figures

Figure 1

Figure 1
<p>In the four-dimensional model of the visual cortex, an occluded contour is completed via the planar projection of a sub-Riemannian length-minimizer in the space <math display="inline"><semantics> <mrow> <msup> <mrow> <mi mathvariant="double-struck">R</mi> </mrow> <mn>2</mn> </msup> <mo>×</mo> <mi>SO</mi> <mrow> <mo>(</mo> <mn>2</mn> <mo>)</mo> </mrow> <mo>×</mo> <mi mathvariant="double-struck">R</mi> <mo>=</mo> <mi>M</mi> <mo>∋</mo> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>,</mo> <mi>θ</mi> <mo>,</mo> <mi>k</mi> <mo>)</mo> </mrow> </mrow> </semantics></math> of positions, orientations, and curvatures. In the left column, we show an example of the image with partially occluded contours and the complete image. In the right column, we show a trajectory that satisfies the given boundary conditions. The curvature is visualized as its reciprocal—the radius of the osculating circle.</p>
Full article ">Figure 2
<p>Orbits of the Poincare map in the space <math display="inline"><semantics> <mrow> <mo>(</mo> <mi>α</mi> <mo>,</mo> <msub> <mi>h</mi> <mn>3</mn> </msub> <mo>,</mo> <msub> <mi>h</mi> <mn>4</mn> </msub> <mo>)</mo> </mrow> </semantics></math> are formed by intersection points of the transversal hyperspace <math display="inline"><semantics> <mrow> <mi>k</mi> <mo>=</mo> <mn>0</mn> </mrow> </semantics></math> with trajectories close to the periodic one (red dot). Different orbits are depicted in different colors. Starting points are indicated for each trajectory.</p>
Full article ">
16 pages, 3707 KiB  
Article
A Pipelined Tracer-Aware Approach for Lesion Segmentation in Breast DCE-MRI
by Antonio Galli, Stefano Marrone, Gabriele Piantadosi, Mario Sansone and Carlo Sansone
J. Imaging 2021, 7(12), 276; https://doi.org/10.3390/jimaging7120276 - 14 Dec 2021
Cited by 7 | Viewed by 2840
Abstract
The recent spread of Deep Learning (DL) in medical imaging is pushing researchers to explore its suitability for lesion segmentation in Dynamic Contrast-Enhanced Magnetic-Resonance Imaging (DCE-MRI), a complementary imaging procedure increasingly used in breast-cancer analysis. Despite some promising proposed solutions, we argue that [...] Read more.
The recent spread of Deep Learning (DL) in medical imaging is pushing researchers to explore its suitability for lesion segmentation in Dynamic Contrast-Enhanced Magnetic-Resonance Imaging (DCE-MRI), a complementary imaging procedure increasingly used in breast-cancer analysis. Despite some promising proposed solutions, we argue that a “naive” use of DL may have limited effectiveness as the presence of a contrast agent results in the acquisition of multimodal 4D images requiring thorough processing before training a DL model. We thus propose a pipelined approach where each stage is intended to deal with or to leverage a peculiar characteristic of breast DCE-MRI data: the use of a breast-masking pre-processing to remove non-breast tissues; the use of Three-Time-Points (3TP) slices to effectively highlight contrast agent time course; the application of a motion-correction technique to deal with patient involuntary movements; the leverage of a modified U-Net architecture tailored on the problem; and the introduction of a new “Eras/Epochs” training strategy to handle the unbalanced dataset while performing a strong data augmentation. We compared our pipelined solution against some literature works. The results show that our approach outperforms the competitors by a large margin (+9.13% over our previous solution) while also showing a higher generalization ability. Full article
Show Figures

Figure 1

Figure 1
<p>A breast DCE-MRI study and an illustrative time intensity curve. On the left (<b>a</b>), the 4D multimodal volume, with the signal intensity variations reflecting the contrast agent flow over time. On the right (<b>b</b>), the corresponding (illustrative) time intensity curve for a single voxel. The vertical line separates the pre-contrast (early) from post-contrast injection instants.</p>
Full article ">Figure 2
<p>Number of studies published between 2001 and the first ten months of 2021 on Google Scholar filtered by “Breast Lesion Segmentation” as topic keyword.</p>
Full article ">Figure 3
<p>The proposed pipelined segmentation schema: in the first stage, all the extraneous tissues and background air are removed; in the second stage, a motion-correction technique is used to register each post-contrast 3D-volume to the pre-contrast one; in the third stage, for each slice, the corresponding 3TP slice (a three-channel image) is generated by concatenating homologous slices identified by the three-time points defined in [<a href="#B11-jimaging-07-00276" class="html-bibr">11</a>]; finally, in the fourth stage, each lesion is segmented by using a modified U-Net to produce the final lesion binary mask.</p>
Full article ">Figure 4
<p>The considered 3TP U-Net architecture. On the left, the encoding side gradually decreases the spatial resolution while increasing the feature size. On the right, the decoding side gradually increases the spatial resolution from the inner embedding to the final output mask. Dotted lines and capital L highlight the network levels, with L5 being the deeper one. Big grey arrows illustrate the sharing of the cropped feature map, within the same layer, from the encoding to the decoding side, happening during the up-sampling. Compared to the classical U-Net architecture, the considered model varies for the use of batch normalization after each ReLU activation (in both encoding and decoding sides), for the use of zero padding, and for the use of a single output map.</p>
Full article ">Figure 5
<p>Eras/epochs training schema: there are as many eras <span class="html-italic">i</span> as needed by the network to converge; there are as many epochs <span class="html-italic">n</span> as the number of chunks; finally, within each epoch <span class="html-italic">n</span>, <span class="html-italic">k</span> batches are built by using the samples from the corresponding chunk <math display="inline"><semantics> <msub> <mi>c</mi> <mi>n</mi> </msub> </semantics></math>.</p>
Full article ">Figure 6
<p>Violin plots for patient-wise (dots in the image) segmentation performance for all the deep-learning-based approaches considered in this work. For a fair comparison, all the plots were generated by setting the kernel density bandwidth to 10.</p>
Full article ">
16 pages, 12889 KiB  
Article
Characterisation of Single-Phase Fluid-Flow Heterogeneity Due to Localised Deformation in a Porous Rock Using Rapid Neutron Tomography
by Maddi Etxegarai, Erika Tudisco, Alessandro Tengattini, Gioacchino Viggiani, Nikolay Kardjilov and Stephen A. Hall
J. Imaging 2021, 7(12), 275; https://doi.org/10.3390/jimaging7120275 - 14 Dec 2021
Cited by 5 | Viewed by 2825
Abstract
The behaviour of subsurface-reservoir porous rocks is a central topic in the resource engineering industry and has relevant applications in hydrocarbon, water production, and CO2 sequestration. One of the key open issues is the effect of deformation on the hydraulic properties of [...] Read more.
The behaviour of subsurface-reservoir porous rocks is a central topic in the resource engineering industry and has relevant applications in hydrocarbon, water production, and CO2 sequestration. One of the key open issues is the effect of deformation on the hydraulic properties of the host rock and, specifically, in saturated environments. This paper presents a novel full-field data set describing the hydro-mechanical properties of porous geomaterials through in situ neutron and X-ray tomography. The use of high-performance neutron imaging facilities such as CONRAD-2 (Helmholtz-Zentrum Berlin) allows the tracking of the fluid front in saturated samples, making use of the differential neutron contrast between “normal” water and heavy water. To quantify the local hydro-mechanical coupling, we applied a number of existing image analysis algorithms and developed an array of bespoke methods to track the water front and calculate the 3D speed maps. The experimental campaign performed revealed that the pressure-driven flow speed decreases, in saturated samples, in the presence of pre-existing low porosity heterogeneities and compactant shear-bands. Furthermore, the observed complex mechanical behaviour of the samples and the associated fluid flow highlight the necessity for 3D imaging and analysis. Full article
(This article belongs to the Special Issue Recent Advances in Image-Based Geotechnics)
Show Figures

Figure 1

Figure 1
<p>(<b>a</b>) Deviator stress vs. axial strain curve of the sample VO03ME at 30 MPa confining pressure. (<b>b</b>) Deviator stress vs. axial strain curve of the sample VO05ME at 40 MPa confining pressure. The discontinuities in the curves are due to the low acquisition frequency of the apparatus, 1 Hz.</p>
Full article ">Figure 2
<p>(<b>a</b>) Experimental setup inside the CONRAD II neutron beam. 1: Cell, 2: Top valve and pressure transducers, 3: bottom valves and pressure transducer, 4: water tanks, 5: pump, 6: water/air interface, 7: pressure transducers. (<b>b</b>) The hydraulic setup shows the tools connected to the cell during the test in order to control and monitor the fluid inside the sample and the pressure applied.</p>
Full article ">Figure 3
<p>(<b>a</b>) Vertical slice of the VO01ME reconstruction at t = 100 min. The sample is saturated with heavy water, and light water is being pushed into the sample from the bottom. (<b>b</b>) Two horizontal slices of the reconstruction shown in (<b>a</b>). The slice of the top is saturated with heavy water and the bottom with light water. (<b>c</b>) Plots of the greyscale of the four lines drawn in images (<b>a</b>,<b>b</b>).</p>
Full article ">Figure 4
<p>Sigmoidal fitting of the greyscale value of a voxel with time in sample VO01ME. Top plots: For position (160,100) in the horizontal different elevations. (<b>a</b>) 400 (<b>b</b>) 250 (<b>c</b>) 100. Bottom plots: For a given elevation, voxels with decreasing distance from the centre: (<b>d</b>) (86,86) (<b>e</b>) (117,117) (<b>f</b>) (149,149).</p>
Full article ">Figure 5
<p>A vertical slice and a horizontal slice of the parameters obtained in the sigmoidal fitting of sample VO01ME (<b>a</b>) A, (<b>b</b>) B, (<b>c</b>) C, (<b>d</b>) D, (<b>e</b>) error.</p>
Full article ">Figure 6
<p>(<b>a</b>) Vertical central slice of the C parameter (representing the number of tomography at which the light water front arrives at a given pixel) of the sample VO01ME. (<b>b</b>) Same slice of the arrival time map after filtering.</p>
Full article ">Figure 7
<p>(<b>a</b>–<b>c</b>) Two vertical slices and a horizontal slice of the speed map of sample VO01ME.</p>
Full article ">Figure 8
<p>Two vertical slices and a horizontal slice of the Digital Volume Correlation for sample VO03ME. (<b>a</b>–<b>c</b>) Shear strain field. (<b>d</b>–<b>f</b>) Volumetric strain field.</p>
Full article ">Figure 9
<p>(<b>a</b>–<b>d</b>) Three vertical slices and a horizontal slice of the speed map from the flow test on sample VO03ME.</p>
Full article ">Figure 10
<p>Twenty-seven equally separated horizontal slices along the top half of the porosity map of VO03ME.</p>
Full article ">Figure 11
<p>Two vertical slices and a horizontal slice through the DVC-derived strain fields for sample VO05ME. (<b>a</b>–<b>c</b>) Shear strain field. (<b>d</b>–<b>f</b>) Volumetric strain field.</p>
Full article ">Figure 12
<p>(<b>a</b>–<b>c</b>) Two vertical and an horizontal slice of the speed map of sample VO05ME.</p>
Full article ">
18 pages, 3219 KiB  
Article
A System for Real-Time, Online Mixed-Reality Visualization of Cardiac Magnetic Resonance Images
by Dominique Franson, Andrew Dupuis, Vikas Gulani, Mark Griswold and Nicole Seiberlich
J. Imaging 2021, 7(12), 274; https://doi.org/10.3390/jimaging7120274 - 14 Dec 2021
Cited by 6 | Viewed by 3817
Abstract
Image-guided cardiovascular interventions are rapidly evolving procedures that necessitate imaging systems capable of rapid data acquisition and low-latency image reconstruction and visualization. Compared to alternative modalities, Magnetic Resonance Imaging (MRI) is attractive for guidance in complex interventional settings thanks to excellent soft tissue [...] Read more.
Image-guided cardiovascular interventions are rapidly evolving procedures that necessitate imaging systems capable of rapid data acquisition and low-latency image reconstruction and visualization. Compared to alternative modalities, Magnetic Resonance Imaging (MRI) is attractive for guidance in complex interventional settings thanks to excellent soft tissue contrast and large fields-of-view without exposure to ionizing radiation. However, most clinically deployed MRI sequences and visualization pipelines exhibit poor latency characteristics, and spatial integration of complex anatomy and device orientation can be challenging on conventional 2D displays. This work demonstrates a proof-of-concept system linking real-time cardiac MR image acquisition, online low-latency reconstruction, and a stereoscopic display to support further development in real-time MR-guided intervention. Data are acquired using an undersampled, radial trajectory and reconstructed via parallelized through-time radial generalized autocalibrating partially parallel acquisition (GRAPPA) implemented on graphics processing units. Images are rendered for display in a stereoscopic mixed-reality head-mounted display. The system is successfully tested by imaging standard cardiac views in healthy volunteers. Datasets comprised of one slice (46 ms), two slices (92 ms), and three slices (138 ms) are collected, with the acquisition time of each listed in parentheses. Images are displayed with latencies of 42 ms/frame or less for all three conditions. Volumetric data are acquired at one volume per heartbeat with acquisition times of 467 ms and 588 ms when 8 and 12 partitions are acquired, respectively. Volumes are displayed with a latency of 286 ms or less. The faster-than-acquisition latencies for both planar and volumetric display enable real-time 3D visualization of the heart. Full article
Show Figures

Figure 1

Figure 1
<p>Concept illustration of the real-time, online mixed-reality visualization system. MR data are collected using an undersampled, radial trajectory for rapid acquisition. Real-time image reconstruction is performed using a parallelized implementation of through-time radial GRAPPA. The final images are rendered in a mixed-reality headset for intuitive display.</p>
Full article ">Figure 2
<p>Schematic of data flow within the online acquisition, reconstruction, and mixed-reality rendering system. A dedicated image reconstruction computer was introduced into the network containing the scanner, the scanner control computer, and the scanner measurement computer via a switch and an Ethernet cable. A second network was set up using a router to contain the reconstruction computer and the mixed-reality headset. Two configurations were possible. In the first, reconstructed data are sent directly to a self-contained HMD capable of handling the entire rendering pipeline. In the second, a dedicated rendering workstation is used. The reconstruction and rendering computers are connected via Ethernet cables, and rendered frame data are transferred wirelessly to the headset. Solid gray lines show component connections, and dashed red lines depict data transfer.</p>
Full article ">Figure 3
<p>Schematic showing the relative timings of the acquisition, reconstruction, and rendering processes and their sub-steps for a two-slice planar scan. Each frame comprises two slices that are acquired sequentially. The removal of readout oversampling is performed on each line of data as it is acquired, as indicated by the diagonal bar across the sub-step rectangle. Data for the slice are buffered, followed by coil compression, GRAPPA, the NUFFT, and export of the images to the rendering system. The rendering processing includes receiving the image data over the TCP socket, parsing the data, and processing the data. Once all slices per frame are processed, as detected by an update flag, the images are rendered to the user on the next headset frame update. The net display latency is considered to be the time between when all of the data for one frame have been acquired and when that frame is displayed to the user. Note that all components of the system may be active at once, operating on different slices and/or frames of data concurrently.</p>
Full article ">Figure 4
<p>Schematic showing the time required for multiple frames of data to pass through the system for three different dataset types: (<b>top</b>) single-slice planar, (<b>middle</b>) two-slice planar, and (<b>bottom</b>) eight-partition volumetric. Planar data are acquired continuously, while volumetric data are ECG-gated. Images are displayed to the user before completing acquisition of the following frame for all cases. Different colors represent different frames.</p>
Full article ">Figure 5
<p>Operation of the full system at the MR scanner. (<b>Left</b>) A user wearing a mixed-reality headset sits at the scanner control computer before scanning a healthy volunteer lying in the scanner bore (red dashed circle). (<b>Middle</b>) During the scan, the user sees a multi-slice rendering of the volunteer’s heart in real-time. (<b>Right</b>) User’s view of the rendering. Note that what appears black in the image appears transparent to the user; the user sees the rendering within the natural environment. A video version of this figure is available in the <a href="#app1-jimaging-07-00274" class="html-app">Supplementary Materials</a>.</p>
Full article ">Figure 6
<p>Multiple still frames from a user’s perspective of a volumetric scan through a mixed-reality headset are shown. The user moved around the rendering, changed the viewing angle, and moved toward and away from it. The stills were captured by a program in the headset that combines the rendering that is projected to the user with a video capture from the camera embedded in the headset. A video version of this figure is available in the <a href="#app1-jimaging-07-00274" class="html-app">Supplementary Materials</a>.</p>
Full article ">Figure 7
<p>Example 2D panel display and corresponding renderings. (<b>Top left</b>) Three slices in short-axis and four-chamber views. Note that the dark bands across the ventricles in the four-chamber view are from saturation of the signal where the short-axis slices intersect. (<b>Top right</b>) Rendering of the slices in the correct spatial positions and orientations. (<b>Bottom left</b>) Panel display of eight partitions in a volumetric dataset centered over the left ventricle. (<b>Bottom right</b>) Rendering of the dataset.</p>
Full article ">
17 pages, 66075 KiB  
Article
Reliable Estimation of Deterioration Levels via Late Fusion Using Multi-View Distress Images for Practical Inspection
by Keisuke Maeda, Naoki Ogawa, Takahiro Ogawa and Miki Haseyama
J. Imaging 2021, 7(12), 273; https://doi.org/10.3390/jimaging7120273 - 9 Dec 2021
Cited by 1 | Viewed by 2416
Abstract
This paper presents reliable estimation of deterioration levels via late fusion using multi-view distress images for practical inspection. The proposed method simultaneously solves the following two problems that are necessary to support the practical inspection. Since maintenance of infrastructures requires a high level [...] Read more.
This paper presents reliable estimation of deterioration levels via late fusion using multi-view distress images for practical inspection. The proposed method simultaneously solves the following two problems that are necessary to support the practical inspection. Since maintenance of infrastructures requires a high level of safety and reliability, this paper proposes a neural network that can generate an attention map from distress images and text data acquired during the inspection. Thus, deterioration level estimation with high interpretability can be realized. In addition, since multi-view distress images are taken for single distress during the actual inspection, it is necessary to estimate the final result from these images. Therefore, the proposed method integrates estimation results obtained from the multi-view images via the late fusion and can derive an appropriate result considering all the images. To the best of our knowledge, no method has been proposed to solve these problems simultaneously, and this achievement is the biggest contribution of this paper. In this paper, we confirm the effectiveness of the proposed method by conducting experiments using data acquired during the actual inspection. Full article
(This article belongs to the Special Issue Intelligent Media Processing)
Show Figures

Figure 1

Figure 1
<p>Examples of multi-view distress images taken by engineers at the practical inspection.</p>
Full article ">Figure 2
<p>Overview of the transformation of text data into text features.</p>
Full article ">Figure 3
<p>Overview of the proposed method. This model consists of the attention and estimation modules. Especially, we obtain reliable output by the estimation module when one distress image is given. Then we compare these reliabilities obtained from multi-view images, and the final estimation result with the attention map is obtained via the late fusion.</p>
Full article ">Figure 4
<p>The examples of distress images in records.</p>
Full article ">Figure 5
<p>The estimation results for each image. Ours w/o LF means the proposed method without the late fusion.</p>
Full article ">Figure 6
<p>The estimation results for each record. LF indicates the late fusion.</p>
Full article ">Figure 6 Cont.
<p>The estimation results for each record. LF indicates the late fusion.</p>
Full article ">Figure 7
<p>The example of estimation results of efflorescence.</p>
Full article ">Figure 8
<p>The example of estimation results of crack.</p>
Full article ">Figure 9
<p>Verification of estimation performance using different late fusion.</p>
Full article ">Figure 10
<p>The example of the case that the most confident result is likely to be wrong.</p>
Full article ">Figure 11
<p>The example of the case that many images in the record are not necessarily related to the distress.</p>
Full article ">Figure 12
<p>The example of incorrectly estimated results. The estimation performance for the image and record is degraded when the attention map is generated outside of the distress region.</p>
Full article ">
25 pages, 12342 KiB  
Article
Methodology for the Automated Visual Detection of Bird and Bat Collision Fatalities at Onshore Wind Turbines
by Christof Happ, Alexander Sutor and Klaus Hochradel
J. Imaging 2021, 7(12), 272; https://doi.org/10.3390/jimaging7120272 - 9 Dec 2021
Cited by 2 | Viewed by 2884
Abstract
The number of collision fatalities is one of the main quantification measures for research concerning wind power impacts on birds and bats. Despite being integral in ongoing investigations as well as regulatory approvals, the state-of-the-art method for the detection of fatalities remains a [...] Read more.
The number of collision fatalities is one of the main quantification measures for research concerning wind power impacts on birds and bats. Despite being integral in ongoing investigations as well as regulatory approvals, the state-of-the-art method for the detection of fatalities remains a manual search by humans or dogs. This is expensive, time consuming and the efficiency varies greatly among different studies. Therefore, we developed a methodology for the automatic detection using visual/near-infrared cameras for daytime and thermal cameras for nighttime. The cameras can be installed in the nacelle of wind turbines and monitor the area below. The methodology is centered around software that analyzes the images in real time using pixel-wise and region-based methods. We found that the structural similarity is the most important measure for the decision about a detection. Phantom drop tests in the actual wind test field with the system installed on 75 m above the ground resulted in a sensitivity of 75.6% for the nighttime detection and 84.3% for the daylight detection. The night camera detected 2.47 false positives per hour using a time window designed for our phantom drop tests. However, in real applications this time window can be extended to eliminate false positives caused by nightly active animals. Excluding these from our data reduced the false positive rate to 0.05. The daylight camera detected 0.20 false positives per hour. Our proposed method has the advantages of being more consistent, more objective, less time consuming, and less expensive than manual search methods. Full article
(This article belongs to the Special Issue Visual Localization)
Show Figures

Figure 1

Figure 1
<p>Main steps of developed image processing software (blue) plus contrast as main hardware requirement (gray).</p>
Full article ">Figure 2
<p>Camera system is mounted on back of nacelle (<b>left</b>) and senses area underneath wind turbine (<b>right</b>). Tower produces a dead corner which cannot be recorded by system (red projection).</p>
Full article ">Figure 3
<p>(<b>a</b>) dimensions of wind turbine; FOV (= field of view) thermal camera (red); FOV spectral camera (blue). (<b>b</b>) Prototype of camera system for collision victim detection.</p>
Full article ">Figure 4
<p>Reflectance of endangered bird species, vegetation, and other possible background surfaces.</p>
Full article ">Figure 5
<p>Measurement positions on the back (<b>a</b>) and on the front side (<b>b</b>) of a Milvus milvus (red kite).</p>
Full article ">Figure 6
<p>Reflectance measurement result of Milvus milvus (<a href="#jimaging-07-00272-f005" class="html-fig">Figure 5</a>) compared to that of grass.</p>
Full article ">Figure 7
<p>Shape of sun spectrum (normed to maximum of amplitude) measured on earth’s surface; measured transmittance of used NIR filter.</p>
Full article ">Figure 8
<p>(<b>a</b>) Picture stack with new (cyan) and old (yellow) images, (<b>b</b>) gray values over time for a potential strike victim, (<b>c</b>) change in brightness over time, (<b>d</b>) alternating gray values, e.g., moving grass.</p>
Full article ">Figure 9
<p>All images show same zoomed part of the scene. In middle of (<b>a</b>) phantom can be seen (middle) laying in the grass. One minute earlier (<b>b</b>) was taken; object was not there and sun was hidden by clouds. (<b>c</b>) is same image as (<b>b</b>) but with median adjusted to (<b>a</b>). (<b>d</b>) is same as (<b>b</b>) after histogram matching with respect to (<b>a</b>).</p>
Full article ">Figure 10
<p>(<b>a</b>) Tool for manual segmentation of inner (pink) and outer (purple) region for analysis. <math display="inline"><semantics> <msub> <mi>t</mi> <mn>0</mn> </msub> </semantics></math> to <math display="inline"><semantics> <msub> <mi>t</mi> <mn>7</mn> </msub> </semantics></math> (newest image) shows image stack according to <a href="#jimaging-07-00272-f008" class="html-fig">Figure 8</a>; images are zoomed to relevant part and are with VIS camera (<b>b</b>) gray values over time of segmented regions starting from <math display="inline"><semantics> <mrow> <msub> <mi>t</mi> <mn>0</mn> </msub> <mo>=</mo> <mn>0</mn> </mrow> </semantics></math> with <math display="inline"><semantics> <mrow> <mo>Δ</mo> <mi>t</mi> <mo>=</mo> <mn>10</mn> <mo> </mo> <mi mathvariant="normal">s</mi> </mrow> </semantics></math>.</p>
Full article ">Figure 11
<p>(<b>a</b>) Fatality score (red) and ground truth; (<b>b</b>) resulting region after thresholding and morphological operations.</p>
Full article ">Figure 12
<p>LUT for LUTshift function in Equation (<a href="#FD1-jimaging-07-00272" class="html-disp-formula">1</a>).</p>
Full article ">Figure 13
<p>Moving people recorded with the LWIR camera; the median helps to get rid of outliers.</p>
Full article ">Figure 14
<p>Region Generation: (<b>a</b>) thresholded image, (<b>b</b>) labeled <span class="html-italic">full</span>, (<b>c</b>) labeled <span class="html-italic">inner</span> and (<b>d</b>) labeled <span class="html-italic">outer</span> areas.</p>
Full article ">Figure 15
<p>Criteria for the decision about a detection.</p>
Full article ">Figure 16
<p>Warm test object (body temperature) with about 5 cm lengths lying in grass; upper left: resulting score from pixel-wise detection; upper right: derived inner (pink) and outer (purple) region; two lower rows: image stack from oldest (<math display="inline"><semantics> <msub> <mi>t</mi> <mn>0</mn> </msub> </semantics></math>) to newest (<math display="inline"><semantics> <msub> <mi>t</mi> <mn>9</mn> </msub> </semantics></math>) image with assumed strike happening between <math display="inline"><semantics> <msub> <mi>t</mi> <mn>4</mn> </msub> </semantics></math> and <math display="inline"><semantics> <msub> <mi>t</mi> <mn>5</mn> </msub> </semantics></math>.</p>
Full article ">Figure 17
<p>Wind leads to a high LWIR score distributed over a big area compared to a potential detection.</p>
Full article ">Figure 18
<p>Grass getting illuminated by sun differently over time, which leads to a false positive detection after preselection, but obviously image content stays same from <math display="inline"><semantics> <msub> <mi>t</mi> <mn>4</mn> </msub> </semantics></math> to <math display="inline"><semantics> <msub> <mi>t</mi> <mn>5</mn> </msub> </semantics></math>.</p>
Full article ">Figure 19
<p>Birds sitting on a guy wire produce false positives without further filtering.</p>
Full article ">Figure 20
<p>Gray value vector of region at time <math display="inline"><semantics> <msub> <mi mathvariant="bold">t</mi> <mi>i</mi> </msub> </semantics></math> on x-axis plotted against vector of the same region at time <math display="inline"><semantics> <msub> <mi mathvariant="bold">t</mi> <mrow> <mi>i</mi> <mo>+</mo> <mn>1</mn> </mrow> </msub> </semantics></math> on y-axis of true positive detection (<a href="#jimaging-07-00272-f016" class="html-fig">Figure 16</a>).</p>
Full article ">Figure 21
<p>Phantom with similar gray values as background.</p>
Full article ">Figure 22
<p>Gray values of vector <math display="inline"><semantics> <msub> <mi mathvariant="bold">t</mi> <mi>i</mi> </msub> </semantics></math> on x-axis and vector <math display="inline"><semantics> <msub> <mi mathvariant="bold">t</mi> <mrow> <mi>i</mi> <mo>+</mo> <mn>1</mn> </mrow> </msub> </semantics></math> on y-axis for image data of <a href="#jimaging-07-00272-f021" class="html-fig">Figure 21</a>.</p>
Full article ">Figure 23
<p>SDIFFs with middle SDIFF in red and the maximum of outer SDIFFs in blue with green distance line. Horizontal lines are the thresholds according to <a href="#jimaging-07-00272-t005" class="html-table">Table 5</a>. (<b>a</b>,<b>b</b>) are true positives, (<b>c</b>,<b>d</b>) are false positives from preselection. (<b>c</b>,<b>d</b>) are correctly dismissed through SDIFF criteria in <a href="#jimaging-07-00272-t005" class="html-table">Table 5</a>.</p>
Full article ">Figure 24
<p>The camera system in the gray box (<b>a</b>) was mounted in 75 m height (<b>b</b>).</p>
Full article ">Figure 25
<p>The bat phantom for the LWIR camera (<b>a</b>) and the red kite phantom for the spectral camera (<b>b</b>–<b>h</b>) were placed on different defined positions including grass with various lengths (<b>b</b>–<b>f</b>,<b>h</b>) and gravel road (<b>g</b>).</p>
Full article ">Figure 26
<p>Positions for phantom drops (VIS and LWIR).</p>
Full article ">Figure 27
<p>Example detection of balloon phantom at about 35 °C in middle high grass at position 2.</p>
Full article ">Figure 28
<p>Two pixel bat phantom cooling down over time.</p>
Full article ">Figure 29
<p>Example detection of red kite phantom at about 35 °C in middle high grass at position 2.</p>
Full article ">Figure 30
<p>Footsteps in high grass can produce false positives.</p>
Full article ">
21 pages, 2435 KiB  
Article
Multi-Frequency Image Completion via a Biologically-Inspired Sub-Riemannian Model with Frequency and Phase
by Emre Baspinar
J. Imaging 2021, 7(12), 271; https://doi.org/10.3390/jimaging7120271 - 9 Dec 2021
Cited by 2 | Viewed by 2359
Abstract
We present a novel cortically-inspired image completion algorithm. It uses five-dimensional sub-Riemannian cortical geometry, modeling the orientation, spatial frequency and phase-selective behavior of the cells in the visual cortex. The algorithm extracts the orientation, frequency and phase information existing in a given two-dimensional [...] Read more.
We present a novel cortically-inspired image completion algorithm. It uses five-dimensional sub-Riemannian cortical geometry, modeling the orientation, spatial frequency and phase-selective behavior of the cells in the visual cortex. The algorithm extracts the orientation, frequency and phase information existing in a given two-dimensional corrupted input image via a Gabor transform and represents those values in terms of cortical cell output responses in the model geometry. Then, it performs completion via a diffusion concentrated in a neighborhood along the neural connections within the model geometry. The diffusion models the activity propagation integrating orientation, frequency and phase features along the neural connections. Finally, the algorithm transforms the diffused and completed output responses back to the two-dimensional image plane. Full article
Show Figures

Figure 1

Figure 1
<p>An example of the law of good continuity. We capture the curve on the bottom left as the curve underlying the aligned fragments on the top left; we do not capture any curve underlying the fragments on the top right due to the abruptly changing orientation angles of the fragments.</p>
Full article ">Figure 2
<p>Two experimental settings from Field, Hayes and Hess [<a href="#B7-jimaging-07-00271" class="html-bibr">7</a>]. A stimuli with aligned patches which we capture (<b>left</b>) and a stimuli plus the background with randomly oriented patches (<b>right</b>) are shown. Abrupt changes in the fragment orientations make it difficult to detect the aligned pattern in the bottom row.</p>
Full article ">Figure 3
<p><b>Top</b>: association fields aligned with a horizontal patch as shown by Field, Hayes and Hess [<a href="#B7-jimaging-07-00271" class="html-bibr">7</a>]. <b>Bottom</b>: solid curves represent the association fields between strongly associated fragments, and the dashed ones imply no fields between weakly associated fragments.</p>
Full article ">Figure 4
<p><b>Left</b>: Kanizsa triangle. There is no direct stimulus, yet we perceive a white triangle on top of the rest (modal completion). We perceive another triangular whose border is marked by the black lines on the bottom layer (amodal completion). <b>Right</b>: Ehrenstein illusion. We perceive a white disk around the center despite the absence of a direct stimulus (modal completion). We recognize that each vertical, horizontal or diagonal line fragment comprises a whole line which is occluded by the white disk (amodal completion).</p>
Full article ">Figure 5
<p><b>Left</b>: real association fields. <b>Right</b>: projections of <math display="inline"><semantics> <mrow> <mi>SE</mi> <mo>(</mo> <mn>2</mn> <mo>)</mo> </mrow> </semantics></math> horizontal integral curves. Figures are adapted from [<a href="#B7-jimaging-07-00271" class="html-bibr">7</a>,<a href="#B33-jimaging-07-00271" class="html-bibr">33</a>].</p>
Full article ">Figure 6
<p>Two Gabor functions with low (<b>left</b> column) and high (<b>right</b> column) spatial frequencies. Top row: even (real) component of the Gabor functions. Bottom row: odd (imaginary) components.</p>
Full article ">Figure 7
<p>A horizontal integral curve along the vector field <math display="inline"><semantics> <mrow> <msub> <mi>X</mi> <mn>1</mn> </msub> <mo>+</mo> <msub> <mi>X</mi> <mn>2</mn> </msub> </mrow> </semantics></math>. It represents an orientation fiber once <span class="html-italic">f</span> and <math display="inline"><semantics> <mi>ϕ</mi> </semantics></math> are fixed. The tangent planes spanned by <math display="inline"><semantics> <mrow> <msub> <mi>X</mi> <mn>1</mn> </msub> <mo>,</mo> </mrow> </semantics></math><math display="inline"><semantics> <msub> <mi>X</mi> <mn>2</mn> </msub> </semantics></math> (<b>left</b>) and <math display="inline"><semantics> <mrow> <msub> <mi>X</mi> <mn>3</mn> </msub> <mo>,</mo> </mrow> </semantics></math><math display="inline"><semantics> <msub> <mi>X</mi> <mn>4</mn> </msub> </semantics></math> (<b>right</b>) are shown at six points on the curve.</p>
Full article ">Figure 8
<p>Horizontal integral curve fans corresponding to <math display="inline"><semantics> <mrow> <msub> <mi>X</mi> <mn>1</mn> </msub> <mo>+</mo> <msub> <mi>c</mi> <mn>2</mn> </msub> <msub> <mi>X</mi> <mn>2</mn> </msub> </mrow> </semantics></math> (<b>left</b>) and <math display="inline"><semantics> <mrow> <msub> <mi>X</mi> <mn>3</mn> </msub> <mo>+</mo> <msub> <mi>c</mi> <mn>4</mn> </msub> <msub> <mi>X</mi> <mn>4</mn> </msub> </mrow> </semantics></math> (<b>right</b>) where <math display="inline"><semantics> <msub> <mi>c</mi> <mn>2</mn> </msub> </semantics></math> and <math display="inline"><semantics> <msub> <mi>c</mi> <mn>4</mn> </msub> </semantics></math> are varied.</p>
Full article ">Figure 9
<p>Illustration of the vectors <math display="inline"><semantics> <msubsup> <mi>e</mi> <mi>ξ</mi> <mi>k</mi> </msubsup> </semantics></math> and <math display="inline"><semantics> <msubsup> <mi>e</mi> <mrow> <mi>η</mi> </mrow> <mi>k</mi> </msubsup> </semantics></math> at <math display="inline"><semantics> <mrow> <mo>(</mo> <mn>0</mn> <mo>,</mo> <mn>0</mn> <mo>)</mo> </mrow> </semantics></math> with <math display="inline"><semantics> <mrow> <mo>Δ</mo> <mi>x</mi> <mo>=</mo> <msub> <mo>Δ</mo> <mi>y</mi> </msub> <mo>=</mo> <mn>1</mn> </mrow> </semantics></math>. The figure was modified and adapted from [<a href="#B44-jimaging-07-00271" class="html-bibr">44</a>].</p>
Full article ">Figure 10
<p>Completion of arcs with sinusoidal pattern. <b>Left</b>: original image. <b>Middle left</b>: corrupted image with occluding arcs. <b>Middle right</b>: completed image via the approximate method. <b>Right</b>: completion via the exact method.</p>
Full article ">Figure 11
<p>Completion of an occluded real texture image by two different arc patterns on the top and bottom rows. <b>Left</b>: Original image. <b>Middle left:</b> Image with occluding arcs. <b>Middle right</b>: Completed image via the approximate framework. <b>Right</b>: Completed image via the exact framework.</p>
Full article ">Figure 12
<p>Completion of a real texture image occluded by two different line patterns on the top and bottom rows. <b>Left</b>: original image. <b>Middle left</b>: image with occluding vertical and horizontal lines. <b>Middle right</b>: completed image via the approximate framework. <b>Right</b>: completed image via the exact framework.</p>
Full article ">Figure 13
<p>Single-frequency completion associated with the bottom row of <a href="#jimaging-07-00271-f011" class="html-fig">Figure 11</a>.</p>
Full article ">Figure 14
<p>Single-frequency completion associated with the bottom row of <a href="#jimaging-07-00271-f012" class="html-fig">Figure 12</a>.</p>
Full article ">Figure 15
<p>Completion of an occluded real texture image by two different arc patterns on the top and bottom rows. Left: original image taken from [<a href="#B68-jimaging-07-00271" class="html-bibr">68</a>]. Middle left: image with occluding arcs. Middle right: completed image via the approximate framework. Right: completed image via the exact framework.</p>
Full article ">Figure 16
<p>Completion of an occluded real texture image by two different line patterns on the top and bottom rows. <b>Left:</b> original image taken from [<a href="#B68-jimaging-07-00271" class="html-bibr">68</a>]. <b>Middle left:</b> image with occluding arcs. <b>Middle right:</b> completed image via the approximate framework. <b>Right:</b> completed image via the exact framework.</p>
Full article ">Figure 17
<p>Single-frequency completion associated with the bottom row of <a href="#jimaging-07-00271-f015" class="html-fig">Figure 15</a>.</p>
Full article ">Figure 18
<p>Single-frequency completion associated with the bottom row of <a href="#jimaging-07-00271-f016" class="html-fig">Figure 16</a>.</p>
Full article ">Figure 19
<p><b>Top left:</b> test image from [<a href="#B69-jimaging-07-00271" class="html-bibr">69</a>]. <b>Top right:</b> image to be completed. <b>Bottom left:</b> completion via our method. <b>Bottom middle:</b> completion via the method in [<a href="#B50-jimaging-07-00271" class="html-bibr">50</a>]. <b>Bottom right:</b> completion via the method in [<a href="#B48-jimaging-07-00271" class="html-bibr">48</a>].</p>
Full article ">
30 pages, 5078 KiB  
Article
A Fast and Accurate Approach to Multiple-Vehicle Localization and Tracking from Monocular Aerial Images
by Daniel Tøttrup, Stinus Lykke Skovgaard, Jonas le Fevre Sejersen and Rui Pimentel de Figueiredo
J. Imaging 2021, 7(12), 270; https://doi.org/10.3390/jimaging7120270 - 8 Dec 2021
Cited by 3 | Viewed by 2571
Abstract
In this work we present a novel end-to-end solution for tracking objects (i.e., vessels), using video streams from aerial drones, in dynamic maritime environments. Our method relies on deep features, which are learned using realistic simulation data, for robust object detection, segmentation and [...] Read more.
In this work we present a novel end-to-end solution for tracking objects (i.e., vessels), using video streams from aerial drones, in dynamic maritime environments. Our method relies on deep features, which are learned using realistic simulation data, for robust object detection, segmentation and tracking. Furthermore, we propose the use of rotated bounding-box representations, which are computed by taking advantage of pixel-level object segmentation, for improved tracking accuracy, by reducing erroneous data associations during tracking, when combined with the appearance-based features. A thorough set of experiments and results obtained in a realistic shipyard simulation environment, demonstrate that our method can accurately, and fast detect and track dynamic objects seen from a top-view. Full article
(This article belongs to the Special Issue Visual Localization)
Show Figures

Figure 1

Figure 1
<p>A high-level illustration of a deep learning (DL) object detection network architecture. Layered squares represent feature maps. The connection between stacked feature maps represents convolutional operations. Elongated rectangles to the right represent feature vectors. The connections between feature vectors represent fully connected operations. This figure is created using the tool from [<a href="#B10-jimaging-07-00270" class="html-bibr">10</a>].</p>
Full article ">Figure 2
<p>A high-level illustration of a DL semantic segmentation network architecture. Layered squares represent feature maps. The connection between stacked feature maps represents convolutional operations. This figure is created using the tool from [<a href="#B10-jimaging-07-00270" class="html-bibr">10</a>].</p>
Full article ">Figure 3
<p>Interface of Airsim in unreal engine.</p>
Full article ">Figure 4
<p>Images from the simulation environment captured at different altitudes.</p>
Full article ">Figure 5
<p>(<b>left</b>) Worst case scenario with axis-aligned bounding boxes. The bounding box of the small ship is inside the bounding box of the large ship. (<b>right</b>) Segmentation mask superimposed on top of the two ships. A rotated bounding box is made based on segmentation, thus representing the vessel’s spatial location more precisely.</p>
Full article ">Figure 6
<p>Rotated bounding box encompassing two instances of the same class.</p>
Full article ">Figure 7
<p>High-level illustration of the architecture of Mask R-CNN.</p>
Full article ">Figure 8
<p>A high-level illustration of input and output of Mask R-CNN.</p>
Full article ">Figure 9
<p><b>Top</b>: Initial bounding-box encompassing a polygon. <b>Bottom</b>: 3 rotations of the “calipers”.</p>
Full article ">Figure 10
<p>(<b>Left</b>) Raw image. (<b>Right</b>) Annotated ground-truth image.</p>
Full article ">Figure 11
<p>Binary mask of a single object instance.</p>
Full article ">Figure 12
<p>High-level illustration of tracking-by-detection framework.</p>
Full article ">Figure 13
<p>High-level pipeline illustration of how Deep SORT is used in the proposed solution.</p>
Full article ">Figure 14
<p>Overview of the different evaluation phases.</p>
Full article ">Figure 15
<p>Examples of images in dataset <span class="html-italic">version 1.1</span>.</p>
Full article ">Figure 16
<p>Example of new images in dataset <span class="html-italic">version 1.2</span>.</p>
Full article ">Figure 17
<p>(<b>Left</b>) AP for bounding boxes for each category and dataset version. (<b>Right</b>) AP for segmentation for each category and dataset version.</p>
Full article ">Figure 18
<p>mAP of bounding box and segmentation for each dataset version.</p>
Full article ">Figure 19
<p>(<b>Left</b>) mAP for the best performing model. The best performing model was found based on the plot on the right. Pre-trained model has been trained for 5 epochs, not pre-trained has been trained for 6 epochs. (<b>Right</b>) mAP when tested on the test set over epochs.</p>
Full article ">Figure 20
<p>Average precision of categories individually with different levels of data augmentation enabled. (<b>Left</b>) AP for bounding boxes. (<b>Right</b>) AP for segmentation.</p>
Full article ">Figure 21
<p>mAP with different levels of data augmentation used.</p>
Full article ">Figure 22
<p>mAP with different levels of data augmentation used over epochs.</p>
Full article ">Figure 23
<p>Mean inference time for the detection stage over number of objects.</p>
Full article ">Figure 24
<p>Mean inference time for MOT stage over number of objects.</p>
Full article ">
16 pages, 3799 KiB  
Article
Brain Tumor Segmentation Based on Deep Learning’s Feature Representation
by Ilyasse Aboussaleh, Jamal Riffi, Adnane Mohamed Mahraz and Hamid Tairi
J. Imaging 2021, 7(12), 269; https://doi.org/10.3390/jimaging7120269 - 8 Dec 2021
Cited by 32 | Viewed by 5431
Abstract
Brain tumor is considered as one of the most serious causes of death in the world. Thus, it is very important to detect it as early as possible. In order to predict and segment the tumor, many approaches have been proposed. However, they [...] Read more.
Brain tumor is considered as one of the most serious causes of death in the world. Thus, it is very important to detect it as early as possible. In order to predict and segment the tumor, many approaches have been proposed. However, they suffer from different problems such as the necessity of the intervention of a specialist, the long required run-time and the choice of the appropriate feature extractor. To address these issues, we proposed an approach based on convolution neural network architecture aiming at predicting and segmenting simultaneously a cerebral tumor. The proposal was divided into two phases. Firstly, aiming at avoiding the use of the labeled image that implies a subject intervention of the specialist, we used a simple binary annotation that reflects the existence of the tumor or not. Secondly, the prepared image data were fed into our deep learning model in which the final classification was obtained; if the classification indicated the existence of the tumor, the brain tumor was segmented based on the feature representations generated by the convolutional neural network architectures. The proposed method was trained on the BraTS 2017 dataset with different types of gliomas. The achieved results show the performance of the proposed approach in terms of accuracy, precision, recall and Dice similarity coefficient. Our model showed an accuracy of 91% in tumor classification and a Dice similarity coefficient of 82.35% in tumor segmentation. Full article
(This article belongs to the Section Medical Imaging)
Show Figures

Figure 1

Figure 1
<p>General architecture of the proposed method.</p>
Full article ">Figure 2
<p>Example of data augmentation: (<b>a</b>) FLAIR image; (<b>b</b>) shift; (<b>c</b>) rotation +90°; (<b>d</b>) rotation -90°. (<b>e</b>) Flip horizontally; (<b>f</b>) flip vertically; (<b>g</b>) noise addition; (<b>h</b>) blur.</p>
Full article ">Figure 3
<p>Different modalities of MRI images: (<b>a</b>) T1; (<b>b</b>) T2; (<b>c</b>) T1c; and (<b>d</b>) FLAIR.</p>
Full article ">Figure 4
<p>CNN architecture used in our approach.</p>
Full article ">Figure 5
<p>Features extracted from the last convolution layer of our CNN model.</p>
Full article ">Figure 6
<p>Illustration of our tumor segmentation process of an MRI image from training BraTS 2017 dataset.</p>
Full article ">Figure 7
<p>Accuracy and loss of the training and validation subsets as a function of number of epochs.</p>
Full article ">Figure 8
<p>ROC curve of our binary classification model to detect the existence of the tumor.</p>
Full article ">Figure 9
<p>Segmentation result of our method on some BraTS 2017 HGG images: (<b>a</b>) original image, (<b>b</b>) segmentation before post-processing, (<b>c</b>) segmentation after post-processing, (<b>d</b>) ground truth.</p>
Full article ">Figure 10
<p>Segmentation result of our method on some BraTS 2017 LGG images: (<b>a</b>) original image, (<b>b</b>) segmentation before post-processing, (<b>c</b>) segmentation after post-processing, (<b>d</b>) ground truth.</p>
Full article ">
13 pages, 3512 KiB  
Article
A Reversible Data Hiding Method in Encrypted Images for Controlling Trade-Off between Hiding Capacity and Compression Efficiency
by Ryota Motomura, Shoko Imaizumi and Hitoshi Kiya
J. Imaging 2021, 7(12), 268; https://doi.org/10.3390/jimaging7120268 - 7 Dec 2021
Cited by 5 | Viewed by 3089
Abstract
In this paper, we propose a new framework for reversible data hiding in encrypted images, where both the hiding capacity and lossless compression efficiency are flexibly controlled. There exist two main purposes; one is to provide highly efficient lossless compression under a required [...] Read more.
In this paper, we propose a new framework for reversible data hiding in encrypted images, where both the hiding capacity and lossless compression efficiency are flexibly controlled. There exist two main purposes; one is to provide highly efficient lossless compression under a required hiding capacity, while the other is to enable us to extract an embedded payload from a decrypted image. The proposed method can decrypt marked encrypted images without data extraction and derive marked images. An original image is arbitrarily divided into two regions. Two different methods for reversible data hiding in encrypted images (RDH-EI) are used in our method, and each one is used for either region. Consequently, one region can be decrypted without data extraction and also losslessly compressed using image coding standards even after the processing. The other region possesses a significantly high hiding rate, around 1 bpp. Experimental results show the effectiveness of the proposed method in terms of hiding capacity and lossless compression efficiency. Full article
(This article belongs to the Special Issue Intelligent Media Processing)
Show Figures

Figure 1

Figure 1
<p>Block diagram of RDH-EtC method [<a href="#B21-jimaging-07-00268" class="html-bibr">21</a>]. <math display="inline"><semantics> <mrow> <mi>P</mi> <mi>P</mi> </mrow> </semantics></math> and <math display="inline"><semantics> <mrow> <mi>Z</mi> <mi>P</mi> </mrow> </semantics></math> are bins with the highest and lowest frequency bins in the original-image histogram.</p>
Full article ">Figure 2
<p>Block diagram of the RDH-MSB method [<a href="#B18-jimaging-07-00268" class="html-bibr">18</a>].</p>
Full article ">Figure 3
<p>Classification for region <math display="inline"><semantics> <mi>α</mi> </semantics></math> and region <math display="inline"><semantics> <mi>β</mi> </semantics></math>. Region <math display="inline"><semantics> <mi>α</mi> </semantics></math> is areas inside red frames, and region <math display="inline"><semantics> <mi>β</mi> </semantics></math> is other areas. (<b>a</b>) kodim09, (<b>b</b>) kodim18.</p>
Full article ">Figure 4
<p>Block diagram of the proposed method.</p>
Full article ">Figure 5
<p>Restoration process of the proposed method.</p>
Full article ">Figure 6
<p>Restoration options for region <math display="inline"><semantics> <mi>α</mi> </semantics></math>. (<b>a</b>) Normal, (<b>b</b>) decryption without data extraction, (<b>c</b>) decryption then data extraction.</p>
Full article ">Figure 7
<p>Restoration process of the RDH-MSB method [<a href="#B18-jimaging-07-00268" class="html-bibr">18</a>].</p>
Full article ">Figure 8
<p>Resulting image with <math display="inline"><semantics> <mi>α</mi> </semantics></math> decryption only (kodim23). (<b>a</b>) Original image, (<b>b</b>) marked image.</p>
Full article ">Figure 9
<p>Examples of test images. (<b>a</b>) kodim09, (<b>b</b>) kodim18.</p>
Full article ">Figure 10
<p>Marked encrypted images with the proposed method (kodim9). (<b>a</b>) Grayscale-based image, (<b>b</b>) <math display="inline"><semantics> <mi>α</mi> </semantics></math>:<math display="inline"><semantics> <mi>β</mi> </semantics></math>=100:0, (<b>c</b>) <math display="inline"><semantics> <mi>α</mi> </semantics></math>:<math display="inline"><semantics> <mi>β</mi> </semantics></math>=75:25, (<b>d</b>) <math display="inline"><semantics> <mi>α</mi> </semantics></math>, <math display="inline"><semantics> <mi>β</mi> </semantics></math>=50:50, (<b>e</b>) <math display="inline"><semantics> <mi>α</mi> </semantics></math>:<math display="inline"><semantics> <mi>β</mi> </semantics></math>=25:75, (<b>f</b>) <math display="inline"><semantics> <mi>α</mi> </semantics></math>:<math display="inline"><semantics> <mi>β</mi> </semantics></math>=0:100.</p>
Full article ">Figure 11
<p>Marked encrypted images with the proposed method (kodim18). (<b>a</b>) Grayscale-based image, (<b>b</b>) <math display="inline"><semantics> <mi>α</mi> </semantics></math>:<math display="inline"><semantics> <mi>β</mi> </semantics></math>=100:0, (<b>c</b>) <math display="inline"><semantics> <mi>α</mi> </semantics></math>:<math display="inline"><semantics> <mi>β</mi> </semantics></math>=75:25, (<b>d</b>) <math display="inline"><semantics> <mi>α</mi> </semantics></math>:<math display="inline"><semantics> <mi>β</mi> </semantics></math>=50:50, (<b>e</b>) <math display="inline"><semantics> <mi>α</mi> </semantics></math>:<math display="inline"><semantics> <mi>β</mi> </semantics></math>=25:75, (<b>f</b>) <math display="inline"><semantics> <mi>α</mi> </semantics></math>:<math display="inline"><semantics> <mi>β</mi> </semantics></math>=0:100.</p>
Full article ">Figure 12
<p>Lossless compression performance. (<b>a</b>) JPEG-LS, (<b>b</b>) JPEG 2000.</p>
Full article ">Figure 13
<p>Data hiding capacity.</p>
Full article ">Figure 14
<p>Transition in lossless compression performance and hiding capacity for three images (kodim11, kodim18, kodim24). (<b>a</b>) Lossless compression performance by JPEG-LS, (<b>b</b>) data hiding capacity.</p>
Full article ">
21 pages, 3963 KiB  
Article
A Semi-Supervised Reduced-Space Method for Hyperspectral Imaging Segmentation
by Giacomo Aletti, Alessandro Benfenati and Giovanni Naldi
J. Imaging 2021, 7(12), 267; https://doi.org/10.3390/jimaging7120267 - 7 Dec 2021
Cited by 8 | Viewed by 3028
Abstract
The development of the hyperspectral remote sensor technology allows the acquisition of images with a very detailed spectral information for each pixel. Because of this, hyperspectral images (HSI) potentially possess larger capabilities in solving many scientific and practical problems in agriculture, biomedical, ecological, [...] Read more.
The development of the hyperspectral remote sensor technology allows the acquisition of images with a very detailed spectral information for each pixel. Because of this, hyperspectral images (HSI) potentially possess larger capabilities in solving many scientific and practical problems in agriculture, biomedical, ecological, geological, hydrological studies. However, their analysis requires developing specialized and fast algorithms for data processing, due the high dimensionality of the data. In this work, we propose a new semi-supervised method for multilabel segmentation of HSI that combines a suitable linear discriminant analysis, a similarity index to compare different spectra, and a random walk based model with a direct label assignment. The user-marked regions are used for the projection of the original high-dimensional feature space to a lower dimensional space, such that the class separation is maximized. This allows to retain in an automatic way the most informative features, lightening the successive computational burden. The part of the random walk is related to a combinatorial Dirichlet problem involving a weighted graph, where the nodes are the projected pixel of the original HSI, and the positive weights depend on the distances between these nodes. We then assign to each pixel of the original image a probability quantifying the likelihood that the pixel (node) belongs to some subregion. The computation of the spectral distance involves both the coordinates in a features space of a pixel and of its neighbors. The final segmentation process is therefore reduced to a suitable optimization problem coupling the probabilities from the random walker computation, and the similarity with respect the initially labeled pixels. We discuss the properties of the new method with experimental results carried on benchmark images. Full article
(This article belongs to the Special Issue Advances in Multi/Hyperspectral Imaging)
Show Figures

Figure 1

Figure 1
<p>Visual representation of the RLDA algorithm and distance computation of the neighborhoods. First row: example with just 3 pixels and their hyperspectral distribution, reduced in their 2D principal components. Second row: RLDA applied to a larger image. Third row: extraction of the 8-neighborhoods of two pixels in each principal component.</p>
Full article ">Figure 2
<p>Flow chart of Algorithm 2.</p>
Full article ">Figure 3
<p>Indian Pines segmentation results. (<b>a</b>): false gray scale image. (<b>b</b>): segmentation result. (<b>c</b>): true labels. (<b>d</b>): accuracy over the classes. The colormaps in (<b>b</b>,<b>c</b>) are different because the segmentation process takes into account 5 labels, while the ground truth contains 16 labels. Each region in the ground truth falls almost entirely in one of the 5 manually selected labels.</p>
Full article ">Figure 4
<p>(<b>a</b>–<b>c</b>): false grayscale image, segmentation result and ground truth labels for the <span class="html-italic">Pavia University</span> dataset. (<b>d</b>–<b>f</b>): false grayscale image, segmentation result and ground truth labels for the <span class="html-italic">Salinas HSI</span> dataset. The former experiments achieves an Overall Accuracy of 0.93, whilst the latter achieves an OA of 0.9696.</p>
Full article ">Figure 5
<p>(<b>a</b>–<b>c</b>): Dependence of the RI wrt <math display="inline"><semantics> <mi>α</mi> </semantics></math> for several values of <math display="inline"><semantics> <mi>λ</mi> </semantics></math>, from 1 to <math display="inline"><semantics> <msup> <mn>10</mn> <mrow> <mo>−</mo> <mn>6</mn> </mrow> </msup> </semantics></math> for <span class="html-italic">Indian Pines</span>, <span class="html-italic">Pavia University</span> and <span class="html-italic">Salinas HSI</span> datasets, respectively. The RI remains high, there is a peak around <math display="inline"><semantics> <mrow> <mi>α</mi> <mo>=</mo> <mn>0.85</mn> </mrow> </semantics></math> in the former cases, while the <span class="html-italic">Salinas HSI</span> datasets achieve its best performance for <math display="inline"><semantics> <mrow> <mi>α</mi> <mo>=</mo> <mn>0.65</mn> </mrow> </semantics></math> when <math display="inline"><semantics> <mi>λ</mi> </semantics></math> is lower than 1. The number of bands obtained by the dimensionality reduction is stable wrt <math display="inline"><semantics> <mi>λ</mi> </semantics></math>.</p>
Full article ">Figure 6
<p>(<b>a</b>) Ground truth labels. (<b>b</b>–<b>d</b>) Random seeds for the proposed procedure, chosen among the ground truth mask. From left to right: square seeds of dimension 3, 5, and 7 pixels. The squares are clipped in order to refer to the correct region.</p>
Full article ">Figure 7
<p>Result of the segmentation of <span class="html-italic">Pavia Center</span> using the ground truth labels as atlas. From left to right: image of the landscape in false colors, labels of the ground truth, final result. The latter panel shows that all the roofs are remarkably recognized, as the shadows they project on the ground. The vegetation is segmented with a very high level of precision. We reported the labels as reported in the database we used for these experiments: there are clearly some errors, since some classes (such as Shadows, Meadows and Bare Soils) refers to objects that are not the ones described by these labels.</p>
Full article ">Figure 8
<p>Segmentation results on the <span class="html-italic">KSC</span> dataset. From left to right: false color image, ground truth labels employed as seeds, and segmentation result. On the bottom of the images the legend associates the color to the labels.</p>
Full article ">Figure 9
<p>Overlay between the segmentation and original image (in grayscale) for the <span class="html-italic">KSC</span> dataset.</p>
Full article ">
33 pages, 3286 KiB  
Review
Off-The-Grid Variational Sparse Spike Recovery: Methods and Algorithms
by Bastien Laville, Laure Blanc-Féraud and Gilles Aubert
J. Imaging 2021, 7(12), 266; https://doi.org/10.3390/jimaging7120266 - 6 Dec 2021
Cited by 7 | Viewed by 3163
Abstract
Gridless sparse spike reconstruction is a rather new research field with significant results for the super-resolution problem, where we want to retrieve fine-scale details from a noisy and filtered acquisition. To tackle this problem, we are interested in optimisation under some prior, typically [...] Read more.
Gridless sparse spike reconstruction is a rather new research field with significant results for the super-resolution problem, where we want to retrieve fine-scale details from a noisy and filtered acquisition. To tackle this problem, we are interested in optimisation under some prior, typically the sparsity i.e., the source is composed of spikes. Following the seminal work on the generalised LASSO for measures called the Beurling-Lasso (BLASSO), we will give a review on the chief theoretical and numerical breakthrough of the off-the-grid inverse problem, as we illustrate its usefulness to the super-resolution problem in Single Molecule Localisation Microscopy (SMLM) through new reconstruction metrics and tests on synthetic and real SMLM data we performed for this review. Full article
(This article belongs to the Special Issue Inverse Problems and Imaging)
Show Figures

Figure 1

Figure 1
<p>(<b>a</b>) Discrete reconstruction, which can be seen as spikes with support constrained on a grid; (<b>b</b>) off-the-grid reconstruction, the spikes can move continuously on the line. The red line is the acquisition <span class="html-italic">y</span>, orange spikes are the source (the cause we want to retrieve), blue spikes are discrete reconstruction constrained on a grid and green can move freely since it is off-the-grid. Note that, when a source spike is between two grid points, two spikes will be recovered in the discrete reconstruction.</p>
Full article ">Figure 2
<p>(<b>a</b>) Certificates associated with acquisition <span class="html-italic">y</span> and noiseless <math display="inline"><semantics> <msub> <mi>y</mi> <mn>0</mn> </msub> </semantics></math>, result of three <math display="inline"><semantics> <mi>δ</mi> </semantics></math>-peaks (in black, plotted with 10 times their ground-truth amplitudes) through a Fourier measurement of cut-off frequency <math display="inline"><semantics> <mrow> <msub> <mi>f</mi> <mi mathvariant="normal">c</mi> </msub> <mo>=</mo> <mn>6</mn> </mrow> </semantics></math>; (<b>b</b>) localisation of the roots of the certificate associated with the dual <span class="html-italic">maximum</span>. All the roots (the three ground-truths and the spurious spike on the right) on the unit circle are interpreted as the support of the <math display="inline"><semantics> <mi>δ</mi> </semantics></math>-peaks.</p>
Full article ">Figure 3
<p>Reconstruction with the Interior Point Method for <math display="inline"><semantics> <mrow> <mi>λ</mi> <mo>=</mo> <mn>1</mn> </mrow> </semantics></math>. The algorithm detected a spurious spike near 0.05; otherwise, amplitudes and positions of the peaks are correctly estimated.</p>
Full article ">Figure 4
<p>Reconstruction by <span class="html-italic">Sliding Frank-Wolfe</span> for a 1D Fourier operator, with the same settings (<span class="html-italic">y</span>, noise realisations, <math display="inline"><semantics> <mrow> <mi>λ</mi> <mo>=</mo> <mn>1</mn> </mrow> </semantics></math>) as the former section. All ground-truth spikes are reconstructed, no spurious spike is detected.</p>
Full article ">Figure 5
<p>(<b>a</b>) First iterate <math display="inline"><semantics> <mrow> <mi>k</mi> <mo>=</mo> <mn>0</mn> </mrow> </semantics></math>; (<b>b</b>) mid-computation <math display="inline"><semantics> <mrow> <mi>k</mi> <mo>=</mo> <mn>1</mn> </mrow> </semantics></math>; (<b>c</b>) end of the computation <math display="inline"><semantics> <mrow> <mi>k</mi> <mo>=</mo> <mn>2</mn> </mrow> </semantics></math>, results for SFW reconstruction on the domain <math display="inline"><semantics> <mrow> <mi mathvariant="script">X</mi> <mo>=</mo> <msup> <mrow> <mo>[</mo> <mn>0</mn> <mo>,</mo> <mn>1</mn> <mo>]</mo> </mrow> <mn>2</mn> </msup> </mrow> </semantics></math> for the Gaussian kernel with spread-factor <math display="inline"><semantics> <mrow> <mi>σ</mi> <mo>=</mo> <mn>0.1</mn> </mrow> </semantics></math> and additive Gaussian noise of variance <math display="inline"><semantics> <mrow> <mn>0.1</mn> </mrow> </semantics></math>. All <math display="inline"><semantics> <mi>δ</mi> </semantics></math>-peaks are successfully recovered only thanks to the acquisition, <math display="inline"><semantics> <mrow> <mi>λ</mi> <mo>=</mo> <mrow> <mn>3</mn> <mo>×</mo> <msup> <mn>10</mn> <mrow> <mo>−</mo> <mn>2</mn> </mrow> </msup> </mrow> </mrow> </semantics></math>.</p>
Full article ">Figure 6
<p>Digest of the important quantities mentioned in [<a href="#B3-jimaging-07-00266" class="html-bibr">3</a>,<a href="#B23-jimaging-07-00266" class="html-bibr">23</a>]: <span style="color:#a35f45">red refers to <math display="inline"><semantics> <mrow> <msup> <mi mathvariant="script">M</mi> <mo>+</mo> </msup> <mfenced open="(" close=")"> <mi mathvariant="script">X</mi> </mfenced> </mrow> </semantics></math> quantities</span>, <span style="color:#6f9e41">green to <math display="inline"><semantics> <mrow> <msup> <mi mathvariant="sans-serif">Ω</mi> <mi>N</mi> </msup> <mover> <mo>=</mo> <mrow> <mi>def</mi> <mo>.</mo> </mrow> </mover> <mrow> <mo>(</mo> <msup> <mi>R</mi> <mo>+</mo> </msup> <mo>×</mo> <mi mathvariant="script">X</mi> <mo>)</mo> </mrow> </mrow> </semantics></math></span> and <span style="color:#332f9e">blue to the Wasserstein space <math display="inline"><semantics> <mrow> <msub> <mi mathvariant="script">P</mi> <mn>2</mn> </msub> <mrow> <mo>(</mo> <mi mathvariant="sans-serif">Ω</mi> <mo>)</mo> </mrow> </mrow> </semantics></math> and theoretical results</span>. Dashed lines correspond to the theoretical section, and continuous lines indicate the numerical part.</p>
Full article ">Figure 7
<p>Reconstruction by <span class="html-italic">Conic Particle Gradient Descent</span> for a 1D Fourier operator in a noiseless setting, with the same ground-truth spikes as the former section. Implementation is an adaptation of [<a href="#B23-jimaging-07-00266" class="html-bibr">23</a>], <math display="inline"><semantics> <mrow> <mi>α</mi> <mo>=</mo> <mi>β</mi> <mo>=</mo> <mrow> <mn>1</mn> <mo>×</mo> <msup> <mn>10</mn> <mrow> <mo>−</mo> <mn>3</mn> </mrow> </msup> </mrow> </mrow> </semantics></math> and <math display="inline"><semantics> <mrow> <mi>λ</mi> <mo>=</mo> <mn>1</mn> </mrow> </semantics></math> for 1000 iterations.</p>
Full article ">Figure 8
<p>(<b>a</b>) Initialisation <math display="inline"><semantics> <mrow> <mi>k</mi> <mo>=</mo> <mn>0</mn> </mrow> </semantics></math>; (<b>b</b>) mid-computation <math display="inline"><semantics> <mrow> <mi>k</mi> <mo>=</mo> <mn>150</mn> </mrow> </semantics></math>; (<b>c</b>) end of the computation <math display="inline"><semantics> <mrow> <mi>k</mi> <mo>=</mo> <mn>1000</mn> </mrow> </semantics></math>. <span class="html-italic">Conic Particle Gradient Descent</span> applied for 2D Gaussian deconvolution, the red dots are the particle measure <math display="inline"><semantics> <msup> <mi>ν</mi> <mrow> <mo>(</mo> <mi>k</mi> <mo>)</mo> </mrow> </msup> </semantics></math> (size of dot proportional with amplitude), the three white dots are the source measure, the image in the background is the noiseless acquisition <math display="inline"><semantics> <msub> <mi>y</mi> <mn>0</mn> </msub> </semantics></math> and the black lines are the paths of the particles <math display="inline"><semantics> <msup> <mi>ν</mi> <mrow> <mo>(</mo> <mi>k</mi> <mo>)</mo> </mrow> </msup> </semantics></math>—all the paths constitute the gradient flow <math display="inline"><semantics> <msub> <mrow> <mo>(</mo> <msub> <mi>ν</mi> <mi>t</mi> </msub> <mo>)</mo> </mrow> <mrow> <mi>t</mi> <mo>≥</mo> <mn>0</mn> </mrow> </msub> </semantics></math>. Implementation is an adaptation of [<a href="#B23-jimaging-07-00266" class="html-bibr">23</a>], <math display="inline"><semantics> <mrow> <mi>α</mi> <mo>=</mo> <mi>β</mi> <mo>=</mo> <mrow> <mn>1</mn> <mo>×</mo> <msup> <mn>10</mn> <mrow> <mo>−</mo> <mn>2</mn> </mrow> </msup> </mrow> </mrow> </semantics></math> and <math display="inline"><semantics> <mrow> <mi>λ</mi> <mo>=</mo> <mn>1</mn> </mrow> </semantics></math>.</p>
Full article ">Figure 9
<p>(<b>a</b>) Ground-truth tubulins, two excerpts of the stack in the square below: convolution + all noise described before; (<b>b</b>) reconstructed measure by <span class="html-italic">Sliding Frank-Wolfe</span> visualised through Gaussian kernel with a smaller <math display="inline"><semantics> <mi>σ</mi> </semantics></math> (see text).</p>
Full article ">Figure 10
<p>(<b>top left</b>) Excerpt of the stack; (<b>top right</b>) mean of the stack; (<b>bottom left</b>) reconstruction by off-the-grid method; (<b>bottom right</b>) Deep-STORM.</p>
Full article ">
20 pages, 2675 KiB  
Review
Using Inertial Sensors to Determine Head Motion—A Review
by Severin Ionut-Cristian and Dobrea Dan-Marius
J. Imaging 2021, 7(12), 265; https://doi.org/10.3390/jimaging7120265 - 6 Dec 2021
Cited by 22 | Viewed by 5421
Abstract
Human activity recognition and classification are some of the most interesting research fields, especially due to the rising popularity of wearable devices, such as mobile phones and smartwatches, which are present in our daily lives. Determining head motion and activities through wearable devices [...] Read more.
Human activity recognition and classification are some of the most interesting research fields, especially due to the rising popularity of wearable devices, such as mobile phones and smartwatches, which are present in our daily lives. Determining head motion and activities through wearable devices has applications in different domains, such as medicine, entertainment, health monitoring, and sports training. In addition, understanding head motion is important for modern-day topics, such as metaverse systems, virtual reality, and touchless systems. The wearability and usability of head motion systems are more technologically advanced than those which use information from a sensor connected to other parts of the human body. The current paper presents an overview of the technical literature from the last decade on state-of-the-art head motion monitoring systems based on inertial sensors. This study provides an overview of the existing solutions used to monitor head motion using inertial sensors. The focus of this study was on determining the acquisition methods, prototype structures, preprocessing steps, computational methods, and techniques used to validate these systems. From a preliminary inspection of the technical literature, we observed that this was the first work which looks specifically at head motion systems based on inertial sensors and their techniques. The research was conducted using four internet databases—IEEE Xplore, Elsevier, MDPI, and Springer. According to this survey, most of the studies focused on analyzing general human activity, and less on a specific activity. In addition, this paper provides a thorough overview of the last decade of approaches and machine learning algorithms used to monitor head motion using inertial sensors. For each method, concept, and final solution, this study provides a comprehensive number of references which help prove the advantages and disadvantages of the inertial sensors used to read head motion. The results of this study help to contextualize emerging inertial sensor technology in relation to broader goals to help people suffering from partial or total paralysis of the body. Full article
(This article belongs to the Special Issue 3D Human Understanding)
Show Figures

Figure 1

Figure 1
<p>Taxonomy of head motion recognition applications.</p>
Full article ">Figure 2
<p>Paper selection steps.</p>
Full article ">Figure 3
<p>Distribution of studies by technical area.</p>
Full article ">Figure 4
<p>Publication count over the years.</p>
Full article ">Figure 5
<p>Distribution of inertial sensors on the head.</p>
Full article ">Figure 6
<p>Example of a CNN computational model for inertial signal classification.</p>
Full article ">Figure 7
<p>Distribution of computational models used in head motion recognition.</p>
Full article ">Figure 8
<p>Distribution of studies related to online vs. offline analyses.</p>
Full article ">Figure 9
<p>Workflow for implementing head motion recognition solutions based on inertial sensors.</p>
Full article ">
16 pages, 1297 KiB  
Article
Skeleton-Based Attention Mask for Pedestrian Attribute Recognition Network
by Sorn Sooksatra and Sitapa Rujikietgumjorn
J. Imaging 2021, 7(12), 264; https://doi.org/10.3390/jimaging7120264 - 4 Dec 2021
Cited by 3 | Viewed by 2807
Abstract
This paper presents an extended model for a pedestrian attribute recognition network utilizing skeleton data as a soft attention model to extract a local feature corresponding to a specific attribute. This technique helped keep valuable information surrounding the target area and handle the [...] Read more.
This paper presents an extended model for a pedestrian attribute recognition network utilizing skeleton data as a soft attention model to extract a local feature corresponding to a specific attribute. This technique helped keep valuable information surrounding the target area and handle the variation of human posture. The attention masks were designed to focus on the partial and the whole-body regions. This research utilized an augmented layer for data augmentation inside the network to reduce over-fitting errors. Our network was evaluated in two datasets (RAP and PETA) with various backbone networks (ResNet-50, Inception V3, and Inception-ResNet V2). The experimental result shows that our network improves overall classification performance with a mean accuracy of about 2–3% in the same backbone network, especially local attributes and various human postures. Full article
(This article belongs to the Special Issue Advances in Human Action Recognition Using Deep Learning)
Show Figures

Figure 1

Figure 1
<p>An example of pedestrian image from RAP dataset [<a href="#B9-jimaging-07-00264" class="html-bibr">9</a>] for PAR network consisting of input image with hard attention mask (Mask-RCNN [<a href="#B7-jimaging-07-00264" class="html-bibr">7</a>]) and with our soft attention mask.</p>
Full article ">Figure 2
<p>The list of skeleton names and indices from OpenPifPaf [<a href="#B32-jimaging-07-00264" class="html-bibr">32</a>] with their locations.</p>
Full article ">Figure 3
<p>The pedestrian image with soft attention masks corresponding to attention masks in different classes where (<b>a</b>) all joints can be localized, (<b>b</b>) some joints can be detected from the partial occlusion, and (<b>c</b>) there is full occlusion.</p>
Full article ">Figure 4
<p>The proposed network architecture with a human-part attention module.</p>
Full article ">Figure 5
<p>The illustration of human-part attention module for (<b>a</b>) separated and (<b>b</b>) single masks.</p>
Full article ">Figure 6
<p>Mean accuracy from training data by using stable binary cross-entropy and focal losses with various values of <math display="inline"><semantics> <mi>γ</mi> </semantics></math>.</p>
Full article ">Figure 7
<p>The illustration of an augmented layer in the proposed network architecture.</p>
Full article ">Figure 8
<p>Learning curve between stable binary cross-entropy loss and epoch while training the network from RAPv2.</p>
Full article ">Figure 9
<p>The examples of the predicted attribute on pedestrian images with normal circumstance with various race, where the wrong attributes are presented as red characters.</p>
Full article ">Figure 10
<p>The examples of predicted attribute from pedestrian images corresponding to full occlusion, partial occlusion, and irregular posture in the first, second, and third rows, respectively, where the wrong attributes are presented as red characters.</p>
Full article ">Figure 10 Cont.
<p>The examples of predicted attribute from pedestrian images corresponding to full occlusion, partial occlusion, and irregular posture in the first, second, and third rows, respectively, where the wrong attributes are presented as red characters.</p>
Full article ">
16 pages, 5518 KiB  
Article
Micro- and Nano-Scales Three-Dimensional Characterisation of Softwood
by Alessandra Patera, Anne Bonnin and Rajmund Mokso
J. Imaging 2021, 7(12), 263; https://doi.org/10.3390/jimaging7120263 - 3 Dec 2021
Cited by 7 | Viewed by 2818
Abstract
Understanding the mechanical response of cellular biological materials to environmental stimuli is of fundamental importance from an engineering perspective in composites. To provide a deep understanding of their behaviour, an exhaustive analytical and experimental protocol is required. Attention is focused on softwood but [...] Read more.
Understanding the mechanical response of cellular biological materials to environmental stimuli is of fundamental importance from an engineering perspective in composites. To provide a deep understanding of their behaviour, an exhaustive analytical and experimental protocol is required. Attention is focused on softwood but the approach can be applied to a range of cellular materials. This work presents a new non-invasive multi-scale approach for the investigation of the hygro-mechanical behaviour of softwood. At the TOMCAT beamline of the Paul Scherrer Institute, in Switzerland, the swelling behaviour of softwood was probed at the cellular and sub-cellular scales by means of 3D high-resolution phase-contrast X-ray imaging. At the cellular scale, new findings in the anisotropic and reversible swelling behaviour of softwood and in the origin of swelling hysteresis of porous materials are explained from a mechanical perspective. However, the mechanical and moisture properties of wood highly depend on sub-cellular features of the wood cell wall, such as bordered pits, yielding local deformations during a full hygroscopic loading protocol. Full article
(This article belongs to the Special Issue X-ray Digital Radiography and Computed Tomography)
Show Figures

Figure 1

Figure 1
<p>Schematic representation of the hierarchical structure of wood: from macro to micro scales, showing: (<b>a</b>) the growth ring, (<b>b</b>) the latewood and the earlywood tissues, (<b>c</b>) singularities in the cell wall, such as pits, (<b>d</b>) the cell wall layers and (<b>e</b>) the organisation of the chemical components.</p>
Full article ">Figure 2
<p>(<b>a</b>) Components of the sample holder, (<b>b</b>) finalized with polyimide foil and with access window for inserting and gluing the wood fiber. (<b>c</b>) A view from the top of the sample holder inserted in the environmental chamber.</p>
Full article ">Figure 3
<p>(<b>a</b>) Photo of the phase-contrast nanotomography setup at TOMCAT, showing (i) the end of the X-ray fly-tube, (ii) the condenser, (iii) the environmental chamber housing the wood sample, (iv) the Fresnel zone plate and (v) the Zernike phase dots. The camera is located at 9 m of distance from the (iv) and is not shown in the picture. (<b>b</b>) Schematic representation of the nanotomographic microscopy, with bordered pits imaged on the detector plane.</p>
Full article ">Figure 4
<p>2D mosaic nanoimaging of the spruce wood. In (<b>a</b>) and (<b>b</b>) are depicted the 2D mosaic projections images, at 0° and 90° respectively.</p>
Full article ">Figure 5
<p>Swelling coefficients of latewood and earlywood for different porosity. The plot collects the results discussed in <a href="#sec3dot2-jimaging-07-00263" class="html-sec">Section 3.2</a> and in previous works taken from different sources on softwood species [<a href="#B23-jimaging-07-00263" class="html-bibr">23</a>,<a href="#B24-jimaging-07-00263" class="html-bibr">24</a>,<a href="#B25-jimaging-07-00263" class="html-bibr">25</a>].</p>
Full article ">Figure 6
<p>Shape-memory effect on the cell undergoing large deformations in a ROI of EWT. The ROI was selected out of 2D reconstructed images and binarized.</p>
Full article ">Figure 7
<p>(<b>a</b>) LWR sample, dry (left) vs. wet (right), with a zoom on the ROI (in blue) where shape-memory is observed, (<b>b</b>) 3D map of the equivalent non-rigid strains, with a slice cut out in the middle of the volume and (<b>c</b>) the components of the non-rigid strains in tangential (xx) and radial (yy) directions.</p>
Full article ">Figure 8
<p>Cross-sectional slice of Sample A at different RH states. In the figure, the three pits analysed in the paper are indicated.</p>
Full article ">Figure 9
<p>Re-slice of the sample A tomographic dataset, with a ROI cutting in the position of pit 1. Bordered pit 1 in section transverse to the pit membrane, showing the k parameters reported in <a href="#jimaging-07-00263-t006" class="html-table">Table 6</a>.</p>
Full article ">Figure 10
<p>Bordered pit in sample B in section transverse to the pit membrane. W is the pit membrane thickness, D the major axis of the ellipse, 2b the minor axis and 2k the pit aperture. Torus was not probed in this case.</p>
Full article ">Figure 11
<p>(<b>a</b>) Cross-sectional view of pit 1 at different RH values in adsorption (10–50–75–90%) and in desorption (75–50–10%) used for the analysis of (<b>b</b>) the tilting angle in adsorption (dashed line) and in desorption (continued line).</p>
Full article ">
12 pages, 940 KiB  
Article
Optical to Planar X-ray Mouse Image Mapping in Preclinical Nuclear Medicine Using Conditional Adversarial Networks
by Eleftherios Fysikopoulos, Maritina Rouchota, Vasilis Eleftheriadis, Christina-Anna Gatsiou, Irinaios Pilatis, Sophia Sarpaki, George Loudos, Spiros Kostopoulos and Dimitrios Glotsos
J. Imaging 2021, 7(12), 262; https://doi.org/10.3390/jimaging7120262 - 3 Dec 2021
Cited by 1 | Viewed by 3536
Abstract
In the current work, a pix2pix conditional generative adversarial network has been evaluated as a potential solution for generating adequately accurate synthesized morphological X-ray images by translating standard photographic images of mice. Such an approach will benefit 2D functional molecular imaging techniques, such [...] Read more.
In the current work, a pix2pix conditional generative adversarial network has been evaluated as a potential solution for generating adequately accurate synthesized morphological X-ray images by translating standard photographic images of mice. Such an approach will benefit 2D functional molecular imaging techniques, such as planar radioisotope and/or fluorescence/bioluminescence imaging, by providing high-resolution information for anatomical mapping, but not for diagnosis, using conventional photographic sensors. Planar functional imaging offers an efficient alternative to biodistribution ex vivo studies and/or 3D high-end molecular imaging systems since it can be effectively used to track new tracers and study the accumulation from zero point in time post-injection. The superimposition of functional information with an artificially produced X-ray image may enhance overall image information in such systems without added complexity and cost. The network has been trained in 700 input (photography)/ground truth (X-ray) paired mouse images and evaluated using a test dataset composed of 80 photographic images and 80 ground truth X-ray images. Performance metrics such as peak signal-to-noise ratio (PSNR), structural similarity index measure (SSIM) and Fréchet inception distance (FID) were used to quantitatively evaluate the proposed approach in the acquired dataset. Full article
(This article belongs to the Special Issue SPECT and PET Imaging of Small Animals)
Show Figures

Figure 1

Figure 1
<p>Indicative optical image acquired using a conventional photographic sensor located in BIOEMTECH eye’s series radioisotope screening tools (<b>a</b>); Indicative X-ray image acquired in a prototype PET/SPECT X-ray system (<b>c</b>) used as ground truth; Aligned pair (<b>b</b>).</p>
Full article ">Figure 2
<p>Aligned image pairs used for training. Photographic input image (<b>left</b>); Corresponding X-ray scan used as ground truth (<b>right</b>).</p>
Full article ">Figure 3
<p>The pix2pix Generator’s training layout. The generator creates output image (y) from input image (x) and random noise vector (z) and improves its performance by receiving feedback from the discriminator, as well as regarding the degree of fakeness of the synthetic image (y) compared to the ground truth (r).</p>
Full article ">Figure 4
<p>The cGAN discriminator’s training layout. The discriminator compares the input(x)/ground truth(r) pair of images and the input(x)/output(y) pair of images and outputs its guess about how realistic they look. The weights vector of the discriminator is then updated based on the classification error of the input/output pair (D fake Loss) and the input/target pair (D real Loss).</p>
Full article ">Figure 5
<p>Training loss curves of the cross entropy, M.S.E. and Wasserstein distance loss function models.</p>
Full article ">Figure 6
<p>Indicative “fake” X-ray images from the pix2pix trained network using different loss functions: Cross Entropy (3rd column); MSE (4th column); Wasserstein distance (5th column). The input photographic images and the corresponding ground truth images are presented in the first two columns.</p>
Full article ">Figure 7
<p><math display="inline"><semantics> <mrow> <msup> <mrow/> <mrow> <mn>99</mn> <mi>m</mi> </mrow> </msup> <mi>T</mi> <mi>c</mi> </mrow> </semantics></math>-MDP-labelled nuclear image of a healthy mouse fused with the optical image provided in the <math display="inline"><semantics> <mi>γ</mi> </semantics></math>-eye scintigraphic system (<b>left</b>) and the X-ray produced from the pix2pix trained network (<b>right</b>).</p>
Full article ">Figure 8
<p><math display="inline"><semantics> <mrow> <msup> <mrow/> <mn>18</mn> </msup> <mi>F</mi> </mrow> </semantics></math>-FDG nuclear image of a healthy mouse fused with the optical image provided in the <math display="inline"><semantics> <mi>β</mi> </semantics></math>-eye planar coincidence imaging system (<b>left</b>) and the X-ray produced from the pix2pix trained network (<b>right</b>).</p>
Full article ">
10 pages, 1766 KiB  
Article
Copper and Trace Elements in Gallbladder form Patients with Wilson’s Disease Imaged and Determined by Synchrotron X-ray Fluorescence
by Wolf Osterode, Gerald Falkenberg and Fritz Wrba
J. Imaging 2021, 7(12), 261; https://doi.org/10.3390/jimaging7120261 - 3 Dec 2021
Cited by 4 | Viewed by 2943
Abstract
Investigations about suspected tissue alterations and the role of gallbladder in Wilson’s disease (WD)—an inherited genetic disease with impaired copper metabolism—are rare. Therefore, tissue from patients with genetically characterised WD was investigated by microscopic synchrotron X-ray fluorescence (µSRXRF). For two-dimensional imaging and quantification [...] Read more.
Investigations about suspected tissue alterations and the role of gallbladder in Wilson’s disease (WD)—an inherited genetic disease with impaired copper metabolism—are rare. Therefore, tissue from patients with genetically characterised WD was investigated by microscopic synchrotron X-ray fluorescence (µSRXRF). For two-dimensional imaging and quantification of elements, X-ray spectra were peak-fitted, and the net peak intensities were normalised to the intensity of the incoming monochromatic beam intensity. Concentrations were calculated by fundamental parameter-based program quant and external standardisation. Copper (Cu), zinc (Zn) and iron (Fe) along with sulphur (S) and phosphorus (P) mappings could be demonstrated in a near histological resolution. All these elements were increased compared to gallbladder tissue from controls. Cu and Zn and Fe in WD-GB were mostly found to be enhanced in the epithelium. We documented a significant linear relationship with Cu, Zn and sulphur. Concentrations of Cu/Zn were roughly 1:1 while S/Cu was about 100:1, depending on the selected areas for investigation. The significant linear relationship with Cu, Zn and sulphur let us assume that metallothioneins, which are sulphur-rich proteins, are increased too. Our data let us suggest that the WD gallbladder is the first in the gastrointestinal tract to reabsorb metals to prevent oxidative damage caused by metal toxicity. Full article
(This article belongs to the Special Issue X-ray Digital Radiography and Computed Tomography)
Show Figures

Figure 1

Figure 1
<p>Accumulated spectra after summation over all pixels with peaks of the investigated elements. Red colour: Control–GB. Black colour: Wilson’s Disease-GB.</p>
Full article ">Figure 2
<p>(<b>a</b>–<b>f</b>) Element distribution in a control gallbladder. (<b>a</b>) Microscope recording of the investigated gallbladder tissue with epithelium (columnar cells), beneath the epithelium the laminar propria (connective tissue) and smooth muscular fibres for contraction. Adventitia (Adv): perimuscular connective tissue that is very dense and connects with the liver. (<b>b</b>–<b>f</b>) copper (Cu), iron (Fe), zinc (Zn), phosphorous (P) and sulphur (S) distribution. The histological section is about 30-µm apart from the investigated one for SRXRF.</p>
Full article ">Figure 3
<p>Correlations between elements in ppm for C-GB (<a href="#jimaging-07-00261-f002" class="html-fig">Figure 2</a>). (<b>a</b>) Cu:Zn = 1.28; r<sub>s</sub> = 0.92. (<b>b</b>) S:Cu = 70; r<sub>s</sub> = 0.92. (<b>c</b>) Correlations between S:Zn = 79, r<sub>s</sub> = 0.88. For all <span class="html-italic">p</span> &lt; 0.001.</p>
Full article ">Figure 4
<p>Distribution of elements in a WD gallbladder. (<b>a</b>) copper (Cu), (<b>b</b>) zinc (Zn), (<b>c</b>) iron (Fe), (<b>d</b>) sulphur (S) and (<b>e</b>) phosphorus (P).</p>
Full article ">Figure 5
<p>Correlations between elements for a WD-GB. (<b>a</b>) Cu/Zn = 1.28; r<sub>s</sub> = 0.94. (<b>b</b>) S/Cu = 119; r<sub>s</sub> = 0.91. (<b>c</b>) Correlations between S/Zn = 81, r<sub>s</sub> = 0.93. For all <span class="html-italic">p</span> &lt; 0.001.</p>
Full article ">
13 pages, 16300 KiB  
Article
HTR for Greek Historical Handwritten Documents
by Lazaros Tsochatzidis, Symeon Symeonidis, Alexandros Papazoglou and Ioannis Pratikakis
J. Imaging 2021, 7(12), 260; https://doi.org/10.3390/jimaging7120260 - 2 Dec 2021
Cited by 9 | Viewed by 3553
Abstract
Offline handwritten text recognition (HTR) for historical documents aims for effective transcription by addressing challenges that originate from the low quality of manuscripts under study as well as from several particularities which are related to the historical period of writing. In this paper, [...] Read more.
Offline handwritten text recognition (HTR) for historical documents aims for effective transcription by addressing challenges that originate from the low quality of manuscripts under study as well as from several particularities which are related to the historical period of writing. In this paper, the challenge in HTR is related to a focused goal of the transcription of Greek historical manuscripts that contain several particularities. To this end, in this paper, a convolutional recurrent neural network architecture is proposed that comprises octave convolution and recurrent units which use effective gated mechanisms. The proposed architecture has been evaluated on three newly created collections from Greek historical handwritten documents that will be made publicly available for research purposes as well as on standard datasets like IAM and RIMES. For evaluation we perform a concise study which shows that compared to state of the art architectures, the proposed one deals effectively with the challenging Greek historical manuscripts. Full article
Show Figures

Figure 1

Figure 1
<p>Schematic diagram of the proposed architecture, consisting of (<b>a</b>) the CNN stage and (<b>b</b>) the recurrent stage.</p>
Full article ">Figure 2
<p>An example text line image (<b>a</b>) before and (<b>b</b>) after the preprocessing stage.</p>
Full article ">Figure 3
<p>Feature maps produced by each layer of the Octave-CNN model.</p>
Full article ">Figure 4
<p>An illustration of the gated recurrent unit (GRU) [<a href="#B8-jimaging-07-00260" class="html-bibr">8</a>].</p>
Full article ">Figure 5
<p>Example document image from the collection <math display="inline"><semantics> <mrow> <mi>χ</mi> <mi>ϕ</mi> </mrow> </semantics></math>53.</p>
Full article ">Figure 6
<p>Example document image from the collection <math display="inline"><semantics> <mrow> <mi>χ</mi> <mi>ϕ</mi> </mrow> </semantics></math>79.</p>
Full article ">Figure 7
<p>Example document image from the collection <math display="inline"><semantics> <mrow> <mi>χ</mi> <mi>ϕ</mi> </mrow> </semantics></math>114.</p>
Full article ">Figure 8
<p>Floating characters appearing at word endings. The floating portion of the word is represented by a rectangle, while the rest of the word is underlined.</p>
Full article ">Figure 9
<p>‘Minuscule’ writing example. Key locations in the text line that correspond to this particularity are underlined.</p>
Full article ">Figure 10
<p>Polytonic orthography example.</p>
Full article ">Figure 11
<p>An example of a correctly predicted text line image along with the corresponding (<b>a</b>) groud-truth and (<b>b</b>) predicted texts.</p>
Full article ">Figure 12
<p>An example of a problematic text line image along with the corresponding (<b>a</b>) groud-truth and (<b>b</b>) predicted texts. The errors concern diacritics (circle), spacing (red line) and abbreviations (square).</p>
Full article ">
10 pages, 2080 KiB  
Article
Infrared Clinical Enamel Crack Detector Based on Silicon CCD and Its Application: A High-Quality and Low-Cost Option
by Yuchen Zheng, Min-Hee Oh, Woo-Sub Song, Ki-Hyun Kim, In-Hee Shin, Min-Seok Kim and Jin-Hyoung Cho
J. Imaging 2021, 7(12), 259; https://doi.org/10.3390/jimaging7120259 - 2 Dec 2021
Cited by 5 | Viewed by 3207
Abstract
Enamel cracks generated in the anterior teeth not only affect the function but also the aesthetics of the teeth. Chair-side tooth enamel crack detection is essential for clinicians to formulate treatment plans and prevent related dental disease. This study aimed to develop a [...] Read more.
Enamel cracks generated in the anterior teeth not only affect the function but also the aesthetics of the teeth. Chair-side tooth enamel crack detection is essential for clinicians to formulate treatment plans and prevent related dental disease. This study aimed to develop a dental imaging system using a near-IR light source to detect enamel cracks and to investigate the relationship between anterior enamel cracks and age in vivo. A total of 68 subjects were divided into three groups according to their age: young, middle, and elderly. Near-infrared radiation of 850 nm was used to identify enamel cracks in anterior teeth. The results of the quantitative examination showed that the number of enamel cracks on the teeth increased considerably with age. For the qualitative examination, the results indicated that there was no significant relationship between the severity of the enamel cracks and age. So, it can be concluded that the prevalence of anterior cracked tooth increased significantly with age in the young and middle age. The length of the anterior enamel cracks tended to increase with age too; however, this result was not significant. The silicon charge-coupled device (CCD) with a wavelength of 850 nm has a good performance in the detection of enamel cracks and has very good clinical practicability. Full article
(This article belongs to the Special Issue New Frontiers of Advanced Imaging in Dentistry)
Show Figures

Figure 1

Figure 1
<p>Device and method of detection. (<bold>a</bold>) Rubber tray, which could emit 850 nm Near-IR in the mouth to highlight the cracks. Red arrows indicate the LED illuminant, whose brightness can be adjusted from level 1 to 5, with 5 being the brightest. (<bold>b</bold>) Near-infrared dental and periodontal imaging device with a screen on the examiner’s side for image capturing and a CCD camera inside for image acquiring. Red arrow indicates a control lever, which is connected to the camera to adjust its direction during the detection process. (<bold>c</bold>) Schematic diagram of the patient position during detection.</p>
Full article ">Figure 2
<p>The three most representative types of teeth in the three age groups. The red arrow points out the enamel cracks. The high contrast of the enamel cracks in the near-infrared image can be clearly seen.</p>
Full article ">Figure 3
<p>Simulated images, intraoral photos, and 850 nm Near-IR images for indicating the differences among the classifications. Red arrows indicate the cracks. (<bold>a</bold>) A simulated Class 1 central incisor crack. (<bold>b</bold>) An intraoral photo of a Class 1 crack on the maxillary central incisor under natural light. (<bold>c</bold>) The image of the tooth in <xref ref-type="fig" rid="jimaging-07-00259-f003">Figure 3</xref>b under Near-IR. (<bold>d</bold>) A simulated Class 2 central incisor crack. (<bold>e</bold>) The intraoral photo of a Class 2 crack on the maxillary central incisor under natural light. (<bold>f</bold>) The image of the tooth in <xref ref-type="fig" rid="jimaging-07-00259-f003">Figure 3</xref>e under Near-IR. (<bold>g</bold>) A simulated Class 3 central incisor crack. (<bold>h</bold>) An intraoral photo of a Class 3 crack on the maxillary central incisor under natural light. (<bold>i</bold>) The image of the tooth in <xref ref-type="fig" rid="jimaging-07-00259-f003">Figure 3</xref>h under Near-IR.</p>
Full article ">Figure 4
<p>The distribution of different tooth types and tooth classifications among the three groups. (<bold>a</bold>) Tooth type distribution. The value of the ordinate indicates the number of cracked teeth. (<bold>b</bold>) Enamel crack classification distribution. The value of the ordinate indicates the number of different classes of cracks.</p>
Full article ">
12 pages, 2722 KiB  
Article
Abdominal Computed Tomography Imaging Findings in Hospitalized COVID-19 Patients: A Year-Long Experience and Associations Revealed by Explainable Artificial Intelligence
by Alice Scarabelli, Massimo Zilocchi, Elena Casiraghi, Pierangelo Fasani, Guido Giovanni Plensich, Andrea Alessandro Esposito, Elvira Stellato, Alessandro Petrini, Justin Reese, Peter Robinson, Giorgio Valentini and Gianpaolo Carrafiello
J. Imaging 2021, 7(12), 258; https://doi.org/10.3390/jimaging7120258 - 1 Dec 2021
Cited by 3 | Viewed by 3171
Abstract
The aim of this retrospective study is to assess any association between abdominal CT findings and the radiological stage of COVID-19 pneumonia, pulmonary embolism and patient outcomes. We included 158 adult hospitalized COVID-19 patients between 1 March 2020 and 1 March 2021 who [...] Read more.
The aim of this retrospective study is to assess any association between abdominal CT findings and the radiological stage of COVID-19 pneumonia, pulmonary embolism and patient outcomes. We included 158 adult hospitalized COVID-19 patients between 1 March 2020 and 1 March 2021 who underwent 206 abdominal CTs. Two radiologists reviewed all CT images. Pathological findings were classified as acute or not. A subset of patients with inflammatory pathology in ACE2 organs (bowel, biliary tract, pancreas, urinary system) was identified. The radiological stage of COVID pneumonia, pulmonary embolism, overall days of hospitalization, ICU admission and outcome were registered. Univariate statistical analysis coupled with explainable artificial intelligence (AI) techniques were used to discover associations between variables. The most frequent acute findings were bowel abnormalities (n = 58), abdominal fluid (n = 42), hematomas (n = 28) and acute urologic conditions (n = 8). According to univariate statistical analysis, pneumonia stage > 2 was significantly associated with increased frequency of hematomas, active bleeding and fluid-filled colon. The presence of at least one hepatobiliary finding was associated with all the COVID-19 stages > 0. Free abdominal fluid, acute pathologies in ACE2 organs and fluid-filled colon were associated with ICU admission; free fluid also presented poor patient outcomes. Hematomas and active bleeding with at least a progressive stage of COVID pneumonia. The explainable AI techniques find no strong relationship between variables. Full article
Show Figures

Figure 1

Figure 1
<p><b>Left</b>: Contrast Enhanced Computed Tomography—CE CT (and CTs) are 94.7% (5.3%) of all the acquired CTs. <b>Right</b>: study indications for each imaging modality show that anemization is the study indication for which most CE CTs were acquired, while CTs were mostly used for follow-ups. Sixteen patients were found to have pulmonary embolism at chest CT performed before or during abdominal examinations. Embolism was not associated with any abdominal finding (<span class="html-italic">p</span> &gt; 0.05).</p>
Full article ">Figure 2
<p>Number of CT abdominal findings, their percentage with respect to all the patients, and the p-value with respect to the patient disease stage, hospitalization history and outcome. Significant values (<span class="html-italic">p</span> &lt; 0.05) are highlighted with light red background.</p>
Full article ">Figure 3
<p>For each variable to be predicted, each sub-table reports the <span class="html-italic">p</span>-values obtained by univariate statistical analysis (red color highlights significant variables—<span class="html-italic">p</span>-value &lt; 0.05), by the individual predictor variable performance, where green cells highlight good performance (&gt;0.7), the performance of the best classifier model (RF or DT) and the variable importance in prediction.</p>
Full article ">
12 pages, 6238 KiB  
Article
Monocular 3D Body Shape Reconstruction under Clothing
by Claudio Ferrari, Leonardo Casini, Stefano Berretti and Alberto Del Bimbo
J. Imaging 2021, 7(12), 257; https://doi.org/10.3390/jimaging7120257 - 30 Nov 2021
Cited by 5 | Viewed by 4205
Abstract
Estimating the 3D shape of objects from monocular images is a well-established and challenging task in the computer vision field. Further challenges arise when highly deformable objects, such as human faces or bodies, are considered. In this work, we address the problem of [...] Read more.
Estimating the 3D shape of objects from monocular images is a well-established and challenging task in the computer vision field. Further challenges arise when highly deformable objects, such as human faces or bodies, are considered. In this work, we address the problem of estimating the 3D shape of a human body from single images. In particular, we provide a solution to the problem of estimating the shape of the body when the subject is wearing clothes. This is a highly challenging scenario as loose clothes might hide the underlying body shape to a large extent. To this aim, we make use of a parametric 3D body model, the SMPL, whose parameters describe the body pose and shape of the body. Our main intuition is that the shape parameters associated with an individual should not change whether the subject is wearing clothes or not. To improve the shape estimation under clothing, we train a deep convolutional network to regress the shape parameters from a single image of a person. To increase the robustness to clothing, we build our training dataset by associating the shape parameters of a “minimally clothed” person to other samples of the same person wearing looser clothes. Experimental validation shows that our approach can more accurately estimate body shape parameters with respect to state-of-the-art approaches, even in the case of loose clothes. Full article
(This article belongs to the Special Issue 3D Human Understanding)
Show Figures

Figure 1

Figure 1
<p>Example of the first 3 principal components of shape variations for the SMPL model.</p>
Full article ">Figure 2
<p><b>Phase 1</b>: the proposed model takes as input an image of a person and regresses the SMPL shape parameters <math display="inline"><semantics> <mi>β</mi> </semantics></math>. <b>Phase 2</b>: the trained model is used to extract <math display="inline"><semantics> <mi>β</mi> </semantics></math> parameters from the minimally clothed images. The parameters are assigned to each image of the same individual and the network if fine-tuned.</p>
Full article ">Figure 3
<p>Samples of a same subject in the People Snapshot dataset. The “minimally clothed” version (with tight clothes) is shown on the <b>left</b>, and is used to estimate the shape parameters. These parameters are then assigned to the other sequences with looser clothes (<b>right</b>).</p>
Full article ">Figure 4
<p>Qualitative examples of body reconstruction from an image of the People Snapshot (<b>left</b>) and an “in the wild” image (<b>right</b>). The two reconstructions are obtained using the HMR method (<b>left</b>) and our approach after Phase 2 (<b>right</b>).</p>
Full article ">Figure 5
<p>Qualitative reconstruction examples from “in the wild” images collected from Internet.</p>
Full article ">
12 pages, 3956 KiB  
Article
Bay Leaf Extract-Based Near-Infrared Fluorescent Probe for Tissue and Cellular Imaging
by Benilde Adriano, Nycol M. Cotto, Neeraj Chauhan, Vinita Karumuru, Meena Jaggi, Subhash C. Chauhan and Murali M. Yallapu
J. Imaging 2021, 7(12), 256; https://doi.org/10.3390/jimaging7120256 - 30 Nov 2021
Viewed by 3635
Abstract
The development of fluorescence dyes for near-infrared (NIR) fluorescence imaging has been a significant interest for deep tissue imaging. Among many imaging fluoroprobes, indocyanine green (ICG) and its analogues have been used in oncology and other medical applications. However, these imaging agents still [...] Read more.
The development of fluorescence dyes for near-infrared (NIR) fluorescence imaging has been a significant interest for deep tissue imaging. Among many imaging fluoroprobes, indocyanine green (ICG) and its analogues have been used in oncology and other medical applications. However, these imaging agents still experience poor imaging capabilities due to low tumor targetability, photostability, and sensitivity in the biological milieu. Thus, developing a biocompatible NIR imaging dye from natural resources holds the potential of facilitating cancer cell/tissue imaging. Chlorophyll (Chl) has been demonstrated to be a potential candidate for imaging purposes due to its natural NIR absorption qualities and its wide availability in plants and green vegetables. Therefore, it was our aim to assess the fluorescence characteristics of twelve dietary leaves as well as the fluorescence of their Chl extractions. Bay leaf extract, a high-fluorescence agent that showed the highest levels of fluorescence, was further evaluated for its tissue contrast and cellular imaging properties. Overall, this study confirms bay-leaf-associated dye as a NIR fluorescence imaging agent that may have important implications for cellular imaging and image-guided cancer surgery. Full article
Show Figures

Figure 1

Figure 1
<p>Chlorophyll measurement with IVIS Imaging System. (<b>A</b>) Twelve different dietary leaves were imaged using the IVIS Imaging System for chlorophyll content. The chlorophyll fluorescence was measured at an excitation and emission of 600/710 nm. Experiment was performed in triplicate. (<b>B</b>) Quantitation of the fluorescence measurement. The figure represents the average of three individual experiments. n = 3.</p>
Full article ">Figure 2
<p>Fluorescence measurement of top six leaf extracts. (<b>A</b>) Based upon whole leaf imaging data, six most fluorescent leaves were selected for further imaging experiments. Bar graph represents chlorophyll fluorescent intensity. Bay leaf exhibited the highest fluorescence levels. (<b>B</b>) Chlorophyll dye was extracted using EtOH from these leaves. Different dilutions of extracted chlorophyll were prepared and a 5 µL drop was imaged. Line graph represents the average fluorescent intensity of extracted chlorophyll dye at different dilutions from three individual experiments. n = 3. (<b>C</b>) Images represent bay leaf extracted chlorophyll concentrations.</p>
Full article ">Figure 3
<p>Optical properties of bay leaf extracts (concentration: 0.16–1.0 mg/mL). (<b>A</b>) Visible absorption spectra of bay leaf (400–800 nm) and (<b>B</b>) fluorescence spectra of bay leaf, excitation: λex, 405 nm and emission: λem, 600–850 nm. Data are representative of a triplicate reading.</p>
Full article ">Figure 4
<p>Chlorophyll measurement in agarose gels. Here, 2.5% agarose was utilized for phantom gel preparation with different dilutions of bay leaf extracted chlorophyll. (<b>A</b>) Gels were imaged using IVIS Imaging System at 600/710 nm. (<b>B</b>) Quantitation of the fluorescence measurement. Data represent the average of three different experiments.</p>
Full article ">Figure 5
<p>Ex Vivo tissue imaging capability of bay leaf extracts. (<b>A</b>) Physical evidence of NIR fluorescence of bay leaf extracts on chicken breast tissue at 0 time and after 24 h, and (<b>B</b>) fluorescence quantification of bay leaf extracts. Data represent the average of three different experiments.</p>
Full article ">Figure 6
<p>Cellular uptake of bay leaf extract in liver and pancreatic cancer cell lines. SK-HEP-1 (liver cancer) and AsPC-1 (pancreatic cancer) cell lines were used for uptake study. Cells were treated with different concentrations of bay leaf extract (contains chlorophyll) for 1 h and imaged using a fluorescent microscope. SK-HEP-1 cells showed higher uptake of chlorophyll. Images were taken at 100×.</p>
Full article ">Figure 7
<p>Characterization of bay leaf chlorophyll extract. Dynamic light scattering (DLS) was used to characterize Chl extracts. (<b>A</b>) Average particle size for bay leaf extracted Chl was 62.72 nm. X-axis scale was set as log. (<b>B</b>) Surface charge was recorded as −24.76 mV. (<b>C</b>) Particle concentration per mL was 1.95 × 10<sup>11</sup>. Data represent the average of three different experiments.</p>
Full article ">Scheme 1
<p>Natural dietary leaf-based NIR fluorescent probe development utilizing leaf chlorophyll extracts and evaluation of their fluorescence properties and proof-of-concept of cellular and tissue imaging capabilities.</p>
Full article ">
17 pages, 45238 KiB  
Article
Evaluation of 2D-/3D-Feet-Detection Methods for Semi-Autonomous Powered Wheelchair Navigation
by Cristian Vilar Giménez, Silvia Krug, Faisal Z. Qureshi and Mattias O’Nils
J. Imaging 2021, 7(12), 255; https://doi.org/10.3390/jimaging7120255 - 30 Nov 2021
Cited by 7 | Viewed by 3082
Abstract
Powered wheelchairs have enhanced the mobility and quality of life of people with special needs. The next step in the development of powered wheelchairs is to incorporate sensors and electronic systems for new control applications and capabilities to improve their usability and the [...] Read more.
Powered wheelchairs have enhanced the mobility and quality of life of people with special needs. The next step in the development of powered wheelchairs is to incorporate sensors and electronic systems for new control applications and capabilities to improve their usability and the safety of their operation, such as obstacle avoidance or autonomous driving. However, autonomous powered wheelchairs require safe navigation in different environments and scenarios, making their development complex. In our research, we propose, instead, to develop contactless control for powered wheelchairs where the position of the caregiver is used as a control reference. Hence, we used a depth camera to recognize the caregiver and measure at the same time their relative distance from the powered wheelchair. In this paper, we compared two different approaches for real-time object recognition using a 3DHOG hand-crafted object descriptor based on a 3D extension of the histogram of oriented gradients (HOG) and a convolutional neural network based on YOLOv4-Tiny. To evaluate both approaches, we constructed Miun-Feet—a custom dataset of images of labeled caregiver’s feet in different scenarios, with backgrounds, objects, and lighting conditions. The experimental results showed that the YOLOv4-Tiny approach outperformed 3DHOG in all the analyzed cases. In addition, the results showed that the recognition accuracy was not improved using the depth channel, enabling the use of a monocular RGB camera only instead of a depth camera and reducing the computational cost and heat dissipation limitations. Hence, the paper proposes an additional method to compute the caregiver’s distance and angle from the Powered Wheelchair (PW) using only the RGB data. This work shows that it is feasible to use the location of the caregiver’s feet as a control signal for the control of a powered wheelchair and that it is possible to use a monocular RGB camera to compute their relative positions. Full article
(This article belongs to the Section Computer Vision and Pattern Recognition)
Show Figures

Figure 1

Figure 1
<p>Power wheelchair setup and description of feet recognition approach. (<b>a</b>) PW with a depth camera below the armrest. (<b>b</b>) Caregiver feet recognition description.</p>
Full article ">Figure 2
<p>Camera placement on a wheelchair armrest and different camera outputs.</p>
Full article ">Figure 3
<p>Experiment definition. (<b>a</b>) Experiment 1: YOLOv4-Tiny. (<b>b</b>) Experiment 2: 3DHOG.</p>
Full article ">Figure 4
<p>Miun-Feet dataset examples in different scenarios and shoes models.</p>
Full article ">Figure 5
<p>Examples of different synthetic objects segmented from the ModelNet40 dataset, including the frontal-view post-processing using a voxel grid of <math display="inline"><semantics> <msup> <mn>20</mn> <mn>3</mn> </msup> </semantics></math> voxels.</p>
Full article ">Figure 6
<p>Point cloud object segmentation using the RGB labels.</p>
Full article ">Figure 7
<p>Caregiver distance and angle measurement.</p>
Full article ">Figure 8
<p>YOLOv4-Tiny results for the different test datasets using 1-channel depth, 3-channel RGB and 4-channel RGB+D input images.</p>
Full article ">Figure 9
<p>YOLOv4-Tiny results for the different test datasets using 1-channel depth, 3-channel RGB and 4-channel RGB+D input images.</p>
Full article ">Figure 10
<p>From top to bottom and left to right: YOLOv4-Tiny results using the 3-channel RGB frames for the 1-Indoor1-Dark, 2-Indoor1-Light, 3-Indoor2-Dark, 4-Indoor2-Light, 5-Outdoor1, 6-Outdoor2, 7-Darkness and 8-Outdoor5 scenarios.</p>
Full article ">Figure 11
<p>3DHOG results for the different test datasets using 1-channel depth, 3-channel RGB and 4-channel RGB+D input images.</p>
Full article ">Figure 12
<p>3DHOG mAP and PCA dimensionality feature reduction.</p>
Full article ">Figure 13
<p>Experiment 3 results to measure the caregiver’s relative distance and angles using the RGB data. (<b>a</b>) Experiment 3: Caregiver’s distance measurement. (<b>b</b>) Experiment 3: Caregiver’s angle measurement.</p>
Full article ">Figure 14
<p>Experiment 3: Ground truth and measured distances using an RGB data.</p>
Full article ">
13 pages, 2216 KiB  
Article
Comparison of Different Image Data Augmentation Approaches
by Loris Nanni, Michelangelo Paci, Sheryl Brahnam and Alessandra Lumini
J. Imaging 2021, 7(12), 254; https://doi.org/10.3390/jimaging7120254 - 27 Nov 2021
Cited by 69 | Viewed by 9020
Abstract
Convolutional neural networks (CNNs) have gained prominence in the research literature on image classification over the last decade. One shortcoming of CNNs, however, is their lack of generalizability and tendency to overfit when presented with small training sets. Augmentation directly confronts this problem [...] Read more.
Convolutional neural networks (CNNs) have gained prominence in the research literature on image classification over the last decade. One shortcoming of CNNs, however, is their lack of generalizability and tendency to overfit when presented with small training sets. Augmentation directly confronts this problem by generating new data points providing additional information. In this paper, we investigate the performance of more than ten different sets of data augmentation methods, with two novel approaches proposed here: one based on the discrete wavelet transform and the other on the constant-Q Gabor transform. Pretrained ResNet50 networks are finetuned on each augmentation method. Combinations of these networks are evaluated and compared across four benchmark data sets of images representing diverse problems and collected by instruments that capture information at different scales: a virus data set, a bark data set, a portrait dataset, and a LIGO glitches data set. Experiments demonstrate the superiority of this approach. The best ensemble proposed in this work achieves state-of-the-art (or comparable) performance across all four data sets. This result shows that varying data augmentation is a feasible way for building an ensemble of classifiers for image classification. Full article
(This article belongs to the Special Issue Color Texture Classification)
Show Figures

Figure 1

Figure 1
<p>Proposed approach. Transfer learning with multiple ResNet50s pretrained on ImageNet using different sets of data augmentation methods, with networks fused by sum rule.</p>
Full article ">Figure 2
<p>Schematic of ResNet50.</p>
Full article ">Figure 3
<p>An example of some traditional augmentation methods on the BARK data set. The left image is the original image.</p>
Full article ">Figure 4
<p>An example image of App5—DCT. The left image is the original image.</p>
Full article ">Figure 5
<p>An example image of App10—DWT. The left image is the original image.</p>
Full article ">Figure 6
<p>An example image of App11—DQT. The left image is the original image.</p>
Full article ">
13 pages, 1645 KiB  
Review
Principles and Perspectives of Radiographic Imaging with Muons
by Luigi Cimmino
J. Imaging 2021, 7(12), 253; https://doi.org/10.3390/jimaging7120253 - 26 Nov 2021
Cited by 7 | Viewed by 4398
Abstract
Radiographic imaging with muons, also called Muography, is based on the measurement of the absorption of muons, generated by the interaction of cosmic rays with the earth’s atmosphere, in matter. Muons are elementary particles with high penetrating power, a characteristic that makes them [...] Read more.
Radiographic imaging with muons, also called Muography, is based on the measurement of the absorption of muons, generated by the interaction of cosmic rays with the earth’s atmosphere, in matter. Muons are elementary particles with high penetrating power, a characteristic that makes them capable of crossing bodies of dimensions of the order of hundreds of meters. The interior of bodies the size of a pyramid or a volcano can be seen directly with the use of this technique, which can rely on highly segmented muon trackers. Since the muon flux is distributed in energy over a wide spectrum that depends on the direction of incidence, the main difference with radiography made with X-rays is in the source. The source of muons is not tunable, neither in energy nor in direction; to improve the signal-to-noise ratio, muography requires large instrumentation, long time data acquisition and high background rejection capacity. Here, we present the principles of the Muography, illustrating how radiographic images can be obtained, starting from the measurement of the attenuation of the muon flux through an object. It will then be discussed how recent technologies regarding artificial intelligence can give an impulse to this methodology in order to improve its results. Full article
(This article belongs to the Special Issue X-ray Digital Radiography and Computed Tomography)
Show Figures

Figure 1

Figure 1
<p>Energy spectra of muons at <math display="inline"><semantics> <msup> <mn>0</mn> <mo>∘</mo> </msup> </semantics></math> (black dots) and at <math display="inline"><semantics> <msup> <mn>75</mn> <mo>∘</mo> </msup> </semantics></math> (white dots) as in [<a href="#B31-jimaging-07-00253" class="html-bibr">31</a>].</p>
Full article ">Figure 2
<p>The angles <math display="inline"><semantics> <mi>θ</mi> </semantics></math> and <math display="inline"><semantics> <msup> <mi>θ</mi> <mo>*</mo> </msup> </semantics></math> formed by an incoming muon, respectively, with the normal axis to the earth surface and with the normal axis to the top of the atmosphere.</p>
Full article ">Figure 3
<p>Trends in the ratio between the flux at different altitudes <span class="html-italic">h</span> and the flux at sea level as a function of the muons energy. The comparison of the curves gives the expected percentage change in the flux of muons at different altitudes.</p>
Full article ">Figure 4
<p>Schematic of the simulated scenario. Detectors (not in scale) placed in the positions A, B and C.</p>
Full article ">Figure 5
<p>A spherical cavity (<b>left</b>) and its 3D reconstruction as seen from the observation point of the detectors C (<b>center</b>) and B (<b>right</b>) as in [<a href="#B48-jimaging-07-00253" class="html-bibr">48</a>].</p>
Full article ">Figure 6
<p>The cube (in gray) predicted by the supervised algorithm compared with the unknown object (left in blue) and with the 3D reconstruction of the unknown object (right in red).</p>
Full article ">Figure 7
<p>Denoising process of the 3D reconstructed object. The XY-view on the right shows the predicted object (blue), the 3D reconstructed object (red) and the reference object (gray) in the space constrained by the viewing angle of the detectors placed in positions A, B and C.</p>
Full article ">
29 pages, 4655 KiB  
Review
Roadmap on Digital Holography-Based Quantitative Phase Imaging
by Vinoth Balasubramani, Małgorzata Kujawińska, Cédric Allier, Vijayakumar Anand, Chau-Jern Cheng, Christian Depeursinge, Nathaniel Hai, Saulius Juodkazis, Jeroen Kalkman, Arkadiusz Kuś, Moosung Lee, Pierre J. Magistretti, Pierre Marquet, Soon Hock Ng, Joseph Rosen, Yong Keun Park and Michał Ziemczonok
J. Imaging 2021, 7(12), 252; https://doi.org/10.3390/jimaging7120252 - 26 Nov 2021
Cited by 49 | Viewed by 8422
Abstract
Quantitative Phase Imaging (QPI) provides unique means for the imaging of biological or technical microstructures, merging beneficial features identified with microscopy, interferometry, holography, and numerical computations. This roadmap article reviews several digital holography-based QPI approaches developed by prominent research groups. It also briefly [...] Read more.
Quantitative Phase Imaging (QPI) provides unique means for the imaging of biological or technical microstructures, merging beneficial features identified with microscopy, interferometry, holography, and numerical computations. This roadmap article reviews several digital holography-based QPI approaches developed by prominent research groups. It also briefly discusses the present and future perspectives of 2D and 3D QPI research based on digital holographic microscopy, holographic tomography, and their applications. Full article
(This article belongs to the Special Issue Digital Holography: Development and Application)
Show Figures

Figure 1

Figure 1
<p>(<b>a</b>) example of optical setup for transmission Digital Holographic Microscopy DHM. (<b>b</b>) Optical setup for Tomographic Diffractive Microscopy: Laser source with controllable coherence length. <span class="html-italic">NF</span> neutral filter, λ/2 plate, <span class="html-italic">PBS</span> polarizing beam splitter, <span class="html-italic">BS</span> beam splitter, <span class="html-italic">BE</span> beam expander, <span class="html-italic">SM</span> steering mirror and <span class="html-italic">M</span> mirror, BF back focal plane, <span class="html-italic">S</span> specimen, <span class="html-italic">C</span> cell, <span class="html-italic">O</span> object wave, R reference wave.</p>
Full article ">Figure 2
<p>3D representation of color-coded optical path difference (OPD) of a living neuronal network. The right part of the image is quasi speckle-free thanks to a polychromatic DHM approach [<a href="#B37-jimaging-07-00252" class="html-bibr">37</a>], allowing to study neuronal processes and network connectivity.</p>
Full article ">Figure 3
<p>Phase image by TDM of an interpenetrated bundle of neuron dendrites.</p>
Full article ">Figure 4
<p>Image of a dendritic spine. Spatial resolution &lt;100 nm [<a href="#B4-jimaging-07-00252" class="html-bibr">4</a>].</p>
Full article ">Figure 5
<p>(<b>a</b>) Schematic of the self-reference on-axis apparatus used for QPI. (<b>b</b>) Flowchart of the phase-shifting process (Adapted from [<a href="#B71-jimaging-07-00252" class="html-bibr">71</a>]). (<b>c</b>) Iterative QPI approach described in the text (Adapted from [<a href="#B73-jimaging-07-00252" class="html-bibr">73</a>]).</p>
Full article ">Figure 6
<p>(<b>a</b>,<b>b2</b>) Reconstructions of thin (OT &lt; λ) phase-only object by the iterative approach when initialized with regular intensity measurement (<b>b1</b>) and phase-contrast measurement (not shown), respectively (Adapted from [<a href="#B73-jimaging-07-00252" class="html-bibr">73</a>]). (<b>c</b>) Theoretical and (<b>d</b>) reconstructed phase of a refractive lens (OT &gt; λ) by using the phase-shifting method (Adapted from [<a href="#B71-jimaging-07-00252" class="html-bibr">71</a>]).</p>
Full article ">Figure 7
<p>The pump (<span class="html-italic">λ</span><sub>1</sub>) two-probes (<span class="html-italic">λ</span><sub>2</sub>) method allows the recovery of permittivity transient, separation of the phase delay, and absorbance contributions at the wavelength of probe (<span class="html-italic">λ</span><sub>2</sub>). Light-matter interaction from the focal region is fully defined by the instantaneous permittivity (square of the complex refractive index [<span class="html-italic">n</span>(<span class="html-italic">t</span>) + <span class="html-italic">ik</span>(<span class="html-italic">t</span>)]<sup>2</sup>). The configuration can be applied to both transmission and reflection modes.</p>
Full article ">Figure 8
<p>AI-based analysis of 3D QPI. (<b>A</b>) Segmentation of immunological synapse formation. (<b>B</b>) Retrieving molecular information from unlabeled cells.</p>
Full article ">Figure 9
<p>(<b>i</b>): Comparison of coherent transfer functions (CTF) of (<b>a</b>) BR, (<b>b</b>) SR, and (<b>c</b>) IDT. (<b>ii</b>): Comparison of slices of 3D refractive index distribution of <span class="html-italic">candida rugosa</span> and the experimentally obtained CTFs corresponds to (<b>a</b>) SR approach, (<b>b</b>) BR approach, and (<b>c</b>) IDT approach. (<b>iii</b>): 3D illustration of <span class="html-italic">candida rugosa</span> at sub-cellular structural views; the different colors represent the different organelles of the cell. Scale bars: 2 µm. Adapted from [<a href="#B130-jimaging-07-00252" class="html-bibr">130</a>].</p>
Full article ">Figure 10
<p>(<b>a</b>) Multi-scale HT zebrafish image [<a href="#B143-jimaging-07-00252" class="html-bibr">143</a>]. (<b>b</b>) Refractive index (red) and polarization contrast (yellow) HT zebrafish tail image [<a href="#B149-jimaging-07-00252" class="html-bibr">149</a>].</p>
Full article ">Figure 11
<p>(<b>a</b>) CAD model of the 3D cell phantom. (<b>b</b>,<b>c</b>) Cross-sections of the 3D RI distribution of the phantom measured using three different holographic tomographs with a limited angle of projections. (<b>d</b>,<b>e</b>) Histograms of the ΔRI in case of full measurement volume (<b>d</b>) and manually segmented single nucleoli (<b>e</b>).</p>
Full article ">Figure 12
<p>Neural microscopy framework for quantitative phase imaging. (<b>a</b>) NN1 performs reconstruction of quantitative phase image from intensity acquisition. Here, as an example, the sample is a culture of adherent cells. (<b>b</b>) NN2 infers a quantitative representation from the phase image. Here, as an example, NN2 maps the phase image of adherent cells into an image that encodes, simultaneously, cell positions and dry mass measurements. (<b>c</b>) NN3 directly infers the quantitative representation from intensity acquisition. Phase image reconstruction is discarded. Such NN could be all-optical (adapted from [<a href="#B176-jimaging-07-00252" class="html-bibr">176</a>]). Light is then transmitted, through the sample, towards a fabricated NN that infers an image of the quantitative representation.</p>
Full article ">
20 pages, 12131 KiB  
Article
Formation of Fused Images of the Land Surface from Radar and Optical Images in Spatially Distributed On-Board Operational Monitoring Systems
by Vadim A. Nenashev and Igor G. Khanykov
J. Imaging 2021, 7(12), 251; https://doi.org/10.3390/jimaging7120251 - 25 Nov 2021
Cited by 12 | Viewed by 2425
Abstract
This paper considers the issues of image fusion in a spatially distributed small-size on-board location system for operational monitoring. The purpose of this research is to develop a new method for the formation of fused images of the land surface based on data [...] Read more.
This paper considers the issues of image fusion in a spatially distributed small-size on-board location system for operational monitoring. The purpose of this research is to develop a new method for the formation of fused images of the land surface based on data obtained from optical and radar devices operated from two-position spatially distributed systems of small aircraft, including unmanned aerial vehicles. The advantages of the method for integrating information from radar and optical information-measuring systems are justified. The combined approach allows removing the limitations of each separate system. The practicality of choosing the integration of information from several widely used variants of heterogeneous sources is shown. An iterative approach is used in the method for combining multi-angle location images. This approach improves the quality of synthesis and increases the accuracy of integration, as well as improves the information content and reliability of the final fused image by using the pixel clustering algorithm, which produces many partitions into clusters. The search for reference points on isolated contours is carried out on a pair of left and right images of the docked image from the selected partition. For these reference points, a functional transformation is determined. Having applied it to the original multi-angle heterogeneous images, the degree of correlation of the fused image is assessed. Both the position of the reference points of the contour and the desired functional transformation itself are refined until the quality assessment of the fusion becomes acceptable. The type of functional transformation is selected based on clustered images and then applied to the original multi-angle heterogeneous images. This process is repeated for clustered images with greater granularity in case if quality assessment of the fusion is considered to be poor. At each iteration, there is a search for pairs of points of the contour of the isolated areas. Areas are isolated with the use of two image segmentation methods. Experiments on the formation of fused images are presented. The result of the research is the proposed method for integrating information obtained from a two-position airborne small-sized radar system and an optical location system. The implemented method can improve the information content, quality, and reliability of the finally established fused image of the land surface. Full article
Show Figures

Figure 1

Figure 1
<p>Block diagram of the radio-optical complex.</p>
Full article ">Figure 2
<p>Scheme of implementation of the method for fused image formation.</p>
Full article ">Figure 3
<p>Optical and radar images of the same area.</p>
Full article ">Figure 4
<p>Part of a series of partitions of the docked optical and radar images into pixel clusters by means of multi-threshold processing.</p>
Full article ">Figure 5
<p>The dependence of the values of the standard deviations σ on the number of clusters in the partition.</p>
Full article ">Figure 6
<p>Boundary isolation by Canny edge detector on piecewise-constant partitions.</p>
Full article ">Figure 7
<p>Search for pairs of reference points of contour on a clustered pair of multi-angle images.</p>
Full article ">Figure 8
<p>Assessment of the quality of the results of fusion of the original pairs of images with a different number of clusters and, the respective reference points in the partition.</p>
Full article ">Figure 8 Cont.
<p>Assessment of the quality of the results of fusion of the original pairs of images with a different number of clusters and, the respective reference points in the partition.</p>
Full article ">Figure 9
<p>Plot of two-dimensional correlation function.</p>
Full article ">Figure 10
<p>The result of the fusion of optical and radar images.</p>
Full article ">Figure 11
<p>Heterogeneous images of the terrain.</p>
Full article ">Figure 12
<p>Fused image.</p>
Full article ">Figure 13
<p>A series of fused images with varying degrees of enhancement of the radar or optical layer.</p>
Full article ">Figure 13 Cont.
<p>A series of fused images with varying degrees of enhancement of the radar or optical layer.</p>
Full article ">
Previous Issue
Next Issue
Back to TopTop