[go: up one dir, main page]

Next Article in Journal
Reinforcement Learning-Based Joint User Pairing and Power Allocation in MIMO-NOMA Systems
Previous Article in Journal
An Analysis of the Reliability of a Bus Safety Structure on Carrying Out the Numerical and Experimental Tests
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Letter

A Novel Approach of Parallel Retina-Like Computational Ghost Imaging

School of Optics and Photonics, Beijing Institute of Technology, Key Laboratory of Biomimetic Robots and Systems, Ministry of Education, Beijing 100081, China
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(24), 7093; https://doi.org/10.3390/s20247093
Submission received: 6 November 2020 / Revised: 4 December 2020 / Accepted: 9 December 2020 / Published: 11 December 2020
(This article belongs to the Section Sensing and Imaging)
Figure 1
<p>Principle of the PRGI (<b>a</b>) PRGI; (<b>b</b>) Original image; (<b>c</b>) Reconstructed image.</p> ">
Figure 2
<p>Measurement principle of the PRGI.</p> ">
Figure 3
<p>Reconstructed results of ‘CSGI’, ‘PGI’, ‘RGI’, and ‘PRGI’ under different sampling rates.</p> ">
Figure 4
<p>Quantitative analysis results of ‘CSGI’, ‘PGI’, ‘RGI’ and ‘PRGI’ under different sampling rates. (<b>a</b>) PSNR; (<b>b</b>) SSIM; (<b>c</b>) Time cost.</p> ">
Figure 5
<p>Reconstructed results of ‘CSGI’, ‘PGI’, ‘RGI’ and ‘PRGI’ under different image sizes.</p> ">
Figure 6
<p>Quantitative analysis results of ‘CSGI’, ‘PGI’, ‘RGI’ and ‘PRGI’ under different image sizes. (<b>a</b>) PSNR; (<b>b</b>) SSIM; (<b>c</b>) Time cost.</p> ">
Figure 7
<p>Reconstructed results of ‘PRGI’ under different image sizes and different block sizes.</p> ">
Figure 8
<p>Quantitative analysis results of ‘PRGI’ under the condition of different image sizes and different block numbers. (<b>a</b>) PSNR; (<b>b</b>) SSIM; (<b>c</b>) Time cost.</p> ">
Figure 9
<p>Imaging results of reconstructing the whole image and reconstructing part of the image. (<b>a</b>) Original image; (<b>b</b>) Sample of retina-like patterns; (<b>c</b>) The result of reconstructing the whole image; (<b>d</b>) The result of reconstructing part of the image.</p> ">
Figure 10
<p>PRGI is applied to multiple objects. (<b>a</b>) Original image contains girl and toys; (<b>b</b>) Sample of retina-like patterns cover whole blocks; (<b>c</b>) Reconstructed image contains the girl and toys; (<b>d</b>) Original image contains three trucks; (<b>e</b>) Sample of retina-like patterns cover part of blocks; (<b>f</b>) Reconstructed image contains the three trucks.</p> ">
Figure 11
<p>System calibration. (<b>a</b>) Ideal detector block; (<b>b</b>) Real detector block; (<b>c</b>) System calibration for 2 × 2 blocks.</p> ">
Versions Notes

Abstract

:
Computational ghost imaging (CGI), with the advantages of wide spectrum, low cost, and robustness to light scattering, has been widely used in many applications. The key issue is long time correlations for acceptable imaging quality. To overcome the issue, we propose parallel retina-like computational ghost imaging (PRGI) method to improve the performance of CGI. In the PRGI scheme, sampling and reconstruction are carried out by using the patterns which are divided into blocks from designed retina-like patterns. Then, the reconstructed image of each block is stitched into the entire image corresponding to the object. The simulations demonstrate that the proposed PRGI method can obtain a sharper image while greatly reducing the time cost than CGI based on compressive sensing (CSGI), parallel architecture (PGI), and retina-like structure (RGI), thereby improving the performance of CGI. The proposed method with reasonable structure design and variable selection may lead to improve performance for similar imaging methods and provide a novel technique for real-time imaging applications.

1. Introduction

Since the first experiment of ghost imaging (GI) was accomplished in 1995 [1], its advantages of nonlocal features, simple structure, high sensitivity etc., has attracted many researchers to study it and it has been applied in many fields [2]. GI provides object information by using a single bucket detector and a set of random patterns produced by pseudo-thermal or thermal light source [3]. Then, the information of scene is obtained by correlating a set of random patterns and the light intensity collected by a single-pixel detector. To simplify the structure, CGI and single-pixel imaging [4,5] are proposed, which employ only one arm, and the other is replaced by spatial light modulator or digital micromirror device (DMD) [6,7]. In CGI, according to the Nyquist sampling criterion, the number of illuminated patterns projecting on the scene and the pixel number of images to be reconstructed are usually required to be identical [8]. Then, a large number of patterns are needed to produce acceptable imaging quality while costing much time [9]. Therefore, the contradiction between imaging quality and imaging efficiency is a challenge of CGI [10,11]. To solve this challenge, the principles and algorithms have been extensively investigated in many previous works [12]. For example, the compressive sensing (CS) technique is proposed to enhance performance of CGI [13]. However, as the image size becomes larger, the required number of illuminated patterns will also increase, resulting in an increase in the sampling time. Meanwhile, a larger image size will increase the computational complexity of the reconstructed algorithm, thereby costing more time for the reconstruction. Therefore, a method of parallel architecture was proposed to improve the performance of high-resolution imaging [14,15,16,17].
Jun Ke et al. [18] presented a block-based compressive imaging system for high-resolution object reconstruction. Moreover, the influence on system parameters was compared under the condition of using different kinds of objects. Later, John P. Dumas et al. [19] proposed a novel CGI structure based on a focal plane array. They reconstructed high-resolution images of binary and grayscale objects using fewer illumination patterns than the traditional method. Yun-Hui Li et al. [20] also proposed a CSGI system based on parallel complementary method and established a mathematical model of processing for blocks. As shown in their results, the peak signal to noise ratio (PSNR) increased under the condition of observed compression ratio increase and the number of blocks is reduced. However, the total amount of data increased with the increase of compression ratio, and the algorithms time consumption reduced with the increase of the number of blocks. Heng Wu et al. [21] proposed a CGI method based on deep learning denoising under the condition of sub-Nyquist sampling ratio. This method can obtain a clear object image and has practical applications in image denoising. Ya-Peng Zhan et al. [22] designed a coprime based on the Eisenstein integer which is used for optimizing the light point pixel arrangement of spatial light modulator. This method enhances the imaging quality, reduces background noise, and avoids periodicity.
To maintain the balance of high resolution and high imaging efficiency, in our previous work [23], we proposed a different three-dimensional CGI method combined with retina-like structure inspired by human eye. Parametric analysis shows that this method can compress redundant information and reduces the imaging time, thus this method can improve imaging efficiency compared to the traditional CGI. In this paper, we propose a novel method based on parallel architecture with retina-like structure, which is named PRGI. In addition to verifying the effectiveness of PRGI by simulation, we also discuss the selection of block number, the design of retina-like patterns, and the system calibration of practical applications. Unlike the traditional imaging system based on parallel architecture, applying retina-like structure can realize variable spatial resolution sampling on DMD without any additional devices, which is highly suitable for CGI. Compared with previous work on retina-like structures [24], the proposed method reconstructs part of the image instead of whole image, which improves the efficiency of the CGI.

2. Theory

The scheme of PRGI is illustrated in Figure 1. A beam of light from the light source is uniformity illuminated on DMD. The element of DMD has two states of ‘0’ and ‘1’, which corresponds to different deflection angles. To realize modulation of reflected light of DMD, the deflection angles are controlled by a computer according to retina-like patterns [21]. The reflected light is collimated by the projecting lens and illuminated on the surface of object. Unlike conventional CGI using a bucket detector or a single-pixel detector, PRGI uses a multi-pixel detector to collect the light intensity reflected by the object after passing through focusing lens. After several measurements, the computer generates the reconstructed image by correlation calculations of retina-like patterns illuminated on DMD and light intensity of multiple acquisition by the multi-pixel detector.
The measurement principle of PRGI is shown in Figure 2, the retina-like patterns based on log-polar transform is denoted by P(x, y, t), where t is the time index. Object is denoted by O(x, y). The multi-pixel detector is used to collect the light intensity, denoted by I(x, y). The remarkable advantages of PRGI are described by two aspects:
  • Higher quality: retina-like patterns have the characteristics of ‘high resolution in center area and low resolution in edge area’, which can effectively improve the image quality of the center area of the reconstructed image.
  • Higher efficiency: the sampling data amount of each block can be sharply decreased after object divided into blocks, and the corresponding sampling time is also reduced. Moreover, in terms of reconstruction algorithm, most of the calculation process is the matrix calculation. As the dimension of the matrix decreases, the amount of data calculation in reconstruction algorithm is also greatly reduced. PRGI calculates the data of each of block rather than the whole image, which improves the efficiency of the reconstruction algorithm.
The measurement principle of each block of the PRGI system can be described as follows:
i n ( t ) = x , y p n ( x , y , t ) o n ( x , y ) ( n = 1 , 2 , N 2 ) .
The block number n here is used to correlate patterns and intensity values, where pn(x,y,t) is the patterns in block n and t is the time index. on(x,y) is the object in block n. in(t) is the total light intensity in the nth block of multi-pixel detector. The retina-like patterns are divided uniformity. It is supposed that the resolution of the object to be imaged is X × X, where N is the divisor of X. When numbering the blocks, we adopt first the row and then the column manner.
P ( x , y , t ) = [ p 1 ( x , y , t ) p 2 ( x , y , t ) p N ( x , y , t ) p N + 1 ( x , y , t ) p N + 2 ( x , y , t ) p 2 N ( x , y , t ) p N 2 N + 1 ( x , y , t ) p N 2 N + 2 ( x , y , t ) p N 2 ( x , y , t ) ] .
In each block of PRGI, the principle of measurement and reconstruction algorithm is same as that of conventional CGI. However, the imaging quality of the reconstructed image of conventional CGI using second-order correlation algorithm does not have a good performance. With the use of CS algorithm [25], the imaging quality and imaging efficiency of CSGI have a better performance than conventional CGI. The current mainstream CS reconstruction algorithms mainly include convex optimization, iterative threshold, and greedy algorithms. The total variation (TV) regularization prior algorithm [26] in the convex optimization algorithm is used to replace the second-order correlation algorithm for image reconstruction. The CS reconstruction algorithm based on TV transforms the image reconstruction problem into a constrained optimization problem. Mathematically, the basis transformation matrix and the corresponding coefficient vector are expressed as H and c, respectively, and the optimization model is as follows:
min c l 1 s . t . H o n = c p n o n = i n
where we use l1 norm to calculate the image’s total variation and H is the gradient calculation matrix. p n R T × m is the nth pattern matrix (there are T patterns and each pattern consists of m = (X/N) × (X/N) pixels). o n R m 1 is the nth block of object. i n R T 1 is nth measurement matrix. Finally, the nth block of reconstructed image o n R m × 1 is obtained by minimizing the Equation (3), and then transforms to o n ( x , y ) R ( X / N ) × ( X / N ) .
The method on reconstructing the image by stitching images of the blocks is same as that of division, the final reconstructed image is obtained as O ( x , y ) R X X .
O ( x , y ) = [ o 1 ( x , y ) o 2 ( x , y ) o N ( x , y ) o N + 1 ( x , y , t ) o N + 2 ( x , y ) o 2 N ( x , y ) o N 2 N + 1 ( x , y ) o N 2 N + 2 ( x , y ) o N 2 ( x , y ) ] .
For comparing the quality of the reconstruct results quantitatively, PSNR [27] and structural similarity index measure (SSIM) [28] are used as the evaluation indexes, defined as:
{ PSNR = 10 log 10 ( 2 k 1 ) 2 MSE M S E = 1 M x , y ( O ( x , y ) O ( x , y ) ) 2 ,
where O′(x,y) and O(x,y) are the illuminated object and reconstructed object; MSE [27] is the mean square error; M is the number of pixels of the whole image; k is the number of bits and set as 8. Note that PSNR is an index that can reflect the imaging quality.
SSIM x , y = ( 2 μ μ + c 1 ) ( 2 w + c 2 ) ( μ 2 + u 2 + c 1 ) ( σ 2 + σ 2 + c 2 ) ,
where μ and μ are the average value of O(x,y) and O′(x,y); σ and σ are the variance of O(x,y) and O′(x,y). w is the covariance between O(x,y) and O′(x,y). c1 = (k1 × L)2 and c2 = (k2 × L)2 are the constant with k1 = 0.01, k2 = 0.03 and L = 1. The value range of SSIM is 0 to 1, the larger the value is, the more similar are the two images.

3. Simulations and Results

3.1. Simulation Setup

In order to verify the system above, two kinds of simulations are carried out under the conditions of different sampling rates and different image sizes. All the following simulations are implemented using Matlab 2019a on the computer with an Intel Core i7-9750H 2.6 GHz CPU, 16G RAM and 64 bits system. There is no optimization for GPU-like computation. The parallel reconstruction algorithm is executed sequentially.
We first clarify several experiment settings. The reconstruction is simulated by one tested image ‘coco’ at the different sampling rates, which means the ratio between the number of measurements and image size. The number of blocks N2 of PGI and PRGI is set as 4. The retina-like patterns are designed to fully cover the object. Besides, we repeat each simulation for 10 times and then use average evaluations to produce final quantitative results. For quantitative comparison, PSNR and SSIM are used to evaluate image quality, and time cost is used to evaluate image efficiency.
According to the acceptable algorithms applied to CGI [26], TV is one of appropriate reconstructed algorithms, compared with other CS algorithms. Therefore, TV is selected to be applied to different methods for image reconstructing in the following simulations. In addition to the conventional CSGI, PGI and our previous RGI [23] are also tested for illustrating the advantages of PRGI. The difference between PGI and PRGI is that the illuminated patterns P(x, y, t) of PGI are random patterns. RGI is similar to CSGI, except the illuminated patterns P(x, y, t) of RGI are retina-like patterns.

3.2. Simulation under Different Sampling Rates

First, the image size X × X is set constant as 64 × 64 pixels in the following simulations. We test the methods on the tested image under different sampling rates ranging from 0.05 to 0.2. The reconstruction results of CSGI, PGI, RGI and PRGI are shown in Figure 3. Obviously, the results of PGI and PRGI are sharper than those of CSGI and RGI. Meanwhile, PRGI has a better performance than PGI under the condition of low sampling rate. With the increases of sampling rate, the imaging quality of four methods are enhanced. Among them, under the condition of sampling rate of 0.2, PRGI and PGI achieves acceptable images, PSNR of which reach 30 dB [29].
The results of PSNR and SSIM of four methods under different sampling rates are shown in Figure 4a,b, in which PSNR and SSIM increase with the increases of sampling rate. The trends of PSNR is similar to that of SSIM, and PRGI shows the best imaging quality, following by PGI, RGI, and CSGI. PSNR of PRGI is 8.09 dB more than that of CSGI under the condition of sampling rate of 0.15. SSIM of PRGI is 0.3 more than that of CSGI under the condition of the sampling rate of 0.15, with an increase of 50.0%. With the respect of PSNR, we observe that the PSNR of PRGI begins to be lower than that of PGI under the condition of sampling rate exceeding 0.15. the imaging quality of PRGI and PGI are well enough as the sampling rate is 0.15, but the imaging quality will not improve obviously with the increases of sampling rate. So PRGI sacrifices information beyond the object, which obtains lower PSNR than PGI on whole image analysis with the increases of the sampling rate increasing as the quality is good enough.
With the respect of time cost, which includes sampling time and reconstruction time and is used to reflect the imaging efficiency, the time cost of four methods under different sampling rates are shown in Figure 4c. We can observe that the time required for CSGI and RGI to generate one image is obviously more than of PRGI and PGI. The time cost of CSGI is about 1.83 s more than PRGI as the sampling rate is 0.15, with a decrease of 75.9%, and is about 1.93 s as the sampling rate is 0.2, with a decrease of 75.1%. Time cost of generating one image by PRGI and PGI is the same order of magnitude, and so is that of CSGI and RGI. The time cost increases with the increases of sampling rate, but the difference between the time of PRGI and CSGI is larger.

3.3. Simulation under Different Image Sizes

Image size is another important factor affecting both reconstruction quality and computational complexity. The sampling rate is set constant as 0.1 in the following simulations. We test the four methods on the tested image under the condition of different image sizes, including 32 × 32, 64 × 64, 96 × 96 and 128 × 128 pixels. The images reconstructed by different methods with the increases of image size are shown in Figure 5, in which PRGI obtains better quality than the other methods.
The results of PSNR and SSIM of four methods under different image sizes are shown in Figure 6a,b. PSNR and SSIM of the four methods increase with the increases of image size. PSNR and SSIM of PRGI are better than any other methods. Meanwhile, PGI and RGI obtain a better performance than CSGI in imaging quality. PSNR of PRGI is 7.54 dB more than that of CSGI under the condition of image size with 128 × 128 pixels. SSIM of PRGI is 0.21 more than that of CSGI under the condition of image size with 128 × 128 pixels, an increase of about 30.9%.
As shown in Figure 6c, as image size increases, the time cost of generating one image also increases. PRGI costs the least time than CSGI and RGI to generate one image. The time cost of CSGI is about 1.77 s more than PRGI under the condition of image size with 64 × 64 pixels, a decrease of 78.3%. The time cost of CSGI is about 71.05 s more than PRGI under the condition of image size with 128 × 128 pixels, a decrease of 86.0%. However, the difference between CSGI and PRGI becomes larger as the sampling rate increases. Compared with Figure 4c, we obverse that time cost increases slower with the sampling rate increasing than with the image size increasing. The details of the time cost are shown in Table 1, we calculate the difference between the time required for CSGI and PRGI. PRGI costs much less time to generate one image as image size becomes larger.

4. Discussions

From the above two simulations, the performances including quality and efficiency of PRGI are the best among the four methods. According to the sampling times of the whole image, each block is equivalent to a higher sampling rate under the smaller image size. At the same time, using retina-like patterns is beneficial for capturing information that we are interested in. However, how to divide the GI system into many blocks and how to design the retina-like pattern are factors that directly affect the performance of PRGI.

4.1. Selection of Block Number

The image quality of GI is improved faster at a low sampling rate and slower gradually as sampling rate increases. The imaging quality tends to be constant with the increases of the sampling rate under the condition that the sampling rate reaches a certain value, and the imaging quality of this time always has a good performance. We carry out the following simulation under different number of blocks and different sampling rates. Image size is constant set as 96 × 96. We test different block numbers of PRGI on tested image under different sampling rates ranging from 0.005 to 0.2. Block numbers are set as 4, 16, 36, 64, and 144. The results are shown in Figure 7, from which we observe that larger block number lead to sharper reconstructed image. The reconstructed image becomes sharp and eventually tends to be have same quality with the increases of sampling rate.
The results of PSNR and SSIM are shown in Figure 8a,b, both are similar differences and trends. PSNR and SSIM are higher with the increases of the sampling rate. PSNR and SSIM tend to be stable under the condition that sampling rate reaches a sampling threshold such as 0.075 and the maximum values of PSNR and SSIM are almost the same. When block number is 36, the sampling rate of the whole image is 0.075, and the sampling rate of each block is 2.7, which is oversampling. As for different block number, the point of obtaining best image quality is that when each block reaches the sampling threshold. After reaching the sampling threshold, the performance of various block number is similar. Besides, we obverse that the larger the number of blocks, the sooner the balance will be reached. Therefore, increasing the number of blocks will not improve the image quality too much, but can reconstruct a high-quality image due to the lesser sampling times.
The results of time cost are shown in Figure 8c, the larger the number of blocks is, the less time that is needed for generating one image. However, we find the smaller the block is divided, the less time that is saved. The time cost of 144 block number even more than that of 64 block number under the condition of sampling rate is 0.05. Although dividing into smaller blocks can improve the efficiency, if the block is too small, it will not improve performance too much, and may lead to poor performance. Besides, dividing too many blocks may make difficulties to realize the actual experiments, for example, it is difficult for a detector to distinguish the intensity value of the corresponding block of object [19].

4.2. Design of Retina-Like Patterns

In the previous simulation, retina-like patterns are set manually depending on the interested region in the tested image; using retina-like patterns can sacrifice edge information to improve the imaging quality of the center. Therefore, the best imaging quality can be obtained by covering the central high-resolution region of patterns just over the region of interest when designing retina-like patterns. PRGI has a good performance on the imaging of the region of interest without obtaining all the information of the scene. PRGI can select the reconstructed part of the scene to obtain higher imaging efficiency when it is used for fast imaging, and the rest part can be replaced by all ‘1′ or ‘0′. As shown in Figure 9, reconstructing the whole image costs more time than part of image which covers the interested area. Moreover, when aiming at an unknown object, few random patterns are projected on the scene and reconstruct a blurred image. Then, the region of the object in the scene can be obtained by Haar wavelet transform [30]. After that, PRGI is used for fast imaging. This can be used in real-time object tracking and object recognition.
In addition to be applied to just one object, PRGI is also suitable for multiple objects that are not in the center of the scene. There are two different schemes to design retina-like patterns when applied to multiple objects. First, only one retina-like pattern cover whole blocks is used for imaging and the high-resolution region cover all objects. As shown in Figure 10b,c, the girl and toys can be imaged clearer than the background region. Second, multiple objects can be imaged by several retina-like patterns which only cover one or several blocks. As shown in Figure 10e,f, we observe that the three trucks are imaged clearer than the other region and the block which contains information beyond the object is reconstructed by low-resolution random patterns. The region of multiple objects can be adaptive covered by several retina-like designed patterns.

4.3. System Calibration to Realize Practical Applications

The ideal optical system can focus the light reflected by the object block to the correspond block on the multi-pixel detector. However, light from each block of the DMD can leak onto blocks of detector neighboring the correspond block according to the distribution of point spread function (PSF). Therefore, for PRGI, if only the intensity value of the ideal block is recorded, there will be a certain error on reconstructed image in practical applications. In previous works, Dumas et al. capture contributions of DMD elements neighboring the geometrically mapped pixel and build a mathematical model for the parallel system to improve image reconstruction [19]. In terms of reducing the noise of PRGI caused by PSF, it is necessary to calibrate the system. The noise of PRGI caused by PSF is shown in Figure 11a,b, the illumination of each block will influence the others’ intensity value.
The system calibration for 2 × 2 blocks is shown in Figure 11c, only the DMD elements of the corresponding block work, and the elements of remaining blocks do not work. Under the condition of elements of the nth block working, the intensity value of remaining blocks is recorded as noise, denoted by ea,n,t, where a is the ath block of DMD working, where n is the intensity value of nth block, where t is the time index. The noise caused by other blocks is denoted by En.
{ e a , n = 1 t 1 t e a , n , t E n = 1 a e a , n ( a n ) .
Therefore, the real intensity value is the difference between the detected value in(t) and the error value En in practical applications. Once the system is calibrated and the average error value En of each block is obtained, this error value will not change under the condition that the system structure and position of devices do not change. Therefore, except the first measurement, proposed method will not increase the measurement time because of the system calibration.

5. Conclusions

The method based on a multi-pixel detector combined with retina-like patterns is proposed to improve the performance of CGI. Sampling and reconstruction use the patterns which are divided into blocks from the designed retina-like pattern. Using to this method, improved imaging quality and efficiency are realized. We design the system of PRGI and compare with conventional CSGI, PGI, and RGI by carrying out a series of simulations to verify the performance of the method. The results show that: (1) considering the imaging quality, the PRGI method has the best performance whether it is PSNR or SSIM under the condition of low sampling rate. PRGI shows the best imaging quality, following by PGI, RGI, and CSGI. The difference of PSNR between PRGI and CSGI is 8.09 dB under the condition of sampling rate of 0.15. SSIM of PRGI is 0.3 more than that of CSGI, with an increase of 50.0%; (2) in terms of imaging efficiency, the PRGI and PGI methods take the least time for generating one image. The time cost of CSGI is about 1.77 s more than PRGI under the condition of image size with 64 × 64 pixels, a decrease of 78.3%. Meanwhile, the time cost of CSGI is about 71.05 s more than PRGI under the condition of image size with 128 × 128 pixels, a decrease of 86.0%. The larger the image size, PRGI save more time than CSGI and RGI. Therefore, the proposed method obtains higher performance in terms of imaging quality and efficiency than CSGI, PGI, and RGI. In addition, PRGI can generate one sharp image with less illuminated patterns, which can reduce the requirement of the DMD’s RAM for high-resolution imaging. It is worth noting that selecting a suitable number of blocks and designing suitable retina-like patterns may initiate a breakthrough for the application of real-time ghost imaging. However, our method also has some shortcomings. First, using multi-pixel detector may reduce the sensitivity of the system compared to single-pixel detector. Second, the center position and the size of high-resolution region of retina-like patterns are set manually in this method. In our next work, the high-resolution region can adaptively correspond to the object of interest. Third, the proposed method is verified by the simulation results in this paper, and practical experiments will be carried out in the next work.

Author Contributions

J.C. conceived and designed the research; J.C. and D.Z. performed the research and completed original draft. H.C.; F.Z. and Y.Z. investigated, Q.H. supervised. All authors have read and agreed to the published version of the manuscript.

Funding

The research was funded by the National Natural Science Foundation of China (NSFC) 61875012 and 61871031; Natural Science Foundation of Beijing Municipality under Grant 4182058, and the funding of foundation enhancement program (2019-JCJQ-JJ-273 and 2020-JCJQ-JJ-030).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Pittman, T.B.; Shih, Y.H.; Strekalov, D.V. Optical imaging by means of two-photon quantum entanglement. Phys. Rev. A 1995, 52, R3429–R3432. [Google Scholar] [CrossRef]
  2. Edgar, M.P.; Gibson, G.M.; Padgett, M.J. Principles and prospects for single-pixel imaging. Nat. Photonics 2018, 13, 13–20. [Google Scholar] [CrossRef]
  3. Erkmen, B.I.; Shapiro, J.H. Ghost imaging: From quantum to classical to computational. Adv. Opt. Photonics 2010, 2, 405–450. [Google Scholar] [CrossRef]
  4. Shapiro, J.H. Computational ghost imaging. Phys. Rev. A 2008, 78, 061802. [Google Scholar] [CrossRef]
  5. Bromberg, Y.; Katz, O.; Silberberg, Y. Ghost imaging with a single detector. Phys. Rev. A 2008, 79, 1744–1747. [Google Scholar] [CrossRef] [Green Version]
  6. Li, Z.; Suo, J.; Hu, X.; Dai, Q. Content-adaptive ghost imaging of dynamic scenes. Opt. Express 2016, 24, 7328–7336. [Google Scholar] [CrossRef] [PubMed]
  7. Sun, M. Image Retrieval in Spatial and Temporal Domains with a Quadrant Detector. IEEE Photonics J. 2017, 9, 1–6. [Google Scholar] [CrossRef]
  8. Sun, M.; Wang, H.; Huang, J. Improving the performance of computational ghost imaging by using a quadrant detector and digital micro-scanning. Sci. Rep. 2019, 9, 4105. [Google Scholar] [CrossRef] [Green Version]
  9. Erkmen, B.I.; Shapiro, J.H. Signal-to-noise ratio of Gaussian-state ghost imaging. Phys. Rev. A 2009, 79, 023833. [Google Scholar] [CrossRef] [Green Version]
  10. Jiang, H.; Zhu, S.; Zhao, H.; Xu, B.; Li, X. Adaptive regional single-pixel imaging based on the Fourier slice theorem. Opt. Express 2017, 25, 15118–15130. [Google Scholar] [CrossRef]
  11. Sun, S. Multi-scale Adaptive Computational Ghost Imaging. Sci. Rep. 2016, 6, 37013. [Google Scholar] [CrossRef] [Green Version]
  12. Zhang, Z. Hadamard single-pixel imaging versus Fourier single-pixel imaging. Opt. Express 2017, 25, 19619. [Google Scholar] [CrossRef]
  13. Welsh, S.S.; Edgar, M.P.; Bowman, R. Fast full-color computational imaging with single-pixel detectors. Opt. Express 2013, 21, 23068–23074. [Google Scholar] [CrossRef] [PubMed]
  14. Wu, Z.; Wang, X. DMD Mask Construction to Suppress Blocky Structural Artifacts for Medium Wave Infrared Focal Plane Array-Based Compressive Imaging. Sensors 2020, 20, 900. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Dumas, J.P.; Lodhi, M.A.; Bajwa, W.U. From modeling to hardware: An experimental evaluation of image plane and Fourier plane coded compressive optical imaging. Opt. Express 2017, 25, 29472. [Google Scholar] [CrossRef]
  16. Wu, Z.; Wang, X. Stray light correction for medium wave infrared focal plane array-based compressive imaging. Opt. Express 2020, 28, 19097–19112. [Google Scholar] [CrossRef]
  17. Neifeld, M.A.; Ke, J. Optical architectures for compressive imaging. Appl. Opt. 2007, 46, 5293–5303. [Google Scholar] [CrossRef]
  18. Ke, J.; Lam, E.Y. Object reconstruction in block-based compressive imaging. Opt. Express 2012, 20, 22102. [Google Scholar] [CrossRef]
  19. Dumas, J.P.; Lodhi, M.A. Computational imaging with a highly parallel image-plane-coded architecture: Challenges and solutions. Opt. Express 2016, 24, 6145–6155. [Google Scholar] [CrossRef]
  20. Li, Y.H.; Wang, X.D. Modeling and image motion analysis of parallel complementary compressive sensing imaging system. Opt. Commun. 2018, 423, 100–110. [Google Scholar] [CrossRef]
  21. Wu, H.; Wang, R.; Zhao, G.; Xiao, H.; Liang, J.; Wang, D.; Tian, X.; Cheng, L.; Zhang, X. Deep-learning denoising computational ghost imaging. Opt. Lasers Eng. 2020, 134, 106183. [Google Scholar] [CrossRef]
  22. Zhan, Y.P.; Liu, J.; Wang, Z.; Yu, Q. Computational Ghost Imaging Based on Light Source Formed by Coprime Array. Sensors 2020, 20, 4508. [Google Scholar] [CrossRef]
  23. Zhang, K. Modeling and Simulations of Retina-Like Three-Dimensional Computational Ghost Imaging. IEEE Photonics J. 2019, 11, 1–13. [Google Scholar] [CrossRef]
  24. Phillips, D.B.; Sun, M.J.; Taylor, J.M. Adaptive foveated single-pixel imaging with dynamic supersampling. Sci. Adv. 2017, 3, e1601782. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  25. Candes, E.; Romberg, J. Sparsity and Incoherence in Compressive Sampling. Inverse Probl. 2007, 23, 969–985. [Google Scholar] [CrossRef] [Green Version]
  26. Bian, L.; Suo, J.; Dai, Q.; Chen, F. Experimental comparison of single-pixel imaging algorithms. J. Opt. Soc. Am. A 2018, 35, 78. [Google Scholar] [CrossRef] [PubMed]
  27. Liu, H.C.; Yang, B.; Guo, Q.; Shi, J.; Guan, C.; Zheng, G.; Mühlenbernd, H.; Li, G.; Zentgraf, P.; Zhang, P. Single-pixel computational ghost imaging with helicity-dependent metasurface hologram. Sci. Adv. 2017, 3, e1701477. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  28. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [Green Version]
  29. Rizvi, S.; Cao, J.; Zhang, K.; Hao, Q. Deringing and Denoising in Extremely Under-sampled Fourier Single Pixel Imaging. Opt. Express 2020, 28, 7360–7374. [Google Scholar] [CrossRef] [PubMed]
  30. Abmann, M.; Bayer, M. Compressive adaptive computational ghost imaging. Sci. Rep. 2013, 3, 405–450. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Principle of the PRGI (a) PRGI; (b) Original image; (c) Reconstructed image.
Figure 1. Principle of the PRGI (a) PRGI; (b) Original image; (c) Reconstructed image.
Sensors 20 07093 g001
Figure 2. Measurement principle of the PRGI.
Figure 2. Measurement principle of the PRGI.
Sensors 20 07093 g002
Figure 3. Reconstructed results of ‘CSGI’, ‘PGI’, ‘RGI’, and ‘PRGI’ under different sampling rates.
Figure 3. Reconstructed results of ‘CSGI’, ‘PGI’, ‘RGI’, and ‘PRGI’ under different sampling rates.
Sensors 20 07093 g003
Figure 4. Quantitative analysis results of ‘CSGI’, ‘PGI’, ‘RGI’ and ‘PRGI’ under different sampling rates. (a) PSNR; (b) SSIM; (c) Time cost.
Figure 4. Quantitative analysis results of ‘CSGI’, ‘PGI’, ‘RGI’ and ‘PRGI’ under different sampling rates. (a) PSNR; (b) SSIM; (c) Time cost.
Sensors 20 07093 g004
Figure 5. Reconstructed results of ‘CSGI’, ‘PGI’, ‘RGI’ and ‘PRGI’ under different image sizes.
Figure 5. Reconstructed results of ‘CSGI’, ‘PGI’, ‘RGI’ and ‘PRGI’ under different image sizes.
Sensors 20 07093 g005
Figure 6. Quantitative analysis results of ‘CSGI’, ‘PGI’, ‘RGI’ and ‘PRGI’ under different image sizes. (a) PSNR; (b) SSIM; (c) Time cost.
Figure 6. Quantitative analysis results of ‘CSGI’, ‘PGI’, ‘RGI’ and ‘PRGI’ under different image sizes. (a) PSNR; (b) SSIM; (c) Time cost.
Sensors 20 07093 g006
Figure 7. Reconstructed results of ‘PRGI’ under different image sizes and different block sizes.
Figure 7. Reconstructed results of ‘PRGI’ under different image sizes and different block sizes.
Sensors 20 07093 g007
Figure 8. Quantitative analysis results of ‘PRGI’ under the condition of different image sizes and different block numbers. (a) PSNR; (b) SSIM; (c) Time cost.
Figure 8. Quantitative analysis results of ‘PRGI’ under the condition of different image sizes and different block numbers. (a) PSNR; (b) SSIM; (c) Time cost.
Sensors 20 07093 g008
Figure 9. Imaging results of reconstructing the whole image and reconstructing part of the image. (a) Original image; (b) Sample of retina-like patterns; (c) The result of reconstructing the whole image; (d) The result of reconstructing part of the image.
Figure 9. Imaging results of reconstructing the whole image and reconstructing part of the image. (a) Original image; (b) Sample of retina-like patterns; (c) The result of reconstructing the whole image; (d) The result of reconstructing part of the image.
Sensors 20 07093 g009
Figure 10. PRGI is applied to multiple objects. (a) Original image contains girl and toys; (b) Sample of retina-like patterns cover whole blocks; (c) Reconstructed image contains the girl and toys; (d) Original image contains three trucks; (e) Sample of retina-like patterns cover part of blocks; (f) Reconstructed image contains the three trucks.
Figure 10. PRGI is applied to multiple objects. (a) Original image contains girl and toys; (b) Sample of retina-like patterns cover whole blocks; (c) Reconstructed image contains the girl and toys; (d) Original image contains three trucks; (e) Sample of retina-like patterns cover part of blocks; (f) Reconstructed image contains the three trucks.
Sensors 20 07093 g010
Figure 11. System calibration. (a) Ideal detector block; (b) Real detector block; (c) System calibration for 2 × 2 blocks.
Figure 11. System calibration. (a) Ideal detector block; (b) Real detector block; (c) System calibration for 2 × 2 blocks.
Sensors 20 07093 g011
Table 1. Time cost(s) of the four methods under different image sizes.
Table 1. Time cost(s) of the four methods under different image sizes.
Image Size32 × 3264 × 6496 × 96128 × 128
CSGI0.152.2616.1882.58
PGI0.060.502.8911.47
RGI0.132.3017.6379.54
PRGI0.040.492.8711.53
tCSGI/tPRGI3.754.615.647.16
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Cao, J.; Zhou, D.; Zhang, F.; Cui, H.; Zhang, Y.; Hao, Q. A Novel Approach of Parallel Retina-Like Computational Ghost Imaging. Sensors 2020, 20, 7093. https://doi.org/10.3390/s20247093

AMA Style

Cao J, Zhou D, Zhang F, Cui H, Zhang Y, Hao Q. A Novel Approach of Parallel Retina-Like Computational Ghost Imaging. Sensors. 2020; 20(24):7093. https://doi.org/10.3390/s20247093

Chicago/Turabian Style

Cao, Jie, Dong Zhou, Fanghua Zhang, Huan Cui, Yingqiang Zhang, and Qun Hao. 2020. "A Novel Approach of Parallel Retina-Like Computational Ghost Imaging" Sensors 20, no. 24: 7093. https://doi.org/10.3390/s20247093

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop