[go: up one dir, main page]

Next Article in Journal
Combined SPRi Sensor for Simultaneous Detection of Nitrate and Ammonium in Wastewater
Previous Article in Journal
Swelling-Based Distributed Chemical Sensing with Standard Acrylate Coated Optical Fibers
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

High-Precision Lensless Microscope on a Chip Based on In-Line Holographic Imaging

Ministry of Education Key Lab of RF Circuits and Systems, Hangzhou Dianzi University, Hangzhou 310018, China
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(3), 720; https://doi.org/10.3390/s21030720
Submission received: 20 December 2020 / Revised: 12 January 2021 / Accepted: 19 January 2021 / Published: 21 January 2021
(This article belongs to the Section Nanosensors)

Abstract

:
The lensless on-chip microscope is an emerging technology in the recent decade that can realize the imaging and analysis of biological samples with a wide field-of-view without huge optical devices and any lenses. Because of its small size, low cost, and being easy to hold and operate, it can be used as an alternative tool for large microscopes in resource-poor or remote areas, which is of great significance for the diagnosis, treatment, and prevention of diseases. To improve the low-resolution characteristics of the existing lensless shadow imaging systems and to meet the high-resolution needs of point-of-care testing, here, we propose a high-precision on-chip microscope based on in-line holographic technology. We demonstrated the ability of the iterative phase recovery algorithm to recover sample information and evaluated it with image quality evaluation algorithms with or without reference. The results showed that the resolution of the holographic image after iterative phase recovery is 1.41 times that of traditional shadow imaging. Moreover, we used machine learning tools to identify and count the mixed samples of mouse ascites tumor cells and micro-particles that were iterative phase recovered. The results showed that the on-chip cell counter had high-precision counting characteristics as compared with manual counting of the microscope reference image. Therefore, the proposed high-precision lensless microscope on a chip based on in-line holographic imaging provides one promising solution for future point-of-care testing (POCT).

1. Introduction

Optical microscopes play a significant role in scientific study and clinical testing in modern medicine and life sciences, as they can expand and identify morphological features of samples (including cells, bacteria, marine microorganisms, etc.) that cannot be enlarged and seen by the naked eyes. With the iterative upgrade of the spatial resolution of optical microscopes, microscopes with various extended functions, including scattering, absorption, refraction, phase modulation, etc., have also emerged so that the morphological characteristics of transparent samples [1,2,3,4,5,6], the subtle internal changes of thick samples [7,8,9], and structural details even smaller than the wavelength of the light source [10,11] can be easily obtained. However, due to large size, high cost, and complicated operation, these microscope detection platforms can only be used in advanced laboratories and medical institutions. Hence, their scope and conditions of application are limited, especially in resource-limited areas where infectious diseases such as tuberculosis, malaria, sickle cell virus infection, and other diseases are common. Therefore, the use of portable optical microscopes in resource-limited areas to improve detection efficiency is vital for preventing the spread of infectious diseases [12,13,14].
To meet such needs, in recent years, the combination of Complementary Metal Oxide Semiconductor (CMOS) image sensor and microfluidic technology provides an opportunity for the development of portable on-chip imaging detection systems, and significant achievements have been made in the field of lensless microscope research in the recent decade [15,16,17,18,19,20]. The reasons are as follows: First, with the advancement of semiconductor processing technology and the replacement of Charge-coupled Device (CCD) image sensors by CMOS image sensors as image capturing elements, the diameter of the pixel pitch is getting smaller and smaller, and the pixel density on the silicon substrate of 20~30 mm2 has become higher. The impact of lower pixel pitch (for example, Sony’s IMX586 color pixel diameter is 0.8 µm) greatly improves the image resolution. Second, lensless microscopes rely heavily on numerical calculations to recover high-resolution images. With the advancement of semiconductor technology, the computing power of processing chips (such as Central Processing Unit (CPU), Graphics Processing Unit (GPU), Digital Signal Processing (DSP), Field Programmable Gate Array (FPGA), etc.) continues to increase, while prices and power consumption continue to decline, and this is conducive to low-cost, portable imaging systems. Third, algorithms to improve image resolution, such as compression and decompression algorithms [21], image synthesis algorithms [22,23], as well as machine learning algorithms to identify samples quickly [24,25], have been successfully applied in lensless microscope systems. According to imaging principles, lensless imaging systems can be divided into shadow imaging, fluorescence imaging, holographic imaging, etc.
The lensless shadow imaging system, which directly uses the shadow pattern of the sample illuminated by the light source for analysis, is the most basic imaging method. For example, our previous work proposed a lensless microfluidic miniaturized flow cytometer with super-resolution processing based on shadow imaging, which uses online machine learning to perform super-resolution reconstruction of a single frame, and compared with the cytometer, the particle counting error is less than 8% [26]. Zheng et al. proposed a sub-pixel perspective scanning microscope (SPSM), which can obtain two shadow images with a small distance by changing the position of the light source, and then high-resolution images are synthesized by sub-pixel shifted low-resolution pictures at a high rate, achieving a resolution of 660 nm [27]. However, the resolution of the shadow imaging method is still limited, and complex algorithms are required for super-resolution reconstruction in the later stages.
The lensless fluorescence imaging system uses light waves of appropriate wavelengths to irradiate dyed fluorescent samples and evaluates the proportion of specimens by comparing different color reactions on the screen. It can be used in white blood cell classification and the detection of rare cells in the blood, such as circulating tumor cells. For example, Coskun and colleagues proposed a large field-of-view (FOV), dual-imaging mode lensless fluorescent on-chip imaging system, which can achieve high-throughput detection and detection of rare cells in a large FOV [28]. Yet, the fluorescence imaging method requires the addition of filters and other parts to the experimental device, which increases the complexity and system cost.
A lensless holographic imaging system uses a coherent or incoherent light source to obtain a ring pattern formed by the interference of the reference light wave and the object wave. Since holographic imaging contains more information about the object, the characteristics of the sample can be reconstructed by the principle of Fresnel integral diffraction, which provides excellent help for sample analysis. Seo et al. used a lensless holographic imaging platform to implement an on-chip cell counter and used a custom pattern recognition algorithm to characterize and count the cells of interest in the sample. Their experiments verified that the lensless holographic imaging not only improved the signal-to-noise ratio but also had a good imaging ability for weakly scattered samples [29]. Vercruysse et al. proposed a lensless online holographic microscope to realize the differential recognition of unlabeled white blood cell subtypes (lymphocytes, monocytes, granulocytes), which was consistent with the results obtained by conventional blood analyzers [30]. Wu et al. used the Generative Adversarial Networks (GAN) network to reconstruct the sample patterns at different levels in the three-dimensional (3D) sample from a single hologram, and the results matched the slice performance of the high Numerical Aperture (NA) bright-field microscope [31]. Fang et al. have developed an on-chip lensless flow cytometer, which uses the holographic imaging principle and the broad FOV of the S-shaped pipe to achieve a cell counting error of less than 2% [32]. Since the experimental device of holographic imaging is simple, and the resolution can be improved through phase recovery, it has been widely studied and applied in recent years [33,34,35].
Therefore, to overcome the shortcomings of shadow imaging and fluorescence imaging, this paper demonstrates an alternative method of lensless on-chip holographic imaging towards telemedicine applications. It is a portable high-precision detection device based on lensless holographic imaging to capture cells on the glass smear, and uses machine learning to classify and count mouse ascites tumor cells and polystyrene particles in holographic phase recovery images. Experiments show that the resolution of the platform is 1.41 times that of the shadow imaging platform, and the accuracy of cell counting is similar to the manual counting results of the microscope reference image. Besides, the platform has no complex optical path system, simple operation, small size, no damage to samples, low cost, and can be applied in resource-poor areas for point-of-care testing (POCT), which has specific application prospects in preventing the spread of diseases and early detection.

2. Methods

2.1. System Setup

This system is built using the imaging principle of in-line holography, and the overall structure is shown in Figure 1. The system mainly consists of a CMOS image sensor (IMX219PQ, Sony, Tokyo, Japan, 1.12 μm pixel size, 3280 H × 2464 V), a yellow Light Emitting Diode (LED) light source (LY-E65B, OSRAM Opto Semiconductors Inc., Regensburg, Germany, λ = 587 nm), and a pinhole (100 μm in diameter). To make full use of the imaging area of the image sensor, a cover glass (with a thickness of 0.5 mm) is directly covered on the image sensor and glued with liquid glue to prevent misalignment due to moving the experimental device. It not only achieves the conditions of static imaging but also controls the distance between the cover glass and the image sensor in a small range so that the magnification ratio of the sample imaging pattern can also be controlled to avoid imaging distortion. In order to obtain a holographic interference pattern on the surface of the image sensor, the LED light is controlled at a distance of 5 cm from above the image sensor. Besides, the pinhole is placed under the LED to improve the coherence of the LED light source. All the experimental structural components are finally assembled by a 3D-printed dark box (6.5 × 6.5 × 7 cm3) to form a miniaturized on-chip lensless holographic imaging detection system. Note that the LED is powered by an external Raspberry Pi circuit board through a power line. The Raspberry Pi board is also connected with the CMOS image sensor board through an Mobile Industry Processor Interface (MIPI) interface. The captured holographic images are then transferred to the PC via the Raspberry Pi for further analysis.

2.2. Sample Preparation

In order to evaluate the imaging effect of the holographic imaging platform, we prepared two different experimental samples. One is the optical positive film resolution plate (USAF-1951, Yaopu Optics, Chengdu, China), which is used to measure the resolution of the imaging system. There are some particular patterns on the resolution board, and its line width and interval have been precisely set. By identifying the most indistinguishable reference line, the resolution capability of the system can be determined. The resolution board was cleaned by wipe and placed above the CMOS image sensor for testing. Another is the mixed solution of mouse ascites tumor cells and polystyrene particles (15 μm). First, some suspended malignant tumor cells were inoculated into the abdominal cavity of mice. A few days later, the cells proliferated in the abdominal cavity of the mice to produce a large number of offspring to form mice with ascites tumors, and ascites were drawn for later use. In order to ensure that clear wavefront information was observed in the coaxial holographic experiment, we diluted the mixed solution, extracted 1 μL of mouse ascites and 1 μL of polystyrene particles to mix, and then extracted 200 μL of phosphate buffered solution (PBS) and mixed with 2 μL of the cell solution. The diluted solution was made into a glass smear for testing.

2.3. Imaging Principle

In the on-chip lensless holographic imaging detection system, a translucent sample is placed above the image sensor, usually z 2 (<1 mm) from the image sensor. An incoherent light source is fixed at a distance z 1 from the sample, and a pinhole is set directly below the light source so that the sample is irradiated by the light source and instantly forms a holographic pattern on the image sensor, and the image sensor encodes the holographic pattern by intensity.
The on-chip lensless holography is derived from the holography invented by Gabor in 1948 [36]. This method not only records the amplitude information of the light field but also records the phase information. Therefore, holography records the complete information of an object. In the holographic recording process, the recording medium records the complex amplitude, that is, simultaneously records the amplitude and phase information of the original object light wave. The intensity change recorded in this way is called a hologram.
In in-line holography, assuming that the object t is semi-transparent, it can be approximated as:
t ( x 0 , y 0 ) = 1 + Δ t ( x 0 , y 0 )
Here, Δ t   is the transmittance fluctuation and Δ t <<1. When an object is locally illuminated by plane wave A , it will coherently propagate z 2 distance (distance from the sample to the image sensor):
P [ z 2 ] { A · t ( x 0 , y 0 ) } = P [ z 2 ] { A } + P [ z 2 ] { A · Δ t ( x 0 , y 0 ) } = A + a ( x , y )
Here, P [ z 2 ] { } is the transfer function acting on z 2 , ε and η are frequency domain coordinates, and λ is the wavelength of the light source.
P z 2 ( ε , η ) = { exp [ j · z 2 · 2 π λ · 1 ( λ ε ) 2 ( λ η ) 2 ] , ε 2 + η 2 < 1 λ 2 0
On the image sensor plane, the object light wave a ( x , y ) passing through the object interferes with the reference light wave A to form a hologram, and the intensity information I ( x , y ) is recorded:
I ( x , y ) = | A + a ( x , y ) | 2 = | A | 2 + A * · a ( x , y ) + A · a * ( x , y ) + | a ( x , y ) | 2
In (4), the second and third items are related to the reconstruction of object information. The first item is the reference light information, which can be removed by a sample-free background image, the fourth item is the object self-interference information, it can be ignored when ∆t<<1, because for holography, it does not include any useful information, and during the holographic reconstruction process, the focused image of the object and the out-of-focus image will overlap, and the twin image will form. In the on-chip holographic imaging, z 2   is usually very small. The twin image has a severe impact on the image of the actual object, so it needs to be removed by related methods.
For digital holographic imaging, according to the angular spectrum theory, sample information, including amplitude and phase, can be decoded by back propagation:
P [ z 2 ] { A * · I ( x , y ) } = | A | 2 · [ 1 + Δ t ( x , y ) + P [ 2 z 2 ] { Δ t * ( x , y ) } ] + P [ z 2 ] { | a ( x , y ) | 2 }
In (5), A * is the conjugate of the plane wave and I ( x , y ) is the recorded hologram. On the right side of the Equation (5), the first term is the background light, the second term is the actual sample, the third term is the twin image interference, and the fourth term is the self-interference term backpropagation interference. For smaller or more scattered samples, the twin image interference will not cause interference to the reconstructed image, and routine analysis can be satisfied by simple backpropagation. For other samples, the existence of the twin image will interfere with the analysis of the results, making the recovered sample information unclear. Therefore, proper methods are needed to reduce the influence of the twin image, such as the phase iterative recovery method.
Regarding the recorded hologram as just ordinary diffraction field intensity information, the phase information is lost. Mudanyali et al. proposed a phase iteration method, which uses a threshold definition method to determine the boundary of the sample object [37]. After backpropagation, the twin image interference information which is outside the edge of the object can be removed iteratively by imposing constraints. The method consists of the following steps:
  • The root mean square of the holographic image information I ( x , y ) recorded by the image sensor is propagated back to the object surface, the phase information on the object surface is assumed to be zero at this time, and the phase information is finally restored on the object surface through an iterative method. At the same time, the threshold segmentation method is used to determine the boundary of the sample object.
    U z 2 1 ( x , y ) = P z 2 [ I ( x , y ) ]
    In Equation (6), U z 2 1 represents the real image field distribution after the first iteration.
  • The information inside the boundary of the object is retained, and the information outside the boundary is replaced by the information after the backpropagation of the sample-free background image:
    U z 2 i + 1 ( x , y ) = { m · D z 2 ( x , y ) , x , y S U z 2 i ( x , y ) , x , y S
    In Equation (7), D z 2 ( x , y )   is obtained by backpropagating the root mean square of the background image without samples, m is the iteration coefficient, and U z 2 i represents the real image field distribution after the ith iteration.
    m = m e a n ( U z 2 i ( x , y ) m e a n ( D z 2 ( x , y ) ) )
  • The reconstructed field after being constrained in the second step is propagated forward to the surface of the image sensor. At this time, the phase value is no longer zero, and the phase value is retained. The amplitude value is determined by the root mean square band of the original recorded holographic image, U 0 0 ( x , y ) , amplitude. The diffraction field, U 0 i ( x , y ) , after the ith iteration is expressed as follows:
    U 0 i ( x , y ) = | U 0 0 ( x , y ) | · exp ( j · ϕ 0 i ( x , y ) )
    where ϕ 0 i ( x , y ) denotes the phase of the field after the ith iteration.
Steps 1 to 3 can be iterated until the phase recovery converges. Typically, the results are obtained with less than 15 iterations, and the results are shown in Figure 2b.

3. Experiments and Results

3.1. Image Quality Assessment

In order to test the performance of the algorithm, we used two imaging methods, shadow imaging and holographic imaging, to shoot experiments on the United States Air Force (USAF) resolution plate. The experimental conditions were set with the same parameters. The center wavelength of the light source is 587 nm, the distance between the USAF resolution plate and the image sensor is 0.5 mm, and the pixel size is 1.12 μm. In addition, we also used a 10× objective microscope to photograph the same area of the USAF resolution plate as a comparison image. The comparison of the holographic image after iterative phase recovered, shadow image, and 10× ocular lens is shown in Figure 3.
In order to more intuitively evaluate the image quality of different imaging methods, we introduced image quality evaluation algorithms, which are divided into non-reference image quality evaluation algorithm [38] and reference-quality evaluation algorithm [39].
  • Non-reference image quality evaluation
In the non-reference image quality evaluation, we used the Brisque algorithm to perform global scoring and evaluation on the holographic restored image and the shadow image. The overall principle of the algorithm is to extract the MSCN (Mean Subtracted Contrast Normalized) coefficients from the normalized image and fit the MSCN coefficients to an asymmetric generalized Gaussian distribution, extract the fitted Gaussian distribution feature, then input it into the support vector machine (SVM) for regression calculation, thereby obtaining the image quality evaluation result, and the score results are shown in Figure 3a,b. Based on the Brisque algorithm, lower score indicates better quality.
  • Reference image quality evaluation
In the reference image quality evaluation, we used two evaluation parameters, namely the blur coefficient ( K b l u r ) and quality index ( Q ). The 10× microscope image is used as the reference image, and the holographic recovered image and shadow image are used as the images to be evaluated. Among them, the blur coefficient is defined as follows:
K b l u r = S i , o u t S i , i n
Among them, the energy characteristic of the oblique edge, S i , is
S i = i j | y f i ( i , j , k ) |
In (11), y f i ( i , j , k )   is the value obtained after the brightness value of the kth frame, the ith row, and the jth column, which are processed by an oblique space differential filter.
It can be seen that K b l u r is the ratio of the output edge energy to the input edge energy. Assuming that the image sequence has no other distortion, the range of value K b l u r is generally between 0 and 1. The closer the K b l u r value is to 1, the higher the image clarity. The result of K b l u r is shown in Table 1. The K b l u r value 0.9738 of the recovered hologram is between 0 and 1 and is closer to 1, so the quality is better.
The quality index is defined as follows:
Q = 4 σ x y x ¯ y ¯ ( σ x 2 + σ y 2 ) [ ( x ¯ ) 2 + ( y ¯ ) 2 ]
In (12), where x and y are the reference image and the image to be tested respectively, x ¯ and y ¯ are pixel mean value, σ x 2 and σ y 2 are pixel variance, and σ x y is the standard deviation. The value range of the quality index Q is usually between [−1, 1], the closer to 1, the higher the image quality. The result of Q is shown in Table 1. The Q value of the shadow image and the recovered hologram image are both between −1 and 1, but the Q value of the recovered hologram image 0.2174 is closer to 1, so the quality of the holographic restored image is higher.

3.2. Cell Analysis

In the miniaturized on-chip biological detection equipment, the cell solution was dropped on a cover glass with a thickness of 0.5 mm through a pipette, and a glass smear was made and placed above the image sensor. Under the irradiation of an incoherent LED light source, the holographic diffraction pattern of the cells was recorded by the image sensor. In order to reduce the volume and cost of the equipment, we used ordinary yellow LED as the light source, and used 3D printing technology to build a dark box with a volume of 6.5 × 6.5 × 7 cm3. The light source was fixed to the pinhole, and the image sensor was placed at the bottom of the box. The smear was attached to the image sensor by liquid glue. After iteratively recovering the captured cell holographic image using the above image recovery algorithm, the Fiji tool is introduced to classify, identify, and count the cells.
Classification and recognition are performed by taking an image containing different cells and use the Trainable Weka Segmentation tool to mark the different cells in color [40]. Trainable Weka Segmentation is a plug-in of Fiji that combines a set of machine learning algorithms and a set of selected image features to generate pixel-based segmentation. Weka includes a set of visualization tools and algorithms for data analysis and predictive modeling and a graphical user interface (GUI) for easy access to this feature. Next, the cell images are binarized, the adhered cells are segmented using the watershed segmentation method, and then the Cell Counter plug-in is used to count the cells of the entire image and display the counting results.

3.3. Discussion

In order to test the imaging performance of the designed holographic imaging system, we captured the holographic images of mixed solution of mouse ascites tumor cells and polystyrene particles (15 μm). The same chosen area was shown for comparison. Figure 4a is a microscopic image under 10× objective. In this figure, the arrows point to the polystyrene particles, and the dotted circles indicate the cells. The holographic imaging experiment process was as follows: First, a background image was taken without any sample solution to use the phase recovery algorithm to remove the background, second, a pipette was used to drop 10 μL of the mixed solution onto a cover glass with a thickness of 0.5 mm and generate the cell smears, from which the holographic imaging device would capture the images of mixed sample solution (Figure 4b). Third, the self-designed holographic phase recovery algorithm based on Matlab was employed to reconstruct the captured holographic image (Figure 4c). The results show that in the cell image after holographic phase recovery, the gray value of the low refractive index cells is generally higher than that of the high refractive index polystyrene particles, and the contour and sharpness of the cell edges is better recovered. The inset in Figure 4d, enclosed with the dashed rectangle, shows the polystyrene particle and cells after iterative phase recovery algorithm, further validating the imaging performance of our holographic microscope.
The image processing flow consists of two steps: The first step is cell identification and segmentation. We used Fiji’s Trainable Weka Segmentation plug-in to train the classifier. A small number of cells, particles, and backgrounds in Figure 4c are first selected and set into three different labels. Each label generally contains sampling information of 3–4 different positions. The classification results can be obtained through training, as shown in Figure 5a. Among them, the red parts represent polystyrene particles, the green parts represent mouse ascites tumor cells, and the purple parts represent the background. The second step is the cell counting. First, the particles and cells with incomplete edges in Figure 5a were removed by setting a threshold, then we binarized the image and filled in the holes, then broke the adherent cells, and finally counted the remaining cells in the image. The counting result is shown in Figure 5b. It can be seen from the results that the number of cells is 62, and the count value is marked inside the cell edge. The count result and manual count result of reference are similar to each other, which preliminarily verifies the accuracy of the on-chip holographic microscope.

4. Conclusions

In this paper, we demonstrated a lensless on-chip holographic imaging platform using in-line holography technology and used an iterative phase recovery algorithm to achieve the phase recovery of the holographic image. The resolution of the reconstructed hologram image was improved by 1.41 times compared with the traditional shadow imaging method, and the image sharpness was also improved. In the reconstructed hologram (Figure 3a), all the bars of group 6, element 4, can be identified (90.5 Line Pairs/mm), but in the traditional shadow image (Figure 3b), only group 6, element 1, can be completely identified (64 Line Pairs/mm), and the edge is out of focus. Moreover, compared with the traditional manual counting of blood smears, this paper realizes the function of automatically identifying and counting cells. Finally, because the system is small in size and low in price, it can be promoted and used in poor and remote areas and has the application prospect of point-of-care testing (POCT).

Author Contributions

Conceptualization, X.H. and Y.L.; Data curation, X.H., Y.L. and W.H.; Formal analysis, X.H. and Y.L.; Funding acquisition, X.H. and L.S.; Investigation, Y.L., X.X., R.W., W.H., M.W. and J.C.; Methodology, X.H., Y.L. and X.X.; Project administration, X.H.; Resources, X.H. and W.X.; Software, J.Y. and M.W.; Supervision, X.H.; Validation, X.X., R.W., J.Y., W.H. and J.C.; Writing—original draft, Y.L.; Writing—review and editing, X.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (Grant No. 61827806), Qianjiang Talent Project Type-D of Zhejiang (Grant No. QJD1802021), Fundamental Research Funds for the Provincial Universities of Zhejiang (Grant No. GK209907299001-305), and 2020 Talent Cultivation Project by Zhejiang Association for Science and Technology (Grant No. CTZB-2020080127-19).

Institutional Review Board Statement

The mouse ascites tumor cell line was obtained from China Center for Type Culture Collection (CCTCC) (Hubei, China).

Informed Consent Statement

Not applicable.

Data Availability Statement

Data available on request due to privacy restrictions.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wang, Z.; Millet, L.; Mir, M.; Ding, H.; Popescu, G. Spatial light interference microscopy (SLIM). Opt. Express 2011, 19, 1016–1026. [Google Scholar] [CrossRef]
  2. Garcia-Sucerquia, J.; Xu, W.; Jericho, S.K.; Klages, P.; Jericho, M.H.; Kreuzer, H.J. Digital in-line holographic microscopy. Appl. Opt. 2006, 45, 836–850. [Google Scholar] [CrossRef] [PubMed]
  3. Cuche, E.; Bevilacqua, F.; Depeursinge, C. Digital holography for quantitative phase-contrast imaging. Opt. Lett. 1999, 24, 291–293. [Google Scholar] [CrossRef] [PubMed]
  4. Xu, W. Digital in-line holography for biological applications. Proc. Natl. Acad. Sci. USA 2001, 98, 11301–11305. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Waller, L.; Luo, Y.; Yang, S.Y.; Barbastathis, G. Transport of intensity phase imaging in a volume holographic microscope. Opt. Lett. 2010, 35, 2961–2963. [Google Scholar] [CrossRef] [PubMed]
  6. Leith, E.N.; Upatnieks, J.; Haines, K.A. Microscopy by Wavefront Reconstruction. JOSA 1965, 55, 981–986. [Google Scholar] [CrossRef]
  7. Schmitt, J.M. Optical coherence tomography (OCT): A review. IEEE J. Sel. Top. Quantum Electron. 2002, 5, 1205–1215. [Google Scholar] [CrossRef] [Green Version]
  8. Fujimoto, G. James, Optical coherence tomography for ultrahigh resolution in vivo imaging. Nat. Biotechnol. 2003, 21, 1361–1367. [Google Scholar] [CrossRef]
  9. Sharpe, J.; Ahlgren, U.; Perry, P.; Hill, B.; Ross, A.; Hecksher-Sørensen, J.; Baldock, R.; Davidson, D. Optical projection tomography as a tool for 3D microscopy and gene expression studies. Science 2002, 296, 541–545. [Google Scholar] [CrossRef] [Green Version]
  10. Hell, S.W.; Wichmann, J. Breaking the diffraction resolution limit by stimulated emission: Stimulated-emission-depletion fluorescence microscopy. Opt. Lett. 1994, 19, 780–782. [Google Scholar] [CrossRef]
  11. Betzig, E.; Patterson, G.; Sougrat, R.; Lindwasser, O.; Olenych, S.; Bonifacino, J.; Davidson, M.; Lippincott-Schwartz, J.; Hess, H. Imaging intracellular fluorescent proteins at nanometer resolution. Science 2006, 313, 1642–1645. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  12. Yager, P.; Domingo, G.J.; Gerdes, J. Point-of-care diagnostics for global health. Annu. Rev. Biomed. Eng. 2008, 10, 107. [Google Scholar] [CrossRef] [PubMed]
  13. Wongsrichanalai, C.; Barcus, M.J.; Muth, S.; Sutamihardja, A.; Wernsdorfer, W.H. A review of malaria diagnostic tools: Microscopy and rapid diagnostic test (RDT). Am. J. Trop. Med. Hyg. 2007, 77, 119–127. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Frenette, P.S.; Atweh, G.F. Sickle cell disease: Old discoveries, new concepts, and future promise. J. Clin. Investig. 2007, 117, 850–858. [Google Scholar] [CrossRef] [PubMed]
  15. Göröcs, Z.; Ozcan, A. On-chip biomedical imaging. IEEE Rev. Biomed. Eng. 2012, 6, 29–46. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Greenbaum, A.; Luo, W.; Su, T.-W.; Göröcs, Z.; Xue, L.; Isikman, S.O.; Coskun, A.F.; Mudanyali, O.; Ozcan, A. Imaging without lenses: Achievements and remaining challenges of wide-field on-chip microscopy. Nat. Methods 2012, 9, 889–895. [Google Scholar] [CrossRef]
  17. Isikman, S.O.; Bishara, W.; Mudanyali, O.; Sencan, I.; Su, T.-W.; Tseng, D.K.; Yaglidere, O.; Sikora, U.; Ozcan, A. Lensfree on-chip microscopy and tomography for biomedical applications. IEEE J. Sel. Top. Quantum Electron. 2011, 18, 1059–1072. [Google Scholar] [CrossRef] [Green Version]
  18. Huang, X.; Xu, D.; Chen, J.; Liu, J.; Li, Y.; Song, J.; Ma, X.; Guo, J. Smartphone-based analytical biosensors. Analyst 2018, 143, 5339–5351. [Google Scholar] [CrossRef]
  19. Huang, X.; Li, Y.; Chen, J.; Liu, J.; Wang, R.; Xu, X.; Yao, J.; Guo, J. Smartphone-based blood lipid data acquisition for cardiovascular disease management in internet of medical things. IEEE Access 2019, 7, 75276–75283. [Google Scholar] [CrossRef]
  20. Xu, D.; Huang, X.; Guo, J.; Ma, X. Automatic smartphone-based microfluidic biosensor system at the point of care. Biosens. Bioelectron. 2018, 110, 78–88. [Google Scholar] [CrossRef]
  21. Yuan, X.; Pu, Y. Parallel lensless compressive imaging via deep convolutional neural networks. Opt. Express 2018, 26, 1962–1977. [Google Scholar] [CrossRef] [PubMed]
  22. Alexandrov, S.A.; Hillman, T.R.; Gutzler, T.; Sampson, D.D. Synthetic aperture Fourier holographic optical microscopy. Phys. Rev. Lett. 2006, 97, 168102. [Google Scholar] [CrossRef] [PubMed]
  23. Luo, W.; Greenbaum, A.; Zhang, Y.; Ozcan, A. Synthetic aperture-based on-chip microscopy. Light Sci. Appl. 2015, 4, e261. [Google Scholar] [CrossRef]
  24. Delacroix, R.; Morel, S.N.A.; Hervé, L.; Bordy, T.; Dinten, J.-M.; Drancourt, M.; Allier, C. Cerebrospinal fluid lens-free microscopy: A new tool for the laboratory diagnosis of meningitis. Sci. Rep. 2017, 7, 39893. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  25. Wei, Q.; McLeod, E.; Qi, H.; Wan, Z.; Sun, R.; Ozcan, A. Lensfree holographic cytometry using plasmonic nanoparticles. In Proceedings of the IEEE Photonics Conference, Bellevue, WA, USA, 8–12 September 2013; pp. 3–4. [Google Scholar]
  26. Huang, X.; Guo, J.; Wang, X.; Yan, M.; Kang, Y.; Yu, H. A contact-imaging based microfluidic cytometer with machine-learning for single-frame super-resolution processing. PLoS ONE 2014, 9, e104539. [Google Scholar] [CrossRef] [Green Version]
  27. Zheng, G.; Lee, S.A.; Antebi, Y.; Elowitz, M.B.; Yang, C. The ePetri dish, an on-chip cell imaging platform based on subpixel perspective sweeping microscopy (SPSM). Proc. Natl. Acad. Sci. USA 2011, 108, 16889–16894. [Google Scholar] [CrossRef] [Green Version]
  28. Coskun, A.F.; Su, T.-W.; Ozcan, A. Wide field-of-view lens-free fluorescent imaging on a chip. Lab Chip 2010, 10, 824–827. [Google Scholar] [CrossRef] [Green Version]
  29. Seo, S.; Su, T.-W.; Tseng, D.K.; Erlinger, A.; Ozcan, A. Lensfree holographic imaging for on-chip cytometry and diagnostics. Lab Chip 2009, 9, 777–787. [Google Scholar] [CrossRef]
  30. Vercruysse, D.; Dusa, A.; Stahl, R.; Vanmeerbeeck, G.; de Wijs, K.; Liu, C.; Prodanov, D.; Peumans, P.; Lagae, L. Three-part differential of unlabeled leukocytes with a compact lens-free imaging flow cytometer. Lab Chip 2015, 15, 1123–1132. [Google Scholar] [CrossRef]
  31. Wu, Y.; Luo, Y.; Chaudhari, G.; Rivenson, Y.; Calis, A.; de Haan, K.; Ozcan, A. Bright-field holography: Cross-modality deep learning enables snapshot 3D imaging with bright-field contrast using a single hologram. Light Sci. Appl. 2019, 8, 1–7. [Google Scholar] [CrossRef] [Green Version]
  32. Fang, Y.; Yu, N.; Jiang, Y.; Dang, C. High-precision lens-less flow cytometer on a chip. Micromachines 2018, 9, 227. [Google Scholar] [CrossRef] [Green Version]
  33. Rivenson, Y.; Wu, Y.; Ozcan, A. Deep learning in holography and coherent imaging. Light Sci. Appl. 2019, 8, 1–8. [Google Scholar] [CrossRef] [Green Version]
  34. Ray, A.; Khalid, M.A.; Demčenko, A.; Daloglu, M.; Tseng, D.; Reboud, J.; Cooper, J.M.; Ozcan, A. Holographic detection of nanoparticles using acoustically actuated nanolenses. Nat. Commun. 2020, 11, 1–10. [Google Scholar] [CrossRef] [PubMed]
  35. Liu, T.; Wei, Z.; Rivenson, Y.; de Haan, K.; Zhang, Y.; Wu, Y.; Ozcan, A. Deep learning-based color holographic microscopy. J. Biophotonics 2019, 12, e201900107. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  36. Gabor, D.; Kock, W.E.; Stroke, G.W. Holography. Science 1971, 173, 11–23. [Google Scholar] [CrossRef] [PubMed]
  37. Mudanyali, O.; Tseng, D.; Oh, C.; Isikman, S.O.; Sencan, I.; Bishara, W.; Oztoprak, C.; Seo, S.; Khademhosseini, B.; Ozcan, A. Compact, light-weight and cost-effective microscope based on lensless incoherent holography for telemedicine applications. Lab Chip 2010, 10, 1417–1428. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  38. Mittal, A.; Moorthy, A.K.; Bovik, A.C. Blind/Referenceless Image Spatial Quality Evaluator. In Proceedings of the 2011 Conference Record of the Forty Fifth Asilomar Conference on Signals, Systems and Computers (ASILOMAR), Pacific Grove, CA, USA, 6–9 November 2011. [Google Scholar]
  39. Lee, C.; Cho, S.; Choe, J.; Jeong, T.; Ahn, W.; Lee, E. Objective video quality assessment. Opt. Eng. 2006, 45, 17004. [Google Scholar] [CrossRef]
  40. Arganda-Carreras, I.; Kaynig, V.; Rueden, C.; Eliceiri, K.W.; Schindelin, J.; Cardona, A.; Seung, H.S. Trainable Weka Segmentation: A machine learning tool for microscopy pixel classification. Bioinformatics 2017, 33, 2424–2426. [Google Scholar] [CrossRef]
Figure 1. Schematic of the lensless on-chip microscope. (a) Photographs of the apparatus, (b) structure of in-line holographic imaging device.
Figure 1. Schematic of the lensless on-chip microscope. (a) Photographs of the apparatus, (b) structure of in-line holographic imaging device.
Sensors 21 00720 g001
Figure 2. The results of iterative phase recovery. (a) Original United States Air Force (USAF) holographic image, (b) iteratively recovered image.
Figure 2. The results of iterative phase recovery. (a) Original United States Air Force (USAF) holographic image, (b) iteratively recovered image.
Sensors 21 00720 g002
Figure 3. Comparison of images taken by different systems. (a) Iteratively recovered image, (b) shadow image, (c) microscope image.
Figure 3. Comparison of images taken by different systems. (a) Iteratively recovered image, (b) shadow image, (c) microscope image.
Sensors 21 00720 g003
Figure 4. Results of the holographic imaging experiment process. (a) Microscope image, (b) raw holographic image, (c) iteratively recovered image, (d) insets of the iteratively recovered image.
Figure 4. Results of the holographic imaging experiment process. (a) Microscope image, (b) raw holographic image, (c) iteratively recovered image, (d) insets of the iteratively recovered image.
Sensors 21 00720 g004
Figure 5. Results of the image process: (a) recognition results, and (b) counting results of the cell.
Figure 5. Results of the image process: (a) recognition results, and (b) counting results of the cell.
Sensors 21 00720 g005
Table 1. The results of reference image quality evaluation.
Table 1. The results of reference image quality evaluation.
Image K b l u r Q
Shadow2.18410.1648
Recovered hologram0.97380.2174
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Huang, X.; Li, Y.; Xu, X.; Wang, R.; Yao, J.; Han, W.; Wei, M.; Chen, J.; Xuan, W.; Sun, L. High-Precision Lensless Microscope on a Chip Based on In-Line Holographic Imaging. Sensors 2021, 21, 720. https://doi.org/10.3390/s21030720

AMA Style

Huang X, Li Y, Xu X, Wang R, Yao J, Han W, Wei M, Chen J, Xuan W, Sun L. High-Precision Lensless Microscope on a Chip Based on In-Line Holographic Imaging. Sensors. 2021; 21(3):720. https://doi.org/10.3390/s21030720

Chicago/Turabian Style

Huang, Xiwei, Yangbo Li, Xuefeng Xu, Renjie Wang, Jiangfan Yao, Wentao Han, Maoyu Wei, Jin Chen, Weipeng Xuan, and Lingling Sun. 2021. "High-Precision Lensless Microscope on a Chip Based on In-Line Holographic Imaging" Sensors 21, no. 3: 720. https://doi.org/10.3390/s21030720

APA Style

Huang, X., Li, Y., Xu, X., Wang, R., Yao, J., Han, W., Wei, M., Chen, J., Xuan, W., & Sun, L. (2021). High-Precision Lensless Microscope on a Chip Based on In-Line Holographic Imaging. Sensors, 21(3), 720. https://doi.org/10.3390/s21030720

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop