EP3105712A1 - On-the-go touchless fingerprint scanner - Google Patents
On-the-go touchless fingerprint scannerInfo
- Publication number
- EP3105712A1 EP3105712A1 EP15749130.9A EP15749130A EP3105712A1 EP 3105712 A1 EP3105712 A1 EP 3105712A1 EP 15749130 A EP15749130 A EP 15749130A EP 3105712 A1 EP3105712 A1 EP 3105712A1
- Authority
- EP
- European Patent Office
- Prior art keywords
- hand
- scanning area
- fingerprint
- camera
- image
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Withdrawn
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/12—Fingerprints or palmprints
- G06V40/13—Sensors therefor
- G06V40/1312—Sensors therefor direct reading, e.g. contactless acquisition
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/74—Image or video pattern matching; Proximity measures in feature spaces
- G06V10/75—Organisation of the matching processes, e.g. simultaneous or sequential comparisons of image or video features; Coarse-fine approaches, e.g. multi-scale approaches; using context analysis; Selection of dictionaries
- G06V10/755—Deformable models or variational models, e.g. snakes or active contours
- G06V10/7553—Deformable models or variational models, e.g. snakes or active contours based on shape, e.g. active shape models [ASM]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/107—Static hand or arm
- G06V40/11—Hand-related biometrics; Hand pose recognition
Definitions
- the present disclosure generally relates to a biometric sensor, and more particularly, to an on-the-go, touchless fingerprint scanner.
- Biometric identity management is a foundational tool used by numerous security and law enforcement agencies to provide secured access as well as to identify and track individuals under surveillance. Additionally, civilian and other organizations looking to provide secured access to physical as well as virtual systems, locations, and data utilize biometric identity management systems (i.e., "biometrics"). To that end, various biometric identity management systems are employed including retinal scanning systems, face and voice recognition systems, DNA recognition, and fingerprint matching and verification systems.
- Fingerprint matching and verification form a foundational part of biometric identity management.
- Traditional fingerprint scanners require a subject to stop, place his hand on a scanner platen, or, an ink pad so that his fingerprints may be scanned. That is, the subject's fingers and/or hand must be static or motionless. Due to this and other constraints, traditional static fingerprint scanners suffer from long processing times.
- An on-the-go fingerprint scanner can include a scanning area and a beam break sensor disposed coincident to the scanning area such that the beam break sensor detects when an object passes through the scanning area.
- the scanner can also include at least one hand camera configured to capture a video image of the scanning area such that when the beam break sensor detects that an object has passed through the scanning area, the hand camera detects whether the object is a human hand and the location of any fingers on said hand, and at least one fingerprint camera configured to receive the location of any fingers within the scanning area from the hand camera and capture a high-resolution image of at least a portion of any of said fingers passing through the scanning area.
- an on-the-go fingerprint scanner comprises a scanning area, a beam break sensor disposed coincident to the scanning area such that the beam break sensor detects when an object passes through the scanning area, and at least one hand camera configured to capture a video image of the scanning area.
- the hand camera can capture the video image such that when the beam break sensor detects that an object has passed through the scanning area, the hand camera acquires an image of the object, detects whether the object is a human hand by determining a contour around the perimeter of the object, applying landmark points to the contour, and comparing the contour to a previously trained Active Shape Model of a hand, and detects the location of any fingers on said detected hand by processing the location of the landmark points, and detects the presence or absence of any fingers using a neural network.
- the scanner can include at least one fingerprint camera configured to receive the location of any fingers within the scanning area from the hand camera and capture a plurality of high-resolution images of at least a portion of any of said fingers passing through the scanning area.
- a method of capturing a fingerprint on-the-go can include determining that an object is passing through a predefined scanning area, and acquiring an image of the object and detecting whether the object is a human hand by determining a contour around the perimeter of the object, applying landmark points to the contour, and comparing the contour to a previously trained Active Shape Model of a hand.
- the method can further include detecting the location of any fingers on said detected hand by processing the location of the landmark points, and detecting the presence or absence of any fingers by applying a neural network, and acquiring at least one high-resolution image of a tip portion of any detected finger.
- FIG. 1 is perspective view of an on-the-go fingerprint scanner
- FIG. 2 is a perspective view of the on-the-go fingerprint scanner of FIG. 1;
- FIG. 3 is a perspective view of a scanning area of the on-the-go fingerprint scanner of FIG. 1;
- FIG. 4 is a close-up view of an imaging subsystem of the on-the-go fingerprint scanner of FIG. 1;
- FIG. 5 is an exploded schematic view of an exemplary illumination source
- FIG. 6 is an example binarized and grey-level fingerprint produced by the on-the-go fingerprint scanner of FIG. 1;
- FIG. 7 is a block diagram of an example hand camera control and processing algorithm
- FIG. 8A is a block diagram of an example hand camera control and processing algorithm utilized to find a set of landmark points for a new image
- FIG. 8B is an example of a binarized image of a hand as used in the block diagram of FIG. 8A;
- FIG. 8C is an example landmark point and k-curvature rendering as used in the block diagram of FIG. 8A;
- FIG. 8D is an example k-Curvature Magnitude chart as used in the block diagram of FIG. 8A;
- FIG. 9 is a block diagram of an example hand camera control and processing algorithm applying an Active Shape Model to a new image
- FIG. 10A is a block diagram of an example hand camera control and processing algorithm to train an Active Shape Model
- FIG. 10B is an example array of training images for use in the block diagram of FIG. 10A;
- FIG. IOC is an example landmarks for use in the block diagram of FIG.
- FIG. 11 is a block diagram of an example hand camera control and processing algorithm to determine finger positions
- FIG. 12 is a block diagram of an example hand camera control and processing algorithm to detect missing fingers
- FIG. 13 is a block diagram of an example fingerprint camera control and processing algorithm
- FIG. 14A is a block diagram of an example fingerprint camera control and processing algorithm for finding the location of the four fingerprints in the first fingerprint camera image
- FIG. 14B is an example image used in the block diagram of FIG. 14A;
- FIG. 14C is an example image used in the block diagram of FIG. 14A;
- FIG. 14D is an example image used in the block diagram of FIG. 14A;
- FIG. 15A is a block diagram of an example fingerprint camera control and processing algorithm for finding the location of the four fingerprints in the subsequent fingerprint camera images
- FIG. 15B is an example image used in the block diagram of FIG. 15 A;
- FIG. 15C is an example image used in the block diagram of FIG. 15 A;
- FIG. 16A is a block diagram of an example fingerprint camera control and processing algorithm for finding the highest focus print for all four fingers
- FIG. 16B is an example image used in the block diagram of FIG. 16A;
- FIG. 17A is a block diagram of an example fingerprint camera control and processing algorithm for determining whether the print is a valid fingerprint
- FIG. 17B is an example image used in the block diagram of FIG. 17A;
- FIG. 18A is a block diagram of an example fingerprint camera control and processing algorithm for creating a binary print
- FIG. 18B is an example image used in the block diagram of FIG. 18 A;
- FIG. 18C is an example binary print produced with the block diagram of FIG. 18 A;
- FIG. 19A is a block diagram of an example fingerprint camera control and processing algorithm for creating a grey- level print
- FIG. 18B is an example image used in the block diagram of FIG. 19A;
- FIG. 18C is an example grey- level print produced with the block diagram of FIG. 19A.
- FIG. 20 is a schematic of an example on-the-go fingerprint scanner.
- like-numbered components generally have similar features, and thus each feature of each like-numbered component is not necessarily fully elaborated upon.
- linear or circular dimensions are used in the description of the disclosed systems, devices, and methods, such dimensions are not intended to limit the types of shapes that can be used in conjunction with such systems, devices, and methods.
- a person skilled in the art will recognize that an equivalent to such linear and circular dimensions can be determined for any geometric shape. Sizes and shapes of the systems and devices, and the components thereof, can depend at least on the size and shape of the components with which the systems and devices will be used, and the methods and procedures in which the systems and devices will be used.
- Biometric scanning is a critical tool utilized by security and law enforcement agencies, such as the Federal Bureau of
- on-the-go fingerprint scanners and methods of capturing fingerprints on-the-go provide fast and reliable scanning.
- "On-the-go” also referred to as “on the go” or “OTG”
- OOG on the go
- a desired hand or finger to be imaged and printed is in motion (i.e., dynamic or not static). That is, a subject can either be moving his or her hand, fingers, or the entire subject can be in motion, such as when a person is walking or being pushed in a wheelchair.
- the FBI and other organizations have established that fingerprint scanners should scan and register a minimum of 30 subjects per minute.
- the on-the-go fingerprint scanners and methods described herein can achieve this minimum of 30 subjects per minute and, in some aspects, the on-the-go fingerprint scanners and methods disclosed herein can reliably and accurately provide fingerprint scans of more than approximately 50 subjects per minute.
- example on-the-go fingerprint scanners can include a scanning area, a beam break sensor, and a variety of cameras for scanning a subject's fingerprints.
- on-the-go fingerprint scanners can include a fingerprint camera (also referred to as "print camera”), one or more range sensors, one or more illumination sources, a second camera (the "hand camera” or “guidance camera”), and a display.
- On-the-go scanners can also include a data acquisition and processing platform, and software and algorithms for image processing.
- FIGS. 1-3 illustrate an on-the-go fingerprint scanner.
- the on- the-go fingerprint scanner 100 includes an entrance frame 102 and an imaging frame 104.
- the entrance frame defines a scanning area 106 and contains a beam break sensor 108 that is disposed coincidentally to the scanning area 106.
- the imaging frame 104 holds a lighting system 110, a fingerprint camera 112, and a hand camera system 114.
- electronic systems can be contained within the frames 102, 104 or otherwise be operatively coupled thereto.
- the entrance frame 102 and the imaging frame 104 can be separated by a frame member 103 such that the entrance frame 102 and the imaging frame 104 are coupled to one another to maintain a specific distance. In other aspects, however, the imaging frame and entrance frame can be separated without being mechanically connected to one another. In either aspect, the imaging frame and the entrance frame can be any desired distance apart so as to allow optimum imaging of the scanning area.
- the imaging frame 104 can be about 5 meters or less from the entrance frame 102. In some aspects, for example, the imaging frame can be about 1 meter from the entrance frame 102.
- the entrance frame 102 can physically define the scanning area 106, as is shown in FIGS. 1-3.
- the entrance frame 102 can optionally hold a visual display 116 and a lane control light 118.
- the visual display 116 can be any display including a computer monitor, television, touch-screen monitor, tablet, or other display device.
- the display 116 can be used to provide instructions to a user of the fingerprint scanner, display advertising, or to display any other visual representation desired.
- the scanning area 106 can be defined by the entrance frame 102, as shown in FIGS. 1-3. That is, the scanning area 106 can be a physical area that is bounded by portions of the entrance frame 102A, 102B, and 102C. Alternatively, the scanning area 106 can be a point, plane, and/or region in space that is not directly defined by the entrance frame 102.
- the fingerprint scanner can be incorporated into a system that lacks a defined entrance frame, such as an x-ray or back-scatter system used in airport security.
- the entrance frame and scanning area can be a portion of the existing x-ray or backscatter machine.
- the scanning area can be the area above a subjects head such that the beam break sensor is triggered as the subject raises his or her hands above their head in the normal course of operating the back-scatter system.
- the beam break sensor 108 can include a commercial beam break system 120, a cueing light 122, and a cover 124.
- the beam break system can be, but is not limited to, a lighting source on one end, and a detector strip on the other.
- the beam break sensor 108 is an infra-red beam break sensor.
- the beam break sensor 108 can be disposed coincidentally to the scanning area. That is, the beam break sensor 108 can be disposed such that the beam break sensor 108 is triggered slightly before a subject's hand is in focus of the hand camera.
- the hand camera can— as is discussed in detail below— determine if the object is a hand, and the fingerprint camera can begin to capture images before the hand is in focus as the hand passes through the scanning area. This can allow the fingerprint camera algorithms to monitor the focus of each finger— as detailed below— identifying which frame is best focused for each finger.
- the beam break sensor 108 is aligned longitudinally to the scanning area.
- the cueing light 122 can be, but is not limited to, an LED strip, with multiple color LEDs.
- the cover 204 can be transparent to allow the cueing light 122 to shine therethrough and soft such that if a user accidentally strikes the cover 124, the user's hand is not injured and does not cause pain. [0056] As shown in FIG.
- the imaging frame 404 can include a mounting plate 426, an illumination source 428, and a plurality of cameras 430, 432. Furthermore, as described above, the imaging frame 404 can include any number of electronic systems—such as a power supply, controller, memory, and/or processor— that are configured to operate the illumination source, cameras, and any other on-the-go fingerprint scanner components described herein.
- a power supply such as a power supply, controller, memory, and/or processor— that are configured to operate the illumination source, cameras, and any other on-the-go fingerprint scanner components described herein.
- the fingerprint camera 432 can be of sufficient frame rate, resolution, and size to acquire sufficient detail of a fingerprint at a range of, for example, 10 meters or less.
- the fingerprint camera can be of sufficient frame rate, resolution, and size to acquire sufficient detail of a fingerprint at a range of about 5 meters or less, about 1 meter or less, about 0.75 meters or less, about 0.5 meters or less, and/or any other desired distance.
- the FBI requires that the fingerprint camera produce at least 500 pixels per inch (“PPI") of object, and that a modulation transfer function of the camera and associated lens provide at sufficient contrast at a spatial frequency of about 9.8 lp/mm on the object.
- the fingerprint camera can produce images having between about 100 pixels per inch (“PPI”) and 10000 PPI, for example 1000 PPI.
- the camera frame rate can be any desired frame rate, for example it can be such that the depth of field of the camera is equal to or greater than the distance traveled by the subject's hand between successive frames. Depth of field is determined by pixel size, distance, and f/#. A larger f/# (smaller aperture) leads to a deeper depth of field (until the diffraction limit is reached). For example and without limitation, with a 5 micron pixel size, 100 cm focal length, at f/8, the depth of field is approximately 0.75 cm.
- a frame rate of 150 frames/second can be used to capture sufficient in-focus images.
- the camera frame rate can be in the range of approximately 20 fps to approximately 300 fps.
- suitable frame rates can include 24 fps, 25 fps, 30 fps, 48 fps, 90 fps, 100 fps, 120 fps, 125 fps, 150 fps, and 240 fps.
- the fingerprint camera i.e., "Print Camera”
- the fingerprint camera is a visible, grayscale, 4096 x 3072 camera, which acquires images at approximately 125 fps.
- a 12 Megapixel camera with pixels of about 5.0 microns in height and width
- fps frames per second
- the camera can include a 100 mm focal length lens operating at f/8 and the plane of focus can be located approximately 750 mm from the entrance pupil of the lens.
- the fingerprint camera can utilize an exposure time short enough to minimize the impact of lateral motion on the modulation transfer function of the system. Additionally, the fingerprint camera can utilize an exposure time short enough to minimize the impact of apparent lateral motion due to magnification change during the exposure time at the limit of the field of view. Lateral motion does, however, blur the image, but it does not affect the system's modulation transfer function as the system is configured so that the lateral motion's blur is lower than the blur caused by the system's modulation transfer function, so that image quality is not affected by lateral motion.
- the fingerprint camera can capture images during the hand traversal of the plane of focus (i.e., as the hand travels through the scanning area).
- the frame rate of the camera can be sufficient such that the best focused image of each finger occurs within a distance equivalent to a change in magnification of one percent. That is, in some aspects, the scale of the fingerprint with respect to the pixel grid is known and the magnification of the image from the object plane (the plane of focus) to the image does not change by more than one percent in order to meet certain standards in fingerprint acquisition. Thus, the plane of focus in object space can be pre-defined and known to better than one percent of the nominal range. In practice, the subject's hand might be presented to the system at an angle to the plane that is in focus.
- the on-the-go fingerprint scanners and methods disclosed herein can ensure that the fingerprint camera takes a plurality of images during the correct window in time by detecting the range of the subject's hand from the plane of focus using one of several means of range detection. For example, when the hand reaches the distal edge of the acquisition region (i.e., scanning area), the range sensor can trigger the illumination and camera frame sequence. Thus, in some aspects, when the hand leaves the region of acquisition, the illumination turns off and the camera ceases frame acquisition.
- the range sensing can be provided by an infrared range sensor, and/or, an acoustic range sensor. In some aspects, the beam brake sensor described above is the range sensor.
- the hand camera 430 can be a video camera and can also include an infrared ("IR") lighting system. Furthermore, in some aspects, the hand camera can be configured to function as a hand guidance system. In this aspect, the hand camera images the hand and displays both the hand and an overlay on the display. The overlay corresponds to the field of view of the fingerprint camera such that if the subject's hand aligns with the overlay, then the hand will be in the acceptable region for data acquisition by the fingerprint camera. Although this aspect is optional and not needed in all aspects, the display with the overlay and hand image allow the subject to self-correct the position of the hand as the subject approaches the data collection region.
- IR infrared
- the illumination source 428 can be configured to illuminate the scanning area 106 such that a subject's hand and fingers are illuminated sufficiently to allow a camera to take an image with acceptable signal to noise ratio.
- the illumination source can include a high power, extended light source 500, a compound hyperbolic concentrator (CHC) 502, a lens 504, and a diffuser 506.
- the illumination subsystem may also include a baffle 508.
- the light source can be narrow band or broad band visible source and could be an LED array, halogen, or other source of sufficient intensity and spatial distribution.
- the illumination is provided by two LED lamps that produce at least 100 klux over an area sufficient to encompass a complete hand.
- the concentrator 502 is a reflecting, imaging form which exhibits particular properties.
- the concentrator 502 collects a high percentage of the light emitted from the source. It creates a virtual image of the emitting source in which the edges of the image are in focus while the interior of the image is out of focus. This virtual image is then imaged by the lens 504 to the field position.
- the diffuser 506 further diffuses the interior image to create uniform illumination over the field while minimizing light distributed outside the intended field.
- the baffle 508 further minimizes the possibility of light scattered from any surface in the illumination subsystem from entering a subject's or bystander's eyes.
- Deconvolution may be used to extend the effective depth of field, but in many optical systems this is limited by uncertainty in the range dependent point spread function.
- this system includes a range sensor, and can, in principle, be configured to track the range of each finger, we also recognize that a range dependent deconvolution operation may be performed in order to extend the depth of field, and consequently reduce the frame rate demands on the camera.
- a large field of view can be desirable. This may be accomplished by several means, the simplest of which is using multiple imagers synced, and stitched such that an aggregate field of view may be obtained. In addition, using multiple cameras, cameras may be set to acquire data at different focal planes, thus trading frame rate for a multiplicity of sensors.
- the on-the-go fingerprint scanner can produce, as shown in FIG. 6, a binarized image 602 of the subject's finger print(s), a grey- level image 604 of the subject's finger print(s), or both.
- an approaching subject passes his/her hand through the scanning area. This activates the beam break sensor.
- the beam break sensor is triggered and sends an electrical signal to the electronic system.
- the electronic system switches the color of the cueing light to red, indicating a need for the next person to wait. After the fingerprint camera completes its tasks, the electronic system can switch the cueing light back to green.
- the beam break sensor also triggers the electronic system to examine the last frame from the hand camera video image. If the electronic system determines that the frame contains a hand, it turns on the lighting system 428 and instructs the fingerprint camera 106 to capture at least one image, or, in some aspects, a sequence of images. The resulting fingerprint image or plurality of images is processed by the electronic system using software described in more detail below.
- the electronic system implements various algorithms to process the images received from the hand camera, and the fingerprint camera if it is determined that a hand and fingers are present.
- these various algorithms are described in more detail in FIGS. 7-19 below, but a person of ordinary skill in the art will understand that these algorithms are by example only and are not limiting.
- the various system components can perform various aspects of the below described algorithm(s) utilizing electronic systems (such as a processor, controller, and memory) that are localized on each individual system component (i.e., the hand camera), or a system-wide controller, processor, and/or memory can operate, control, and perform the algorithm(s) described herein.
- FIG. 1 FIG.
- FIG. 7 illustrates an algorithm used to process the last frame from the hand camera after the beam brake sensor signals an object is in the scanning area.
- the hand camera acquires a pre-processed image of the object from the last frame of the video image.
- the system then rotates this image so that the fingers are pointed in an upward direction and masks out any non-hand guide pixels.
- FIGS. 8A-8D the system obtains landmark points around the contour of the perimeter of the object.
- the algorithm automatically selects landmark points around the perimeter, or contour of the object or hand. These landmark points are then— as shown in FIG. 9— compared to a previously trained Active Shape Model of a hand to determine if the acquired object is a hand.
- An “Active Shape Model” is a set of landmarks for a mean hand shape plus variation about that shape for each principal component.
- the "Active Shape Model” can be preprogramed into the system, or can be created using the hand camera as described in FIGS. lOA-lOC.
- the landmark points also define the locations of the fingertips that are communicated to the Print Camera algorithms.
- the algorithm inputs landmark points into a previously trained neural network to determine if any missing fingers are present.
- the neural network takes as input the principal components acquired while fitting the Active Shape Hand Model to the current image landmarks.
- the output of the neural network is a four element vector indicating if each finger is present or not.
- the algorithm assumes that a finger substantially shorter than normal is "missing."
- the on-the-go scanner can use a neural network, or multiple neural networks, to determine if any missing fingers are present in the hand camera image. In some aspects, though, another way this determination can be made is if the processed fingerprint is exceptionally poor, then the finger might be missing.
- Training the neural network requires using a large set of training imagery.
- the training imagery can be captured using the hand camera, or can be preloaded into the system.
- the training imagery can consist of as many different people's hands as possible.
- the hands should be of normal (no missing fingers) as well as missing fingers. Missing fingers can be simulated by bending a particular finger down so it appears much shorter. Fingers should also be spread as well as merged for a variety of hand configurations.
- the training imagery is run through a landmark detection algorithm as well as a hand model fitting algorithm to create a ten element primary component vector for each training image. This vector is the "input".
- the expected output can be manually annotated from the training imagery.
- the expected or target output can be a four element vector containing a "0" for a present finger and a "1" for a missing finger.
- a large number of hand images are collected. Contours and landmark points are applied to each training image.
- the collection of landmark points from all training images is processed using Principal Component Analysis.
- Principal component analysis is a standard statistical technique that can take a collection of correlated variables (the Landmark points) and convert them into a set of linearly uncorrected variables, also called principal components.
- Each principal component can be seen as describing a kind of motion a typical hand can do. For example, a single principal component might describe the motion of the thumb. A value of -1 might describe a thumb which is spread far from the hand. A value of +1 might describe a thumb which is up against the pointing finger.
- Another principal component might describe when the fingers are spread apart, vs. when they are close together.
- Another principal component might describe a long ring finger, vs. a very short (missing) ring finger.
- the principal component analysis does not care about what the data represents, it just finds the relationships between all the data and sorts the data into a minimum number of linearly independent components describing all possible hand positions represented in the training data set.
- a ten element primary component vector can be used to describe the motion of hands.
- these ten principal components sufficiently describe the position, configuration, and finger length of a hand, they can be used to train a neural network to detect missing fingers by, for example, dividing the principal component vectors for all of the training images with additional descriptors indicating if a particular finger is missing or not. This is enough information to train the neural network to perform the same task.
- the fingerprint camera can be configured to operate using the algorithm(s) shown in FIGS. 13-19.
- the fingerprint camera (“Print Camera”) algorithm acquires an image from the fingerprint camera, and fingertip locations from the hand camera algorithm. If the frame is not dark, the prints can be segmented from the image. To determine if the frame is dark (“Dark Frame Detection"), the algorithm determines whether a sum of all pixels is above a threshold. If yes, then the image is passed to the Segmentation Algorithm. If it is below threshold for three consecutive frames, then processing stops.
- the algorithm processes these segmented prints to find a more accurate print location for the next frame to be acquired.
- the fingerprint image(s) are also evaluated using a focus algorithm. The prints with the highest focus metric are retained for further processing after image acquisition is completed.
- the hand camera and fingerprint camera do not capture images at exactly the same time, hand motion perpendicular to the cameras' axis can induces a position error between the hand camera and the print camera finger positions. If the error is too high, it can be difficult to draw correspondences between fingers captured in both cameras. It can be important to draw these correspondences because, for example, the hand camera algorithm is responsible for positively numbering fingers. To account for this position error, the fingerprint camera image(s) are processed in the following way:
- the algorithm can obtain an initial box size that is used for tracking the size of the print and increasing the box size as necessary as the print grows. Further aspects include segmenting the fingers from one another. To do so, the algorithm can receive the four finger positions from the finger position correction algorithm (described above) and then proceeds to segment four prints from the Print Camera image. In some aspects, the segmented image can have a box size of about 512x768.
- Binarize Print Image block As well as the Focus Metric block.
- the Binarize Print Image block and the next in the chain can further refine the actual position of each print so that a good segmentation can be performed even in the presence of noise, bad lighting, and neighboring fingers. Binarization of the print involves the following steps:
- the algorithm can use curvature to again update the fingertip location in each image.
- the algorithm further processes the processing began by the binarization process by performing the following steps:
- the algorithm can additionally apply a focus metric.
- the purpose of this block is to find the single frame that contains the best focused print for each of the four fingers.
- Each segmented print is processed in the following way to create four focus metrics for each Print Camera image:
- the algorithm can retain only the image that has the highest focus metric for each finger identified. That is, if there are 4 fingers present, the image having the highest focus metric for each individual finger may be one image or could be up to four images.
- the algorithm stores in memory all of focus metric scores for each print segmentation, and a processor then compares the metrics and keeps the highest one for each finger. When image acquisition stops, this block outputs the four prints that have the highest focus metrics for that finger. These four focused prints are passed to the next section of the Print Camera algorithm for post processing. Additionally, this block can also keep track of when to stop image acquisition, stopping acquisition when the peak focus point of all fingers has been acquired. For example, the algorithm can 1) set a flag when an increase in focus has been detected for a single finger, and 2) stop image acquisition when focus decreases for all fingers three consecutive frames.
- the algorithm further processes the image(s) having the highest focus metric as described.
- the lighting used to illuminate the four fingers is not perfect. Therefore, during calibration of the on-the-go fingerprint scanners and methods described herein, a brightness map can be acquired of the lighting. This brightness map can then be used to calculate a brightness correction. Essentially, darker areas of the image(s) are gained up so that they approximately match the exposure of the bright areas.
- the algorithm can mask off neighboring fingers that do not belong to the individual print being created (i.e., if focusing on the index finger, the pinky finger can be masked off). For example, if fingers are close together, this block can remove any parts of neighboring fingers currently in the segmented print. This block can also resize the print window to better fit the particular finger.
- the following processing can be performed: 1) Binarize;
- the image(s) can be downsampled to a lower pixel per inch.
- Image resampling is simply a way to smoothly interpolate the data contained in an image, and optimally produce an image of a different size containing fewer pixels, yet still look the same.
- Downsampling can be performed by any known method, such as bilinear interpolation, and typical image resampling (i.e., downsampling) algorithms include nearest neighbor, bilinear, bicubic, as well as more advanced algorithms such as Lanczos resampling.
- the particular PPI that the image is downsampled to is dependent upon the particular application and can be any PPI desired.
- the image can be downsampled to between about 100 pixels per inch (“PPI") and about 2000 PPI.
- the image can be downsampled to about 500 PPI.
- each print can be evaluated for "print likeness.”
- Print likeness can be evaluated using a variety of methods, including but not limited to an NFIQ score and a simplified ridge frequency detection algorithm.
- NFIQ is a complex algorithm that takes into account more aspects of the print ridges. Because of this complexity, it is better at excluding non-finger- like objects than other methods. These objects might be the wrong side of the hand, a band-aid obscuring the actual print, or an intentionally smudged finger.
- all prints receiving an NFIQ score of 5 are marked as "Not a Print.”
- the second print- likeness algorithm is based on ridge frequencies. This algorithm is designed to find where in the print segmentation box the print-like object actually is. This location can be used for things like refining the print segmentation, and excluding long fingernails.
- the ridge frequency algorithm performs the following steps.
- each print is processed for both binary and grey-level output. Prior to creation of grey-level and binary prints, however, each print is cropped and centered. To do this, the algorithm can perform the following steps:
- the algorithm in some aspects, also flips all prints from left to right so that the prints match those captured by a traditional print reader.
- Grey-level (or "gray-level”) processing can be performed to create a desired grey-level fingerprint output (such as that shown in FIG. 6 and FIG. 19C).
- Grey-line processing can be performed in any manner suitable.
- a suitable processing method includes: 1) Locally normalize raw grayscale print image;
- the image(s) can be processed to form binarized prints, such as is shown in FIG. 6 and FIG. 18C.
- any suitable method can be used to produce binarized prints.
- a baseline method or an optional method can be utilized.
- a baseline method includes utilizing 1) a low pass Gaussian blur, and 2) adaptive threshold binarization.
- An optional method includes utilizing:
- the scanner can output said prints to an output device or user.
- the scanner can cross- reference the prints against a database of known biometric data—including fingerprints— to either obtain the identity of the subject or otherwise provide or obtain information about the subject. Additionally, the scanner can build a local database of prints and other biometric and identifying data.
- scanners can be utilized to confirm security access to buildings and other secured areas, confirm that the identity of a passenger is the same as the identity corresponding to a particular ticket, act as a form of identification or payment wherein the payment information (i.e., bank account and routing numbers) are keyed to your individual fingerprints, or any other application where a set of data is cross-referenced or checked against a scanned fingerprint.
- payment information i.e., bank account and routing numbers
- FIG. 20 shows an on-the-go fingerprint scanner having a frame 2000 holding beam break sensor 2002 that is replaced with a mirror 2004, held in place by a mounting apparatus 2006.
- the beam break sensor 2002 is now located on frame 2000, and optical, imaging subsystem 2008 (including the illumination system, hand camera, and fingerprint camera) is in accordance with the above description.
- optical, imaging subsystem 2008 including the illumination system, hand camera, and fingerprint camera
- the mounting apparatus 2006 could be any of a number of
- the beam break sensor 2002 initiates the operation. Instead of a direct optical path from optical subsystem 2008 to the hand, mirror 2004 reflects the optical path back to the location of beam break sensor 2002, now located close to optical subsystem 2008.
- the lighting system, fingerprint camera, and hand camera systems all perform as described above.
- mirror 2008 is passive.
- Mirror 2004 is located at approximately half (1/2) the distance that hand break sensor 2002 was located in the aspects described above.
- mirror 2008 can be placed further away, allowing increased focal length of fingerprint camera and increased depth of field.
- the mounting apparatus 2006 can provide a folding operation to allow a more compact non- operational mode, for ease in transportation, or to reduce footprint when not operating.
- the mounting apparatus should, however, provide a stable positioning of the mirror, which can be provided by a variety of means.
- a group of items linked with the conjunction "and” should not be read as requiring that each and every one of those items be present in the grouping, but rather should be read as “and/or” unless expressly stated otherwise.
- a group of items linked with the conjunction "or” should not be read as requiring mutual exclusivity among that group, but rather should also be read as “and/or” unless expressly stated otherwise.
- items, elements or components of the invention may be described or claimed in the singular, the plural is contemplated to be within the scope thereof unless limitation to the singular is explicitly stated.
- module does not imply that the components or functionality described or claimed as part of the module are all configured in a common package. Indeed, any or all of the various components of a module, whether control logic or other components, may be combined in a single package or separately maintained and may further be distributed across multiple locations.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Human Computer Interaction (AREA)
- Software Systems (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Health & Medical Sciences (AREA)
- Artificial Intelligence (AREA)
- Computing Systems (AREA)
- Databases & Information Systems (AREA)
- Evolutionary Computation (AREA)
- General Health & Medical Sciences (AREA)
- Medical Informatics (AREA)
- Image Input (AREA)
- Collating Specific Patterns (AREA)
Abstract
Description
Claims
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201461938748P | 2014-02-12 | 2014-02-12 | |
PCT/US2015/015538 WO2015123374A1 (en) | 2014-02-12 | 2015-02-12 | On-the-go touchless fingerprint scanner |
Publications (2)
Publication Number | Publication Date |
---|---|
EP3105712A1 true EP3105712A1 (en) | 2016-12-21 |
EP3105712A4 EP3105712A4 (en) | 2017-09-20 |
Family
ID=53800610
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
EP15749130.9A Withdrawn EP3105712A4 (en) | 2014-02-12 | 2015-02-12 | On-the-go touchless fingerprint scanner |
Country Status (2)
Country | Link |
---|---|
EP (1) | EP3105712A4 (en) |
WO (1) | WO2015123374A1 (en) |
Families Citing this family (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
ES2902448T3 (en) * | 2015-09-09 | 2022-03-28 | Thales Dis France Sas | Non-contact friction ridge pickup device |
WO2017044343A1 (en) | 2015-09-09 | 2017-03-16 | 3M Innovative Properties Company | Non-contact friction ridge capture device |
DE102020200569A1 (en) | 2020-01-17 | 2021-07-22 | Brno University of Technology | Device for biometric identification with the aid of fingerprints and / or hand characteristics and methods for biometric identification with the aid of these characteristics |
Family Cites Families (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7394919B2 (en) * | 2004-06-01 | 2008-07-01 | Lumidigm, Inc. | Multispectral biometric imaging |
US8525876B2 (en) * | 2010-05-12 | 2013-09-03 | Visionbrite Technologies Inc. | Real-time embedded vision-based human hand detection |
US8600123B2 (en) * | 2010-09-24 | 2013-12-03 | General Electric Company | System and method for contactless multi-fingerprint collection |
CA2851225A1 (en) * | 2010-10-08 | 2012-04-12 | Advanced Optical Systems, Inc. | Contactless fingerprint acquisition and processing |
-
2015
- 2015-02-12 EP EP15749130.9A patent/EP3105712A4/en not_active Withdrawn
- 2015-02-12 WO PCT/US2015/015538 patent/WO2015123374A1/en active Application Filing
Also Published As
Publication number | Publication date |
---|---|
EP3105712A4 (en) | 2017-09-20 |
WO2015123374A1 (en) | 2015-08-20 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9489560B2 (en) | On-the go touchless fingerprint scanner | |
KR102587193B1 (en) | System and method for performing fingerprint-based user authentication using images captured using a mobile device | |
KR101683712B1 (en) | An iris and ocular recognition system using trace transforms | |
US8803963B2 (en) | Vein pattern recognition based biometric system and methods thereof | |
JP2020501264A (en) | System and method for performing fingerprint authentication using images captured with a mobile device | |
JP2009544108A (en) | Multispectral image for multiple biometric authentication | |
CN107368774A (en) | Gaze detection equipment and gaze detection method | |
JP2006260583A (en) | Automated, non-invasive iris recognition system and method | |
JP2000259814A (en) | Image processor and method therefor | |
GB2400714A (en) | Combined optical fingerprint recogniser and navigation control | |
WO2008131201A1 (en) | Method and system for biometric recognition | |
CN112232155B (en) | Non-contact fingerprint identification method and device, terminal and storage medium | |
KR102679397B1 (en) | Biometric authentication apparatus and biometric authentication method | |
CN112232163B (en) | Fingerprint acquisition method and device, fingerprint comparison method and device, and equipment | |
JP5556663B2 (en) | Verification device, verification method, and program | |
JP2003150942A (en) | Eye position tracing method | |
Matey et al. | Iris recognition in less constrained environments | |
Parziale et al. | Advanced technologies for touchless fingerprint recognition | |
US20230386249A1 (en) | Presentation attack detection | |
US11080511B2 (en) | Contactless rolled fingerprints | |
WO2015123374A1 (en) | On-the-go touchless fingerprint scanner | |
CN112668539A (en) | Biological characteristic acquisition and identification system and method, terminal equipment and storage medium | |
CN112232157B (en) | Fingerprint area detection method, device, equipment and storage medium | |
Bashir et al. | Video surveillance for biometrics: long-range multi-biometric system | |
Rao et al. | Iris image segmentation and localization using dynamic reconfigurable processor |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PUAI | Public reference made under article 153(3) epc to a published international application that has entered the european phase |
Free format text: ORIGINAL CODE: 0009012 |
|
17P | Request for examination filed |
Effective date: 20160906 |
|
AK | Designated contracting states |
Kind code of ref document: A1 Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR |
|
AX | Request for extension of the european patent |
Extension state: BA ME |
|
DAX | Request for extension of the european patent (deleted) | ||
A4 | Supplementary search report drawn up and despatched |
Effective date: 20170823 |
|
RIC1 | Information provided on ipc code assigned before grant |
Ipc: G06K 9/00 20060101AFI20170817BHEP Ipc: G06K 9/62 20060101ALI20170817BHEP |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: THE APPLICATION IS DEEMED TO BE WITHDRAWN |
|
18D | Application deemed to be withdrawn |
Effective date: 20190110 |