CA2972821A1 - Detecting facial liveliness - Google Patents
Detecting facial liveliness Download PDFInfo
- Publication number
- CA2972821A1 CA2972821A1 CA2972821A CA2972821A CA2972821A1 CA 2972821 A1 CA2972821 A1 CA 2972821A1 CA 2972821 A CA2972821 A CA 2972821A CA 2972821 A CA2972821 A CA 2972821A CA 2972821 A1 CA2972821 A1 CA 2972821A1
- Authority
- CA
- Canada
- Prior art keywords
- illumination
- subject
- illuminator
- facial
- change
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 230000001815 facial effect Effects 0.000 title claims abstract description 188
- 238000005286 illumination Methods 0.000 claims abstract description 206
- 230000033001 locomotion Effects 0.000 claims abstract description 132
- 230000008859 change Effects 0.000 claims abstract description 114
- 238000000034 method Methods 0.000 claims abstract description 99
- 238000012545 processing Methods 0.000 claims abstract description 26
- 238000012795 verification Methods 0.000 claims description 57
- 230000002123 temporal effect Effects 0.000 claims description 34
- 230000004044 response Effects 0.000 claims description 26
- 239000003086 colorant Substances 0.000 claims description 14
- 230000003068 static effect Effects 0.000 claims description 13
- 230000003213 activating effect Effects 0.000 claims description 8
- 230000009471 action Effects 0.000 abstract description 32
- 210000001508 eye Anatomy 0.000 description 21
- 230000007246 mechanism Effects 0.000 description 15
- 230000008569 process Effects 0.000 description 15
- 238000004891 communication Methods 0.000 description 11
- 238000004590 computer program Methods 0.000 description 10
- 230000002596 correlated effect Effects 0.000 description 9
- 238000010586 diagram Methods 0.000 description 9
- 230000003993 interaction Effects 0.000 description 9
- 238000004458 analytical method Methods 0.000 description 7
- 210000004087 cornea Anatomy 0.000 description 7
- 210000003128 head Anatomy 0.000 description 6
- 230000004913 activation Effects 0.000 description 4
- 230000008901 benefit Effects 0.000 description 4
- 238000004422 calculation algorithm Methods 0.000 description 4
- 238000001514 detection method Methods 0.000 description 4
- 230000005540 biological transmission Effects 0.000 description 3
- 238000013478 data encryption standard Methods 0.000 description 3
- 239000004973 liquid crystal related substance Substances 0.000 description 3
- 230000003287 optical effect Effects 0.000 description 3
- 230000011218 segmentation Effects 0.000 description 3
- 230000004075 alteration Effects 0.000 description 2
- 210000000887 face Anatomy 0.000 description 2
- 230000006870 function Effects 0.000 description 2
- 230000005021 gait Effects 0.000 description 2
- 238000003384 imaging method Methods 0.000 description 2
- 210000001525 retina Anatomy 0.000 description 2
- 238000013515 script Methods 0.000 description 2
- 238000000926 separation method Methods 0.000 description 2
- 230000001755 vocal effect Effects 0.000 description 2
- 241001289753 Graphium sarpedon Species 0.000 description 1
- OAICVXFJPJFONN-UHFFFAOYSA-N Phosphorus Chemical compound [P] OAICVXFJPJFONN-UHFFFAOYSA-N 0.000 description 1
- 230000003044 adaptive effect Effects 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 238000012790 confirmation Methods 0.000 description 1
- 235000014510 cooky Nutrition 0.000 description 1
- 230000000875 corresponding effect Effects 0.000 description 1
- 230000001419 dependent effect Effects 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 210000004709 eyebrow Anatomy 0.000 description 1
- 210000001061 forehead Anatomy 0.000 description 1
- 230000002452 interceptive effect Effects 0.000 description 1
- 239000011159 matrix material Substances 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 230000035755 proliferation Effects 0.000 description 1
- 238000005070 sampling Methods 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 230000001953 sensory effect Effects 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 239000000758 substrate Substances 0.000 description 1
- 238000012546 transfer Methods 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Classifications
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B3/00—Apparatus for testing the eyes; Instruments for examining the eyes
- A61B3/10—Objective types, i.e. instruments for examining the eyes independent of the patients' perceptions or reactions
- A61B3/117—Objective types, i.e. instruments for examining the eyes independent of the patients' perceptions or reactions for examining the anterior chamber or the anterior chamber angle, e.g. gonioscopes
- A61B3/1173—Objective types, i.e. instruments for examining the eyes independent of the patients' perceptions or reactions for examining the anterior chamber or the anterior chamber angle, e.g. gonioscopes for examining the eye lens
- A61B3/1176—Objective types, i.e. instruments for examining the eyes independent of the patients' perceptions or reactions for examining the anterior chamber or the anterior chamber angle, e.g. gonioscopes for examining the eye lens for determining lens opacity, e.g. cataract
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/117—Identification of persons
- A61B5/1171—Identification of persons based on the shapes or appearances of their bodies or parts thereof
- A61B5/1176—Recognition of faces
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B3/00—Apparatus for testing the eyes; Instruments for examining the eyes
- A61B3/10—Objective types, i.e. instruments for examining the eyes independent of the patients' perceptions or reactions
- A61B3/107—Objective types, i.e. instruments for examining the eyes independent of the patients' perceptions or reactions for determining the shape or measuring the curvature of the cornea
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B3/00—Apparatus for testing the eyes; Instruments for examining the eyes
- A61B3/10—Objective types, i.e. instruments for examining the eyes independent of the patients' perceptions or reactions
- A61B3/14—Arrangements specially adapted for eye photography
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B3/00—Apparatus for testing the eyes; Instruments for examining the eyes
- A61B3/10—Objective types, i.e. instruments for examining the eyes independent of the patients' perceptions or reactions
- A61B3/14—Arrangements specially adapted for eye photography
- A61B3/15—Arrangements specially adapted for eye photography with means for aligning, spacing or blocking spurious reflection ; with means for relaxing
- A61B3/152—Arrangements specially adapted for eye photography with means for aligning, spacing or blocking spurious reflection ; with means for relaxing for aligning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/18—Eye characteristics, e.g. of the iris
- G06V40/19—Sensors therefor
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/40—Spoof detection, e.g. liveness detection
- G06V40/45—Detection of the body part being alive
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/10—Image acquisition
- G06V10/12—Details of acquisition arrangements; Constructional details thereof
- G06V10/14—Optical characteristics of the device performing the acquisition or on the illumination arrangements
- G06V10/141—Control of illumination
Landscapes
- Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Health & Medical Sciences (AREA)
- Biophysics (AREA)
- Medical Informatics (AREA)
- Veterinary Medicine (AREA)
- Public Health (AREA)
- Animal Behavior & Ethology (AREA)
- Ophthalmology & Optometry (AREA)
- Surgery (AREA)
- Biomedical Technology (AREA)
- Heart & Thoracic Surgery (AREA)
- Molecular Biology (AREA)
- General Physics & Mathematics (AREA)
- Human Computer Interaction (AREA)
- Theoretical Computer Science (AREA)
- Multimedia (AREA)
- Pathology (AREA)
- Collating Specific Patterns (AREA)
- Image Input (AREA)
- Image Analysis (AREA)
Abstract
Methods, systems, and computer-readable storage mediums for detecting facial liveliness are provided. Implementations include actions of processing first and second facial images of a subject to determine first and second corneal reflections of an object, the first and second facial images being captured at first and second sequential time points, determining a corneal reflection change of the object based on the determined first and second corneal reflections, comparing the determined corneal reflection change of the object to a known change associated with the first and second time points, and determining facial liveliness of the subject based on a result of the comparison. The known change can be a motion of a camera device or an illumination change of an illuminator.
Description
Detecting Facial Liveliness CROSS-REFERENCE TO RELATED APPLICATION
[00011 This application claims priority under 35 U SC 119(e) to U.S.
Provisional Patent Application Serial No. 62/098,575, filed on 12/31/2014, and U.S.
Provisional Patent Application Serial No. 62/098,596, filed on 12/31/2014, the entire contents of which is hereby incorporated by reference.
FIELD
[00021 This disclosure generally relates to computer-based authentication.
BACKGROUND
[OM] Transactions between a consumer and a merchant may be subject to risks of identity theft, identity fraud, spoofing, phishing, etc., all of which may potentially hinder the flow of conunerce.
SUMMARY
NON] Implementations of the present disclosure include computer-implemented methods for detecting facial liveliness. In some implementations, methods include actions of processing first and second facial images of a subject to determine first and second corneal reflections of an object, the first and second facial images being captured at first and second sequential time points; determining a corneal reflection change of the object based on the determined first and second corneal reflections; comparing the determined corneal reflection change of the object to a known change associated with the first and second time points; and determining facial liveliness of the subject based on a result of the comparison. The known change can be a motion of a camera device or an illtunination change of an illtuninator.
NM] Implementations of the present disclosure can also include computer-implemented methods for detecting facial liveliness that include actions of processing first and second facial images of a subject to determine first and second corneal reflections of an object, the first and second facial images being captured at first and second sequential time points; determining a corneal reflection change of the object based on the determined first and second corneal reflections; comparing the determined corneal reflection change of the object to a motion associated with the first and second time points; and determining facial liveliness of the subject based on a result of the comparison.
[0006] These and other implementations can each optionally include one or more of the following features: comparing the determined corneal reflection change of the object to a motion includes correlating the determined corneal reflection change of the object to the motion; scoring a matching quality based on a result of the correlation; and comparing the scored matching quality to a predetermined threshold. Determining facial liveliness of the subject includes determining that the face of the subject is live in response to determining that the scored matching quality is beyond the predetermined threshold. The actions further includes determining an expected corneal reflection of the object at the second time point based on the determined first corneal reflection and the motion; and assessing the facial liveliness of the subject by determining likelihood between the expected corneal reflection and the determined second corneal reflection.
[0007] In some implementations, the corneal reflection change of the object includes a position change of the corneal refection of the object. The motion can be associated with a movement of the object between first and second positions, and the first facial image can be captured at the first time point when the object is at the first position and the second facial image can be captured at the second time point when the object is at the second position. The actions can further include receiving information of the movement of the object measured by a sensor; and determining the motion based on the received information of the movement of the object. In some cases, the actions include prompting the subject to move the object. In some cases, the movement of the object is associated with a natural motion of the subject.
[0008] The actions are executed by one or more processors that can be included in a computing device. The computing device can include the object and the sensor, and the object can be a camera device and the sensor can include at least one of an accelerometer, a gyroscope, or a global positioning system (GPS). The actions can further include determining, by using the sensor, that the object is moving; and requesting the camera device to capture facial images of the subject.
[0009] In some examples, the motion is a predetermined motion for the object, and the object is moved from the first position to the second position based on the predetermined motion. In some examples; the first and second facial images include a second object that is static during the movement of the object, and the actions further include determining the motion based on a position change of the second object in the first and second facial images.
[00011 This application claims priority under 35 U SC 119(e) to U.S.
Provisional Patent Application Serial No. 62/098,575, filed on 12/31/2014, and U.S.
Provisional Patent Application Serial No. 62/098,596, filed on 12/31/2014, the entire contents of which is hereby incorporated by reference.
FIELD
[00021 This disclosure generally relates to computer-based authentication.
BACKGROUND
[OM] Transactions between a consumer and a merchant may be subject to risks of identity theft, identity fraud, spoofing, phishing, etc., all of which may potentially hinder the flow of conunerce.
SUMMARY
NON] Implementations of the present disclosure include computer-implemented methods for detecting facial liveliness. In some implementations, methods include actions of processing first and second facial images of a subject to determine first and second corneal reflections of an object, the first and second facial images being captured at first and second sequential time points; determining a corneal reflection change of the object based on the determined first and second corneal reflections; comparing the determined corneal reflection change of the object to a known change associated with the first and second time points; and determining facial liveliness of the subject based on a result of the comparison. The known change can be a motion of a camera device or an illtunination change of an illtuninator.
NM] Implementations of the present disclosure can also include computer-implemented methods for detecting facial liveliness that include actions of processing first and second facial images of a subject to determine first and second corneal reflections of an object, the first and second facial images being captured at first and second sequential time points; determining a corneal reflection change of the object based on the determined first and second corneal reflections; comparing the determined corneal reflection change of the object to a motion associated with the first and second time points; and determining facial liveliness of the subject based on a result of the comparison.
[0006] These and other implementations can each optionally include one or more of the following features: comparing the determined corneal reflection change of the object to a motion includes correlating the determined corneal reflection change of the object to the motion; scoring a matching quality based on a result of the correlation; and comparing the scored matching quality to a predetermined threshold. Determining facial liveliness of the subject includes determining that the face of the subject is live in response to determining that the scored matching quality is beyond the predetermined threshold. The actions further includes determining an expected corneal reflection of the object at the second time point based on the determined first corneal reflection and the motion; and assessing the facial liveliness of the subject by determining likelihood between the expected corneal reflection and the determined second corneal reflection.
[0007] In some implementations, the corneal reflection change of the object includes a position change of the corneal refection of the object. The motion can be associated with a movement of the object between first and second positions, and the first facial image can be captured at the first time point when the object is at the first position and the second facial image can be captured at the second time point when the object is at the second position. The actions can further include receiving information of the movement of the object measured by a sensor; and determining the motion based on the received information of the movement of the object. In some cases, the actions include prompting the subject to move the object. In some cases, the movement of the object is associated with a natural motion of the subject.
[0008] The actions are executed by one or more processors that can be included in a computing device. The computing device can include the object and the sensor, and the object can be a camera device and the sensor can include at least one of an accelerometer, a gyroscope, or a global positioning system (GPS). The actions can further include determining, by using the sensor, that the object is moving; and requesting the camera device to capture facial images of the subject.
[0009] In some examples, the motion is a predetermined motion for the object, and the object is moved from the first position to the second position based on the predetermined motion. In some examples; the first and second facial images include a second object that is static during the movement of the object, and the actions further include determining the motion based on a position change of the second object in the first and second facial images.
2 100101 The actions can include transmitting a command to a controller coupled to the object, the command indicating the controller to move the object. The command can include a predetermined motion for the object, and wherein the predetermined motion is different from a previous predetermined motion for the object.
100111 In some implementations, the actions include transmitting a command to a camera device to capture facial images of the subject at sequential time points including the first and second time points; and receiving the captured facial images from the camera device. Receiving the captured facial images from the camera device can include receiving a video stream feed from the camera device, the video stream feed comprising the captured facial images. In some examples, the camera device is moved from a first position to a second position between the first and second time points. The first facial image of the subject is captured at the first time point when the camera device is at the first position, and the second facial image of the subject is captured at the second time point when the camera device is at the second position, and the motion is based on a movement of the camera device between the first and second positions. In some examples, the actions include transmitting a second command to a controller coupled to the object, the second command indicating the controller to move the object at the first and second time points, and the motion is based on the movement of the camera device and the movement of the object.
100121 The object can include one of a camera device, an illumination device, or an object brighter than ambient environment. The corneal reflection change of the object can include a first illumination change on the object in the first and second corneal reflections. In some cases, the actions further include obtaining a second illumination change on the object for the first and second time points; and determining a matching quality between the obtained second illumination change on the object and the first illumination change in the first and second corneal reflections. Determining facial liveliness of the subject can include determining the facial liveliness of the subject based on the determined matching quality and the result of the comparison.
100131 The actions can include determining that the subject chooses to use liveliness verification for biometric authentication. The actions can also include transmitting, to a computing device at an identity provider, an authentication request of the subject for accessing an account managed by a relying party different from the identity provider; the authentication request including biometric data of the subject and the determined facial liveliness of the subject.
100111 In some implementations, the actions include transmitting a command to a camera device to capture facial images of the subject at sequential time points including the first and second time points; and receiving the captured facial images from the camera device. Receiving the captured facial images from the camera device can include receiving a video stream feed from the camera device, the video stream feed comprising the captured facial images. In some examples, the camera device is moved from a first position to a second position between the first and second time points. The first facial image of the subject is captured at the first time point when the camera device is at the first position, and the second facial image of the subject is captured at the second time point when the camera device is at the second position, and the motion is based on a movement of the camera device between the first and second positions. In some examples, the actions include transmitting a second command to a controller coupled to the object, the second command indicating the controller to move the object at the first and second time points, and the motion is based on the movement of the camera device and the movement of the object.
100121 The object can include one of a camera device, an illumination device, or an object brighter than ambient environment. The corneal reflection change of the object can include a first illumination change on the object in the first and second corneal reflections. In some cases, the actions further include obtaining a second illumination change on the object for the first and second time points; and determining a matching quality between the obtained second illumination change on the object and the first illumination change in the first and second corneal reflections. Determining facial liveliness of the subject can include determining the facial liveliness of the subject based on the determined matching quality and the result of the comparison.
100131 The actions can include determining that the subject chooses to use liveliness verification for biometric authentication. The actions can also include transmitting, to a computing device at an identity provider, an authentication request of the subject for accessing an account managed by a relying party different from the identity provider; the authentication request including biometric data of the subject and the determined facial liveliness of the subject.
3 [0014] Implementations of the present disclosure can also include computer-implemented methods for detecting facial liveliness that include actions of processing first and second facial images of a subject to determine first and second poses of a face of the subject, the first and second facial images being captured at first and second sequential time points; determining a change in pose of the face based on the determined first and second poses; comparing the determined change in pose of the face to a motion associated with the first and second time points: and determining facial liveliness of the subject based on a result of the comparison.
[0015] These and other implementations can each optionally include one or more of the following features: comparing the determined change in pose of the face to a motion includes correlating the determined change in pose of the face to the motion:
scoring a matching quality based on a result of the correlation; and comparing the scored matching quality to a predetermined threshold. Determining facial liveliness of the subject includes determining that the face of the subject is live in response to determining that the scored matching quality is beyond the predetermined threshold. Determining first and second poses of the face can include determining at least one facial landmark of the face.
[0016] In some examples, the actions include calculating a pose of the face at the second time point based on the determined first pose and the motion; and assessing facial liveliness by determining likelihood between the calculated pose of the face to the determined second pose of the face from the second facial image.
[0017] The motion can be associated with a relative movement between the face and a camera device configured to capture facial images of the subject. In some examples, the actions include prompting the subject to move the camera device relative to the face of the subject. In some examples, the actions include receiving information of the movement of the camera device measured by a sensor; and determining the motion based on the received information of the movement of the camera device.
[0018] The actions are executed by one or more processors that can be included in a computing device, and the computing device can include the camera device and the sensor, and the sensor includes at least one of an accelerometer, a gyroscope, or a global positioning system (GPS).
[0019] Implementations of the present disclosure can also provide computer-implemented methods for detecting facial liveliness that include actions of processing a facial image of a subject to determine a corneal reflection of an illuminator adjacent to the subject, the facial image being captured at a time point, determining an expected corneal reflection of
[0015] These and other implementations can each optionally include one or more of the following features: comparing the determined change in pose of the face to a motion includes correlating the determined change in pose of the face to the motion:
scoring a matching quality based on a result of the correlation; and comparing the scored matching quality to a predetermined threshold. Determining facial liveliness of the subject includes determining that the face of the subject is live in response to determining that the scored matching quality is beyond the predetermined threshold. Determining first and second poses of the face can include determining at least one facial landmark of the face.
[0016] In some examples, the actions include calculating a pose of the face at the second time point based on the determined first pose and the motion; and assessing facial liveliness by determining likelihood between the calculated pose of the face to the determined second pose of the face from the second facial image.
[0017] The motion can be associated with a relative movement between the face and a camera device configured to capture facial images of the subject. In some examples, the actions include prompting the subject to move the camera device relative to the face of the subject. In some examples, the actions include receiving information of the movement of the camera device measured by a sensor; and determining the motion based on the received information of the movement of the camera device.
[0018] The actions are executed by one or more processors that can be included in a computing device, and the computing device can include the camera device and the sensor, and the sensor includes at least one of an accelerometer, a gyroscope, or a global positioning system (GPS).
[0019] Implementations of the present disclosure can also provide computer-implemented methods for detecting facial liveliness that include actions of processing a facial image of a subject to determine a corneal reflection of an illuminator adjacent to the subject, the facial image being captured at a time point, determining an expected corneal reflection of
4 the illuminator based on an illumination of the illuminator at the time point, comparing the determined corneal reflection of the illuminator to the expected corneal reflection of the illuminator to obtain a comparison result, and determining facial liveliness of the subject based on the comparison result.
[0020] These and other implementations can each optionally include one or more of the following features: comparing the determined corneal reflection of the illuminator to the expected corneal reflection of the illuminator can include: correlating the determined corneal reflection of the illuminator to the expected corneal reflection of the illuminator; scoring a matching quality based on a result of the correlation; and comparing the scored matching quality to a predetermined threshold. Determining facial liveliness of the subject can include determining that the face of the subject is live in response to determining that the scored matching quality is beyond the predetermined threshold. Determining a corneal reflection of the illuminator can include determining the corneal reflection of the illuminator in at least one of: color, shape, position, or pattern.
[0021] The actions can include activating the illuminator to illuminate based on an illumination pattern. In some examples, the actions include generating the illumination pattern for the illuminator. In some examples, activating the illuminator comprises:
activating the illuminator in response to determining that the subject chooses to use liveliness verification for biometric authentication.
[0022] The illuminator can include two or more illumination sources with respective colors and at respective positions, respectively. In some implementations, the illumination of the illuminator is based on a temporal illumination pattern of the two or more illumination sources. The two or more illumination sources can generate illumination in bursts each with respective temporal durations. At least one of the two or more illumination sources can have different illumination shapes for different temporal durations. In some implementations, the illumination of the illuminator is based on a spatial illumination pattern of the two or more ilhunination sources. Each of the two or more illumination sources can have an illumination on or off state.
[0023] In some examples, the illuminator includes one or more physical light sources each at respective light wavelengths. In some examples, the illuminator includes one or more digital illumination sources displayed on a screen to provide respective illumination colors.
The actions can include generating the digital illumination sources. The illuminator can be positioned such that illumination from the illuminator is incident on an eye of the subject at an angle. The actions can include prompting the subject to change a relative position of the illuminator and the eye of the subject.
[0024] The actions can include receiving at least a portion of a video stream feed from a camera device, the portion of the video stream feed including facial images of the subject captured by the camera device at sequential time points. The action can further include: processing a second facial image of the subject to determine a second corneal reflection of the illuminator, the second facial image being captured at a second time point sequential to the time point for the first facial image; determining a second expected corneal reflection of the illuminator based on a second illumination of the illuminator at the second time point, the second illumination being different from the first illumination; comparing the determined second corneal reflection of the illuminator to the expected second corneal reflection of the illuminator to obtain a second comparison result; and determining facial liveliness of the subject based on the comparison result and the second comparison result.
[0025] The actions can include transmitting the determined facial liveliness of the subject to a computing device at identity provider. The actions can also include transmitting, to a computing device at identity provider, an authentication request of the subject for accessing an account managed by a relying party different from the identity provider, the authentication request including biometric data of the subject and the determined facial liveliness of the subject.
[0026] Implementations of the present disclosure can also provide computer-implemented methods for detecting facial liveliness that include actions of processing first and second facial images of a subject to determine first and second corneal reflections of an illuminator adjacent to the subject, the first and second facial images being captured at first and second sequential time points, respectively: determining a corneal reflection change of the illuminator based on the determined first and second corneal reflections;
comparing the determined corneal reflection change of the illuminator to an illumination change of the ilhuninator associated with the first and second sequential time points; and determining facial liveliness of the subject based on a result of the comparison.
[0027] These and other implementations can each optionally include one or more of the following features: the actions can include activating the illuminator to change illumination of the illuminator based on the illumination change. The illumination change of the illuminator can include at least one of: an illumination color change, an illumination shape change, an illumination position change, an illumination on/off state change, or an ilhunination temporal duration change. Determining a corneal reflection change of the illuminator can include determining a change of the corneal reflection of the illuminator in at least one of: color, shape, position, on/off state, or temporal duration.
100281 The illuminator can include at least one of: one or more physical light sources each at respective light wavelengths, or one or more digital illumination sources displayed on a screen to provide respective illumination colors. In some implementations, the illuminator includes first and second illumination sources with first and second colors and at first and second positions, respectively. In some examples, the illumination change of the illuminator includes a change of a temporal illumination pattern of the first and second illumination sources, and the first and second illumination sources can generate illumination in bursts each with respective temporal durations. At least one of the first illumination source or the second illumination source can change an illumination shape for different temporal durations. In some examples, the illumination change of the illuminator includes a change of a spatial illtunination pattern of the first and second illumination sources over time, and at least one of the first illumination source or the second illumination source can change an illumination on/off state over time.
100291 The present disclosure also provides one or more non-transitory computer-readable storage media coupled to one or more processors and having instructions stored thereon which, when executed by the one or more processors, cause the one or more processors to perform operations in accordance with implementations of the methods provided herein.
100301 The present disclosure further provides a system for implementing the methods provided herein. The system includes one or more processors, and a computer-readable storage medium coupled to the one or more processors having instructions stored thereon which, when executed by the one or more processors, cause the one or more processors to perform operations in accordance with implementations of the methods provided herein.
100311 It is appreciated that methods in accordance with the present disclosure can include any combination of the aspects and features described herein. That is, methods in accordance with the present disclosure are not limited to the combinations of aspects and features specifically described herein, but also include any combination of the aspects and features provided.
100321 The details of one or more implementations of the present disclosure are set forth in the accompanying drawings and the description below. Other features and advantages of the present disclosure will be apparent from the description and drawings, and from the claims.
DESCRIPTION OF DRAWINGS
[0033] Fig. IA is a diagram showing an example web site capable of authenticating a user based on a biometric identity of the user according to some implementations.
[0034] Fig. 1B is a diagram showing an example interaction among a financial institution customer, a financial institution, and a third-party biometric verification service to authenticate a user based on a biometric identity of the user according to some implementations.
[0035] Fig. 2A is a timing diagram showing an example interaction among a consumer, a relying party, and an identity provider in authenticating the consumer when the consumer has enrolled at the identity provider according to some implementations.
[0036] Fig. 2B is a timing diagram showing an example interaction among a consumer, a relying party, and an identity provider in authenticating the consumer when the consumer has not yet enrolled at the identity provider according to some implementations.
[0037] Figs. 3A and 3B illustrate the changed positions of the corneal reflection of an object.
[0038] Fig. 4 illustrates an object that has changed position by virtue of arm adjustments.
[0039] Fig. 5 is a diagram showing an example verification based on corneal reflection from illumination sources.
[0040] Like reference symbols in the various drawings indicate like elements.
DETAILED DESCRIPTION
[0041] Proof of identity may present a perennial challenge in our daily transactions.
With the advent of the Internet, comes the age of e-commerce in which on-line transactions may replace in-person transactions. However, the sheer volume and complexity of these on-line transactions may give rise to a digital world fraught with peril, including, for example, identity theft, identity fraud, spoofmg, phishing, etc. Notably, such risks may not be new in the Internet age, although the Internet may have amplified such risks. As the society moves towards cloud computing, more and more databases may become accessible.
Identity data in some databases may be more reliable and robust than others, based on history or tradition.
As connectivity becomes ubiquitous and as more identity databases become available on accessible platforms, identity-related data housed therein can be accessed to increase the confidence in the quality of transactions conducted either online and in person. In addition, biometric data submitted on-line can be verified for liveliness to deter spoofing attempts. In some implementations, the liveliness verification can be leverage analysis of real-time corneal reflection and/or real-time face pose to increase confidence in the biometric data being submitted on-line for authentication. Implementations disclosed herein may be extended to enterprise entities other than financial institutions. This application is directed to systems and methods to detect liveliness of a facial representation during, for example, an on-line session between two parties over a network. Note that the terms "liveliness" and liveness" can be used interchangeably here.
[0042] Fig. IA is a diagram showing an example web site capable of authenticating a user based on a biometric identity of the user according to some implementations. As illustrated, a user name and password may be required for logging into an account administered by the server hosting the web-site. For example, the server may run an implementation of a hypertext transmission protocol secure (https). In the Internet age, a consumer user may have accounts at dozens of more web-sites, each with disparate requirements of user names and passwords. The consumer user may have a hard time keeping track of user names and passwords at each of the web-sites. Losing track of the registered username or password can hinder on-line access convenience. The consumer user may store the usernames and passwords for each account at a central file. But, access to the central file may be breached, leading to comprises in the stored usernames and passwords. In fact, simple on-line identities including user name and password may be subject to identity theft and identity fraud. A recent survey revealed that identity theft in the United States rose to a three-year high in 2012, with more than 5 percent of the adult population, or 12.6 million people, falling victim to such crimes. The numbers are up from 4.9 percent in 2011 and 4.35 percent in 2010. The incidence of identity theft is only expected to rise. To mitigate the risks arising from identity theft in the context of e-commerce, some implementations, as disclosed herein may choose to authenticate biometric data presented the user on-line.
Such biometric data may include, but are not limited to, facial pattern, finger print, palm print, retina scan, iris scan, DNA pattern, voice characteristics, gait analysis. Generally speaking, such biometric data submitted on-line may be in a digital form to facilitate electronic verification.
[0043] As illustrated by Fig. 1A, biometric login window 102 may be provided by a third-party, different from, for example, the entity that runs bigbank.com.
The login window 102 may present the captured facial biometric of the user as the user attempts to log in based on the user's biometric. When the user's biometric data is transmitted for on-line authentication, the biometric data may be encrypted in accordance with industry standard including, for example, data encryption standard (DES), triple-DES, advanced encryption standard (AES), Rivest-Shamir-Adlema (RSA), open pretty good privacy (POP), etc. In some implementations, the encryptions may be performed on a session-by-session basis in which the session encryption and decryption keys are constantly updated or refreshed to thwart attacks.
100441 At least two issues still remain. The first issue is regarding the vetting process in which a user may register a biometric of the user at a trusted server, for example, a third party server. The second issue is regarding the authentication process in which the liveliness of a biometric presented on-line can be verified to thwart spoofing attempts (for example, someone using a video or photo of the registered user to pretend to be the registered user), or man-in-the-middle attacks. To these ends, Fig. 1B is a diagram showing an example interaction among a financial institution customer, a financial institution, and a third-party biometric verification service to authenticate a user based on a biometric identity of the user according to some implementations. As disclosed herein, the financial institution may include a bank, and may also be known as the relying party. The third-party service provider may also be known as the identity provider.
100451 When a financial institution customer 104 attempts to access an account through the financial institution website 106 (110A), customer 104 may be greeted with the biometric authentication window 102. The financial institution customer 104 may choose to use the biometric authentication. In response to receiving indication from customer 104 that the customer 104 would like to use the biometric authentication 102, the server running the financial institution website 106 may verify the user's authentication request using biometric data at biometric verification service 108 (110B). In some implementations, if customer 104 has never registered a biometric data at the biometric verification service 108, biometric verification service 108 may engage customer 104 to have the customer enrolled (110C).
The enrollment process may generally include asking customer 104 to provide a biometric data as the biometric identity data stored/managed by biometric verification service 108. The enrollment process may become part of the vetting process in which biometric verification service further verifies the identity of the customer 104 at other authorities, including, for example, Department of Motor Vehicles (DMV) at the state level and the State Department at the federal level. The authorities may include a credit verification agency, such as, for example, Equifax, Experian, LexisNexis. The credit verification agency may serve as a surrogate of a government-administered authority. After the customer 104 obtains an enrolled biometric identity at the biometric verification service, customer 104 may interact with biometric verification service 108 to conduct routine management of the biometric identity (110C). Routine management may include, for example, replacement of a facial pattern, revocation of an existing facial pattern, etc. Hence, the first issue of a vetting process to register an enrolled biometric data of customer 104 may be addressed.
100461 Regarding the second issue of the liveliness of the biometric data being presented on-line, the biometric login window 102 may authenticate customer 104 by analyzing real-time corneal reflection of an object (e.g., a camera device or an illumination source) and/or analyzing real-time face pose from customer 104. As discussed herein, a financial institution may be generalized as a relying party (RP) who may rely on a third party to authenticate that a customer is who the customer purports to be and that the customer is presenting a live facial biometric. The third party, known as the biometric verification service 108 in Fig. 1B, may be generalized as an identity provider (IdP).
100471 With the above generalization, Fig. 2A is a timing diagram showing an example interaction among a consumer 202, a relying party (RP) 204, and an identity provider (IdP) 206 in authenticating the consumer when the consumer has enrolled at the identity provider 206 according to some implementations. The process may initiate with consumer 202 send an authentication request to access an account managed by relying party 204 (208). In some implementations, the submission of the authentication request may correspond to consumer 202 choosing biometric authentication 102 to access an account managed by, for example, a financial institution.
[00481 When the constuner submits the authentication request, the constuner may also choose a verified liveliness login for biometric authentication (210). The verified liveliness login, as disclosed herein, may address the liveliness of biometric data being submitted on-line, e.g., by verifying facial liveliness of the consumer when the consumer submits the biometric data.
100491 For context, using a third-party intermediary to provide an authentication bridge way may be implemented in accordance with industry standards, for example, Security' Assertion Markup Language (SAML) and Open standard for Authentication (0Auth).
[00501 SAML addresses web browser single sign-on (SSO). Single sign-on solutions may be implemented at the intranet level using, for example, cookies.
Extending such single sign-on solutions beyond the intranet has been problematic and has led to the proliferation of non-interoperable proprietary technologies. In the use case addressed by SAML, the user (also known as the principal) may request a service from the relying party (also known as a service provider). The service provider then requests and obtains an identity assertion from the identity provider. On the basis of this identity assertion, the service provider can make an access control decision -deciding whether to perform some service for the connected principal.
[0051] In the case of 0Auth (including more recent 0Auth 2.0), the web application may specifically request a limited access 0Auth Token (also known as the valet key) to access the Application Program Interfaces (APIs) on behalf of the user. The valet token may explicitly name the particular rights requested, and the user may not be required to enter a credential. If the user can grant that access, the web application can retrieve the unique identifier for establishing the profile (identity) using the APIs.
[0052] Hence, authentication standards may generally provide a single sign-on so that a consumer can use, for example, the constuner's Facebook or Google login credentials to access other accounts (for example, an amazon account, a linked-in account, etc.), typically of the same consumer. Some authentication methodologies as disclosed herein may provide a single authentication proof of identity by using biometric data, for example, using facial biometric.
[0053] Referring to Fig. 2A, in response to receiving the consumer choice of verified liveliness login, a server at the relying party may redirect consumer to proceed with verified liveliness (212). Thereafter, constuner 202 may be redirected to a server at the identity provider 206. For example, consumer 202 may submit a request to use verified liveliness by using a credential token in the form of consumer@relying_party.com (214). As illustrated in Fig. 2A, the request may be submitted at a server at identity provider 206.
[0054] Upon receiving the request to use verified liveliness, the server at identity provider 206 may perform verified liveliness authentication to verify the biometric submitted by the consumer (216). The verification may include that (i) the biometric submitted by the consumer provides affirmative proof of identity for the consumer; and (ii) the biometric submitted by the consumer on-line is fresh and live and not is being forged, for example, in a spoofing attempt.
100551 As an initial matter, the server at identity provider 206 may retrieve, for example, a facial biometric from the data submitted by consumer 202. In some implementations, the retrieved facial biometric may be sent, from a computing device of consumer 202 to the server at identity provider 206, in an encrypted format based on an encryption algorithm. The computing device may include a video input component to provide a video stream feed. In some instances, the video input component includes a webcam. Images in the video feed may be captured at a frame rate of, for example, 24 frames per second. The frames generally are digitized for digital processing.
In some instances, the images may be at a spatial resolution of, for example, 800x600 pixels, 1024x768 pixels, 1152x864 pixels, or 1280x1024 pixels. The computing device may include an imaging device (e.g., a camera device) that captures images of consumer 202. The captured images include facial biometric of consumer 202, and can be transmitted by the computing device to the server at identity provider 206.
[00561 In some implementations, the retrieved facial biometric is sent from a computing device associated with the relying party 204 or the identity provider 206. The computing device can internally include or be externally coupled to a camera device that is local to consumer 202 when consumer 202 is seeking authentication. For example, the camera device can be adjacent to an automatic teller machine (ATM) facing constuner 202 or within a registration office of a bank. The ATM or the office may be a property of the relying party 204 with whom consumer 202 seeks to perform a transaction. The camera device can capture images or a video feed that includes facial biometric of consumer 202 and transmit to the computing device. The computing device can process the captured images or video feeds and transmit the captured images or video feeds and/or the processed data to the server at identity provider 206.
100571 The server at identity provider 206 may compare the retrieved facial biometric with biometric identity data of consumer 202. The biometric identity data of consumer 202 may be pre-stored at the server at identity provider 206. The comparison may be conducted in a manner consistent with the context-dependent rigidity requirement of the authentication.
For example, if the amount in transaction is less than a particular threshold, such as $50, then the degree of matching between the facial biometric as submitted by consumer 202 on-line and the pre-stored facial biometric at the server at the identity provider 206 may be reduced.
On the other hand, however, the if amount in transaction is higher than a particular threshold, such as, for example, $1,000, then the degree of matching between the facial biometric as submitted by consumer 202 on-line and the pre-stored facial biometric at the server at the identity provider 205 may be heightened. In a related example, if the transaction is in the nature of a mortgage application, a credit card application, the degree of matching may also be heightened. A scoring mechanism may be implemented to generate a score to quantify the degree of resemblance between the facial biometric as submitted by consumer 202 on-line and the pre-stored facial biometric at the server of identity provider 206.
The score of resemblance may be used by the server at identity provider 206 to determine whether the retrieved facial biometric as submitted by consumer 202 on-line is sufficient to prove that the submitting consumer 202 is the person whose biometric data is stored at the server of identity provider 206.
100581 Further, the server at identity provider 206 may determine whether the retrieved facial biometric being submitted by consumer 202 on-line is a live facial biometric.
One risk associated with on-line authentication has been the prospect of a man-in-the-middle attack. For example, during spoofing, a third party may impersonate consumer 202 by replaying an authentication proof as submitted by constuner 202 earlier and was obtained by the third party. Some encryption mechanism may counter the man-in-the-middle threat by refreshing or updating the encryption key used to encrypt the stream of communication sessions between consumer 202 and identity provider 206. Generally, however, such encryption mechanism may not address the staleness of the biometric data itself. In other words, some encryption mechanism may be inadequate if a third party obtains an instance of the biometric data and attempts to replay the instance of biometric data to gain access to the account of consumer 202. Such risks may be amplified with the combination of usemame and password login because the text strings for usemame and password may not be refreshed at the granularity of each login session. Specifically, refreshing username and password for each login session (or at a sub-session level at, for example, every few seconds) may not be practical.
100591 Some implementations as disclosed herein may verify the liveliness of the biometric data as submitted by consumer 202, for example, by performing verification of facial liveliness of consumer 202. In some implementations, the server at identity provider 206 performs liveliness verification based on the submitted images or video feeds. In some implementations, a computing device associated with consumer 202 performs liveliness verification based on the captured images or video feeds and submits a result of liveliness verification to the server at identity provider 206, e.g., in an authentication request including the biometric data of consumer 202. The server at identity provider 206 can then seamlessly determine that the biometric data submitted by consumer 202 is live. In some implementations, a computing device associated with relying party 204 or identity provider 206 that obtains images or video feeds from a camera device local to consumer 202, performs liveliness verification based on the captured images or video feeds and submits a result of liveliness verification to the server at identity provider 206.
[0060] In some implementations, liveliness verification may include analysis of a corneal reflection from consumer 202. Initially, the face of consumer 202 may be detected.
Referring to Figs. 3A and 3B, the eye 301 of consumer 202 may then be detected in a video feed or images. Generally, such detections include segmentation techniques that leverage face or eye template data. Such template data may be morphed through various shape transforms to identify face or eye area in the video feed or images. Once eye 301 is detected, cornea region 302 may then be determined in the video feed or images.
Reflections of the corneal region in the video frames or images may then be analyzed.
[0061] First, an object at the location of constuner 202 may cause corneal reflection from consumer 202. Here, the object may include an illumination source such as a light source, a camera device such as a webcam, or a device or subject (e.g., a window or a table) brighter than the ambient environment. In some examples, the object is a camera in a mobile device of consumer 202. Consumer 202 uses the camera to capture facial images or biometric of consumer 202. In some examples, the object may be in a place local to consumer 202. The place may be associated with relying party 204 with whom consumer 202 seeks to perform a transaction. For example, the place includes an automatic teller machine (ATM) facing consumer 202. The object can be mounted adjacent to the ATM. In some instances, the object may be fixed or positioned to the locale where consumer 202 is seeking authentication. Generally, the object is readily visible under normal illumination condition and can cause specular reflections from consumer 202 on-site. The specular reflection off the cornea of consumer 202 may be leveraged to determine liveliness of the authentication.
[0062] In some implementations, if the object has moved, the specular reflections of such object in the corneal of consumer 202 are expected to move accordingly.
If the specular reflection of such object moves in concert with the movement of the object on-site, then consumer 202 can be authenticated as live. In some cases, the movement of the object is performed by consumer 202. For example, consumer 202 is prompted to move a mobile phone including a camera device, and the corneal reflection of the camera device is moved accordingly. The movement of the camera device on-site can be measured by an on-board accelerometer or gyroscope in the mobile device. In some cases, the object is moved in a manner unpredictable by consumer 202 (or would-be hackers as man in the middle).
[0063] For example, the object can be moved from position 1 to position 2.
When the object is at position 1, a first image is captured. The location, shape, and color/wavelength of the reflected object may be detected by analyzing the corneal reflections in the captured first image. The detection can leverage segmentation techniques to automatically zero in on the corneal reflection 302A of the object. Thereafter, when the object is moved to position 2, a second image is captured. The location, shape, and color/wavelength of the reflected object may be detected anew as corneal reflection 302B by analyzing the captured second image. In some cases, the trajectory of the motion of object may be tracked by monitoring the reflected object. As illustrated, changes in the location of the detected object in corneal reflections 302A and 302B may be compared to the changes of object locations, e.g., programmed or measured location changes of the object. By way of illustration, some implementations may register the position of the reflected object relative to reflections of other structures adjacent to consumer 202 in a room or in front of an ATM where consumer 202 is seeking authentication. Some implementations may register the position of the reflected object relation to other bodily structures in or around the eye of consumer 202.
[0064] Fig. 4 illustrates an example object camera 403 that has been changed position by virtue of ann adjustments of consumer 202. Camera 403 is held on arm 402 of consumer 202, and faces head 401 of consumer 202. In position 1, arm 402 is more extended, for example, to be closer to the face of consumer 202. In position 2, arm 402 is more recoiled, for example, to move further away from the face of consumer 202. Such positional changes of camera 403 are expected to cause changes in the reflected camera from cornea of consumer 202. In some examples, the object may rely on a mechanism that leads a changed position of the object. The mechanism can be pan, zoom, tilt (PZT), or any other suitable mechanism.
[0065] Motions of an object can cause more interesting changed location and shape of the reflected object in corneal reflections. The object may be programmed to move by virtue of a controller such as on-board accelerometer or gyro motors. In some cases, the object is moved in a predetermined (or preprogrammed) motion. The predetermined motion can be previously embedded in the object. The predetermined motion can be also transmitted to the controller, e.g., from the server at identity provider 206 or a computing device coupled to the controller. Each time the predetermined motion can be different from previous predetermined motions. In some cases, the object is moved randomly. The actual movement of the object can be measured or detected by sensors such as on-board accelerometer or gyro motors.
[0066] The changes of the object may not be limited to mere positional changes manifested as translational or rotational adjustments. The variety' of such changes in position and change can be leveraged to provide an enriched analysis of corneal reflections. In one example, the object can even change to a position to cause the wavelength of corneal reflection to change. By way of illustration, the object can be translated to deflect illumination from a different illumination source facing consumer 202, e.g., a flashlight of a mobile device of consumer 202.
100671 Notably, even though the head or face of consumer 202 may or may not move, the object moves, in a pattern known to the identity provider but unknown to consumer 202 (or would-be hackers as man in the middle) beforehand, to cause changes in cornea reflection of the object. In some cases, the object can be moved from a first position to a second position in relatively fast speed compared to the movement of the head or face of consumer 202, such that the movement of the head or face of consumer 202 can be considered as relatively static. A camera device can capture a first facial image of consumer 202 when the object is at the first position at a first time point and capture a second facial image of consumer 202 when the object is at the second position at a second time point.
An interval between the first and second time points can be short, e.g., less than I
second. In some cases, the movement of the head or face of consumer 202 can be leveraged by analysis of movement of static or unmoved subjects (e.g., a window or a table adjacent to consumer 202) in the captured images. If consumer 202 is prompted to move the object, consumer 202 can be notified not to move the head or face of consumer 202.
100681 A camera device is configured to capture the images and/or video feeds of consumer 202. The camera device can be positioned facing a face of consumer 202. In some examples, the camera device can be internally included in a mobile device of consumer 202.
Consumer 202 can hold the mobile device such that the camera device faces the face of consumer 202 to capture facial images. The camera device can be positioned such that the camera device is at a tilted angle to the eye of consumer 202 or opposite to the eye of consumer 202. In some examples, consumer 202 is seeking authentication before an ATM.
The camera device can be fixed on the ATM or adjacent to the ATM.
100691 The camera device can capture the images and/or video feeds for verifying facial liveliness of consumer 202, e.g., without interaction with consumer 202 or without notice to consumer 202. In some cases, the movement of the camera device can be associated with a natural motion of consumer 202. A sensor, e.g., an accelerometer, a gyroscope, or a global positioning system (GPS) on a mobile device including the camera device, can measure the movement of the camera device. The mobile device can determine that the camera device is moving by using the sensor and then transmit a request to the camera device to capture facial images of consumer 202.
[0070] In some implementations, upon receiving the request to use verified liveliness, the server at identity provider 206 may transmit a request to the camera device, e.g., directly or through a computing device coupled to camera device and in communication with the server. In response to the request, the camera device can automatically capture facial images of consumer 202 when the object is moved between positions. The object can be moved based on the request. For example, a computing device in communication with the server receives the request and transmits a command to a controller of the object for moving the object. The object can be moved between positions in synchronization with the camera device capturing facial images of consumer 202.
[0071] The movement of the object can be associated with respective time points.
For example, at a first time point, the object is at a first position; at a second time point, the object is moved to a second position. The camera device can capture facial images or video feeds of consumer 202 during the movement of the object and/or before and after the movement of the object. The captured facial images are also associated with respective time points. The time points can have a high accuracy, e.g., less than 1 millisecond. For example, the camera device captures a first facial image when the object is at the first position at the first time point, and a second facial image when the object is at the second position at the second time point. Based on the time points, the captured facial images of consumer 202 can be associated with the positions of the object. Thus, the corneal reflection of the object in the captured images or video feeds can be associated with the movement of the object, which can be used to verify facial liveliness of consumer 202.
100721 The camera device can also capture images and/or video feeds of consumer 202 that include biometric data submitted by consumer 202 to the server at identity provider 206 for authentication, e.g., by interacting with consumer 202 to obtain the biometric data. In some cases, an additional imaging device or a video input component is used to capture images and/or video feeds that include biometric data of consumer 202 for authentication.
[0073] In some implementations, the object is the camera device. As noted above, the camera device can be moved by consumer 202 or by a mechanism in a controlled manner or randomly. The motion (and/or orientation and/or relative position) of the camera device can be measured, e.g., by on-board accelerometer or gyroscope. For verification of facial liveliness of consumer 202, the detected position changes in the corneal reflections of the camera device may be correlated to the predetermined or measured motion of the camera device.
100741 In some implementations, the object is a static subject adjacent to consumer 202, e.g., a door behind consumer 202 or a window besides consumer 202. The camera device is moved between positions and captures facial images of consumer 202 at these positions. The facial images include corneal reflections of the object. For verification of facial liveliness of consumer 202, the detected position changes in the corneal reflections of the object may be correlated to a predetermined or measured motion of the camera device.
[0075] In some implementations, the object is a subject configured to be moved between positions in a motion for liveliness verification, while the camera device is static and captures facial images of consumer 202 when the object is at different positions. The facial images include corneal reflections of the object. For verification of facial liveliness of consumer 202, the detected position changes in the corneal reflections of the object may be correlated to a predetermined or measured motion of the object.
[0076] In some implementations, the object is a subject configured to be moved between positions in a first motion pattern, while the camera device is also moved between positions in a second motion pattern. Moving both the object and the camera device for liveliness verification can increase security against hackers. The camera device captures facial images of consumer 202 when the object is at different positions and the camera device is also at different positions. The facial images include corneal reflections of the object. For verification of facial liveliness of consumer 202, the detected position changes in the corneal reflections of the object may be correlated to a motion pattern based on the first motion pattern (predetermined or measured) of the object and the second motion pattern (predetermined or measured) of the camera device. The motion pattern can be a relative motion pattern for the object and the camera device.
[0077] In some implementations, the object is static, and the camera device is also static. An illumination on the object from an illuminator can change at different time points, e.g., changing illumination colors or patterns, and the corneal reflection of the object can change correspondingly. For verification of facial liveliness of consumer 202, the detected illumination change on the object in the corneal reflections of the object may be correlated to a known illumination change of the illuminator.
100781 In some implementations, the object is an illuminator, e.g., a light source and/or a digital illumination source displayed on a screen, e.g., a liquid crystal display (LCD) of a mobile device. The camera device can capture facial images of consumer 202 that include corneal reflections of the object. For verification of facial liveliness of consumer 202, the detected illuminator in the corneal reflections of the object may be correlated to a predetermined or measured illumination change of the illuminator. Using illumination for liveliness verification is described with respect to Fig. 5 below.
[0079] In some implementations, the camera device is static and/or the object is also static. A computing device coupled to the camera device can prompt consumer 202 to move the head or the face in a motion pattern, e.g., turning from left to right.
The camera device can capture facial images of consumer 202 during the movement of consumer 202.
For verification of facial liveliness of consumer 202, the detected position changes in the corneal reflections of the object may be correlated to a motion of consumer 202.
100801 In some implementations, the object and/or the camera device can be moved in a motion pattern, and an illumination on the object can also change at different time points, e.g., changing illumination colors or patterns. Verification of facial liveliness of consumer 202 can be based on: 1) the correlation between the detected position changes in the corneal reflections of the object and the motion pattern of the object and/or the camera device, and 2) the correlation between the detected illumination change on the object in the corneal reflections of the object and the predetermined or measured illumination change.
100811 Initially, the images during the movement of the object and/or the camera device may be captured in a video stream at a full frame rate (for example, as determined by Nation Television System Committee¨NTSC, or as determined by the video communication standard H.261). In some implementations, for efficiency of communication, only a portion of the captured video stream may be transmitted, for example, in accordance with adaptive video codec specification of H.261. When changes in corneal reflections are about to initiate, the video feed may be digitalized and transmitted to the server at identity provider 206 at a rate faster than a floor-level rate (when consumer 202 remains stationary). In some examples, the video feed or images are transmitted to the server at identity provider 206 by the camera device. In some examples, a computing device coupled to the camera device receives the images or video feed from the camera device and transmits the images or video feed to the server at identity provider 206.
100821 As noted above, the server at identity provider 206 can perform liveliness verification based on the transmitted images or video feed. The server can analyze the images or video feed to determine corneal reflections of the object. Based on the determined corneal reflections, the server can determine a position change in the corneal reflections of the object. The server can compare the position change in the corneal reflections of the object to a motion pattern and determine facial liveliness of consumer 202 based on a result of the comparison. In some examples, the motion pattern is a predetermined motion pattern of the object or the camera device or both when moved between positions. The server can pre-store the predetermined motion pattern. In some examples, the motion pattern is a measured motion pattern of the object or the camera device or both when moved between positions.
The motion pattern can be transmitted to the server together with the captured images or video feed. In some examples, the server at identity provider 206 can analyze the received images and/or video feeds to determine the motion pattern, e.g., by analyzing a positon change of a static subject in the captured images and/or video feeds.
100831 In some implementations, a computing device coupled to the camera device performs liveliness verification based on the captured images or video feeds and submits a result of liveliness verification to the server at identity provider 206, e.g., with or without transmitting the captured images or video feeds. In some cases, the computing device receives biometric data of consumer 202 for authentication and transmits the result of liveliness verification together with the biometric data of consumer 202 to the server at identity provider 206.
[0084] In some examples, the computing device is associated with relying party 204 or identity provider 206. For example, consumer 202 is seeking authentication using an ATM associated with relying party 204. The computing device can be local to the ATM and coupled to the camera device that faces consumer 202. The computing device can receive from the server at identity provider 206 a request to verify liveliness of consumer 202. In response to the request, the computing device can transmit commands to the camera device and/or the object to ask the camera device to capture images and/or video feeds when the camera device and/or the object is moved between positions in a motion pattern, e.g., unknown to consumer 202 and without input from consumer 202. In some cases, the computing device predetermines the motion pattern and transmits the motion pattern to the camera device and/or the object. The computing device can change the motion pattern each time when performing liveliness verification. In some cases, the motion pattern is a preprogrammed or predetermined motion pattern. The computing device can pre-store the motion pattern. In some cases, the motion pattern is measured by sensors when the camera device and/or the object is moved with positions. The computing device can receive the measured motion pattern from the sensors.
[0085] in some examples, the computing device is associated with consumer 202, e.g., a smartphone or a computer. For example, consumer 202 is seeking authentication at home. The computing device can internally include a camera device or externally be coupled to a camera device. In some cases, an application (or an App when the computing device is a smartphone) provided by identity provider 204 in installed in the computing device. Upon Neel ing a request from the server at identity provider 204 or determining consumer 202 chooses to use liveliness verification for biometric authentication, the application can solicit consumer 202 to move the camera device in relative to the face of consumer 202. By using a sensor (e.g., an accelerometer or gyroscope, the computing device can determine a movement of the camera device. In some examples, the computing device can analyze the captured images and/or video feeds to determine the motion of the camera device, e.g., by analyzing a positon change of a static subject in the captured images and/or video feeds.
The images of the static subject can be corneal reflections or outside of the corneal area.
The computing device can perform liveliness verification based on the motion pattern and the corneal reflections of the camera device in captured images and/or video feeds. In some cases, the application in the computing device controls an illumination source such as a flashlight of the camera device to change illumination color or pattern unknown to consumer 202.
The application can perform liveliness verification based on the changed color or pattern of the illumination source and the corneal reflections of the camera device in captured images and/or video feeds.
[0086] Similar to facial authentication, a scoring mechanism may be implemented to quantify the degree of match between the response corneal reflection of the object and the expected change (e.g., a motion pattern or an illumination change). The scoring mechanism may generate a numerical value as a measure for determining liveliness of the face of consumer 202. The degree of match may depend on the context and hence can vary from application to application. For example, the degree of match may depend on the nature of the underlying transaction or the dollar amount involved in the underlying transaction.
100871 In some implementations, corneal reflections of the object in sequential facial images of consumer 202 are detected. The sequential facial images are captured at sequential time points by a camera device. An expected corneal reflection of the object at a particular time point can be determined based on a determined corneal reflection in a previous time point and a motion between the previous time point and the particular time point. The facial liveliness of consumer 202 can be assessed by determining likelihood between the expected corneal reflection and the determined corneal reflection from the captured facial images at the particular time point.
[0088] Corneal reflection may proffer benefits as non-intrusive and capable of handling a wide band of operations (for example, from infra-red to visible light).
Particularly, corneal reflection tends to be more specular than skin reflections, which tend to be more scattered. Implementations disclosed herein may not include verbal directives to prompt consumer 202 to change facial gestures or other bodily positions.
[0089] Further, pose of the face of consumer 202 may be estimated, for example, using an automatic algorithm. The automatic algorithm can detect the pose of the face relative to the camera device in each capture image, e.g., from a video feed.
Additionally, an expected pose or change in pose of a face may be calculated based on the measured or known positions of the camera device, e.g., by on-board accelerometer or gyro motors or from analyzing the corneal reflections of the camera device. In some implementations, consumer 202 is prompted to move the camera device in relative to the face of consumer 202. Such calculation can be performed for each additional image captured from the video feed. More interesting, the expected pose or changes in pose, as calculated from the measured locations of the camera device and the detected pose of the face in the capture image may be correlated in a quantitative manner using a scoring mechanism. The numerical value as determined from the correlation can be used to gauge the liveliness of the face.
[0090] In some implementations, pose of the face of consumer 202 is detected or estimated by analyzing one or more facial landmarks of consumer 202. The facial landmark features can include the nose, the forehead, the eye brows, the eyes, the eye corners, the lips, and/or the mouths. Poses of the face or facial portraits from a live session reflect the face of the subject at various angles. Due to slight changes in viewing angles of the pose or facial portrait, the apparent orientation and size of each facial landmark varies.
Orientation and size of one or more facial landmark features can be expected to vary. Each landmark feature may be associated with one conunensurate change in orientation and size. The threshold of expected change may be a combination of all expected changes for the landmark features. As noted above, the threshold may be application-specific. For example, for transaction involving dollar amounts under a particular level, the comparison may be less sophisticated in order to determine that the session is live and the submitted facial biometric is genuine.
Otherwise, when the transaction involves a paramount identification document, for example, renewal of passport, the comparison may involve a more comprehensively determined threshold.
[0091] In some implementations, verification of facial liveliness of consumer 202 is based on both the corneal reflection of the object and the pose change of the face. A scoring mechanism may be implemented to quantify the degree of match between the response corneal reflection of the object and the known motion pattern to determine a first numerical value. The scorning mechanism can be also implemented to quantify the degree of match between the expected pose or changes in pose (e.g., calculated from the measured locations of the camera device) and the detected pose in a capture image to determine a second numerical value. Then an overall score can be determined based on the first and second matching values, e.g., with a weighting factor. The weighting factors of the first and second matching values can be, for example, 50% and 50% or 60% and 40%. The additional verification may increase a confidence level that consumer 202 is indeed alive and is the person whose biometric is stored at the server of identity provider 206.
100921 In some implementations, liveliness verification includes analysis of a corneal reflection of an illuminator from constuner 202. Initially, the face of consumer 202 may be detected. Referring to Fig. 5, an eye 501 of consumer 202 may then be detected in a video feed (or an image). Generally, such detections include segmentation techniques that leverage face or eye template data. Such template data may be morphed through various shape transforms to identify face or eye area in a video feed. Once eye 501 is detected, cornea region 502 may then be determined in the video feed. Reflections of illuminations on the corneal region in the video frame may then be analyzed.
100931 First, illuminator 506 adjacent to consumer 202 may be activated.
For example, illuminator 506 can be activated in response to determining that consumer 202 chooses a verified liveliness login for biometric authentication, e.g., 210 of FIG. 2A. As an example, illuminator 506 in Fig. 5 includes illumination source 506A and illumination source 506B with two distinct colors (e.g., blue or red). An illumination source can be a physical light source or a digital illumination source. In some implementations, illuminator 506 includes physical light sources, e.g., light-emitting diodes (LEDs) or light lamps. Illuminator 506 may include one or more light sources at distinct light wavelengths thus with one or more distinct colors. In some implementations, illuminator 306 includes one or more digital illumination sources displayed on a screen, e.g., on a laser-powered phosphor display (LPD), a liquid-crystal display (LCD) screen, a light-emitting diode (LED) screen such as an active matrix organic LED screen, or any suitable display screen. A digital illtunination source can be a digital color pattern displayed on a screen, e.g., a LCD screen of a mobile device. The digital illumination sources can be provided by an application (or an App) installed on a computing device of consumer 202 or provided by the server at identity provider 206, e.g., in a user interface displayed on the computing device of consumer 202. In some implementations, illuminator 306 includes one or more physically light sources and one or more digital illumination sources.
100941 Illuminator 506 may be included or displayed on a consumer electronic device, such as a smartphone, a mobile phone, a mobile computing device, or a desktop computer, from which consumer 202 attempts to authenticate his or her identity and liveliness. Illuminator 506 may be included on the console of an automatic teller machine (ATM) of a relying party or adjacent to the ATM. A computing device associated with the relying party or identity provider 206 may be configured to control illuminator 506.
Consumer 202 may attempt to authenticate his or her identity and liveliness through the computing device.
[0095] Once activated, illuminator 506 may illuminate consumer 202. The illumination from illuminator 506 can be straight to the eye of consumer 202 or to the eye of consumer 202 with an incident angle. In some examples, illuminator 506 is positioned so that the illumination from illuminator 506 is incident on the eye of consumer 202 with an angle. In some examples, consumer 202 can be prompted to move illtuninator 506 (e.g., a flashlight on his/her smartphone) to a side of the face (or eyes) or move his/her face (or eyes) so that the illumination from illuminator 506 is incident on the eye of consumer 202 with an angle.
[0096] Information of the illumination (e.g., a position, a shape, a duration, a color or wavelength, or an illumination pattern) can be known or predetermined. In some examples, a computing device of consumer 202 performs liveliness verification. The information of illtunination can be known to the computing device but unknown to consumer 202 (or would-be hackers as man in the middle) beforehand. The information of illumination can be generated by the computing device or provided by the server at identity provider 206.
[0097] The corneal reflection immediately following the activation of illuminator 506 may be captured, e.g., by a camera device adjacent to consumer 202. The camera device can keep capturing facial images of consumer 202 before, during, and after the activation of illuminator 506. In some examples, the captured image frame may be analyzed to ascertain whether the corneal reflection is in agreement with the expected (known or predetermined) illumination. In analyzing the corneal reflections, the position, shape, and color/wavelength of the reflected illumination in the cornea of consumer 202 may be detected and compared to the expectation illumination. In one example, the position of the corneal reflection, for example, reflection spot 504A or 504B may be correlated with the position of illumination sources 506A or 506B. In another example, the shape of the reflection spot 504A or 504B
may be compared to the shape of illumination sources 506A or 506B. In yet another example, the wavelength (or color) of reflected light from reflection spot 504A or 504B may be compared to the wavelengths (or colors) of illumination sources 506A or 506B.
100981 In some examples, illumination pattern by illuminator 506 (e.g., an illumination source 506A or 506B) may be modulated over time to form a temporal illumination pattern. For example, illumination source 506A may generate illumination in bursts, each burst with a same or different temporal duration. For example, a first temporal duration can be 1 second and a second temporal duration can be 2 seconds. The temporal illumination pattern can also include a change in illumination shape, e.g., varying between a square shape, a triangle shape, and/or a circle shape, and/or a change in illumination color, during different temporal durations or at different time points. In one example, illuminator 506 provides a red square at a first temporal duration, a red circle at a second temporal duration, no illumination at a third temporal duration, and a blue triangle at a fourth temporal duration. The corneal reflection in captured sequential images may be analyzed to determine whether the corneal reflection contains the same temporal pattern. The corneal reflection in captured sequential images can be also individually analyzed to determine whether the corneal reflection in the captured sequential images matches the individually expected corneal reflections of the illuminator 506. Determination of facial liveliness of consumer 202 can be based on the match results of the captured sequential images.
[0099i In some examples, illumination pattern by two or more illumination sources (e.g., ilkunination sources 506A and 506B) may be modulated over time to form a temporal illumination pattern. The illumination sources can illuminate same or different colors. For example, illumination sources 506A and 506B may generate illumination in bursts, each with a same or distinct temporal duration and/or each with a same or different distinct illumination shape during the temporal duration. In an example, at a first time point, illumination source 506A illuminates a red square and illumination source 506B illuminates a blue square. At a second time point, illumination source 506A illuminates a red circle and illumination source 506B illuminates a blue square. At a third time point, illumination source 506 illuminates a red circle and illumination source 506 illuminates a blue circle. At a fourth time point, illumination source 506 illuminates a red rectangle and illumination source 506 illuminates a red circle. The corneal reflection in sequential images may be analyzed to determine whether the corneal reflection contains the same temporal pattern.
1001001 In some examples, the illumination pattern by two or more illumination sources (e.g., illumination sources 506A and 506B) may be modulated spatially to form a spatial illumination pattern. For example, illumination sources 506A and 506B
may be activated to illuminate cornea 502 in an interleaved manner. Generally, the illumination pattern from illumination sources 506A and 506B may be varied by turning on or off illumination source 506A, illumination source 506A, or the combination of illumination sources 506A and 506B at different time points. In response, the corneal reflection may be analyzed to determine whether the reflection contains the same spatial pattern of illumination.
As discussed herein, the temporal and spatial aspects of illumination may be combined to further buttress the variety of matches that improves the confidence of a liveliness authentication session.
[001011 The images in response to activated illiunination may be captured in a video stream at a full frame rate. In some implementations, as noted above, the server at identity provider 206 performs liveliness verification based on the captured video stream. For efficiency of communication, only a portion of the captured video stream may be transmitted.
When corneal reflections are expected to change, the video stream may be digitalized and transmitted to the server at identity provider 206 at a rate faster than a floor-level rate (when consumer 202 remains stationaty). In fact, the transmitted video stream may be analyzed by the server at identity provider 206 to determine whether the changes in corneal reflections of the video stream from consumer 202 matches the known changes in illumination of illuminator 506 or whether the corneal reflections of video frames at particular time points matches the expected corneal reflections based on known illuminations of illuminator 506 at the particular time points.
1001021 Similar to facial authentication, a scoring mechanism may be implemented to quantify the degree of match between the response corneal reflection being received at the server of the identity provider 206 in the captured images and the expected illumination (e.g., position, shape, and wavelength or color, or temporal and/or spatial pattern).
The degree of match may depend on the context and hence can vary from application to application. For example, the degree of match may depend on the nature of the underlying transaction or the dollar amount involved in the underlying transaction.
1001031 In some implementations, a computing device receives the captured video stream, e.g., by a camera device included in or coupled to the computing device, and performs liveliness verification based on the captured video stream. The computing device can be a computing device of consumer 202 from which consumer 202 attempts to authenticate his or her identity and liveliness. For example, the computing device can be a smartphone which includes a flashlight as illuminator 506 and a camera as the camera device.
A mobile app installed on the smartphone can also provide one or more digital illiunination sources as illuminator 506. The flashlight and the digital illumination sources can be combined to be used as illuminator 506. The computing device can control illuminator 506 to illuminate based on a predetermined or known illumination pattern. The illumination pattern can be generated by the computing device or transmitted from the server at identity provider 206. The illumination of illuminator 506 can be also measured by a detector that provides the measured illumination to the computing device. The known illtunination of illuminator 506 can be stored in the computing device. The computing device can also control the camera device to capture a video stream during activation of the illumination pattern of illtuninator 506. In some other implementations, the computing device is associated with relying party or identity provider 206.
1001041 As the camera device is local to the computing device, the transmission of the captured video stream (e.g., a portion) can be fast. The computing device can analyze the transmitted video stream to determine the illumination (or illumination changes) in corneal reflections of the video stream (e.g., sequential video frames) matches the known illumination (or illumination changes) of illuminator 506. The computing device can use a scoring mechanism to quantify the degree of match between the response corneal reflection in the captured images and the expected illumination. The degree of match may depend on the context and hence can vary from application to application. Based on the scored match, the computing device can verify facial liveliness of consumer 202. The computing device can then submit the result of the verification to the server at identity provider 206, e.g., together with biometric data of consumer 202. Then the server at identity provider 206 can seamlessly determine whether the submitted biometric data is a live facial biometric of consumer 202 based on the submitted result of the verification from the computing device.
1001051 Implementations disclosed herein detect corneal reflections in response to illuminations, including illumination patterns temporally and/or spatially.
Corneal reflection may proffer benefits as non-intrusive and capable of handling a wide band of operations (for example, from infra-red to visible light). Particularly, corneal reflection tends to be more specular than skin reflections, which tend to be more scattered.
Implementations disclosed herein may not include verbal directives to prompt consumer 202 to change facial gestures or other bodily positions. Some implementations disclosed herein can detect liveliness seamlessly and without user awareness and input. For example, activation of an illumination pattern and capturing sequential images can be performed without user's notice, e.g., when facial biometric of a consumer is captured. The liveliness verification can be also completed during capturing facial biometric, e.g., by a computing device of consumer 202.
[001061 Some implementations may additionally include a biometric recognition feature to authenticate consumer 202. By way of example, server at identity provider 206 may conduct a facial recognition to compare the facial biometric of an enrolled consumer with the facial biometric of consumer 202 submitting the authentication request. If the comparison yields a matching result, the facial recognition may prove that consumer 202 is the person whose facial biometric is stored at the server at identity provider 206. In other examples, the biometric recognition may include voice recognition, speaker recognition, gait analysis, etc. Some implementations disclosed herein can detected liveliness seamlessly and without user awareness (e.g., directing the user to make any specific moves during the image capturing process). The stealth feature can be leveraged to further isolate liveliness detection from an on-line identity management session.
[00107) Referring back to FIG. 2A, after verifying that consumer 202 is alive and is the person whose biometric has been stored at identity provider, server at identity provider 206 may provide a signal that authentication is successful. If, however, consumer 202 cannot be verified as alive or as the person whose biometric has been stored at identity provider 206 in association with the account at relying party 204, server at identity provider 206 may provide a signal that authentication has failed. The signal may be embedded in a message to consumer 202 to redirect consumer 202 back to relying party 204 (218).
[00108) Consumer 202 may then return to the relying party 204 (220). If the embedded signal indicates that the verification is successful, consumer 202 may proceed to access the account at relying party 204. In returning to relying party 204, consumer 202 may attach a success signal backed up by a credential of identity provider 206.
The credential of the identity provider 206 may include, for example, a digital private key of identity provider 206, a digital watermark of identity provider 206. If, however, the embedded signal indicates that the verification has failed, consumer 202 may not access the account at relying party 204.
Moreover, consumer 202 may not even be redirected back to the relying party 204.
[00109] Fig. 2B is a timing diagram showing an example interaction among a consumer, a relying party, and an identity provider in authenticating the consumer when the consumer has not yet enrolled at the identity provider according to some implementations.
As discussed above in association with Fig. 2A, the process may initiate with consumer 202 send an authentication request to access an account managed by relying party 204 (208). In some implementations, the submission of the authentication request may correspond to consumer 202 choosing biometric authentication 102 to access an account managed by, for example, a financial institution. When the consumer submits the authentication request, the consumer may also choose to a verified liveliness login for biometric authentication (210). In response to receiving the consumer choice of verified liveliness login, a server at the relying party may redirect consumer to proceed with verified liveliness (212).
Thereafter, consumer 202 may be redirected to a server at the identity provider 206. For example, consumer 202 may submit a request to use verified liveliness by using a credential token in the form of consumer@relying_party.com (214). As illustrated in Fig. 2B, the request may be submitted at a server at identity provider 206.
[001101 The server at identity provider 206 may notice that consumer 202 has not yet registered for the verified liveliness service at identity provider 206. In response, the server at identity provider 206 may inquire the server at the relying party 204 whether the identity provider 206 may enroll consumer 202 in the verified liveliness service, as requested by consumer 202. Some relying parties may fmancially compensate identity provider 206 for the verified liveliness. In turn, these relying parties may stratify consumers into various classes with corresponding account benefits. By way of illustration, consumers with holdings above a threshold value of, for example, $50k may have the verified liveliness service as a complimentary service. Consumers with less account value or shorter membership history may need to pay a fee to relying party 204 in order to have the verified liveliness service.
[0011.11 Upon determination that consumer 202 may enroll in the verified liveliness service, server at relying party 204 may indicate to the server at identity provider 206 that the enrollment for constuner 202 may proceed (224). If, however, server at relying party 204 determines that consumer 202 may not enroll in the verified liveliness service, server at relying party 204 may inform the server at identity provider 206 that consumer 202 does not have permission to enroll in the requested service and the server at identity provider 206 may drop the enrollment request initiated from consumer 202. Particularly, server at identity provider 206 may direct consumer 202 back to relying party 204 with the flag that enrollment has failed (224).
[00112] If the server at identity provider 206 receives confirmation that identity provider 206 may proceed with enrolling consumer 202, server at identity provider 206 may administer an enrollment procedure for consumer 202 (226). Specifically, identity provider 206 may request consumer 202 to register a biometric. The biometric may include, for example, a facial biometric, a finger-print, a palm-print, an iris scan, a retina scan. In the case of a facial biometric, server at identity provider 206 may request consumer 202 to pose in front of, for example, a web camera, a video camera, or a smart-phone camera. The camera device may reside on, for example, a computing device of consumer 202, or an ATM
or in registration office of a bank. The camera device or a computing device coupled to the camera device may initially transmit a capture frame to server at identity provider 206. This captured frame may be stored at identity provider 206 as biometric information identifying consumer 202. By way of example, server at identity provider 206 or the computing device may verify the liveliness of the enrollment session as discussed above in association with Figs. 2A, 3A, 3B, 4 and 5.
[00113] Consistent with the disclosure herein, server at identity provider 206 may also initiate illumination patterns on consumer 202 to increase the confidence level that a live consumer is attempting to enroll in the service. As discussed herein, multiple illumination patterns for different and separate spatial and temporal modulations may obviate the ambiguities associated with a single illumination response. For example, when consumer 202 may not be responding to the fullest extent as expected from one single illumination, or when sampling error during video transmission may have missed the frame that contains the exact response reflection, a single response illumination may be insufficient to determine with confidence that consumer 202 is a live person attempting to enroll in the service.
[00114] When consumer 202 has been determined as a live person and has been enrolled in the verified liveliness service, the server at identity provider 206 may redirect consumer 202 back to relying party 204 with a signal that the authentication is successful (228). If consumer 202 has been successfully enrolled in the program, the server at identity provider 206 may provide a signal that the authentication has succeeded. If, however, consumer 202 cannot be verified as alive or as the person and cannot be enrolled in the verified service program, the server at identity provider 206 may provide a signal that authentication has failed. The signal may be embedded in a message to consumer 202 to redirect consumer 202 back to relying party 204 (218).
[00115] Consumer 202 may then return to the relying party 204 (230). If the embedded signal indicates that the authentication has succeeded, consumer 202 may proceed to access the account at relying party 204. In returning to relying party 204, consumer 202 may attach a success signal backed up by a credential of identity provider 206. The credential of the identity provider 206 may include, for example, a digital private key of identity provider 206, a digital watermark of identity provider 206. If, however, the embedded signal indicates that the authentication has failed, consumer 202 may not access the account at relying party 204. Moreover, consumer 202 may not even be redirected back to the relying party 204.
[00116] Notably, in some implementations, a first-time enrollee may be sent to the server at identity party 206 after the first-time enrollee has successfully logged into an account of the first-time enrollee at the relying party 204. The enrollment may provide consumers with additional security features to guard against spoofing attacks.
The security features may also include biometric recognition, as detailed above. In fact, some configurations may use the biometric recognition as a form of password authentication. In other words, the biometric recognition may be used as a surrogate for the password login.
[001171 In some implementations, however, identity provider 206 may obtain a database for a number of consumers attempting to enroll in a verified liveliness program.
During the enrollment process, the server at identity provider 206 may collect a biometric from a consumer attempting to enroll. After verifying that the consumer is a live person making a legitimate enrollment request, the server at identity provider 206 may enroll the consumer. Hence, in some implementations, the server at identity provider 206 may have a copy of a digital biometric of consumer 202 when consumer 202 has been redirected from relying party 204. When a copy of the digital biometric of consumer 202 is available, the server at identity provider 206 may only need to determine that consumer 202 is a live person.
[00118] Implementations of the subject matter and the functional operations described in this specification can be implemented in digital electronic circuitry, in tangibly-implemented computer software or firmware, in computer hardware, including the structures disclosed in this specification and their structural equivalents, or in combinations of one or more of them. Implementations of the subject matter described in this specification can be implemented as one or more computer programs, i.e., one or more modules of computer program instructions encoded on a tangible non transitory program carrier for execution by, or to control the operation of, data processing apparatus. The computer storage medium can be a machine-readable storage device, a machine-readable storage substrate, a random or serial access memory device, or a combination of one or more of them.
[001191 The tenn "data processing apparatus" refers to data processing hardware and encompasses all kinds of apparatus, devices, and machines for processing data, including, by way of example, a programmable processor, a computer, or multiple processors or computers.
The apparatus can also be or further include special purpose logic circuitry, e.g., a central processing unit (CPU), a FPGA (field programmable gate array), or an ASIC
(application specific integrated circuit). In some implementations, the data processing apparatus and/or special purpose logic circuitry may be hardware-based and/or software-based.
The apparatus can optionally include code that creates an execution environment for computer programs, e.g., code that constitutes processor firtnware, a protocol stack, a database management system, an operating system, or a combination of one or more of them. The present disclosure contemplates the use of data processing apparatuses with or without conventional operating systems, for example Linux, UNIX, Windows, Mac OS, Android, iOS or any other suitable conventional operating system.
1001201 A computer program, which may also be referred to or described as a program, software, a software application, a module, a software module, a script, or code, can be written in any form of progranuning language, including compiled or interpreted languages, or declarative or procedural languages, and it can be deployed in any form, including as a stand-alone program or as a module, component, subroutine, or other unit suitable for use in a computing environment. A computer program may, but need not, correspond to a file in a file system. A program can be stored in a portion of a file that holds other programs or data, e.g., one or more scripts stored in a markup language document, in a single file dedicated to the program in question, or in multiple coordinated files, e.g., files that store one or more modules, sub programs, or portions of code. A computer program can be deployed to be executed on one computer or on multiple computers that are located at one site or distributed across multiple sites and interconnected by a communication network.
While portions of the programs illustrated in the various figures are shown as individual modules that implement the various features and functionality through various objects, methods, or other processes, the programs may instead include a number of sub-modules, third party services, components, libraries, and such, as appropriate.
Conversely, the features and functionality of various components can be combined into single components as appropriate.
1001211 The processes and logic flows described in this specification can be performed by one or more programmable computers executing one or more computer programs to perform functions by operating on input data and generating output. The processes and logic flows can also be performed by, and apparatus can also be implemented as, special purpose logic circuitry, e.g., a central processing unit (CPU), a FPGA (field programmable gate array), or an ASIC (application specific integrated circuit).
1001221 Computers suitable for the execution of a computer program include, by way of example, can be based on general or special purpose microprocessors or both, or any other kind of central processing unit. Generally, a central processing unit will receive instructions and data from a read only memory or a random access memory or both. The essential elements of a computer are a central processing, unit for performing or executing instructions and one or more memory devices for storing instructions and data. Generally, a computer will also include, or be operatively coupled to receive data from or transfer data to, or both, one or more mass storage devices for storing data, e.g., magnetic, magneto optical disks, or optical disks. However, a computer need not have such devices. Moreover, a computer can be embedded in another device, e.g., a mobile telephone, a personal digital assistant (PDA), a mobile audio or video player, a game console, a Global Positioning System (GPS) receiver, or a portable storage device, e.g., a universal serial bus (USB) flash drive, to name just a few.
[001231 Computer readable media (transitory or non-transitory, as appropriate) suitable for storing computer program instructions and data include all forms of non volatile memory, media and memory devices, including by way of example semiconductor memory devices, e.g., EPROM, EEPROM, and flash memory devices; magnetic disks, e.g., internal hard disks or removable disks; magneto optical disks; and CD ROM and DVD-ROM disks. The memory may store various objects or data, including caches, classes, frameworks, applications, backup data, jobs, web pages, web page templates, database tables, repositories storing business and/or dynamic information, and any other appropriate information including any parameters, variables, algorithms, instructions, rules, constraints, or references thereto. Additionally, the memory may include any other appropriate data, such as logs, policies, security or access data, reporting files, as well as others. The processor and the memory can be supplemented by, or incorporated in, special purpose logic circuitry.
1001241 To provide for interaction with a user, implementations of the subject matter described in this specification can be implemented on a computer having a display device, e.g., a CRT (cathode ray tube), LCD (liquid crystal display), or plasma monitor, for displaying information to the user and a keyboard and a pointing device, e.g., a mouse or a trackball, by which the user can provide input to the computer. Other kinds of devices can be used to provide for interaction with a user as well; for example, feedback provided to the user can be any form of sensory feedback, e.g., visual feedback, auditory feedback, or tactile feedback; and input from the user can be received in any form, including acoustic, speech, or tactile input. In addition, a computer can interact with a user by sending documents to and receiving documents from a device that is used by the user; for example, by sending web pages to a web browser on a user's client device in response to requests received from the web browser.
I00125) The term "graphical user interface," or GUI, may be used in the singular or the plural to describe one or more graphical user interfaces and each of the displays of a particular graphical user interface. Therefore, a GUI may represent any graphical user interface, including but not limited to, a web browser, a touch screen, or a command line interface (CLI) that processes information and efficiently presents the information results to the user. In general, a GUI may include a plurality of user interface (UI) elements, some or all associated with a web browser, such as interactive fields, pull-down lists, and buttons operable by the business suite user. These and other UI elements may be related to or represent the functions of the web browser.
1001261 Implementations of the subject matter described in this specification can be implemented in a computing system that includes a back end component, e.g., as a data server, or that includes a middlevvare component, e.g., an application server, or that includes a front end component, e.g., a client computer having a graphical user interface or a Web browser through which a user can interact with an implementation of the subject matter described in this specification, or any combination of one or more such back end, middleware, or front end components. The components of the system can be interconnected by any form or medium of digital data communication, e.g., a communication network.
Examples of communication networks include a local area network (LAN), a wide area network (WAN), e.g., the Internet, and a wireless local area network (WLAN).
1001271 The computing system can include clients and servers. A client and server are generally remote from each other and typically interact through a communication network.
The relationship of client and server arises by virtue of computer programs running on the respective computers and having a client-server relationship to each other.
1001281 While this specification contains many specific implementation details, these should not be construed as limitations on the scope of any invention or on the scope of what may be claimed, but rather as descriptions of features that may be specific to particular implementations of particular inventions. Certain features that are described in this specification in the context of separate implementations can also be implemented in combination in a single implementation. Conversely, various features that are described in the context of a single implementation can also be implemented in multiple implementations separately or in any suitable sub-combination. Moreover, although features may be described above as acting in certain combinations and even initially claimed as such, one or more features from a claimed combination can in some cases be excised from the combination, and the claimed combination may be directed to a sub-combination or variation of a sub-combinations.
1001291 Similarly, while operations are depicted in the drawings in a particular order, this should not be understood as requiring that such operations be performed in the particular order shown or in sequential order, or that all illustrated operations be perfonned, to achieve desirable results. In certain circumstances, multitasking and parallel processing may be helpful. Moreover, the separation of various system modules and components in the implementations described above should not be understood as requiring such separation in all implementations, and it should be understood that the described program components and systems can generally be integrated together in a single software product or packaged into multiple software products.
[00130j Particular implementations of the subject matter have been described. Other implementations, alterations, and permutations of the described implementations are within the scope of the following claims as will be apparent to those skilled in the art. For example, the actions recited in the claims can be performed in a different order and still achieve desirable results.
1001311 Accordingly, the above description of example implementations does not define or constrain this disclosure. Other changes, substitutions, and alterations are also possible without departing from the spirit and scope of this disclosure.
[0020] These and other implementations can each optionally include one or more of the following features: comparing the determined corneal reflection of the illuminator to the expected corneal reflection of the illuminator can include: correlating the determined corneal reflection of the illuminator to the expected corneal reflection of the illuminator; scoring a matching quality based on a result of the correlation; and comparing the scored matching quality to a predetermined threshold. Determining facial liveliness of the subject can include determining that the face of the subject is live in response to determining that the scored matching quality is beyond the predetermined threshold. Determining a corneal reflection of the illuminator can include determining the corneal reflection of the illuminator in at least one of: color, shape, position, or pattern.
[0021] The actions can include activating the illuminator to illuminate based on an illumination pattern. In some examples, the actions include generating the illumination pattern for the illuminator. In some examples, activating the illuminator comprises:
activating the illuminator in response to determining that the subject chooses to use liveliness verification for biometric authentication.
[0022] The illuminator can include two or more illumination sources with respective colors and at respective positions, respectively. In some implementations, the illumination of the illuminator is based on a temporal illumination pattern of the two or more illumination sources. The two or more illumination sources can generate illumination in bursts each with respective temporal durations. At least one of the two or more illumination sources can have different illumination shapes for different temporal durations. In some implementations, the illumination of the illuminator is based on a spatial illumination pattern of the two or more ilhunination sources. Each of the two or more illumination sources can have an illumination on or off state.
[0023] In some examples, the illuminator includes one or more physical light sources each at respective light wavelengths. In some examples, the illuminator includes one or more digital illumination sources displayed on a screen to provide respective illumination colors.
The actions can include generating the digital illumination sources. The illuminator can be positioned such that illumination from the illuminator is incident on an eye of the subject at an angle. The actions can include prompting the subject to change a relative position of the illuminator and the eye of the subject.
[0024] The actions can include receiving at least a portion of a video stream feed from a camera device, the portion of the video stream feed including facial images of the subject captured by the camera device at sequential time points. The action can further include: processing a second facial image of the subject to determine a second corneal reflection of the illuminator, the second facial image being captured at a second time point sequential to the time point for the first facial image; determining a second expected corneal reflection of the illuminator based on a second illumination of the illuminator at the second time point, the second illumination being different from the first illumination; comparing the determined second corneal reflection of the illuminator to the expected second corneal reflection of the illuminator to obtain a second comparison result; and determining facial liveliness of the subject based on the comparison result and the second comparison result.
[0025] The actions can include transmitting the determined facial liveliness of the subject to a computing device at identity provider. The actions can also include transmitting, to a computing device at identity provider, an authentication request of the subject for accessing an account managed by a relying party different from the identity provider, the authentication request including biometric data of the subject and the determined facial liveliness of the subject.
[0026] Implementations of the present disclosure can also provide computer-implemented methods for detecting facial liveliness that include actions of processing first and second facial images of a subject to determine first and second corneal reflections of an illuminator adjacent to the subject, the first and second facial images being captured at first and second sequential time points, respectively: determining a corneal reflection change of the illuminator based on the determined first and second corneal reflections;
comparing the determined corneal reflection change of the illuminator to an illumination change of the ilhuninator associated with the first and second sequential time points; and determining facial liveliness of the subject based on a result of the comparison.
[0027] These and other implementations can each optionally include one or more of the following features: the actions can include activating the illuminator to change illumination of the illuminator based on the illumination change. The illumination change of the illuminator can include at least one of: an illumination color change, an illumination shape change, an illumination position change, an illumination on/off state change, or an ilhunination temporal duration change. Determining a corneal reflection change of the illuminator can include determining a change of the corneal reflection of the illuminator in at least one of: color, shape, position, on/off state, or temporal duration.
100281 The illuminator can include at least one of: one or more physical light sources each at respective light wavelengths, or one or more digital illumination sources displayed on a screen to provide respective illumination colors. In some implementations, the illuminator includes first and second illumination sources with first and second colors and at first and second positions, respectively. In some examples, the illumination change of the illuminator includes a change of a temporal illumination pattern of the first and second illumination sources, and the first and second illumination sources can generate illumination in bursts each with respective temporal durations. At least one of the first illumination source or the second illumination source can change an illumination shape for different temporal durations. In some examples, the illumination change of the illuminator includes a change of a spatial illtunination pattern of the first and second illumination sources over time, and at least one of the first illumination source or the second illumination source can change an illumination on/off state over time.
100291 The present disclosure also provides one or more non-transitory computer-readable storage media coupled to one or more processors and having instructions stored thereon which, when executed by the one or more processors, cause the one or more processors to perform operations in accordance with implementations of the methods provided herein.
100301 The present disclosure further provides a system for implementing the methods provided herein. The system includes one or more processors, and a computer-readable storage medium coupled to the one or more processors having instructions stored thereon which, when executed by the one or more processors, cause the one or more processors to perform operations in accordance with implementations of the methods provided herein.
100311 It is appreciated that methods in accordance with the present disclosure can include any combination of the aspects and features described herein. That is, methods in accordance with the present disclosure are not limited to the combinations of aspects and features specifically described herein, but also include any combination of the aspects and features provided.
100321 The details of one or more implementations of the present disclosure are set forth in the accompanying drawings and the description below. Other features and advantages of the present disclosure will be apparent from the description and drawings, and from the claims.
DESCRIPTION OF DRAWINGS
[0033] Fig. IA is a diagram showing an example web site capable of authenticating a user based on a biometric identity of the user according to some implementations.
[0034] Fig. 1B is a diagram showing an example interaction among a financial institution customer, a financial institution, and a third-party biometric verification service to authenticate a user based on a biometric identity of the user according to some implementations.
[0035] Fig. 2A is a timing diagram showing an example interaction among a consumer, a relying party, and an identity provider in authenticating the consumer when the consumer has enrolled at the identity provider according to some implementations.
[0036] Fig. 2B is a timing diagram showing an example interaction among a consumer, a relying party, and an identity provider in authenticating the consumer when the consumer has not yet enrolled at the identity provider according to some implementations.
[0037] Figs. 3A and 3B illustrate the changed positions of the corneal reflection of an object.
[0038] Fig. 4 illustrates an object that has changed position by virtue of arm adjustments.
[0039] Fig. 5 is a diagram showing an example verification based on corneal reflection from illumination sources.
[0040] Like reference symbols in the various drawings indicate like elements.
DETAILED DESCRIPTION
[0041] Proof of identity may present a perennial challenge in our daily transactions.
With the advent of the Internet, comes the age of e-commerce in which on-line transactions may replace in-person transactions. However, the sheer volume and complexity of these on-line transactions may give rise to a digital world fraught with peril, including, for example, identity theft, identity fraud, spoofmg, phishing, etc. Notably, such risks may not be new in the Internet age, although the Internet may have amplified such risks. As the society moves towards cloud computing, more and more databases may become accessible.
Identity data in some databases may be more reliable and robust than others, based on history or tradition.
As connectivity becomes ubiquitous and as more identity databases become available on accessible platforms, identity-related data housed therein can be accessed to increase the confidence in the quality of transactions conducted either online and in person. In addition, biometric data submitted on-line can be verified for liveliness to deter spoofing attempts. In some implementations, the liveliness verification can be leverage analysis of real-time corneal reflection and/or real-time face pose to increase confidence in the biometric data being submitted on-line for authentication. Implementations disclosed herein may be extended to enterprise entities other than financial institutions. This application is directed to systems and methods to detect liveliness of a facial representation during, for example, an on-line session between two parties over a network. Note that the terms "liveliness" and liveness" can be used interchangeably here.
[0042] Fig. IA is a diagram showing an example web site capable of authenticating a user based on a biometric identity of the user according to some implementations. As illustrated, a user name and password may be required for logging into an account administered by the server hosting the web-site. For example, the server may run an implementation of a hypertext transmission protocol secure (https). In the Internet age, a consumer user may have accounts at dozens of more web-sites, each with disparate requirements of user names and passwords. The consumer user may have a hard time keeping track of user names and passwords at each of the web-sites. Losing track of the registered username or password can hinder on-line access convenience. The consumer user may store the usernames and passwords for each account at a central file. But, access to the central file may be breached, leading to comprises in the stored usernames and passwords. In fact, simple on-line identities including user name and password may be subject to identity theft and identity fraud. A recent survey revealed that identity theft in the United States rose to a three-year high in 2012, with more than 5 percent of the adult population, or 12.6 million people, falling victim to such crimes. The numbers are up from 4.9 percent in 2011 and 4.35 percent in 2010. The incidence of identity theft is only expected to rise. To mitigate the risks arising from identity theft in the context of e-commerce, some implementations, as disclosed herein may choose to authenticate biometric data presented the user on-line.
Such biometric data may include, but are not limited to, facial pattern, finger print, palm print, retina scan, iris scan, DNA pattern, voice characteristics, gait analysis. Generally speaking, such biometric data submitted on-line may be in a digital form to facilitate electronic verification.
[0043] As illustrated by Fig. 1A, biometric login window 102 may be provided by a third-party, different from, for example, the entity that runs bigbank.com.
The login window 102 may present the captured facial biometric of the user as the user attempts to log in based on the user's biometric. When the user's biometric data is transmitted for on-line authentication, the biometric data may be encrypted in accordance with industry standard including, for example, data encryption standard (DES), triple-DES, advanced encryption standard (AES), Rivest-Shamir-Adlema (RSA), open pretty good privacy (POP), etc. In some implementations, the encryptions may be performed on a session-by-session basis in which the session encryption and decryption keys are constantly updated or refreshed to thwart attacks.
100441 At least two issues still remain. The first issue is regarding the vetting process in which a user may register a biometric of the user at a trusted server, for example, a third party server. The second issue is regarding the authentication process in which the liveliness of a biometric presented on-line can be verified to thwart spoofing attempts (for example, someone using a video or photo of the registered user to pretend to be the registered user), or man-in-the-middle attacks. To these ends, Fig. 1B is a diagram showing an example interaction among a financial institution customer, a financial institution, and a third-party biometric verification service to authenticate a user based on a biometric identity of the user according to some implementations. As disclosed herein, the financial institution may include a bank, and may also be known as the relying party. The third-party service provider may also be known as the identity provider.
100451 When a financial institution customer 104 attempts to access an account through the financial institution website 106 (110A), customer 104 may be greeted with the biometric authentication window 102. The financial institution customer 104 may choose to use the biometric authentication. In response to receiving indication from customer 104 that the customer 104 would like to use the biometric authentication 102, the server running the financial institution website 106 may verify the user's authentication request using biometric data at biometric verification service 108 (110B). In some implementations, if customer 104 has never registered a biometric data at the biometric verification service 108, biometric verification service 108 may engage customer 104 to have the customer enrolled (110C).
The enrollment process may generally include asking customer 104 to provide a biometric data as the biometric identity data stored/managed by biometric verification service 108. The enrollment process may become part of the vetting process in which biometric verification service further verifies the identity of the customer 104 at other authorities, including, for example, Department of Motor Vehicles (DMV) at the state level and the State Department at the federal level. The authorities may include a credit verification agency, such as, for example, Equifax, Experian, LexisNexis. The credit verification agency may serve as a surrogate of a government-administered authority. After the customer 104 obtains an enrolled biometric identity at the biometric verification service, customer 104 may interact with biometric verification service 108 to conduct routine management of the biometric identity (110C). Routine management may include, for example, replacement of a facial pattern, revocation of an existing facial pattern, etc. Hence, the first issue of a vetting process to register an enrolled biometric data of customer 104 may be addressed.
100461 Regarding the second issue of the liveliness of the biometric data being presented on-line, the biometric login window 102 may authenticate customer 104 by analyzing real-time corneal reflection of an object (e.g., a camera device or an illumination source) and/or analyzing real-time face pose from customer 104. As discussed herein, a financial institution may be generalized as a relying party (RP) who may rely on a third party to authenticate that a customer is who the customer purports to be and that the customer is presenting a live facial biometric. The third party, known as the biometric verification service 108 in Fig. 1B, may be generalized as an identity provider (IdP).
100471 With the above generalization, Fig. 2A is a timing diagram showing an example interaction among a consumer 202, a relying party (RP) 204, and an identity provider (IdP) 206 in authenticating the consumer when the consumer has enrolled at the identity provider 206 according to some implementations. The process may initiate with consumer 202 send an authentication request to access an account managed by relying party 204 (208). In some implementations, the submission of the authentication request may correspond to consumer 202 choosing biometric authentication 102 to access an account managed by, for example, a financial institution.
[00481 When the constuner submits the authentication request, the constuner may also choose a verified liveliness login for biometric authentication (210). The verified liveliness login, as disclosed herein, may address the liveliness of biometric data being submitted on-line, e.g., by verifying facial liveliness of the consumer when the consumer submits the biometric data.
100491 For context, using a third-party intermediary to provide an authentication bridge way may be implemented in accordance with industry standards, for example, Security' Assertion Markup Language (SAML) and Open standard for Authentication (0Auth).
[00501 SAML addresses web browser single sign-on (SSO). Single sign-on solutions may be implemented at the intranet level using, for example, cookies.
Extending such single sign-on solutions beyond the intranet has been problematic and has led to the proliferation of non-interoperable proprietary technologies. In the use case addressed by SAML, the user (also known as the principal) may request a service from the relying party (also known as a service provider). The service provider then requests and obtains an identity assertion from the identity provider. On the basis of this identity assertion, the service provider can make an access control decision -deciding whether to perform some service for the connected principal.
[0051] In the case of 0Auth (including more recent 0Auth 2.0), the web application may specifically request a limited access 0Auth Token (also known as the valet key) to access the Application Program Interfaces (APIs) on behalf of the user. The valet token may explicitly name the particular rights requested, and the user may not be required to enter a credential. If the user can grant that access, the web application can retrieve the unique identifier for establishing the profile (identity) using the APIs.
[0052] Hence, authentication standards may generally provide a single sign-on so that a consumer can use, for example, the constuner's Facebook or Google login credentials to access other accounts (for example, an amazon account, a linked-in account, etc.), typically of the same consumer. Some authentication methodologies as disclosed herein may provide a single authentication proof of identity by using biometric data, for example, using facial biometric.
[0053] Referring to Fig. 2A, in response to receiving the consumer choice of verified liveliness login, a server at the relying party may redirect consumer to proceed with verified liveliness (212). Thereafter, constuner 202 may be redirected to a server at the identity provider 206. For example, consumer 202 may submit a request to use verified liveliness by using a credential token in the form of consumer@relying_party.com (214). As illustrated in Fig. 2A, the request may be submitted at a server at identity provider 206.
[0054] Upon receiving the request to use verified liveliness, the server at identity provider 206 may perform verified liveliness authentication to verify the biometric submitted by the consumer (216). The verification may include that (i) the biometric submitted by the consumer provides affirmative proof of identity for the consumer; and (ii) the biometric submitted by the consumer on-line is fresh and live and not is being forged, for example, in a spoofing attempt.
100551 As an initial matter, the server at identity provider 206 may retrieve, for example, a facial biometric from the data submitted by consumer 202. In some implementations, the retrieved facial biometric may be sent, from a computing device of consumer 202 to the server at identity provider 206, in an encrypted format based on an encryption algorithm. The computing device may include a video input component to provide a video stream feed. In some instances, the video input component includes a webcam. Images in the video feed may be captured at a frame rate of, for example, 24 frames per second. The frames generally are digitized for digital processing.
In some instances, the images may be at a spatial resolution of, for example, 800x600 pixels, 1024x768 pixels, 1152x864 pixels, or 1280x1024 pixels. The computing device may include an imaging device (e.g., a camera device) that captures images of consumer 202. The captured images include facial biometric of consumer 202, and can be transmitted by the computing device to the server at identity provider 206.
[00561 In some implementations, the retrieved facial biometric is sent from a computing device associated with the relying party 204 or the identity provider 206. The computing device can internally include or be externally coupled to a camera device that is local to consumer 202 when consumer 202 is seeking authentication. For example, the camera device can be adjacent to an automatic teller machine (ATM) facing constuner 202 or within a registration office of a bank. The ATM or the office may be a property of the relying party 204 with whom consumer 202 seeks to perform a transaction. The camera device can capture images or a video feed that includes facial biometric of consumer 202 and transmit to the computing device. The computing device can process the captured images or video feeds and transmit the captured images or video feeds and/or the processed data to the server at identity provider 206.
100571 The server at identity provider 206 may compare the retrieved facial biometric with biometric identity data of consumer 202. The biometric identity data of consumer 202 may be pre-stored at the server at identity provider 206. The comparison may be conducted in a manner consistent with the context-dependent rigidity requirement of the authentication.
For example, if the amount in transaction is less than a particular threshold, such as $50, then the degree of matching between the facial biometric as submitted by consumer 202 on-line and the pre-stored facial biometric at the server at the identity provider 206 may be reduced.
On the other hand, however, the if amount in transaction is higher than a particular threshold, such as, for example, $1,000, then the degree of matching between the facial biometric as submitted by consumer 202 on-line and the pre-stored facial biometric at the server at the identity provider 205 may be heightened. In a related example, if the transaction is in the nature of a mortgage application, a credit card application, the degree of matching may also be heightened. A scoring mechanism may be implemented to generate a score to quantify the degree of resemblance between the facial biometric as submitted by consumer 202 on-line and the pre-stored facial biometric at the server of identity provider 206.
The score of resemblance may be used by the server at identity provider 206 to determine whether the retrieved facial biometric as submitted by consumer 202 on-line is sufficient to prove that the submitting consumer 202 is the person whose biometric data is stored at the server of identity provider 206.
100581 Further, the server at identity provider 206 may determine whether the retrieved facial biometric being submitted by consumer 202 on-line is a live facial biometric.
One risk associated with on-line authentication has been the prospect of a man-in-the-middle attack. For example, during spoofing, a third party may impersonate consumer 202 by replaying an authentication proof as submitted by constuner 202 earlier and was obtained by the third party. Some encryption mechanism may counter the man-in-the-middle threat by refreshing or updating the encryption key used to encrypt the stream of communication sessions between consumer 202 and identity provider 206. Generally, however, such encryption mechanism may not address the staleness of the biometric data itself. In other words, some encryption mechanism may be inadequate if a third party obtains an instance of the biometric data and attempts to replay the instance of biometric data to gain access to the account of consumer 202. Such risks may be amplified with the combination of usemame and password login because the text strings for usemame and password may not be refreshed at the granularity of each login session. Specifically, refreshing username and password for each login session (or at a sub-session level at, for example, every few seconds) may not be practical.
100591 Some implementations as disclosed herein may verify the liveliness of the biometric data as submitted by consumer 202, for example, by performing verification of facial liveliness of consumer 202. In some implementations, the server at identity provider 206 performs liveliness verification based on the submitted images or video feeds. In some implementations, a computing device associated with consumer 202 performs liveliness verification based on the captured images or video feeds and submits a result of liveliness verification to the server at identity provider 206, e.g., in an authentication request including the biometric data of consumer 202. The server at identity provider 206 can then seamlessly determine that the biometric data submitted by consumer 202 is live. In some implementations, a computing device associated with relying party 204 or identity provider 206 that obtains images or video feeds from a camera device local to consumer 202, performs liveliness verification based on the captured images or video feeds and submits a result of liveliness verification to the server at identity provider 206.
[0060] In some implementations, liveliness verification may include analysis of a corneal reflection from consumer 202. Initially, the face of consumer 202 may be detected.
Referring to Figs. 3A and 3B, the eye 301 of consumer 202 may then be detected in a video feed or images. Generally, such detections include segmentation techniques that leverage face or eye template data. Such template data may be morphed through various shape transforms to identify face or eye area in the video feed or images. Once eye 301 is detected, cornea region 302 may then be determined in the video feed or images.
Reflections of the corneal region in the video frames or images may then be analyzed.
[0061] First, an object at the location of constuner 202 may cause corneal reflection from consumer 202. Here, the object may include an illumination source such as a light source, a camera device such as a webcam, or a device or subject (e.g., a window or a table) brighter than the ambient environment. In some examples, the object is a camera in a mobile device of consumer 202. Consumer 202 uses the camera to capture facial images or biometric of consumer 202. In some examples, the object may be in a place local to consumer 202. The place may be associated with relying party 204 with whom consumer 202 seeks to perform a transaction. For example, the place includes an automatic teller machine (ATM) facing consumer 202. The object can be mounted adjacent to the ATM. In some instances, the object may be fixed or positioned to the locale where consumer 202 is seeking authentication. Generally, the object is readily visible under normal illumination condition and can cause specular reflections from consumer 202 on-site. The specular reflection off the cornea of consumer 202 may be leveraged to determine liveliness of the authentication.
[0062] In some implementations, if the object has moved, the specular reflections of such object in the corneal of consumer 202 are expected to move accordingly.
If the specular reflection of such object moves in concert with the movement of the object on-site, then consumer 202 can be authenticated as live. In some cases, the movement of the object is performed by consumer 202. For example, consumer 202 is prompted to move a mobile phone including a camera device, and the corneal reflection of the camera device is moved accordingly. The movement of the camera device on-site can be measured by an on-board accelerometer or gyroscope in the mobile device. In some cases, the object is moved in a manner unpredictable by consumer 202 (or would-be hackers as man in the middle).
[0063] For example, the object can be moved from position 1 to position 2.
When the object is at position 1, a first image is captured. The location, shape, and color/wavelength of the reflected object may be detected by analyzing the corneal reflections in the captured first image. The detection can leverage segmentation techniques to automatically zero in on the corneal reflection 302A of the object. Thereafter, when the object is moved to position 2, a second image is captured. The location, shape, and color/wavelength of the reflected object may be detected anew as corneal reflection 302B by analyzing the captured second image. In some cases, the trajectory of the motion of object may be tracked by monitoring the reflected object. As illustrated, changes in the location of the detected object in corneal reflections 302A and 302B may be compared to the changes of object locations, e.g., programmed or measured location changes of the object. By way of illustration, some implementations may register the position of the reflected object relative to reflections of other structures adjacent to consumer 202 in a room or in front of an ATM where consumer 202 is seeking authentication. Some implementations may register the position of the reflected object relation to other bodily structures in or around the eye of consumer 202.
[0064] Fig. 4 illustrates an example object camera 403 that has been changed position by virtue of ann adjustments of consumer 202. Camera 403 is held on arm 402 of consumer 202, and faces head 401 of consumer 202. In position 1, arm 402 is more extended, for example, to be closer to the face of consumer 202. In position 2, arm 402 is more recoiled, for example, to move further away from the face of consumer 202. Such positional changes of camera 403 are expected to cause changes in the reflected camera from cornea of consumer 202. In some examples, the object may rely on a mechanism that leads a changed position of the object. The mechanism can be pan, zoom, tilt (PZT), or any other suitable mechanism.
[0065] Motions of an object can cause more interesting changed location and shape of the reflected object in corneal reflections. The object may be programmed to move by virtue of a controller such as on-board accelerometer or gyro motors. In some cases, the object is moved in a predetermined (or preprogrammed) motion. The predetermined motion can be previously embedded in the object. The predetermined motion can be also transmitted to the controller, e.g., from the server at identity provider 206 or a computing device coupled to the controller. Each time the predetermined motion can be different from previous predetermined motions. In some cases, the object is moved randomly. The actual movement of the object can be measured or detected by sensors such as on-board accelerometer or gyro motors.
[0066] The changes of the object may not be limited to mere positional changes manifested as translational or rotational adjustments. The variety' of such changes in position and change can be leveraged to provide an enriched analysis of corneal reflections. In one example, the object can even change to a position to cause the wavelength of corneal reflection to change. By way of illustration, the object can be translated to deflect illumination from a different illumination source facing consumer 202, e.g., a flashlight of a mobile device of consumer 202.
100671 Notably, even though the head or face of consumer 202 may or may not move, the object moves, in a pattern known to the identity provider but unknown to consumer 202 (or would-be hackers as man in the middle) beforehand, to cause changes in cornea reflection of the object. In some cases, the object can be moved from a first position to a second position in relatively fast speed compared to the movement of the head or face of consumer 202, such that the movement of the head or face of consumer 202 can be considered as relatively static. A camera device can capture a first facial image of consumer 202 when the object is at the first position at a first time point and capture a second facial image of consumer 202 when the object is at the second position at a second time point.
An interval between the first and second time points can be short, e.g., less than I
second. In some cases, the movement of the head or face of consumer 202 can be leveraged by analysis of movement of static or unmoved subjects (e.g., a window or a table adjacent to consumer 202) in the captured images. If consumer 202 is prompted to move the object, consumer 202 can be notified not to move the head or face of consumer 202.
100681 A camera device is configured to capture the images and/or video feeds of consumer 202. The camera device can be positioned facing a face of consumer 202. In some examples, the camera device can be internally included in a mobile device of consumer 202.
Consumer 202 can hold the mobile device such that the camera device faces the face of consumer 202 to capture facial images. The camera device can be positioned such that the camera device is at a tilted angle to the eye of consumer 202 or opposite to the eye of consumer 202. In some examples, consumer 202 is seeking authentication before an ATM.
The camera device can be fixed on the ATM or adjacent to the ATM.
100691 The camera device can capture the images and/or video feeds for verifying facial liveliness of consumer 202, e.g., without interaction with consumer 202 or without notice to consumer 202. In some cases, the movement of the camera device can be associated with a natural motion of consumer 202. A sensor, e.g., an accelerometer, a gyroscope, or a global positioning system (GPS) on a mobile device including the camera device, can measure the movement of the camera device. The mobile device can determine that the camera device is moving by using the sensor and then transmit a request to the camera device to capture facial images of consumer 202.
[0070] In some implementations, upon receiving the request to use verified liveliness, the server at identity provider 206 may transmit a request to the camera device, e.g., directly or through a computing device coupled to camera device and in communication with the server. In response to the request, the camera device can automatically capture facial images of consumer 202 when the object is moved between positions. The object can be moved based on the request. For example, a computing device in communication with the server receives the request and transmits a command to a controller of the object for moving the object. The object can be moved between positions in synchronization with the camera device capturing facial images of consumer 202.
[0071] The movement of the object can be associated with respective time points.
For example, at a first time point, the object is at a first position; at a second time point, the object is moved to a second position. The camera device can capture facial images or video feeds of consumer 202 during the movement of the object and/or before and after the movement of the object. The captured facial images are also associated with respective time points. The time points can have a high accuracy, e.g., less than 1 millisecond. For example, the camera device captures a first facial image when the object is at the first position at the first time point, and a second facial image when the object is at the second position at the second time point. Based on the time points, the captured facial images of consumer 202 can be associated with the positions of the object. Thus, the corneal reflection of the object in the captured images or video feeds can be associated with the movement of the object, which can be used to verify facial liveliness of consumer 202.
100721 The camera device can also capture images and/or video feeds of consumer 202 that include biometric data submitted by consumer 202 to the server at identity provider 206 for authentication, e.g., by interacting with consumer 202 to obtain the biometric data. In some cases, an additional imaging device or a video input component is used to capture images and/or video feeds that include biometric data of consumer 202 for authentication.
[0073] In some implementations, the object is the camera device. As noted above, the camera device can be moved by consumer 202 or by a mechanism in a controlled manner or randomly. The motion (and/or orientation and/or relative position) of the camera device can be measured, e.g., by on-board accelerometer or gyroscope. For verification of facial liveliness of consumer 202, the detected position changes in the corneal reflections of the camera device may be correlated to the predetermined or measured motion of the camera device.
100741 In some implementations, the object is a static subject adjacent to consumer 202, e.g., a door behind consumer 202 or a window besides consumer 202. The camera device is moved between positions and captures facial images of consumer 202 at these positions. The facial images include corneal reflections of the object. For verification of facial liveliness of consumer 202, the detected position changes in the corneal reflections of the object may be correlated to a predetermined or measured motion of the camera device.
[0075] In some implementations, the object is a subject configured to be moved between positions in a motion for liveliness verification, while the camera device is static and captures facial images of consumer 202 when the object is at different positions. The facial images include corneal reflections of the object. For verification of facial liveliness of consumer 202, the detected position changes in the corneal reflections of the object may be correlated to a predetermined or measured motion of the object.
[0076] In some implementations, the object is a subject configured to be moved between positions in a first motion pattern, while the camera device is also moved between positions in a second motion pattern. Moving both the object and the camera device for liveliness verification can increase security against hackers. The camera device captures facial images of consumer 202 when the object is at different positions and the camera device is also at different positions. The facial images include corneal reflections of the object. For verification of facial liveliness of consumer 202, the detected position changes in the corneal reflections of the object may be correlated to a motion pattern based on the first motion pattern (predetermined or measured) of the object and the second motion pattern (predetermined or measured) of the camera device. The motion pattern can be a relative motion pattern for the object and the camera device.
[0077] In some implementations, the object is static, and the camera device is also static. An illumination on the object from an illuminator can change at different time points, e.g., changing illumination colors or patterns, and the corneal reflection of the object can change correspondingly. For verification of facial liveliness of consumer 202, the detected illumination change on the object in the corneal reflections of the object may be correlated to a known illumination change of the illuminator.
100781 In some implementations, the object is an illuminator, e.g., a light source and/or a digital illumination source displayed on a screen, e.g., a liquid crystal display (LCD) of a mobile device. The camera device can capture facial images of consumer 202 that include corneal reflections of the object. For verification of facial liveliness of consumer 202, the detected illuminator in the corneal reflections of the object may be correlated to a predetermined or measured illumination change of the illuminator. Using illumination for liveliness verification is described with respect to Fig. 5 below.
[0079] In some implementations, the camera device is static and/or the object is also static. A computing device coupled to the camera device can prompt consumer 202 to move the head or the face in a motion pattern, e.g., turning from left to right.
The camera device can capture facial images of consumer 202 during the movement of consumer 202.
For verification of facial liveliness of consumer 202, the detected position changes in the corneal reflections of the object may be correlated to a motion of consumer 202.
100801 In some implementations, the object and/or the camera device can be moved in a motion pattern, and an illumination on the object can also change at different time points, e.g., changing illumination colors or patterns. Verification of facial liveliness of consumer 202 can be based on: 1) the correlation between the detected position changes in the corneal reflections of the object and the motion pattern of the object and/or the camera device, and 2) the correlation between the detected illumination change on the object in the corneal reflections of the object and the predetermined or measured illumination change.
100811 Initially, the images during the movement of the object and/or the camera device may be captured in a video stream at a full frame rate (for example, as determined by Nation Television System Committee¨NTSC, or as determined by the video communication standard H.261). In some implementations, for efficiency of communication, only a portion of the captured video stream may be transmitted, for example, in accordance with adaptive video codec specification of H.261. When changes in corneal reflections are about to initiate, the video feed may be digitalized and transmitted to the server at identity provider 206 at a rate faster than a floor-level rate (when consumer 202 remains stationary). In some examples, the video feed or images are transmitted to the server at identity provider 206 by the camera device. In some examples, a computing device coupled to the camera device receives the images or video feed from the camera device and transmits the images or video feed to the server at identity provider 206.
100821 As noted above, the server at identity provider 206 can perform liveliness verification based on the transmitted images or video feed. The server can analyze the images or video feed to determine corneal reflections of the object. Based on the determined corneal reflections, the server can determine a position change in the corneal reflections of the object. The server can compare the position change in the corneal reflections of the object to a motion pattern and determine facial liveliness of consumer 202 based on a result of the comparison. In some examples, the motion pattern is a predetermined motion pattern of the object or the camera device or both when moved between positions. The server can pre-store the predetermined motion pattern. In some examples, the motion pattern is a measured motion pattern of the object or the camera device or both when moved between positions.
The motion pattern can be transmitted to the server together with the captured images or video feed. In some examples, the server at identity provider 206 can analyze the received images and/or video feeds to determine the motion pattern, e.g., by analyzing a positon change of a static subject in the captured images and/or video feeds.
100831 In some implementations, a computing device coupled to the camera device performs liveliness verification based on the captured images or video feeds and submits a result of liveliness verification to the server at identity provider 206, e.g., with or without transmitting the captured images or video feeds. In some cases, the computing device receives biometric data of consumer 202 for authentication and transmits the result of liveliness verification together with the biometric data of consumer 202 to the server at identity provider 206.
[0084] In some examples, the computing device is associated with relying party 204 or identity provider 206. For example, consumer 202 is seeking authentication using an ATM associated with relying party 204. The computing device can be local to the ATM and coupled to the camera device that faces consumer 202. The computing device can receive from the server at identity provider 206 a request to verify liveliness of consumer 202. In response to the request, the computing device can transmit commands to the camera device and/or the object to ask the camera device to capture images and/or video feeds when the camera device and/or the object is moved between positions in a motion pattern, e.g., unknown to consumer 202 and without input from consumer 202. In some cases, the computing device predetermines the motion pattern and transmits the motion pattern to the camera device and/or the object. The computing device can change the motion pattern each time when performing liveliness verification. In some cases, the motion pattern is a preprogrammed or predetermined motion pattern. The computing device can pre-store the motion pattern. In some cases, the motion pattern is measured by sensors when the camera device and/or the object is moved with positions. The computing device can receive the measured motion pattern from the sensors.
[0085] in some examples, the computing device is associated with consumer 202, e.g., a smartphone or a computer. For example, consumer 202 is seeking authentication at home. The computing device can internally include a camera device or externally be coupled to a camera device. In some cases, an application (or an App when the computing device is a smartphone) provided by identity provider 204 in installed in the computing device. Upon Neel ing a request from the server at identity provider 204 or determining consumer 202 chooses to use liveliness verification for biometric authentication, the application can solicit consumer 202 to move the camera device in relative to the face of consumer 202. By using a sensor (e.g., an accelerometer or gyroscope, the computing device can determine a movement of the camera device. In some examples, the computing device can analyze the captured images and/or video feeds to determine the motion of the camera device, e.g., by analyzing a positon change of a static subject in the captured images and/or video feeds.
The images of the static subject can be corneal reflections or outside of the corneal area.
The computing device can perform liveliness verification based on the motion pattern and the corneal reflections of the camera device in captured images and/or video feeds. In some cases, the application in the computing device controls an illumination source such as a flashlight of the camera device to change illumination color or pattern unknown to consumer 202.
The application can perform liveliness verification based on the changed color or pattern of the illumination source and the corneal reflections of the camera device in captured images and/or video feeds.
[0086] Similar to facial authentication, a scoring mechanism may be implemented to quantify the degree of match between the response corneal reflection of the object and the expected change (e.g., a motion pattern or an illumination change). The scoring mechanism may generate a numerical value as a measure for determining liveliness of the face of consumer 202. The degree of match may depend on the context and hence can vary from application to application. For example, the degree of match may depend on the nature of the underlying transaction or the dollar amount involved in the underlying transaction.
100871 In some implementations, corneal reflections of the object in sequential facial images of consumer 202 are detected. The sequential facial images are captured at sequential time points by a camera device. An expected corneal reflection of the object at a particular time point can be determined based on a determined corneal reflection in a previous time point and a motion between the previous time point and the particular time point. The facial liveliness of consumer 202 can be assessed by determining likelihood between the expected corneal reflection and the determined corneal reflection from the captured facial images at the particular time point.
[0088] Corneal reflection may proffer benefits as non-intrusive and capable of handling a wide band of operations (for example, from infra-red to visible light).
Particularly, corneal reflection tends to be more specular than skin reflections, which tend to be more scattered. Implementations disclosed herein may not include verbal directives to prompt consumer 202 to change facial gestures or other bodily positions.
[0089] Further, pose of the face of consumer 202 may be estimated, for example, using an automatic algorithm. The automatic algorithm can detect the pose of the face relative to the camera device in each capture image, e.g., from a video feed.
Additionally, an expected pose or change in pose of a face may be calculated based on the measured or known positions of the camera device, e.g., by on-board accelerometer or gyro motors or from analyzing the corneal reflections of the camera device. In some implementations, consumer 202 is prompted to move the camera device in relative to the face of consumer 202. Such calculation can be performed for each additional image captured from the video feed. More interesting, the expected pose or changes in pose, as calculated from the measured locations of the camera device and the detected pose of the face in the capture image may be correlated in a quantitative manner using a scoring mechanism. The numerical value as determined from the correlation can be used to gauge the liveliness of the face.
[0090] In some implementations, pose of the face of consumer 202 is detected or estimated by analyzing one or more facial landmarks of consumer 202. The facial landmark features can include the nose, the forehead, the eye brows, the eyes, the eye corners, the lips, and/or the mouths. Poses of the face or facial portraits from a live session reflect the face of the subject at various angles. Due to slight changes in viewing angles of the pose or facial portrait, the apparent orientation and size of each facial landmark varies.
Orientation and size of one or more facial landmark features can be expected to vary. Each landmark feature may be associated with one conunensurate change in orientation and size. The threshold of expected change may be a combination of all expected changes for the landmark features. As noted above, the threshold may be application-specific. For example, for transaction involving dollar amounts under a particular level, the comparison may be less sophisticated in order to determine that the session is live and the submitted facial biometric is genuine.
Otherwise, when the transaction involves a paramount identification document, for example, renewal of passport, the comparison may involve a more comprehensively determined threshold.
[0091] In some implementations, verification of facial liveliness of consumer 202 is based on both the corneal reflection of the object and the pose change of the face. A scoring mechanism may be implemented to quantify the degree of match between the response corneal reflection of the object and the known motion pattern to determine a first numerical value. The scorning mechanism can be also implemented to quantify the degree of match between the expected pose or changes in pose (e.g., calculated from the measured locations of the camera device) and the detected pose in a capture image to determine a second numerical value. Then an overall score can be determined based on the first and second matching values, e.g., with a weighting factor. The weighting factors of the first and second matching values can be, for example, 50% and 50% or 60% and 40%. The additional verification may increase a confidence level that consumer 202 is indeed alive and is the person whose biometric is stored at the server of identity provider 206.
100921 In some implementations, liveliness verification includes analysis of a corneal reflection of an illuminator from constuner 202. Initially, the face of consumer 202 may be detected. Referring to Fig. 5, an eye 501 of consumer 202 may then be detected in a video feed (or an image). Generally, such detections include segmentation techniques that leverage face or eye template data. Such template data may be morphed through various shape transforms to identify face or eye area in a video feed. Once eye 501 is detected, cornea region 502 may then be determined in the video feed. Reflections of illuminations on the corneal region in the video frame may then be analyzed.
100931 First, illuminator 506 adjacent to consumer 202 may be activated.
For example, illuminator 506 can be activated in response to determining that consumer 202 chooses a verified liveliness login for biometric authentication, e.g., 210 of FIG. 2A. As an example, illuminator 506 in Fig. 5 includes illumination source 506A and illumination source 506B with two distinct colors (e.g., blue or red). An illumination source can be a physical light source or a digital illumination source. In some implementations, illuminator 506 includes physical light sources, e.g., light-emitting diodes (LEDs) or light lamps. Illuminator 506 may include one or more light sources at distinct light wavelengths thus with one or more distinct colors. In some implementations, illuminator 306 includes one or more digital illumination sources displayed on a screen, e.g., on a laser-powered phosphor display (LPD), a liquid-crystal display (LCD) screen, a light-emitting diode (LED) screen such as an active matrix organic LED screen, or any suitable display screen. A digital illtunination source can be a digital color pattern displayed on a screen, e.g., a LCD screen of a mobile device. The digital illumination sources can be provided by an application (or an App) installed on a computing device of consumer 202 or provided by the server at identity provider 206, e.g., in a user interface displayed on the computing device of consumer 202. In some implementations, illuminator 306 includes one or more physically light sources and one or more digital illumination sources.
100941 Illuminator 506 may be included or displayed on a consumer electronic device, such as a smartphone, a mobile phone, a mobile computing device, or a desktop computer, from which consumer 202 attempts to authenticate his or her identity and liveliness. Illuminator 506 may be included on the console of an automatic teller machine (ATM) of a relying party or adjacent to the ATM. A computing device associated with the relying party or identity provider 206 may be configured to control illuminator 506.
Consumer 202 may attempt to authenticate his or her identity and liveliness through the computing device.
[0095] Once activated, illuminator 506 may illuminate consumer 202. The illumination from illuminator 506 can be straight to the eye of consumer 202 or to the eye of consumer 202 with an incident angle. In some examples, illuminator 506 is positioned so that the illumination from illuminator 506 is incident on the eye of consumer 202 with an angle. In some examples, consumer 202 can be prompted to move illtuninator 506 (e.g., a flashlight on his/her smartphone) to a side of the face (or eyes) or move his/her face (or eyes) so that the illumination from illuminator 506 is incident on the eye of consumer 202 with an angle.
[0096] Information of the illumination (e.g., a position, a shape, a duration, a color or wavelength, or an illumination pattern) can be known or predetermined. In some examples, a computing device of consumer 202 performs liveliness verification. The information of illtunination can be known to the computing device but unknown to consumer 202 (or would-be hackers as man in the middle) beforehand. The information of illumination can be generated by the computing device or provided by the server at identity provider 206.
[0097] The corneal reflection immediately following the activation of illuminator 506 may be captured, e.g., by a camera device adjacent to consumer 202. The camera device can keep capturing facial images of consumer 202 before, during, and after the activation of illuminator 506. In some examples, the captured image frame may be analyzed to ascertain whether the corneal reflection is in agreement with the expected (known or predetermined) illumination. In analyzing the corneal reflections, the position, shape, and color/wavelength of the reflected illumination in the cornea of consumer 202 may be detected and compared to the expectation illumination. In one example, the position of the corneal reflection, for example, reflection spot 504A or 504B may be correlated with the position of illumination sources 506A or 506B. In another example, the shape of the reflection spot 504A or 504B
may be compared to the shape of illumination sources 506A or 506B. In yet another example, the wavelength (or color) of reflected light from reflection spot 504A or 504B may be compared to the wavelengths (or colors) of illumination sources 506A or 506B.
100981 In some examples, illumination pattern by illuminator 506 (e.g., an illumination source 506A or 506B) may be modulated over time to form a temporal illumination pattern. For example, illumination source 506A may generate illumination in bursts, each burst with a same or different temporal duration. For example, a first temporal duration can be 1 second and a second temporal duration can be 2 seconds. The temporal illumination pattern can also include a change in illumination shape, e.g., varying between a square shape, a triangle shape, and/or a circle shape, and/or a change in illumination color, during different temporal durations or at different time points. In one example, illuminator 506 provides a red square at a first temporal duration, a red circle at a second temporal duration, no illumination at a third temporal duration, and a blue triangle at a fourth temporal duration. The corneal reflection in captured sequential images may be analyzed to determine whether the corneal reflection contains the same temporal pattern. The corneal reflection in captured sequential images can be also individually analyzed to determine whether the corneal reflection in the captured sequential images matches the individually expected corneal reflections of the illuminator 506. Determination of facial liveliness of consumer 202 can be based on the match results of the captured sequential images.
[0099i In some examples, illumination pattern by two or more illumination sources (e.g., ilkunination sources 506A and 506B) may be modulated over time to form a temporal illumination pattern. The illumination sources can illuminate same or different colors. For example, illumination sources 506A and 506B may generate illumination in bursts, each with a same or distinct temporal duration and/or each with a same or different distinct illumination shape during the temporal duration. In an example, at a first time point, illumination source 506A illuminates a red square and illumination source 506B illuminates a blue square. At a second time point, illumination source 506A illuminates a red circle and illumination source 506B illuminates a blue square. At a third time point, illumination source 506 illuminates a red circle and illumination source 506 illuminates a blue circle. At a fourth time point, illumination source 506 illuminates a red rectangle and illumination source 506 illuminates a red circle. The corneal reflection in sequential images may be analyzed to determine whether the corneal reflection contains the same temporal pattern.
1001001 In some examples, the illumination pattern by two or more illumination sources (e.g., illumination sources 506A and 506B) may be modulated spatially to form a spatial illumination pattern. For example, illumination sources 506A and 506B
may be activated to illuminate cornea 502 in an interleaved manner. Generally, the illumination pattern from illumination sources 506A and 506B may be varied by turning on or off illumination source 506A, illumination source 506A, or the combination of illumination sources 506A and 506B at different time points. In response, the corneal reflection may be analyzed to determine whether the reflection contains the same spatial pattern of illumination.
As discussed herein, the temporal and spatial aspects of illumination may be combined to further buttress the variety of matches that improves the confidence of a liveliness authentication session.
[001011 The images in response to activated illiunination may be captured in a video stream at a full frame rate. In some implementations, as noted above, the server at identity provider 206 performs liveliness verification based on the captured video stream. For efficiency of communication, only a portion of the captured video stream may be transmitted.
When corneal reflections are expected to change, the video stream may be digitalized and transmitted to the server at identity provider 206 at a rate faster than a floor-level rate (when consumer 202 remains stationaty). In fact, the transmitted video stream may be analyzed by the server at identity provider 206 to determine whether the changes in corneal reflections of the video stream from consumer 202 matches the known changes in illumination of illuminator 506 or whether the corneal reflections of video frames at particular time points matches the expected corneal reflections based on known illuminations of illuminator 506 at the particular time points.
1001021 Similar to facial authentication, a scoring mechanism may be implemented to quantify the degree of match between the response corneal reflection being received at the server of the identity provider 206 in the captured images and the expected illumination (e.g., position, shape, and wavelength or color, or temporal and/or spatial pattern).
The degree of match may depend on the context and hence can vary from application to application. For example, the degree of match may depend on the nature of the underlying transaction or the dollar amount involved in the underlying transaction.
1001031 In some implementations, a computing device receives the captured video stream, e.g., by a camera device included in or coupled to the computing device, and performs liveliness verification based on the captured video stream. The computing device can be a computing device of consumer 202 from which consumer 202 attempts to authenticate his or her identity and liveliness. For example, the computing device can be a smartphone which includes a flashlight as illuminator 506 and a camera as the camera device.
A mobile app installed on the smartphone can also provide one or more digital illiunination sources as illuminator 506. The flashlight and the digital illumination sources can be combined to be used as illuminator 506. The computing device can control illuminator 506 to illuminate based on a predetermined or known illumination pattern. The illumination pattern can be generated by the computing device or transmitted from the server at identity provider 206. The illumination of illuminator 506 can be also measured by a detector that provides the measured illumination to the computing device. The known illtunination of illuminator 506 can be stored in the computing device. The computing device can also control the camera device to capture a video stream during activation of the illumination pattern of illtuninator 506. In some other implementations, the computing device is associated with relying party or identity provider 206.
1001041 As the camera device is local to the computing device, the transmission of the captured video stream (e.g., a portion) can be fast. The computing device can analyze the transmitted video stream to determine the illumination (or illumination changes) in corneal reflections of the video stream (e.g., sequential video frames) matches the known illumination (or illumination changes) of illuminator 506. The computing device can use a scoring mechanism to quantify the degree of match between the response corneal reflection in the captured images and the expected illumination. The degree of match may depend on the context and hence can vary from application to application. Based on the scored match, the computing device can verify facial liveliness of consumer 202. The computing device can then submit the result of the verification to the server at identity provider 206, e.g., together with biometric data of consumer 202. Then the server at identity provider 206 can seamlessly determine whether the submitted biometric data is a live facial biometric of consumer 202 based on the submitted result of the verification from the computing device.
1001051 Implementations disclosed herein detect corneal reflections in response to illuminations, including illumination patterns temporally and/or spatially.
Corneal reflection may proffer benefits as non-intrusive and capable of handling a wide band of operations (for example, from infra-red to visible light). Particularly, corneal reflection tends to be more specular than skin reflections, which tend to be more scattered.
Implementations disclosed herein may not include verbal directives to prompt consumer 202 to change facial gestures or other bodily positions. Some implementations disclosed herein can detect liveliness seamlessly and without user awareness and input. For example, activation of an illumination pattern and capturing sequential images can be performed without user's notice, e.g., when facial biometric of a consumer is captured. The liveliness verification can be also completed during capturing facial biometric, e.g., by a computing device of consumer 202.
[001061 Some implementations may additionally include a biometric recognition feature to authenticate consumer 202. By way of example, server at identity provider 206 may conduct a facial recognition to compare the facial biometric of an enrolled consumer with the facial biometric of consumer 202 submitting the authentication request. If the comparison yields a matching result, the facial recognition may prove that consumer 202 is the person whose facial biometric is stored at the server at identity provider 206. In other examples, the biometric recognition may include voice recognition, speaker recognition, gait analysis, etc. Some implementations disclosed herein can detected liveliness seamlessly and without user awareness (e.g., directing the user to make any specific moves during the image capturing process). The stealth feature can be leveraged to further isolate liveliness detection from an on-line identity management session.
[00107) Referring back to FIG. 2A, after verifying that consumer 202 is alive and is the person whose biometric has been stored at identity provider, server at identity provider 206 may provide a signal that authentication is successful. If, however, consumer 202 cannot be verified as alive or as the person whose biometric has been stored at identity provider 206 in association with the account at relying party 204, server at identity provider 206 may provide a signal that authentication has failed. The signal may be embedded in a message to consumer 202 to redirect consumer 202 back to relying party 204 (218).
[00108) Consumer 202 may then return to the relying party 204 (220). If the embedded signal indicates that the verification is successful, consumer 202 may proceed to access the account at relying party 204. In returning to relying party 204, consumer 202 may attach a success signal backed up by a credential of identity provider 206.
The credential of the identity provider 206 may include, for example, a digital private key of identity provider 206, a digital watermark of identity provider 206. If, however, the embedded signal indicates that the verification has failed, consumer 202 may not access the account at relying party 204.
Moreover, consumer 202 may not even be redirected back to the relying party 204.
[00109] Fig. 2B is a timing diagram showing an example interaction among a consumer, a relying party, and an identity provider in authenticating the consumer when the consumer has not yet enrolled at the identity provider according to some implementations.
As discussed above in association with Fig. 2A, the process may initiate with consumer 202 send an authentication request to access an account managed by relying party 204 (208). In some implementations, the submission of the authentication request may correspond to consumer 202 choosing biometric authentication 102 to access an account managed by, for example, a financial institution. When the consumer submits the authentication request, the consumer may also choose to a verified liveliness login for biometric authentication (210). In response to receiving the consumer choice of verified liveliness login, a server at the relying party may redirect consumer to proceed with verified liveliness (212).
Thereafter, consumer 202 may be redirected to a server at the identity provider 206. For example, consumer 202 may submit a request to use verified liveliness by using a credential token in the form of consumer@relying_party.com (214). As illustrated in Fig. 2B, the request may be submitted at a server at identity provider 206.
[001101 The server at identity provider 206 may notice that consumer 202 has not yet registered for the verified liveliness service at identity provider 206. In response, the server at identity provider 206 may inquire the server at the relying party 204 whether the identity provider 206 may enroll consumer 202 in the verified liveliness service, as requested by consumer 202. Some relying parties may fmancially compensate identity provider 206 for the verified liveliness. In turn, these relying parties may stratify consumers into various classes with corresponding account benefits. By way of illustration, consumers with holdings above a threshold value of, for example, $50k may have the verified liveliness service as a complimentary service. Consumers with less account value or shorter membership history may need to pay a fee to relying party 204 in order to have the verified liveliness service.
[0011.11 Upon determination that consumer 202 may enroll in the verified liveliness service, server at relying party 204 may indicate to the server at identity provider 206 that the enrollment for constuner 202 may proceed (224). If, however, server at relying party 204 determines that consumer 202 may not enroll in the verified liveliness service, server at relying party 204 may inform the server at identity provider 206 that consumer 202 does not have permission to enroll in the requested service and the server at identity provider 206 may drop the enrollment request initiated from consumer 202. Particularly, server at identity provider 206 may direct consumer 202 back to relying party 204 with the flag that enrollment has failed (224).
[00112] If the server at identity provider 206 receives confirmation that identity provider 206 may proceed with enrolling consumer 202, server at identity provider 206 may administer an enrollment procedure for consumer 202 (226). Specifically, identity provider 206 may request consumer 202 to register a biometric. The biometric may include, for example, a facial biometric, a finger-print, a palm-print, an iris scan, a retina scan. In the case of a facial biometric, server at identity provider 206 may request consumer 202 to pose in front of, for example, a web camera, a video camera, or a smart-phone camera. The camera device may reside on, for example, a computing device of consumer 202, or an ATM
or in registration office of a bank. The camera device or a computing device coupled to the camera device may initially transmit a capture frame to server at identity provider 206. This captured frame may be stored at identity provider 206 as biometric information identifying consumer 202. By way of example, server at identity provider 206 or the computing device may verify the liveliness of the enrollment session as discussed above in association with Figs. 2A, 3A, 3B, 4 and 5.
[00113] Consistent with the disclosure herein, server at identity provider 206 may also initiate illumination patterns on consumer 202 to increase the confidence level that a live consumer is attempting to enroll in the service. As discussed herein, multiple illumination patterns for different and separate spatial and temporal modulations may obviate the ambiguities associated with a single illumination response. For example, when consumer 202 may not be responding to the fullest extent as expected from one single illumination, or when sampling error during video transmission may have missed the frame that contains the exact response reflection, a single response illumination may be insufficient to determine with confidence that consumer 202 is a live person attempting to enroll in the service.
[00114] When consumer 202 has been determined as a live person and has been enrolled in the verified liveliness service, the server at identity provider 206 may redirect consumer 202 back to relying party 204 with a signal that the authentication is successful (228). If consumer 202 has been successfully enrolled in the program, the server at identity provider 206 may provide a signal that the authentication has succeeded. If, however, consumer 202 cannot be verified as alive or as the person and cannot be enrolled in the verified service program, the server at identity provider 206 may provide a signal that authentication has failed. The signal may be embedded in a message to consumer 202 to redirect consumer 202 back to relying party 204 (218).
[00115] Consumer 202 may then return to the relying party 204 (230). If the embedded signal indicates that the authentication has succeeded, consumer 202 may proceed to access the account at relying party 204. In returning to relying party 204, consumer 202 may attach a success signal backed up by a credential of identity provider 206. The credential of the identity provider 206 may include, for example, a digital private key of identity provider 206, a digital watermark of identity provider 206. If, however, the embedded signal indicates that the authentication has failed, consumer 202 may not access the account at relying party 204. Moreover, consumer 202 may not even be redirected back to the relying party 204.
[00116] Notably, in some implementations, a first-time enrollee may be sent to the server at identity party 206 after the first-time enrollee has successfully logged into an account of the first-time enrollee at the relying party 204. The enrollment may provide consumers with additional security features to guard against spoofing attacks.
The security features may also include biometric recognition, as detailed above. In fact, some configurations may use the biometric recognition as a form of password authentication. In other words, the biometric recognition may be used as a surrogate for the password login.
[001171 In some implementations, however, identity provider 206 may obtain a database for a number of consumers attempting to enroll in a verified liveliness program.
During the enrollment process, the server at identity provider 206 may collect a biometric from a consumer attempting to enroll. After verifying that the consumer is a live person making a legitimate enrollment request, the server at identity provider 206 may enroll the consumer. Hence, in some implementations, the server at identity provider 206 may have a copy of a digital biometric of consumer 202 when consumer 202 has been redirected from relying party 204. When a copy of the digital biometric of consumer 202 is available, the server at identity provider 206 may only need to determine that consumer 202 is a live person.
[00118] Implementations of the subject matter and the functional operations described in this specification can be implemented in digital electronic circuitry, in tangibly-implemented computer software or firmware, in computer hardware, including the structures disclosed in this specification and their structural equivalents, or in combinations of one or more of them. Implementations of the subject matter described in this specification can be implemented as one or more computer programs, i.e., one or more modules of computer program instructions encoded on a tangible non transitory program carrier for execution by, or to control the operation of, data processing apparatus. The computer storage medium can be a machine-readable storage device, a machine-readable storage substrate, a random or serial access memory device, or a combination of one or more of them.
[001191 The tenn "data processing apparatus" refers to data processing hardware and encompasses all kinds of apparatus, devices, and machines for processing data, including, by way of example, a programmable processor, a computer, or multiple processors or computers.
The apparatus can also be or further include special purpose logic circuitry, e.g., a central processing unit (CPU), a FPGA (field programmable gate array), or an ASIC
(application specific integrated circuit). In some implementations, the data processing apparatus and/or special purpose logic circuitry may be hardware-based and/or software-based.
The apparatus can optionally include code that creates an execution environment for computer programs, e.g., code that constitutes processor firtnware, a protocol stack, a database management system, an operating system, or a combination of one or more of them. The present disclosure contemplates the use of data processing apparatuses with or without conventional operating systems, for example Linux, UNIX, Windows, Mac OS, Android, iOS or any other suitable conventional operating system.
1001201 A computer program, which may also be referred to or described as a program, software, a software application, a module, a software module, a script, or code, can be written in any form of progranuning language, including compiled or interpreted languages, or declarative or procedural languages, and it can be deployed in any form, including as a stand-alone program or as a module, component, subroutine, or other unit suitable for use in a computing environment. A computer program may, but need not, correspond to a file in a file system. A program can be stored in a portion of a file that holds other programs or data, e.g., one or more scripts stored in a markup language document, in a single file dedicated to the program in question, or in multiple coordinated files, e.g., files that store one or more modules, sub programs, or portions of code. A computer program can be deployed to be executed on one computer or on multiple computers that are located at one site or distributed across multiple sites and interconnected by a communication network.
While portions of the programs illustrated in the various figures are shown as individual modules that implement the various features and functionality through various objects, methods, or other processes, the programs may instead include a number of sub-modules, third party services, components, libraries, and such, as appropriate.
Conversely, the features and functionality of various components can be combined into single components as appropriate.
1001211 The processes and logic flows described in this specification can be performed by one or more programmable computers executing one or more computer programs to perform functions by operating on input data and generating output. The processes and logic flows can also be performed by, and apparatus can also be implemented as, special purpose logic circuitry, e.g., a central processing unit (CPU), a FPGA (field programmable gate array), or an ASIC (application specific integrated circuit).
1001221 Computers suitable for the execution of a computer program include, by way of example, can be based on general or special purpose microprocessors or both, or any other kind of central processing unit. Generally, a central processing unit will receive instructions and data from a read only memory or a random access memory or both. The essential elements of a computer are a central processing, unit for performing or executing instructions and one or more memory devices for storing instructions and data. Generally, a computer will also include, or be operatively coupled to receive data from or transfer data to, or both, one or more mass storage devices for storing data, e.g., magnetic, magneto optical disks, or optical disks. However, a computer need not have such devices. Moreover, a computer can be embedded in another device, e.g., a mobile telephone, a personal digital assistant (PDA), a mobile audio or video player, a game console, a Global Positioning System (GPS) receiver, or a portable storage device, e.g., a universal serial bus (USB) flash drive, to name just a few.
[001231 Computer readable media (transitory or non-transitory, as appropriate) suitable for storing computer program instructions and data include all forms of non volatile memory, media and memory devices, including by way of example semiconductor memory devices, e.g., EPROM, EEPROM, and flash memory devices; magnetic disks, e.g., internal hard disks or removable disks; magneto optical disks; and CD ROM and DVD-ROM disks. The memory may store various objects or data, including caches, classes, frameworks, applications, backup data, jobs, web pages, web page templates, database tables, repositories storing business and/or dynamic information, and any other appropriate information including any parameters, variables, algorithms, instructions, rules, constraints, or references thereto. Additionally, the memory may include any other appropriate data, such as logs, policies, security or access data, reporting files, as well as others. The processor and the memory can be supplemented by, or incorporated in, special purpose logic circuitry.
1001241 To provide for interaction with a user, implementations of the subject matter described in this specification can be implemented on a computer having a display device, e.g., a CRT (cathode ray tube), LCD (liquid crystal display), or plasma monitor, for displaying information to the user and a keyboard and a pointing device, e.g., a mouse or a trackball, by which the user can provide input to the computer. Other kinds of devices can be used to provide for interaction with a user as well; for example, feedback provided to the user can be any form of sensory feedback, e.g., visual feedback, auditory feedback, or tactile feedback; and input from the user can be received in any form, including acoustic, speech, or tactile input. In addition, a computer can interact with a user by sending documents to and receiving documents from a device that is used by the user; for example, by sending web pages to a web browser on a user's client device in response to requests received from the web browser.
I00125) The term "graphical user interface," or GUI, may be used in the singular or the plural to describe one or more graphical user interfaces and each of the displays of a particular graphical user interface. Therefore, a GUI may represent any graphical user interface, including but not limited to, a web browser, a touch screen, or a command line interface (CLI) that processes information and efficiently presents the information results to the user. In general, a GUI may include a plurality of user interface (UI) elements, some or all associated with a web browser, such as interactive fields, pull-down lists, and buttons operable by the business suite user. These and other UI elements may be related to or represent the functions of the web browser.
1001261 Implementations of the subject matter described in this specification can be implemented in a computing system that includes a back end component, e.g., as a data server, or that includes a middlevvare component, e.g., an application server, or that includes a front end component, e.g., a client computer having a graphical user interface or a Web browser through which a user can interact with an implementation of the subject matter described in this specification, or any combination of one or more such back end, middleware, or front end components. The components of the system can be interconnected by any form or medium of digital data communication, e.g., a communication network.
Examples of communication networks include a local area network (LAN), a wide area network (WAN), e.g., the Internet, and a wireless local area network (WLAN).
1001271 The computing system can include clients and servers. A client and server are generally remote from each other and typically interact through a communication network.
The relationship of client and server arises by virtue of computer programs running on the respective computers and having a client-server relationship to each other.
1001281 While this specification contains many specific implementation details, these should not be construed as limitations on the scope of any invention or on the scope of what may be claimed, but rather as descriptions of features that may be specific to particular implementations of particular inventions. Certain features that are described in this specification in the context of separate implementations can also be implemented in combination in a single implementation. Conversely, various features that are described in the context of a single implementation can also be implemented in multiple implementations separately or in any suitable sub-combination. Moreover, although features may be described above as acting in certain combinations and even initially claimed as such, one or more features from a claimed combination can in some cases be excised from the combination, and the claimed combination may be directed to a sub-combination or variation of a sub-combinations.
1001291 Similarly, while operations are depicted in the drawings in a particular order, this should not be understood as requiring that such operations be performed in the particular order shown or in sequential order, or that all illustrated operations be perfonned, to achieve desirable results. In certain circumstances, multitasking and parallel processing may be helpful. Moreover, the separation of various system modules and components in the implementations described above should not be understood as requiring such separation in all implementations, and it should be understood that the described program components and systems can generally be integrated together in a single software product or packaged into multiple software products.
[00130j Particular implementations of the subject matter have been described. Other implementations, alterations, and permutations of the described implementations are within the scope of the following claims as will be apparent to those skilled in the art. For example, the actions recited in the claims can be performed in a different order and still achieve desirable results.
1001311 Accordingly, the above description of example implementations does not define or constrain this disclosure. Other changes, substitutions, and alterations are also possible without departing from the spirit and scope of this disclosure.
Claims (63)
1. A computer-implemented method for detecting facial liveliness, comprising:
processing, by one or more processors, first and second facial images of a subject to determine first and second corneal reflections of an object, the first and second facial images being captured at first and second sequential time points;
determining, by the one or more processors, a corneal reflection change of the object based on the determined first and second corneal reflections;
comparing, by the one or more processors, the determined corneal reflection change of the object to a motion associated with the first and second time points;
and determining, by the one or more processors, facial liveliness of the subject based on a result of the comparison.
processing, by one or more processors, first and second facial images of a subject to determine first and second corneal reflections of an object, the first and second facial images being captured at first and second sequential time points;
determining, by the one or more processors, a corneal reflection change of the object based on the determined first and second corneal reflections;
comparing, by the one or more processors, the determined corneal reflection change of the object to a motion associated with the first and second time points;
and determining, by the one or more processors, facial liveliness of the subject based on a result of the comparison.
2. The method of claim 1, wherein comparing the determined corneal reflection change of the object to a motion comprises:
correlating the determined corneal reflection change of the object to the motion;
scoring a matching quality based on a result of the correlation; and comparing the scored matching quality to a predetermined threshold, and wherein determining facial liveliness of the subject comprises:
determining that the face of the subject is live in response to determining that the scored matching quality is beyond the predetermined threshold.
correlating the determined corneal reflection change of the object to the motion;
scoring a matching quality based on a result of the correlation; and comparing the scored matching quality to a predetermined threshold, and wherein determining facial liveliness of the subject comprises:
determining that the face of the subject is live in response to determining that the scored matching quality is beyond the predetermined threshold.
3. The method of claim 1, further comprising:
determining an expected corneal reflection of the object at the second time point based on the determined first corneal reflection and the motion; and assessing the facial liveliness of the subject by determining likelihood between the expected corneal reflection and the determined second corneal reflection.
determining an expected corneal reflection of the object at the second time point based on the determined first corneal reflection and the motion; and assessing the facial liveliness of the subject by determining likelihood between the expected corneal reflection and the determined second corneal reflection.
4. The method of claim 1, wherein the corneal reflection change of the object includes a position change of the corneal refection of the object.
5. The method of claim 4, wherein the motion is associated with a movement of the object between first and second positions, and wherein the first facial image is captured at the first time point when the object is at the first position and the second facial image is captured at the second time point when the object is at the second position.
6. The method of claim 5, further comprising:
receiving information of the movement of the object measured by a sensor; and determining the motion based on the received information of the movement of the object.
receiving information of the movement of the object measured by a sensor; and determining the motion based on the received information of the movement of the object.
7. The method of claim 6, further comprising:
prompting the subject to move the object.
prompting the subject to move the object.
8. The method of claim 6, wherein the movement of the object is associated with a natural motion of the subject.
9. The method of claim 6, wherein the one or more processors are included in a computing device, and wherein the computing device includes the object and the sensor, and the object is a camera device and the sensor includes at least one of an accelerometer, a gyroscope, or a global positioning system (GPS).
10. The method of claim 9, further comprising:
determining, by using the sensor, that the object is moving; and requesting the camera device to capture facial images of the subject.
determining, by using the sensor, that the object is moving; and requesting the camera device to capture facial images of the subject.
11. The method of claim 5, wherein the motion is a predetermined motion for the object, and wherein the object is moved from the first position to the second position based on the predetermined motion.
12. The method of claim 5, wherein the first and second facial images include a second object that is static during the movement of the object, and wherein the method further comprises determining the motion based on a position change of the second object in the first and second facial images.
13. The method of claim 5, further comprising:
transmitting a command to a controller coupled to the object, the command indicating the controller to move the object.
transmitting a command to a controller coupled to the object, the command indicating the controller to move the object.
14. The method of claim 13, wherein the command comprises a predetermined motion for the object, and wherein the predetermined motion is different from a previous predetermined motion for the object.
15. The method of claim 4, further comprising:
transmitting a command to a camera device to capture facial images of the subject at sequential time points including the first and second time points; and receiving the captured facial images from the camera device.
transmitting a command to a camera device to capture facial images of the subject at sequential time points including the first and second time points; and receiving the captured facial images from the camera device.
16. The method of claim 15, wherein receiving the captured facial images from the camera device comprises receiving a video stream feed from the camera device, the video stream feed comprising the captured facial images.
17. The method of claim 15, wherein the camera device is moved from a first position to a second position between the first and second time points, wherein the first facial image of the subject is captured at the first time point when the camera device is at the first position, and the second facial image of the subject is captured at the second time point when the camera device is at the second position, and wherein the motion is based on a movement of the camera device between the first and second positions.
18. The method of claim 17, further comprising:
transmitting a second command to a controller coupled to the object, the second command indicating the controller to move the object at the first and second time points, wherein the motion is based on the movement of the camera device and the movement of the object.
transmitting a second command to a controller coupled to the object, the second command indicating the controller to move the object at the first and second time points, wherein the motion is based on the movement of the camera device and the movement of the object.
19. The method of claim 1, wherein the object includes one of a camera device, an illumination device, or an object brighter than ambient environment.
20. The method of claim 1, wherein the corneal reflection change of the object includes a first illumination change on the object in the first and second corneal reflections, and wherein the method further comprises:
obtaining a second illumination change on the object for the first and second time points; and determining a matching quality between the obtained second illumination change on the object and the first illumination change in the first and second corneal reflections; and wherein determining facial liveliness of the subject comprises:
determining the facial liveliness of the subject based on the determined matching quality and the result of the comparison.
obtaining a second illumination change on the object for the first and second time points; and determining a matching quality between the obtained second illumination change on the object and the first illumination change in the first and second corneal reflections; and wherein determining facial liveliness of the subject comprises:
determining the facial liveliness of the subject based on the determined matching quality and the result of the comparison.
21. The method of claim 1, further comprising:
determining that the subject chooses to use liveliness verification for biometric authentication.
determining that the subject chooses to use liveliness verification for biometric authentication.
22. The method of claim 21, further comprising:
transmitting, to a computing device at an identity provider, an authentication request of the subject for accessing an account managed by a relying party different from the identity provider, the authentication request including biometric data of the subject and the determined facial liveliness of the subject.
transmitting, to a computing device at an identity provider, an authentication request of the subject for accessing an account managed by a relying party different from the identity provider, the authentication request including biometric data of the subject and the determined facial liveliness of the subject.
23. A computer-implemented method for detecting facial liveliness, comprising:
processing, by one or more processors, first and second facial images of a subject to determine first and second poses of a face of the subject, the first and second facial images being captured at first and second sequential time points;
determining, by the one or more processors, a change in pose of the face based on the determined first and second poses;
comparing, by the one or more processors, the determined change in pose of the face to a motion associated with the first and second time points; and determining, by the one or more processors, facial liveliness of the subject based on a result of the comparison.
processing, by one or more processors, first and second facial images of a subject to determine first and second poses of a face of the subject, the first and second facial images being captured at first and second sequential time points;
determining, by the one or more processors, a change in pose of the face based on the determined first and second poses;
comparing, by the one or more processors, the determined change in pose of the face to a motion associated with the first and second time points; and determining, by the one or more processors, facial liveliness of the subject based on a result of the comparison.
24. The method of claim 23, wherein comparing the determined change in pose of the face to a motion comprises:
correlating the determined change in pose of the face to the motion;
scoring a matching quality based on a result of the correlation; and comparing the scored matching quality to a predetermined threshold, and wherein determining facial liveliness of the subject comprises:
determining that the face of the subject is live in response to determining that the scored matching quality is beyond the predetermined threshold.
correlating the determined change in pose of the face to the motion;
scoring a matching quality based on a result of the correlation; and comparing the scored matching quality to a predetermined threshold, and wherein determining facial liveliness of the subject comprises:
determining that the face of the subject is live in response to determining that the scored matching quality is beyond the predetermined threshold.
25. The method of claim 23, further comprising:
calculating a pose of the face at the second time point based on the determined first pose and the motion; and assessing facial liveliness by determining likelihood between the calculated pose of the face to the determined second pose of the face from the second facial image.
calculating a pose of the face at the second time point based on the determined first pose and the motion; and assessing facial liveliness by determining likelihood between the calculated pose of the face to the determined second pose of the face from the second facial image.
26. The method of claim 23, wherein the motion is associated with a relative movement between the face and a camera device configured to capture facial images of the subject.
27. The method of claim 26, further comprising:
prompting the subject to move the camera device relative to the face of the subject.
prompting the subject to move the camera device relative to the face of the subject.
28. The method of claim 27, further comprising:
receiving information of the movement of the camera device measured by a sensor;
and determining the motion based on the received information of the movement of the camera device.
receiving information of the movement of the camera device measured by a sensor;
and determining the motion based on the received information of the movement of the camera device.
29. The method of claim 26, wherein the one or more processors are included in a computing device, and wherein the computing device includes the camera device and the sensor, and the sensor includes at least one of an accelerometer, a gyroscope, or a global positioning system (GPS).
30. The method of claim 23, wherein determining first and second poses of the face comprises determining at least one facial landmark of the face.
31. A computer-implemented method for detecting facial liveliness, comprising:
processing, by one or more processors, a facial image of a subject to determine a corneal reflection of an illuminator adjacent to the subject, the facial image being captured at a time point;
determining, by the one or more processors, an expected corneal reflection of the illuminator based on an illumination of the illuminator at the time point;
comparing, by the one or more processors, the determined corneal reflection of the illuminator to the expected corneal reflection of the illuminator to obtain a comparison result;
and determining, by the one or more processors, facial liveliness of the subject based on the comparison result.
processing, by one or more processors, a facial image of a subject to determine a corneal reflection of an illuminator adjacent to the subject, the facial image being captured at a time point;
determining, by the one or more processors, an expected corneal reflection of the illuminator based on an illumination of the illuminator at the time point;
comparing, by the one or more processors, the determined corneal reflection of the illuminator to the expected corneal reflection of the illuminator to obtain a comparison result;
and determining, by the one or more processors, facial liveliness of the subject based on the comparison result.
32. The method of claim 31, wherein comparing the determined corneal reflection of the illuminator to the expected corneal reflection of the illuminator comprises:
correlating the determined corneal reflection of the illuminator to the expected corneal reflection of the illuminator;
scoring a matching quality based on a result of the correlation; and comparing the scored matching quality to a predetermined threshold, and wherein determining facial liveliness of the subject comprises:
determining that the face of the subject is live in response to determining that the scored matching quality is beyond the predetermined threshold.
correlating the determined corneal reflection of the illuminator to the expected corneal reflection of the illuminator;
scoring a matching quality based on a result of the correlation; and comparing the scored matching quality to a predetermined threshold, and wherein determining facial liveliness of the subject comprises:
determining that the face of the subject is live in response to determining that the scored matching quality is beyond the predetermined threshold.
33. The method of claim 31, further comprising:
activating the illuminator to illuminate based on an illumination pattem.
activating the illuminator to illuminate based on an illumination pattem.
34. The method of claim 33, further comprising:
generating the illumination pattern for the illuminator.
generating the illumination pattern for the illuminator.
35. The method of claim 33, wherein activating the illuminator comprises:
activating the illuminator in response to determining that the subject chooses to use liveliness verification for biometric authentication.
activating the illuminator in response to determining that the subject chooses to use liveliness verification for biometric authentication.
36. The method of claim 31, wherein determining a corneal reflection of the illuminator comprises determining the corneal reflection of the illuminator in at least one of: color, shape, position, or pattern.
37. The method of claim 31, wherein the illuminator comprises two or more illumination sources with respective colors and at respective positions, respectively.
38. The method of claim 37, wherein the illumination of the illuminator is based on a temporal illumination pattern of the two or more illumination sources.
39. The method of claim 38, wherein the two or more illumination sources generate illumination in bursts each with respective temporal durations.
40. The method of claim 39, wherein at least one of the two or more illumination sources has different illumination shapes for different temporal durations.
41. The method of claim 37, wherein the illumination of the illuminator is based on a spatial illumination pattern of the two or more illumination sources.
42. The method of claim 41, wherein each of the two or more illumination sources has an illumination on or off state.
43. The method of claim 31, wherein the illuminator includes one or more physical light sources each at respective light wavelengths.
44. The method of claim 31, wherein the illuminator includes one or more digital illumination sources displayed on a screen to provide respective illumination colors.
45. The method of claim 44, further comprising generating the digital illumination sources.
46. The method of claim 31, wherein the illuminator is positioned such that illumination from the illuminator is incident on an eye of the subject at an angle.
47. The method of claim 46, further comprising:
prompting the subject to change a relative position of the illuminator and the eye of the subject.
prompting the subject to change a relative position of the illuminator and the eye of the subject.
48. The method of claim 31, further comprising:
receiving at least a portion of a video stream feed from a camera device, the portion of the video stream feed including facial images of the subject captured by the camera device at sequential time points.
receiving at least a portion of a video stream feed from a camera device, the portion of the video stream feed including facial images of the subject captured by the camera device at sequential time points.
49. The method of claim 48, further comprising:
processing a second facial image of the subject to determine a second corneal reflection of the illuminator, the second facial image being captured at a second time point sequential to the time point for the first facial image;
determining a second expected corneal reflection of the illuminator based on a second illumination of the illuminator at the second time point, the second illumination being different from the first illumination;
comparing the determined second corneal reflection of the illuminator to the expected second corneal reflection of the illuminator to obtain a second comparison result; and determining facial liveliness of the subject based on the comparison result and the second comparison result.
processing a second facial image of the subject to determine a second corneal reflection of the illuminator, the second facial image being captured at a second time point sequential to the time point for the first facial image;
determining a second expected corneal reflection of the illuminator based on a second illumination of the illuminator at the second time point, the second illumination being different from the first illumination;
comparing the determined second corneal reflection of the illuminator to the expected second corneal reflection of the illuminator to obtain a second comparison result; and determining facial liveliness of the subject based on the comparison result and the second comparison result.
50. The method of claim 31, further comprising:
transmitting the determined facial liveliness of the subject to a computing device at identity provider.
transmitting the determined facial liveliness of the subject to a computing device at identity provider.
51. The method of claim 31, further comprising:
transmitting, to a computing device at identity provider, an authentication request of the subject for accessing an account managed by a relying party different from the identity provider, the authentication request including biometric data of the subject and the determined facial liveliness of the subject.
transmitting, to a computing device at identity provider, an authentication request of the subject for accessing an account managed by a relying party different from the identity provider, the authentication request including biometric data of the subject and the determined facial liveliness of the subject.
52. A computer-implemented method for detecting facial liveliness, comprising:
processing, by one or more processors, first and second facial images of a subject to determine first and second corneal reflections of an illuminator adjacent to the subject, the first and second facial images being captured at first and second sequential time points, respectively;
determining, by the one or more processors, a corneal reflection change of the illuminator based on the determined first and second corneal reflections;
comparing, by the one or more processors, the determined corneal reflection change of the illuminator to an illumination change of the illuminator associated with the first and second sequential time points; and determining, by the one or more processors, facial liveliness of the subject based on a result of the comparison.
processing, by one or more processors, first and second facial images of a subject to determine first and second corneal reflections of an illuminator adjacent to the subject, the first and second facial images being captured at first and second sequential time points, respectively;
determining, by the one or more processors, a corneal reflection change of the illuminator based on the determined first and second corneal reflections;
comparing, by the one or more processors, the determined corneal reflection change of the illuminator to an illumination change of the illuminator associated with the first and second sequential time points; and determining, by the one or more processors, facial liveliness of the subject based on a result of the comparison.
53. The method of claim 52, further comprising:
activating the illuminator to change illumination of the illuminator based on the illumination change.
activating the illuminator to change illumination of the illuminator based on the illumination change.
54. The method of claim 52, wherein the illumination change of the illuminator comprises at least one of: an illumination color change, an illumination shape change, an illumination position change, an illumination on/off state change, or an illumination temporal duration change.
55. The method of claim 54, wherein determining a corneal reflection change of the illuminator comprises determining a change of the corneal reflection of the illuminator in at least one of: color, shape, position, on/off state, or temporal duration.
56. The method of claim 52, wherein the illuminator comprises first and second illumination sources with first and second colors and at first and second positions, respectively.
57. The method of claim 56, wherein the illumination change of the illuminator comprises a change of a temporal illumination pattern of the first and second illumination sources, and wherein the first and second illumination sources generate illumination in bursts each with respective temporal durations.
58. The method of claim 57, wherein at least one of the first illumination source or the second illumination source changes an illumination shape for different temporal durations.
59. The method of claim 56, wherein the illumination change of the illuminator comprises a change of a spatial illumination pattern of the first and second illumination sources over time, and wherein at least one of the first illumination source or the second illumination source changes an illumination on/off state over time.
60. The method of claim 52, wherein the illuminator includes at least one of one or more physical light sources each at respective light wavelengths, or one or more digital illumination sources displayed on a screen to provide respective illumination colors.
61. A computer-implemented method for detecting facial liveliness, comprising:
processing, by one or more processors, first and second facial images of a subject to determine first and second corneal reflections of an object, the first and second facial images being captured at first and second sequential time points;
determining, by the one or more processors, a corneal reflection change of the object based on the determined first and second corneal reflections:
comparing, by the one or more processors, the determined corneal reflection change of the object to a known change associated with the first and second time points; and determining, by the one or more processors, facial liveliness of the subject based on a result of the comparison.
processing, by one or more processors, first and second facial images of a subject to determine first and second corneal reflections of an object, the first and second facial images being captured at first and second sequential time points;
determining, by the one or more processors, a corneal reflection change of the object based on the determined first and second corneal reflections:
comparing, by the one or more processors, the determined corneal reflection change of the object to a known change associated with the first and second time points; and determining, by the one or more processors, facial liveliness of the subject based on a result of the comparison.
62. The method of claim 61, wherein the known change is a motion of a camera device.
63. The method of claim 61, wherein the known change is an illumination change of an illuminator.
Applications Claiming Priority (5)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201462098575P | 2014-12-31 | 2014-12-31 | |
US201462098596P | 2014-12-31 | 2014-12-31 | |
US62/098,596 | 2014-12-31 | ||
US62/098,575 | 2014-12-31 | ||
PCT/US2015/068357 WO2016109841A1 (en) | 2014-12-31 | 2015-12-31 | Detecting facial liveliness |
Publications (1)
Publication Number | Publication Date |
---|---|
CA2972821A1 true CA2972821A1 (en) | 2016-07-07 |
Family
ID=56285092
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CA2972821A Abandoned CA2972821A1 (en) | 2014-12-31 | 2015-12-31 | Detecting facial liveliness |
Country Status (6)
Country | Link |
---|---|
EP (1) | EP3240467A4 (en) |
JP (1) | JP2018504703A (en) |
KR (1) | KR20170126444A (en) |
AU (1) | AU2015373894B2 (en) |
CA (1) | CA2972821A1 (en) |
WO (1) | WO2016109841A1 (en) |
Families Citing this family (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
GB2560340A (en) * | 2017-03-07 | 2018-09-12 | Eyn Ltd | Verification method and system |
CN108537111A (en) * | 2018-02-26 | 2018-09-14 | 阿里巴巴集团控股有限公司 | A kind of method, apparatus and equipment of In vivo detection |
CN108629298A (en) * | 2018-04-23 | 2018-10-09 | Oppo广东移动通信有限公司 | Face recognition method, electronic device and non-volatile computer-readable storage medium |
JP6446676B1 (en) * | 2018-04-24 | 2019-01-09 | 株式会社サイバーウェア | Identity authentication system, method and program |
GB2576139A (en) * | 2018-07-20 | 2020-02-12 | Ophthalmic Surgical Services Ltd | Ocular assessment |
WO2020022014A1 (en) | 2018-07-25 | 2020-01-30 | 日本電気株式会社 | Information processing device, information processing method, and information processing program |
CN112437947A (en) | 2018-07-25 | 2021-03-02 | 日本电气株式会社 | Information processing apparatus, information processing method, and information processing program |
CN109086731A (en) * | 2018-08-15 | 2018-12-25 | 深圳市烽焌信息科技有限公司 | It is a kind of for carrying out the robot and storage medium of behavior monitoring |
CN109145804A (en) * | 2018-08-15 | 2019-01-04 | 深圳市烽焌信息科技有限公司 | Behavior monitoring method and robot |
CN109271954B (en) * | 2018-09-29 | 2022-04-22 | 北京百度网讯科技有限公司 | Method and device for detecting reliability of model |
CN110135370B (en) * | 2019-05-20 | 2022-09-09 | 北京百度网讯科技有限公司 | Method and device for detecting living human face, electronic equipment and computer readable medium |
WO2020246141A1 (en) | 2019-06-06 | 2020-12-10 | 日本電気株式会社 | Processing system, processing method, and program |
EP3772699A1 (en) * | 2019-08-09 | 2021-02-10 | Siemens Aktiengesellschaft | Method for user verification, communication device and computer program |
FR3121773B1 (en) * | 2021-04-13 | 2024-03-08 | Idemia Identity & Security France | identity verification method using random lighting conditions |
Family Cites Families (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP3315648B2 (en) * | 1998-07-17 | 2002-08-19 | 沖電気工業株式会社 | Iris code generation device and iris recognition system |
JP2000194854A (en) * | 1998-12-25 | 2000-07-14 | Oki Electric Ind Co Ltd | Individual identification device |
JP4521086B2 (en) * | 2000-03-13 | 2010-08-11 | 株式会社東芝 | Face image recognition apparatus and face image recognition method |
JP3825222B2 (en) * | 2000-03-24 | 2006-09-27 | 松下電器産業株式会社 | Personal authentication device, personal authentication system, and electronic payment system |
JP2003178306A (en) * | 2001-12-12 | 2003-06-27 | Toshiba Corp | Personal identification device and personal identification method |
JP4009173B2 (en) * | 2002-09-30 | 2007-11-14 | 富士フイルム株式会社 | Digital camera |
JP2006126899A (en) * | 2004-10-26 | 2006-05-18 | Matsushita Electric Ind Co Ltd | Biological discrimination device, biological discrimination method, and authentication system using the same |
JP4609253B2 (en) * | 2005-09-08 | 2011-01-12 | オムロン株式会社 | Impersonation detection device and face authentication device |
US7801335B2 (en) * | 2005-11-11 | 2010-09-21 | Global Rainmakers Inc. | Apparatus and methods for detecting the presence of a human eye |
US8260008B2 (en) * | 2005-11-11 | 2012-09-04 | Eyelock, Inc. | Methods for performing biometric recognition of a human eye and corroboration of same |
US20130212655A1 (en) * | 2006-10-02 | 2013-08-15 | Hector T. Hoyos | Efficient prevention fraud |
JP5207776B2 (en) * | 2008-03-05 | 2013-06-12 | エヌ・ティ・ティ・コミュニケーションズ株式会社 | Authentication system, information device, authentication method, and program |
US9208492B2 (en) * | 2013-05-13 | 2015-12-08 | Hoyos Labs Corp. | Systems and methods for biometric authentication of transactions |
US8542879B1 (en) * | 2012-06-26 | 2013-09-24 | Google Inc. | Facial recognition |
US8437513B1 (en) * | 2012-08-10 | 2013-05-07 | EyeVerify LLC | Spoof detection for biometric authentication |
US9058519B2 (en) * | 2012-12-17 | 2015-06-16 | Qualcomm Incorporated | System and method for passive live person verification using real-time eye reflection |
-
2015
- 2015-12-31 AU AU2015373894A patent/AU2015373894B2/en not_active Expired - Fee Related
- 2015-12-31 CA CA2972821A patent/CA2972821A1/en not_active Abandoned
- 2015-12-31 EP EP15876387.0A patent/EP3240467A4/en not_active Withdrawn
- 2015-12-31 KR KR1020177021245A patent/KR20170126444A/en not_active Withdrawn
- 2015-12-31 JP JP2017535830A patent/JP2018504703A/en active Pending
- 2015-12-31 WO PCT/US2015/068357 patent/WO2016109841A1/en active Application Filing
Also Published As
Publication number | Publication date |
---|---|
EP3240467A1 (en) | 2017-11-08 |
WO2016109841A1 (en) | 2016-07-07 |
JP2018504703A (en) | 2018-02-15 |
AU2015373894A1 (en) | 2017-07-20 |
KR20170126444A (en) | 2017-11-17 |
AU2015373894B2 (en) | 2020-07-09 |
EP3240467A4 (en) | 2018-12-26 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10346990B2 (en) | Detecting facial liveliness | |
US10055662B2 (en) | Detecting facial liveliness | |
AU2015373894B2 (en) | Detecting facial liveliness | |
US11528268B2 (en) | System and method for verifying liveliness | |
US10515264B2 (en) | Systems and methods for authenticating a user based on captured image data | |
US10268910B1 (en) | Authentication based on heartbeat detection and facial recognition in video data | |
JP6641511B2 (en) | System and method for authorizing access to an access controlled environment | |
US10896248B2 (en) | Systems and methods for authenticating user identity based on user defined image data | |
US10326761B2 (en) | Web-based user authentication techniques and applications | |
US20160269411A1 (en) | System and Method for Anonymous Biometric Access Control | |
US10733275B1 (en) | Access control through head imaging and biometric authentication | |
US10217009B2 (en) | Methods and systems for enhancing user liveness detection | |
US10956544B1 (en) | Access control through head imaging and biometric authentication | |
US11093770B2 (en) | System and method for liveness detection | |
CA2910929C (en) | Systems and methods for authenticating user identity based on user-defined image data | |
US10277595B2 (en) | Identity recognition with living signatures from multiple devices |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
EEER | Examination request |
Effective date: 20201223 |
|
FZDE | Discontinued |
Effective date: 20220630 |
|
FZDE | Discontinued |
Effective date: 20220630 |