CN111222433B - Automatic face auditing method, system, equipment and readable storage medium - Google Patents
Automatic face auditing method, system, equipment and readable storage medium Download PDFInfo
- Publication number
- CN111222433B CN111222433B CN201911387766.9A CN201911387766A CN111222433B CN 111222433 B CN111222433 B CN 111222433B CN 201911387766 A CN201911387766 A CN 201911387766A CN 111222433 B CN111222433 B CN 111222433B
- Authority
- CN
- China
- Prior art keywords
- face
- pictures
- picture
- quality
- image
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000000034 method Methods 0.000 title claims abstract description 38
- 239000013598 vector Substances 0.000 claims abstract description 45
- 238000001514 detection method Methods 0.000 claims abstract description 30
- 238000001914 filtration Methods 0.000 claims abstract description 26
- 238000013441 quality evaluation Methods 0.000 claims abstract description 16
- 238000013528 artificial neural network Methods 0.000 claims abstract description 15
- 238000012216 screening Methods 0.000 claims abstract description 13
- 238000013527 convolutional neural network Methods 0.000 claims abstract description 11
- 230000006870 function Effects 0.000 claims description 38
- 210000001508 eye Anatomy 0.000 claims description 19
- 230000008569 process Effects 0.000 claims description 9
- 238000013135 deep learning Methods 0.000 claims description 8
- 238000004590 computer program Methods 0.000 claims description 7
- 238000010586 diagram Methods 0.000 claims description 6
- 238000007667 floating Methods 0.000 claims description 6
- 239000011159 matrix material Substances 0.000 claims description 6
- 238000010606 normalization Methods 0.000 claims description 6
- 238000001303 quality assessment method Methods 0.000 claims description 6
- 230000009466 transformation Effects 0.000 claims description 6
- 238000001727 in vivo Methods 0.000 claims description 4
- 238000012545 processing Methods 0.000 claims description 4
- 238000012549 training Methods 0.000 claims description 4
- 230000001815 facial effect Effects 0.000 claims 1
- 238000005516 engineering process Methods 0.000 description 3
- 238000004458 analytical method Methods 0.000 description 2
- 238000000605 extraction Methods 0.000 description 2
- 239000011521 glass Substances 0.000 description 2
- 238000012795 verification Methods 0.000 description 2
- 230000004075 alteration Effects 0.000 description 1
- 238000001574 biopsy Methods 0.000 description 1
- 239000003795 chemical substances by application Substances 0.000 description 1
- 238000010276 construction Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 210000004709 eyebrow Anatomy 0.000 description 1
- 210000000887 face Anatomy 0.000 description 1
- 238000010191 image analysis Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 210000002569 neuron Anatomy 0.000 description 1
- 210000001747 pupil Anatomy 0.000 description 1
- 210000000697 sensory organ Anatomy 0.000 description 1
- 239000004575 stone Substances 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/168—Feature extraction; Face representation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/22—Matching criteria, e.g. proximity measures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q50/00—Information and communication technology [ICT] specially adapted for implementation of business processes of specific business sectors, e.g. utilities or tourism
- G06Q50/10—Services
- G06Q50/26—Government or public services
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/172—Classification, e.g. identification
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/174—Facial expression recognition
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/40—Spoof detection, e.g. liveness detection
- G06V40/45—Detection of the body part being alive
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Oral & Maxillofacial Surgery (AREA)
- Human Computer Interaction (AREA)
- Data Mining & Analysis (AREA)
- Multimedia (AREA)
- Business, Economics & Management (AREA)
- Evolutionary Computation (AREA)
- Artificial Intelligence (AREA)
- General Engineering & Computer Science (AREA)
- Life Sciences & Earth Sciences (AREA)
- Biophysics (AREA)
- Software Systems (AREA)
- Mathematical Physics (AREA)
- Computing Systems (AREA)
- Tourism & Hospitality (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Molecular Biology (AREA)
- Computational Linguistics (AREA)
- Biomedical Technology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Evolutionary Biology (AREA)
- Development Economics (AREA)
- Educational Administration (AREA)
- Bioinformatics & Computational Biology (AREA)
- Economics (AREA)
- Human Resources & Organizations (AREA)
- Marketing (AREA)
- Primary Health Care (AREA)
- Strategic Management (AREA)
- General Business, Economics & Management (AREA)
- Collating Specific Patterns (AREA)
- Image Analysis (AREA)
Abstract
The invention discloses an automatic face auditing method, which comprises the following steps: face detection, namely detecting coordinates of a rectangular frame of the face and coordinates of key points of the face through a cascade neural network algorithm; face quality evaluation and screening, namely performing face quality evaluation on quality attributes of a plurality of face pictures, and screening high-quality pictures; detecting living bodies, namely detecting whether the pictures are true persons or not by using a double-flow convolutional neural network, and filtering and judging the pictures to be non-true persons; and (3) face comparison authentication, namely extracting face feature vectors of the pictures, comparing the similarity degree between the face feature vectors of the pictures and the face feature vectors of the standard pictures, and outputting a comparison result. The invention realizes full automation of picture auditing, does not need manual operation, reduces labor cost, filters pictures with low similarity degree, and reduces manual errors. In addition, each invention uses a cascading picture filtering mode, so that the speed is higher.
Description
Technical Field
The present invention relates to the field of image recognition processing technologies, and in particular, to a method, a system, an apparatus, and a readable storage medium for automatic face auditing.
Background
With the rapid development of mobile devices and mobile internet, many government platforms and enterprise platforms are increasingly beginning to merge offline and online services for the convenience and convenience of mass offices. For example, governments in various places have accelerated the construction of on-line government service platforms, mobile operators have opened on-line business halls, and so on. Many online services need to upload photos and keep a bottom according to requirements, for example, mobile phone operators need to upload personal photos for online handling of mobile phone numbers, front photos and back photos of identity cards, and independent application of shops for services such as kittens, treasures and the like need to upload personal handheld photos. In the prior art, manual verification and judgment are mainly adopted for personal photos uploaded by users. However, depending on manual operation to examine whether the personal photo uploaded by the user is compliant, whether the personal photo is compliant, etc., the working efficiency is low, the labor cost is high, human errors are inevitably generated in the process, and feedback to the user cannot be timely obtained.
Disclosure of Invention
The invention aims to provide a quick and efficient automatic face auditing method, system and device and a readable storage medium.
In order to solve the technical problems, the technical scheme of the invention is as follows:
in a first aspect, the present invention provides an automatic face auditing method, including the steps of:
face detection, namely detecting coordinates of a rectangular frame of the face and coordinates of key points of the face through a cascade neural network algorithm;
face quality evaluation and screening, namely performing face quality evaluation on quality attributes of a plurality of face pictures, and screening high-quality pictures;
detecting living bodies, namely detecting whether the pictures are true persons or not by using a double-flow convolutional neural network, and filtering and judging the pictures to be non-true persons;
and (3) face comparison authentication, namely extracting face feature vectors of the pictures, comparing the similarity degree between the face feature vectors of the pictures and the face feature vectors of the standard pictures, and filtering the pictures with low similarity degree.
Preferably, before face detection, a picture with a filtering resolution outside the threshold range is further included.
Preferably, after the face detection, the method further comprises:
filtering out the pictures with the size of the face rectangular frame accounting for the size of the picture and lower than a threshold value;
and filtering out pictures with the distance between the two eyes of the human face lower than a threshold value.
Preferably, after the face detection, the method further comprises:
and aligning the human face, calculating a transformation matrix between the key point coordinates of the human face of a picture and the key point coordinates of a pre-stored standard human face, and acting the transformation matrix on the picture to obtain an aligned human face image.
Preferably, the face comparison authentication process is as follows:
extracting the high-quality picture, outputting a floating point vector with 512 dimensions by using a 50-layer ResNet neural network, and recording the floating point vector as a face feature vector;
by comparing the similarity degree between the face feature vector of the current picture and the face feature vector of the standard picture, the formula is as follows:
wherein S is i Is the face feature vector of the current picture, S j The face feature vector is the face feature vector of the standard picture;
if the similarity is lower than a threshold, judging that the person evidence is not uniform; and if the similarity is higher than a threshold value, judging that the person and the evidence are uniform.
Preferably, the quality attributes used for face quality assessment include: the quality attributes used for face quality assessment comprise face gestures, eye states, mouth states, makeup states, overall brightness, left and right face brightness differences, ambiguity and shielding;
the face pose, the eye state, the mouth state, the dressing state, the ambiguity and the occlusion all adopt a MobileFaceNet structure as a main body to construct a multi-task convolutional neural network, and a plurality of tasks output corresponding to the quality attributes of the face respectively.
Eye state, mouth state, makeup state and face shielding are used as classification tasks, and a softmax loss function is used as an objective function;
the face gesture, the illuminance of the image and the image ambiguity are used as regression tasks, and a Euclidean loss function is used as an objective function;
the total objective function of the network training comprises a combination of a plurality of Softmax loss functions and a Euclidean loss function, and when a plurality of tasks are learned together, the total objective function is a linear combination of a plurality of loss functions.
Preferably, the in vivo detection process is:
acquiring a depth image, and carrying out normalization processing on a face region in the image to obtain a processed face depth image;
inputting RGB face images with a face ID of a preset frame number and the face depth map into a deep learning network for detection, and obtaining a living body judgment result of each frame of picture;
voting all living body judgment results of the face ID, when the number of frames of the living body is judged to be large, determining the object as a living body, and when the number of frames of the attack is judged to be large, determining the object as a non-living body.
Preferably, the in vivo detection process is:
the face is intercepted from the original image, RGB channels are converted into HSV and YCbCr spaces, and the converted HSV and YCbCr images are overlapped to obtain an overlapped image; extracting Sobel features from the face region through a Sobel operator, and obtaining a Sobel feature map;
respectively inputting the Sobel characteristic diagram and the superposition diagram of a preset frame number of a face ID from two input channels of a double-flow neural network to obtain a living body judgment result of each frame of picture;
voting all living body judgment results of the face ID, when the number of frames of the living body is judged to be large, the object is judged to be living body, and when the number of frames of the attack is judged to be large, the object is judged to be non-living body.
In a second aspect, the present invention further provides an automatic face auditing system, including:
the human face detection module is used for detecting a human face rectangular frame and human face key points through a cascade neural network algorithm;
the face quality evaluation module is used for carrying out face quality evaluation through quality attributes of a plurality of face pictures and screening high-quality pictures;
the living body detection module is used for detecting whether the pictures are true persons or not by using a double-flow convolutional neural network and filtering pictures without the true persons;
and a face comparison module for extracting face feature vectors of the pictures, comparing the similarity between the face feature vectors of the pictures and the face feature vectors of the pre-stored credentials, and filtering the pictures with low similarity.
In a third aspect, the present invention further provides an automatic face auditing apparatus, including a memory, a processor, and a computer program stored in the memory and capable of running on the processor, where the processor implements the steps of the automatic face auditing method described above when executing the program.
In a fourth aspect, the present invention further proposes a readable storage medium having stored thereon a computer program which, when executed by a processor, implements the steps of the above-described automatic face auditing method.
The technical scheme of the invention is an automatic face auditing method combining face detection, face quality analysis, face living body detection and face recognition technologies. The method examines whether the quality of the personal photo uploaded by the user is compliant, whether the personal photo is a true person or not, and whether the personal photo is authenticated. In the technical scheme, whether the quality of the personal photo is in compliance mainly considers whether the face in the picture is high in quality and is convenient to identify; the living body detection mainly considers whether the photo is a flip or fake; whether the personal certificate is unified mainly considers whether the certificate photograph and the personal photo are the same person. The invention realizes full automation of picture auditing, does not need manual operation, reduces labor cost, has stable algorithm and reduces manual error. In addition, each invention uses a cascading picture filtering mode, so that the speed is higher.
Drawings
Fig. 1 is a flowchart illustrating steps of an embodiment of an automatic face auditing method according to the present invention.
Detailed Description
The following describes the embodiments of the present invention further with reference to the drawings. The description of these embodiments is provided to assist understanding of the present invention, but is not intended to limit the present invention. In addition, the technical features of the embodiments of the present invention described below may be combined with each other as long as they do not collide with each other.
Referring to fig. 1, the invention provides an automatic face auditing method, which comprises the following steps:
filtering pictures with resolution outside the threshold range, and taking pictures with resolution lower than the threshold in the vertical direction or lower than the threshold in the horizontal direction as pictures with resolution not meeting the requirements. In the embodiment of the invention, the photos with the resolution lower than 640 in the vertical direction or the resolution lower than 480 in the horizontal direction are filtered and deleted.
S10: and detecting the coordinates of the rectangular frame of the human face and the coordinates of key points of the human face through a cascade neural network algorithm.
And predicting the coordinates of the face frame and the coordinates of the key points of the face in the image by using a cascade neural network algorithm. The face frame coordinates refer to a rectangular face frame containing a face area; the face key point coordinates refer to positions of 106 key points of the face area, and cover eyebrows, glasses, nose, mouth and face contours of the face area.
And calculating the size of the face frame according to the face frame coordinates. When the face duty ratio is greater than the threshold, the face duty ratio does not meet the requirement. In this embodiment, when the face ratio is greater than 0.4, the proportion of the face to the whole image is considered to be too large.
Face duty ratio = face frame size/image size
And calculating the distance between pupils of two eyes according to the key points of the human face, namely the number of pixels between the centers of the two eyes. When the interocular distance is smaller than the threshold value, the interocular distance does not satisfy the requirement. For example, when the distance between the left eye and the right eye is less than 40, the inter-eye distance is too small.
Face alignment: and aiming at each face, calculating a transformation matrix between the extracted face key point coordinates and the standard face key point coordinates, and applying the transformation matrix to the initial face image to obtain an aligned face image. The distribution of the aligned face key point coordinates is more consistent, and the face is corrected.
S20: face quality evaluation and screening, namely performing face quality evaluation on quality attributes of a plurality of face pictures, and screening high-quality pictures.
The face quality evaluation algorithm adopts a mode of combining deep learning and a traditional image analysis algorithm, and achieves the quality attributes of face brightness, left-right face brightness difference, face angle (yaw) around the y-axis direction, face angle (pitch) around the x-axis direction, face angle (roll) around the z-axis direction, expression classification, glasses classification, mask classification, eye state classification, mouth state classification, makeup state classification, face reality (the distinguishing of the items is stone statue image, CG face and real face), face ambiguity, face shielding degree and the like according to the face characteristics of the face image.
The face brightness and the left and right face brightness difference adopt a traditional algorithm, specifically, three RGB channels of a face image are converted into gray images according to a certain proportion, each region of the face is divided according to 106 key points of the face, the face brightness is calculated according to the gray average value of the face region, and the left and right face brightness difference is calculated according to the gray average value of the left and right faces.
The other attributes are realized by adopting a deep learning method, a lightweight MobileFaceNet structure is adopted as a main body to construct a multi-task convolutional neural network, and a plurality of tasks output each quality attribute corresponding to a face. The quality judgment of eye state, mouth state, dressing state, face shielding, mask classification and the like belongs to classification tasks, and a softmax loss function is adopted as an objective function; face pose angles, image ambiguities and the like belong to regression tasks, and a Euclidean loss function is adopted as an objective function. The total objective function of the network training is a combination of a plurality of Softmax loss functions and a Euclidean loss function, and when a plurality of tasks are learned together, the total objective function is a linear combination of a plurality of loss functions.
Calculate Softmax loss L: l= -log (p i ),
Wherein p is i Normalized probabilities calculated for each attribute class, i.ex i Representing the ith neuron output, N representing the total number of categories;
After face quality evaluation, face quality screening is also required, and reference factors for screening include the following:
face ratio: and calculating the size of the face frame according to the face frame coordinates, wherein when the face ratio is larger than a threshold value, the face ratio does not meet the requirement. For example: when the face ratio is greater than 0.4, the proportion of the face to the whole image is too large.
Face brightness: the brightness of the human face should be within a reasonable range. For example: the human face brightness value is between 0 and 1, and the reasonable human face brightness is more than 0.3 and less than 0.8.
Left-right face brightness difference: the left-right face luminance difference should be less than the threshold. For example: when the difference of the left and right face brightness is between 0 and 1, the reasonable difference of the left and right face brightness is less than 0.4.
Face pose: the face angle (yaw) around the y-axis, the face angle (pitch) around the x-axis, and the face angle (roll) around the z-axis should be within a reasonable range. For example, within + -10 deg..
Ambiguity: the ambiguity should be less than a threshold. For example: when the ambiguity is between 0 and 1, the face ambiguity should be less than 0.6.
Shielding: if the face map is judged to have shielding of the five sense organs and the outlines, including a sunglasses or a mask, filtering is carried out.
Expression: if the face image is judged to be an exaggerated expression, the eyes are closed, and the mouth is enlarged, filtering is performed.
Degree of realism: the degree of realism should be greater than a threshold, if the degree of realism is small, this may indicate that the face is a statue face/cartoon face, etc. For example: when the value of the reality is between 0 and 1, the reality of the human face is larger than 0.6.
And filtering out pictures which do not meet the quality requirements according to the requirements.
S30: and detecting living bodies, namely detecting whether the pictures are true persons or not by using a double-flow convolutional neural network, and filtering and judging the pictures to be non-true persons.
The following two methods can be used in performing the biopsy:
the first living body detection process is as follows:
acquiring a depth image, and carrying out normalization processing on a face region in the image to obtain a processed face depth image;
and inputting the RGB face image of the picture and the face depth image into a deep learning network for detection, and obtaining a living body judgment result of the picture.
Specifically, a deep learning network for performing living body judgment on a picture uses Resnet as a basic network, the deep learning network adopts a double-input channel of a face image and a face depth image, after feature extraction is performed on two input branches respectively, the features extracted by the two branches are selectively excited and fused through a se-module, and then feature extraction is performed on the fused features through a plurality of layers of convolution, so that a living body judgment result is obtained.
Specifically, the objective function of the deep learning network is a focal loss function.
Specifically, the actual depths of all the eyes and mouth angles in the key points of the human face are calculated, the average value of the actual depths of the points is calculated, the upper limit of normalization is taken as the average value plus a fixed value, the lower limit is taken as the average value minus the fixed value, and the depth of the human face area is normalized into a gray scale map with the pixel value in the interval of 0-255.
The gray value for the actual depth greater than the upper limit and less than the lower limit is set to 0.
Wherein, the normalization formula is:
wherein V is a gray value after depth normalization, the range is 0-255, dreal is the actual depth of the face region, dmax is the upper limit of the actual depth of the face, and Dmin is the lower limit of the actual depth of the face.
The second living body detection process is as follows:
the face is intercepted from the original image, RGB channels are converted into HSV and YCbCr spaces, and the converted HSV and YCbCr images are overlapped to obtain an overlapped image; extracting Sobel features from the face region through a Sobel operator to obtain a Sobel feature map;
and respectively inputting the Sobel characteristic diagram and the superposition diagram of the picture from two input channels of the double-flow neural network to obtain a living body judgment result of the picture.
Specifically, for each input image a, gx and Gy are convolved with image a to obtain AGx, AGy, respectively, and then image AG is output, the value of each pixel is:where Gx represents the convolution kernel in the x-direction and Gy represents the convolution kernel in the y-direction.
S40: and (3) face comparison authentication, namely extracting face feature vectors of the pictures, comparing the similarity degree between the face feature vectors of the pictures and the face feature vectors of the standard pictures, and outputting a comparison result.
Extracting a high-quality picture, outputting a floating point vector with 512 dimensions by using a 50-layer ResNet neural network, and recording the floating point vector as a face feature vector;
by comparing the similarity degree between the face feature vector of the current picture and the face feature vector of the standard picture, the formula is as follows:
wherein S is i Is the face feature vector of the current picture, S j The face feature vector is the face feature vector of the standard picture;
if the similarity is lower than the threshold, judging that the credentials are not uniform; if the similarity is higher than the threshold, judging that the people are unified
The technical scheme of the invention is a face auditing method combining face detection, face quality analysis, face living body detection and face recognition technology, and is used for checking whether the quality of the personal photo uploaded by the user is compliant, whether the personal photo is a true person or not and whether the personal photo is authenticated. Whether the quality of the personal photo is compliant mainly considers whether the face in the picture is high in quality and is convenient to identify; the living body detection mainly considers whether the photo is a flip or fake; whether the personal certificate is unified mainly considers whether the certificate photograph and the personal photo are the same person. The invention realizes full automation of picture auditing, does not need manual operation, reduces labor cost, has stable algorithm and reduces manual error. In addition, each invention uses a cascading picture filtering mode, so that the speed is higher.
On the other hand, the invention also provides an automatic face auditing system, which comprises the following steps:
the human face detection module is used for detecting a human face rectangular frame and human face key points through a cascade neural network algorithm;
the face quality evaluation module is used for carrying out face quality evaluation through quality attributes of a plurality of face pictures and screening high-quality pictures;
the living body detection module is used for detecting whether the pictures are true persons or not by using a double-flow convolutional neural network and filtering pictures without the true persons;
and a face comparison module for extracting face feature vectors of the pictures, comparing the similarity between the face feature vectors of the pictures and the face feature vectors of the pre-stored credentials, and filtering the pictures with low similarity.
In still another aspect, the present invention further provides an automatic face auditing apparatus, including a memory, a processor, and a computer program stored in the memory and capable of running on the processor, where the processor implements the steps of the automatic face auditing method described above when executing the program.
In yet another aspect, the present invention further provides a readable storage medium having stored thereon a computer program which, when executed by a processor, performs the steps of the above-described automatic face auditing method.
Aiming at websites and applications requiring uploading images to meet certain standard requirements, the invention provides an automatic face auditing method and system based on a deep neural network. The method can be effectively used in the information verification category, and realizes rapid face filtering and person evidence comparison.
The embodiments of the present invention have been described in detail above with reference to the accompanying drawings, but the present invention is not limited to the described embodiments. It will be apparent to those skilled in the art that various changes, modifications, substitutions and alterations can be made to these embodiments without departing from the principles and spirit of the invention, and yet fall within the scope of the invention.
Claims (9)
1. An automatic face auditing method is characterized by comprising the following steps:
face detection, namely detecting coordinates of a rectangular frame of the face and coordinates of key points of the face through a cascade neural network algorithm;
face quality evaluation and screening, namely performing face quality evaluation on quality attributes of a plurality of face pictures, and screening high-quality pictures;
detecting living bodies, namely detecting whether the pictures are true persons or not by using a double-flow convolutional neural network, and filtering and judging the pictures to be non-true persons;
face comparison and authentication, namely extracting a face feature vector of a picture, comparing the similarity degree between the face feature vector of the picture and the face feature vector of a standard picture, and filtering a picture with low similarity degree; before face detection, filtering pictures with resolution outside a threshold range;
the quality attributes used for face quality assessment include: the quality attributes used for face quality assessment comprise face gestures, eye states, mouth states, makeup states, overall brightness, left and right face brightness differences, ambiguity and shielding;
the face pose, eye state, mouth state, dressing state, ambiguity and occlusion all adopt a MobileFaceNet structure as a main body to construct a multi-task convolutional neural network, and a plurality of task outputs respectively correspond to each quality attribute of the face;
eye state, mouth state, makeup state and face shielding are used as classification tasks, and a softmax loss function is used as an objective function;
the face gesture, the illuminance of the image and the image ambiguity are used as regression tasks, and a Euclidean loss function is used as an objective function;
the total objective function of the network training comprises a combination of a plurality of Softmax loss functions and a Euclidean loss function, and when a plurality of tasks are learned together, the total objective function is a linear combination of a plurality of loss functions.
2. The automatic face auditing method of claim 1, characterized by: after face detection, further comprising:
filtering out the pictures with the size of the face rectangular frame accounting for the size of the picture and lower than a threshold value;
and filtering out pictures with the distance between the two eyes of the human face lower than a threshold value.
3. The automatic face auditing method according to claim 1, further comprising, after face detection:
and aligning the human face, calculating a transformation matrix between the coordinates of the key points of the human face of a picture and the coordinates of the key points of the pre-stored standard human face, and acting the transformation matrix on the picture to obtain an aligned human face image.
4. A method of automatic face auditing according to any of claims 1 to 3, in which the process of face comparison authentication is:
extracting the high-quality picture, outputting a floating point vector with 512 dimensions by using a 50-layer ResNet neural network, and recording the floating point vector as a face feature vector;
by comparing the similarity degree between the face feature vector of the current picture and the face feature vector of the standard picture, the formula is as follows:
wherein S is i Is the face feature vector of the current picture, S j The face feature vector is the face feature vector of the standard picture;
if the similarity is lower than a threshold, judging that the person evidence is not uniform; and if the similarity is higher than a threshold value, judging that the person and the evidence are unified.
5. A method of automatic face auditing according to any of claims 1 to 3, in which the process of in vivo detection is:
acquiring a depth image, and carrying out normalization processing on a face region in the image to obtain a processed face depth image;
and inputting the RGB face image of the picture and the face depth image into a deep learning network for detection, and obtaining a living body judgment result of the picture.
6. A method of automatic face auditing according to any of claims 1 to 3, in which the process of in vivo detection is:
the face is intercepted from the original image, RGB channels are converted into HSV and YCbCr spaces, and the converted HSV and YCbCr images are overlapped to obtain an overlapped image; extracting Sobel features from the face region through a Sobel operator to obtain a Sobel feature map;
and respectively inputting the Sobel characteristic diagram and the superposition diagram of the picture from two input channels of the double-flow neural network to obtain a living body judgment result of the picture.
7. An automated face auditing system, comprising:
the human face detection module is used for detecting a human face rectangular frame and human face key points through a cascade neural network algorithm;
the face quality evaluation module is used for carrying out face quality evaluation through quality attributes of a plurality of face pictures and screening high-quality pictures;
the living body detection module is used for detecting whether the pictures are true persons or not by using a double-flow convolutional neural network and filtering pictures without the true persons;
the face comparison module is used for extracting face feature vectors of the pictures, comparing the similarity degree between the face feature vectors of the pictures and the face feature vectors of the pre-stored credentials, and filtering the pictures with low similarity degree;
the quality attributes used for face quality assessment include: the quality attributes used for face quality assessment comprise face gestures, eye states, mouth states, makeup states, overall brightness, left and right face brightness differences, ambiguity and shielding;
the face pose, eye state, mouth state, dressing state, ambiguity and occlusion all adopt a MobileFaceNet structure as a main body to construct a multi-task convolutional neural network, and a plurality of task outputs respectively correspond to each quality attribute of the face;
eye state, mouth state, makeup state and face shielding are used as classification tasks, and a softmax loss function is used as an objective function;
the face gesture, the illuminance of the image and the image ambiguity are used as regression tasks, and a Euclidean loss function is used as an objective function;
the total objective function of the network training comprises a combination of a plurality of Softmax loss functions and a Euclidean loss function, and when a plurality of tasks are learned together, the total objective function is a linear combination of a plurality of loss functions.
8. An automated facial auditing apparatus comprising a memory, a processor, and a computer program stored on the memory and executable on the processor, characterized by: the processor, when executing the program, implements the steps of the automatic face auditing method of any of claims 1-6.
9. A readable storage medium having stored thereon a computer program for automated face auditing, characterized by: the computer program when executed by a processor implements the steps of the automatic face auditing method of any of claims 1-6.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201911387766.9A CN111222433B (en) | 2019-12-30 | 2019-12-30 | Automatic face auditing method, system, equipment and readable storage medium |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201911387766.9A CN111222433B (en) | 2019-12-30 | 2019-12-30 | Automatic face auditing method, system, equipment and readable storage medium |
Publications (2)
Publication Number | Publication Date |
---|---|
CN111222433A CN111222433A (en) | 2020-06-02 |
CN111222433B true CN111222433B (en) | 2023-06-20 |
Family
ID=70829143
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201911387766.9A Active CN111222433B (en) | 2019-12-30 | 2019-12-30 | Automatic face auditing method, system, equipment and readable storage medium |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN111222433B (en) |
Families Citing this family (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113807144A (en) * | 2020-06-15 | 2021-12-17 | 福建新大陆支付技术有限公司 | Testing method of living body detection equipment |
CN111753923A (en) * | 2020-07-02 | 2020-10-09 | 携程计算机技术(上海)有限公司 | Intelligent photo album clustering method, system, equipment and storage medium based on human face |
CN112200108A (en) * | 2020-10-16 | 2021-01-08 | 深圳市华付信息技术有限公司 | Mask face recognition method |
CN112329638A (en) * | 2020-11-06 | 2021-02-05 | 上海优扬新媒信息技术有限公司 | Image scoring method, device and system |
CN112528939B (en) * | 2020-12-22 | 2024-09-06 | 广州海格星航信息科技有限公司 | Quality evaluation method and device for face image |
CN113282894A (en) * | 2021-01-26 | 2021-08-20 | 上海欧冶金融信息服务股份有限公司 | Identity verification method and system for wind-control full-pitch |
CN114093004B (en) * | 2021-11-25 | 2023-05-02 | 成都智元汇信息技术股份有限公司 | Face fusion comparison method and device based on multiple cameras |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102622588A (en) * | 2012-03-08 | 2012-08-01 | 无锡数字奥森科技有限公司 | Dual-certification face anti-counterfeit method and device |
CN104143086A (en) * | 2014-07-18 | 2014-11-12 | 吴建忠 | Application technology of portrait comparison to mobile terminal operating system |
CN108280399A (en) * | 2017-12-27 | 2018-07-13 | 武汉普利商用机器有限公司 | A kind of scene adaptive face identification method |
CN109815826A (en) * | 2018-12-28 | 2019-05-28 | 新大陆数字技术股份有限公司 | The generation method and device of face character model |
WO2019100608A1 (en) * | 2017-11-21 | 2019-05-31 | 平安科技(深圳)有限公司 | Video capturing device, face recognition method, system, and computer-readable storage medium |
CN110580445A (en) * | 2019-07-12 | 2019-12-17 | 西北工业大学 | An Improved Face Keypoint Detection Method Based on GIoU and Weighted NMS |
-
2019
- 2019-12-30 CN CN201911387766.9A patent/CN111222433B/en active Active
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102622588A (en) * | 2012-03-08 | 2012-08-01 | 无锡数字奥森科技有限公司 | Dual-certification face anti-counterfeit method and device |
CN104143086A (en) * | 2014-07-18 | 2014-11-12 | 吴建忠 | Application technology of portrait comparison to mobile terminal operating system |
WO2019100608A1 (en) * | 2017-11-21 | 2019-05-31 | 平安科技(深圳)有限公司 | Video capturing device, face recognition method, system, and computer-readable storage medium |
CN108280399A (en) * | 2017-12-27 | 2018-07-13 | 武汉普利商用机器有限公司 | A kind of scene adaptive face identification method |
CN109815826A (en) * | 2018-12-28 | 2019-05-28 | 新大陆数字技术股份有限公司 | The generation method and device of face character model |
CN110580445A (en) * | 2019-07-12 | 2019-12-17 | 西北工业大学 | An Improved Face Keypoint Detection Method Based on GIoU and Weighted NMS |
Also Published As
Publication number | Publication date |
---|---|
CN111222433A (en) | 2020-06-02 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN111222433B (en) | Automatic face auditing method, system, equipment and readable storage medium | |
WO2022161286A1 (en) | Image detection method, model training method, device, medium, and program product | |
US11830230B2 (en) | Living body detection method based on facial recognition, and electronic device and storage medium | |
KR102554724B1 (en) | Method for identifying an object in an image and mobile device for practicing the method | |
CN109952594B (en) | Image processing method, device, terminal and storage medium | |
US11403874B2 (en) | Virtual avatar generation method and apparatus for generating virtual avatar including user selected face property, and storage medium | |
CN106096582B (en) | Distinguish real and flat surfaces | |
WO2021196389A1 (en) | Facial action unit recognition method and apparatus, electronic device, and storage medium | |
CN108090511B (en) | Image classification method and device, electronic equipment and readable storage medium | |
CN110728209A (en) | Gesture recognition method and device, electronic equipment and storage medium | |
CN109871845B (en) | Certificate image extraction method and terminal equipment | |
CN109522883A (en) | A kind of method for detecting human face, system, device and storage medium | |
CN112396050B (en) | Image processing method, device and storage medium | |
CN115115552B (en) | Image correction model training method, image correction device and computer equipment | |
CN115019364A (en) | Identity authentication method and device based on face recognition, electronic equipment and medium | |
CN107369086A (en) | A kind of identity card stamp system and method | |
CN112613457B (en) | Image acquisition mode detection method, device, computer equipment and storage medium | |
Shukla et al. | Deep Learning Model to Identify Hide Images using CNN Algorithm | |
CN117496019B (en) | Image animation processing method and system for driving static image | |
CN114140839A (en) | Image sending method, device and equipment for face recognition and storage medium | |
CN113434912B (en) | Material compliance verification method and device | |
CN112822393B (en) | Image processing method and device and electronic equipment | |
CN108764033A (en) | Auth method and device, electronic equipment, computer program and storage medium | |
CN113221606A (en) | Face recognition method based on IMS video conference login | |
CN112132107A (en) | Image processing method, image processing device, electronic equipment and computer readable storage medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |