[go: up one dir, main page]

CN107066943B - A face detection method and device - Google Patents

A face detection method and device Download PDF

Info

Publication number
CN107066943B
CN107066943B CN201710127367.3A CN201710127367A CN107066943B CN 107066943 B CN107066943 B CN 107066943B CN 201710127367 A CN201710127367 A CN 201710127367A CN 107066943 B CN107066943 B CN 107066943B
Authority
CN
China
Prior art keywords
face
feature
candidate
masked
dictionary
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201710127367.3A
Other languages
Chinese (zh)
Other versions
CN107066943A (en
Inventor
葛仕明
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Institute of Information Engineering of CAS
Original Assignee
Institute of Information Engineering of CAS
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Institute of Information Engineering of CAS filed Critical Institute of Information Engineering of CAS
Priority to CN201710127367.3A priority Critical patent/CN107066943B/en
Publication of CN107066943A publication Critical patent/CN107066943A/en
Application granted granted Critical
Publication of CN107066943B publication Critical patent/CN107066943B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/161Detection; Localisation; Normalisation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/213Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Evolutionary Computation (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Evolutionary Biology (AREA)
  • General Health & Medical Sciences (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Human Computer Interaction (AREA)
  • Multimedia (AREA)
  • Collating Specific Patterns (AREA)
  • Image Analysis (AREA)

Abstract

本发明公开了一种人脸检测方法及装置。本发明的方法为:1)从待处理的图像中检测出候选人脸,并提取该候选人脸的候选特征;2)将每一所述候选特征在预先构建的传统外部特征空间或近似外部特征空间进行投影变换,得到对应的传统或近似嵌入特征;其中,该近似外部特征空间是从参考人脸特征字典和非人脸特征字典中选择具有代表性的特征组成的字典;3)对所述嵌入特征进行验证,判别出所述嵌入特征对应的候选人脸是否为人脸。本发明的人脸检测装置包括候选模块、嵌入模块和验证模块。本发明能够得到精度更高的人脸检测性能;对有遮挡情况下,也具备良好的人脸检测能力。

The invention discloses a face detection method and device. The method of the present invention is: 1) detect the candidate face from the image to be processed, and extract the candidate features of the candidate face; The feature space is projected and transformed to obtain the corresponding traditional or approximate embedded features; wherein, the approximate external feature space is a dictionary composed of representative features selected from the reference face feature dictionary and non-face feature dictionary; 3) for all The embedding feature is verified to determine whether the candidate face corresponding to the embedding feature is a human face. The face detection device of the present invention includes a candidate module, an embedding module and a verification module. The present invention can obtain higher precision human face detection performance, and also has good human face detection ability in the case of occlusion.

Description

一种人脸检测方法及装置A face detection method and device

技术领域technical field

本发明属于计算机视觉和深度学习领域,尤其涉及一种针对遮挡条件下的人脸检测方法及装置。The invention belongs to the field of computer vision and deep learning, and in particular relates to a face detection method and device under occlusion conditions.

背景技术Background technique

人脸检测技术可应用于相机自动对焦、人机交互、照片管理、城市安防监控、智能驾驶等诸多领域。当前,人脸检测在开放环境条件下的实际应用中,由于遮挡的普遍存在(如人群密集情况下),人脸检测性能受到严重的挑战,因而遮挡条件下的人脸检测性能问题有待解决。另外,研究蒙面遮挡条件下的人脸检测具有重要的现实意义,例如:视频监控中用于发现可疑人员从而提供警告、通过检测蒙面人脸的分布规律进行天气状况预测等。传统的人脸检测方法在遮挡情况下遭遇严重的性能下降,原因在于检测过程中,被遮挡部分的人脸线索已经无效,从而造成在特征提取过程中不可避免地引入噪声。总之,不完整且不准确的特征使蒙面遮挡的人脸检测成为一个极具挑战的问题。Face detection technology can be applied to camera autofocus, human-computer interaction, photo management, urban security monitoring, intelligent driving and many other fields. At present, in the practical application of face detection in open environment conditions, due to the ubiquity of occlusion (such as dense crowds), the performance of face detection is seriously challenged, so the performance of face detection under occlusion conditions needs to be solved. In addition, it is of great practical significance to study face detection under masked conditions, for example: in video surveillance, it is used to find suspicious persons to provide warnings, and to predict weather conditions by detecting the distribution of masked faces. Traditional face detection methods suffer from severe performance degradation under occlusion, because the face clues of the occluded part are invalid during the detection process, which inevitably introduces noise in the feature extraction process. In conclusion, incomplete and inaccurate features make masked face detection an extremely challenging problem.

近几年来,在这一领域也研究了一些方法,现有技术是先检测出人脸候选,然后再对人脸候选分类确认。其中一种方法通过训练多个神经网络获得人脸多个部件的响应来检测人脸候选,然后再训练一个新的神经网络进行人脸候选的分类确认(参见:S.Yang,P.Luo,C.C.Loy,and X.Tang.From facial parts responses to face detection:A deeplearning approach.In:IEEE ICCV,2015)。另一种方法则通过选择部分特征比较来计算损失从而确认人脸候选(参见:M.Opitz,G.Waltner,G.Poier,H.Possegger,andH.Bischo.Grid loss:Detecting occluded faces.In ECCV,2016),该方法能够较好地处理部分遮挡情况的人脸检测问题。上述方法在一定程度上缓解了严重遮挡(如蒙面遮挡)情况下的人脸检测问题,但仍未能获得完全解决。当人脸部件被遮挡,通过多个部件响应来检测人脸候选的人脸检测方法,在遮挡区域的部件会引入噪声或错误,从而可能导致人脸分类确认错误;当遮挡严重的时候,通过选择部分特征比较计算损失确认人脸候选的人脸检测方法,计算得到的损失误差较大,从而导致人脸检测失败。In recent years, some methods have also been studied in this field. The existing technology is to detect face candidates first, and then classify and confirm the face candidates. One of the methods detects face candidates by training multiple neural networks to obtain the responses of multiple parts of the face, and then trains a new neural network to classify and confirm face candidates (see: S.Yang, P.Luo, C. C. Loy, and X. Tang. From facial parts responses to face detection: A deep learning approach. In: IEEE ICCV, 2015). Another method is to confirm the face candidate by selecting some feature comparisons to calculate the loss (see: M.Opitz, G.Waltner, G.Poier, H.Possegger, and H.Bischo. Grid loss: Detecting occluded faces.In ECCV , 2016), this method can better deal with the face detection problem of partial occlusion. The above method alleviates the problem of face detection in the case of severe occlusion (such as mask occlusion) to a certain extent, but it still cannot be completely solved. When the face parts are occluded, the face detection method that detects face candidates through the response of multiple parts will introduce noise or errors in the parts in the occluded area, which may lead to errors in face classification confirmation; when the occlusion is serious, The face detection method that confirms the face candidate by selecting some features and comparing the calculation loss to confirm the face candidate, the calculated loss has a large error, which leads to the failure of face detection.

发明内容Contents of the invention

为克服现有技术的不足,本发明提供了一种人脸检测方法及装置,该方法通过卷积神经网络检测候选人脸和提取高维深度特征(即候选特征),然后通过局部线性嵌入进行特征投影来消除蒙面遮挡带来的特征不完全和不精确,再采用多任务卷积神经网络(即CNN-V)验证候选人脸,从而获得更加精确的人脸检测性能。同时,本发明还提出了一种近似外部特征空间的构造方法,通过从外部的数据库中寻找最相似的参考人脸和差异最大的参考非人脸,进行近似外部特征空间构造,利用近似外部特征空间对候选特征进行嵌入变换,从而修正候选特征。本发明通过以下技术方案来实现。In order to overcome the deficiencies in the prior art, the present invention provides a face detection method and device, which detects candidate faces and extracts high-dimensional depth features (i.e. candidate features) through convolutional neural networks, and then performs local linear embedding. Feature projection is used to eliminate the incomplete and inaccurate features caused by mask occlusion, and then the multi-task convolutional neural network (ie CNN-V) is used to verify the candidate face, so as to obtain more accurate face detection performance. At the same time, the present invention also proposes a method for constructing an approximate external feature space, by searching for the most similar reference face and the most different reference non-face from an external database, and constructing an approximate external feature space, using the approximate external feature Embedding transforms the candidate features in space to modify the candidate features. The present invention is realized through the following technical solutions.

本发明的一种人脸检测方法,其步骤包括:A kind of human face detection method of the present invention, its step comprises:

1)对待检测图像进行候选人脸检测,得到候选人脸图像;1) Perform candidate face detection on the image to be detected to obtain a candidate face image;

2)对所述候选人脸图像进行候选特征提取,得到候选特征;2) Carry out candidate feature extraction to described candidate face image, obtain candidate feature;

3)对所述候选特征进行嵌入变换,得到传统嵌入特征或近似嵌入特征,所述嵌入特征能够恢复人脸线索并移除遮挡带来的噪声;3) performing embedding transformation on the candidate features to obtain traditional embedding features or approximate embedding features, the embedding features can restore face clues and remove noise caused by occlusion;

4)对所述传统嵌入特征或近似嵌入特征,通过分类与回归算法进行验证,得到检测结果。4) The traditional embedded features or approximate embedded features are verified through classification and regression algorithms to obtain detection results.

进一步的,候选特征通过一个预先构建好的外部特征空间进行嵌入变换后,得到传统嵌入特征或近似嵌入特征;外部特征空间为传统外部特征空间或近似外部特征空间。Further, the candidate features are embedded and transformed through a pre-built external feature space to obtain traditional embedded features or approximate embedded features; the external feature space is traditional external feature space or approximate external feature space.

进一步的,嵌入变换采用传统的局部线性嵌入方法或快速近似局部线性嵌入方法实现;传统的局部线性嵌入方法利用传统外部特征空间对带噪声的候选特征进行嵌入变换,得到传统嵌入特征;快速近似局部线性嵌入利用近似外部特征空间对带噪声的候选特征进行嵌入变换,得到近似嵌入特征。Further, the embedding transformation is realized by the traditional local linear embedding method or the fast approximate local linear embedding method; the traditional local linear embedding method uses the traditional external feature space to perform embedding transformation on the candidate features with noise to obtain the traditional embedded features; the fast approximate local linear embedding method Linear embedding uses the approximate external feature space to perform embedding transformation on the candidate features with noise to obtain approximate embedded features.

进一步的,快速近似局部线性嵌入方法中近似外部特征空间的构造方法,包括以下步骤:Further, the method for constructing an approximate external feature space in the fast approximate local linear embedding method includes the following steps:

a)对标注好的参考人脸数据集进行候选人脸检测及候选特征提取,判断候选特征属于人脸特征还是非人脸特征,将这些候选特征分别存入参考人脸特征字典和参考非人脸特征字典;a) Perform candidate face detection and candidate feature extraction on the marked reference face dataset, judge whether the candidate features belong to human face features or non-human face features, and store these candidate features in the reference face feature dictionary and reference non-human feature dictionary respectively face feature dictionary;

b)对标注好的蒙面人脸数据集进行候选人脸检测及候选特征提取,判断候选特征属于蒙面人脸特征还是蒙面非人脸特征,将这些候选特征分别存入蒙面人脸特征字典和蒙面非人脸特征字典;b) Perform candidate face detection and candidate feature extraction on the marked masked face dataset, determine whether the candidate features belong to masked face features or masked non-face features, and store these candidate features into masked faces Feature dictionary and masked non-face feature dictionary;

c)从上述参考人脸特征字典中选择具有代表性的能够代表上述蒙面人脸特征字典的参考人脸特征字典;c) select a representative reference face feature dictionary that can represent the above-mentioned masked face feature dictionary from the above-mentioned reference face feature dictionary;

d)从上述参考非人脸特征字典中选择具有代表性的能够代表上述蒙面非人脸特征字典的参考非人脸特征字典;D) select a representative reference non-face feature dictionary that can represent the above-mentioned masked non-face feature dictionary from the above-mentioned reference non-face feature dictionary;

e)合并上述具有代表性的参考人脸特征字典和具有代表性的参考非人脸特征字典,得到近似外部特征空间。e) Merge the above-mentioned representative reference face feature dictionary and representative reference non-face feature dictionary to obtain an approximate external feature space.

进一步的,步骤a)中,通过计算该候选特征对应的候选人脸位置与标注好的人脸位置之间的重叠度来确定,其重叠度用交并比来度量,其中,交并比大于0.7则判断候选特征为参考人脸的特征,交并比小于0.3则判断候选特征为参考非人脸的特征。Further, in step a), it is determined by calculating the degree of overlap between the candidate face position corresponding to the candidate feature and the marked face position, and the degree of overlap is measured by the intersection ratio, wherein the intersection ratio is greater than If it is 0.7, it is judged that the candidate feature is the feature of the reference face, and if the intersection ratio is less than 0.3, it is judged that the candidate feature is the feature of the reference non-face.

进一步的,步骤b)中,通过计算该候选特征对应的候选人脸位置与标注好的人脸位置之间的重叠度来确定,其重叠度用交并比来度量,其中,交并比大于0.6则判断候选特征为蒙面人脸的特征,交并比小于0.4则判断候选特征为蒙面非人脸的特征。Further, in step b), it is determined by calculating the degree of overlap between the candidate face position corresponding to the candidate feature and the marked face position, and the degree of overlap is measured by the intersection ratio, wherein the intersection ratio is greater than If it is 0.6, it is judged that the candidate feature is a masked face feature, and if the intersection ratio is less than 0.4, it is judged that the candidate feature is a masked non-face feature.

进一步的,步骤c)中采用贪婪算法从参考人脸特征字典中选择具有代表性的参考人脸特征字典;所述贪婪算法是指计算参考人脸特征字典中每个参考人脸特征的损失,得到按损失由小到大升序排列的参考人脸特征列表,取该列表最前面的参考人脸特征来代表蒙面人脸特征;其中所述损失是指每个参考人脸特征与蒙面人脸特征字典的最近邻特征的距离和每个参考人脸特征与蒙面非人脸特征字典的最近邻特征的距离之差。Further, in step c), adopt greedy algorithm to select representative reference face feature dictionary from reference face feature dictionary; Described greedy algorithm refers to calculating the loss of each reference face feature in the reference face feature dictionary, Obtain the list of reference human face features arranged in ascending order by loss, and get the reference human face feature at the top of the list to represent the masked human face feature; The distance between the nearest neighbor feature of the face feature dictionary and the distance between each reference face feature and the nearest neighbor feature of the masked non-face feature dictionary.

进一步的,步骤d)中采用贪婪算法从参考非人脸特征字典中选择具有代表性的参考非人脸特征字典;所述贪婪算法是指计算参考非人脸特征字典中每个参考非人脸特征的损失,得到按损失由小到大升序排列的参考非人脸特征列表,取该列表最前面的参考非人脸特征来代表蒙面非人脸特征;其中所述损失是指每个参考非人脸特征与蒙面非人脸特征字典的最近邻特征的距离和每个参考非人脸特征与蒙面人脸特征字典的最近邻特征的距离之差。Further, in step d), adopt greedy algorithm to select representative reference non-face feature dictionary from reference non-face feature dictionary; The loss of features, get the list of reference non-face features arranged in ascending order according to the loss, take the reference non-face features at the top of the list to represent masked non-face features; wherein the loss refers to each reference The difference between the distance of the non-face feature to the nearest neighbor feature of the masked non-face feature dictionary and the distance of each reference non-face feature to the nearest neighbor feature of the masked face feature dictionary.

本发明还涉及一种人脸检测装置,包括候选模块、嵌入模块和验证模块。候选模块用于对待检测图像进行候选人脸检测并提取候选特征;嵌入模块用于对候选特征进行嵌入变换,得到传统嵌入特征或近似嵌入特征,嵌入特征能够恢复人脸线索并移除遮挡带来的噪声;验证模块用于对传统嵌入特征或近似嵌入特征,通过分类与回归算法进行验证,以得到最后的检测结果。候选模块得到多个候选特征,然后在嵌入模块中通过一个预先构建好的外部特征空间进行嵌入变换后,得到传统嵌入特征或近似嵌入特征;外部特征空间为传统外部特征空间或近似外部特征空间;嵌入变换采用传统的局部线性嵌入方法或快速近似局部线性嵌入方法实现。The invention also relates to a face detection device, which includes a candidate module, an embedding module and a verification module. The candidate module is used to detect candidate faces of the image to be detected and extract candidate features; the embedding module is used to embed and transform the candidate features to obtain traditional embedded features or approximate embedded features. The embedded features can restore face clues and remove occlusions. The noise; the verification module is used to verify the traditional embedded features or approximate embedded features through classification and regression algorithms to obtain the final detection results. The candidate module obtains multiple candidate features, and then performs embedding transformation through a pre-built external feature space in the embedding module to obtain traditional embedded features or approximate embedded features; the external feature space is traditional external feature space or approximate external feature space; The embedding transformation is realized by traditional local linear embedding method or fast approximate local linear embedding method.

本发明的有益效果在于:The beneficial effects of the present invention are:

针对遮挡条件下的人脸检测问题,尤其是严重蒙面遮挡条件下的人脸检测问题,本发明的检测方法及装置具有相对较好的性能;对无遮挡情况下的人脸,本发明的人脸检测方法及装置也具备良好的处理能力。Aiming at the face detection problem under occlusion conditions, especially the face detection problem under severe mask occlusion conditions, the detection method and device of the present invention have relatively good performance; The face detection method and device also have good processing capabilities.

附图说明Description of drawings

图1为本发明一种人脸检测方法的流程图;Fig. 1 is the flowchart of a kind of face detection method of the present invention;

图2为本发明装置候选模块流程示意图;Fig. 2 is a schematic flow diagram of a candidate module of the device of the present invention;

图3为本发明装置嵌入模块流程示意图;Fig. 3 is a schematic flow chart of the embedding module of the device of the present invention;

图4为本发明装置验证模块流程示意图;Fig. 4 is a schematic flow chart of the verification module of the device of the present invention;

图5为本发明的近似外部特征空间构造流程示意图。Fig. 5 is a schematic diagram of the construction process of the approximate external feature space of the present invention.

具体实施方式Detailed ways

为使本发明的上述方案和有益效果更明显易懂,下文通过实施例,并配合附图作详细说明如下。In order to make the above solutions and beneficial effects of the present invention more comprehensible, the following will be described in detail through the examples and accompanying drawings.

本发明提供一种人脸检测方法及装置,该装置包括候选模块、嵌入模块和验证模块;该方法的流程图如图1所示,其步骤包括:The present invention provides a kind of human face detection method and device, and this device comprises candidate module, embedding module and verification module; The flowchart of this method is as shown in Figure 1, and its steps include:

1)接收图像。所述图像既可以是遮挡条件下的人脸图像或者严重蒙面遮挡条件下的人脸图像,也可以是无遮挡情况下的人脸图像,也可以是不包含人脸的图像。1) Receive an image. The image can be a face image under occlusion conditions or a face image under severe mask occlusion conditions, or a face image without occlusion, or an image without a face.

2)通过候选模块检测出候选人脸并提取候选人脸的高维深度特征,即候选特征。2) Detect the candidate face through the candidate module and extract the high-dimensional deep feature of the candidate face, that is, the candidate feature.

在候选模块中,先进行候选人脸检测,接着判断是否检测到候选人脸,如果未检测到候选人脸则结束;如果检测到候选人脸则进行候选特征提取,得到候选特征。In the candidate module, the candidate face detection is performed first, and then it is judged whether the candidate face is detected, and if the candidate face is not detected, it ends; if the candidate face is detected, the candidate feature extraction is performed to obtain the candidate feature.

请参考图2,所述候选模块主要包含两个卷积神经网络:一个是小的卷积神经网络(称为候选卷积神经网络,简称CNN-P),该网络用于实现候选人脸检测;另外一个大的卷积神经网络(称为特征卷积神经网络,简称CNN-F),用于实现候选特征提取。首先,接收到的图像通过候选卷积神经网络,进行候选人脸检测,接着判断是否检测到候选人脸,如果未检测到候选人脸,则结束;如果检测到候选人脸,则先进行候选人脸归一化处理,再通过特征卷积神经网络进行候选特征提取,得到候选特征。Please refer to Figure 2, the candidate module mainly includes two convolutional neural networks: one is a small convolutional neural network (called candidate convolutional neural network, referred to as CNN-P), which is used to implement candidate face detection ; Another large convolutional neural network (called feature convolutional neural network, referred to as CNN-F) is used to implement candidate feature extraction. First, the received image passes through the candidate convolutional neural network for candidate face detection, and then judges whether the candidate face is detected. If the candidate face is not detected, it ends; if the candidate face is detected, the candidate face is detected first. The face is normalized, and then the candidate features are extracted through the feature convolutional neural network to obtain the candidate features.

3)通过嵌入模块进行候选特征嵌入,得到嵌入变换后的特征,即传统嵌入特征或近似嵌入特征(统称为嵌入特征)。3) Embedding candidate features through the embedding module to obtain features after embedding transformation, that is, traditional embedding features or approximate embedding features (collectively referred to as embedding features).

由于蒙面遮挡会造成人脸线索缺失及特征噪声,从而导致特征不完整和不精确。针对该问题,本发明技术方案中的嵌入模块实现从候选特征中恢复人脸线索并移除噪声。嵌入模块处理的优点是获得的嵌入特征能够很好地表征蒙面遮挡人脸,从而能够提升检测精度。Mask occlusion will cause missing face clues and feature noise, resulting in incomplete and inaccurate features. To solve this problem, the embedding module in the technical solution of the present invention realizes recovering face clues and removing noise from candidate features. The advantage of embedding module processing is that the obtained embedding features can well represent the masked face, which can improve the detection accuracy.

请参考图3,在嵌入模块中,候选特征通过一个预先构建好的外部特征空间,进行嵌入变换后,得到传统嵌入特征或近似嵌入特征。所述嵌入变换主要采用LLE(LocalLinear Embedding)方法实现。LLE是一种针对非线性数据的降维方法,处理后的低维数据能够保持原有的拓扑关系,已经广泛应用于图像数据的分类与聚类、多维数据的可视化以及生物信息学等领域。本发明使用传统的LLE方法和快速近似LLE方法实现嵌入变换。Please refer to Figure 3. In the embedding module, the candidate features pass through a pre-built external feature space, and after embedding transformation, traditional embedded features or approximate embedded features are obtained. The embedding transformation is mainly implemented by the LLE (LocalLinear Embedding) method. LLE is a dimensionality reduction method for nonlinear data. The processed low-dimensional data can maintain the original topological relationship. It has been widely used in the classification and clustering of image data, the visualization of multidimensional data, and bioinformatics. The invention uses the traditional LLE method and the fast approximate LLE method to realize the embedded transformation.

4)通过验证模块,进行传统嵌入特征或近似嵌入特征验证,判断每个传统嵌入特征或近似嵌入特征对应的候选人脸是否属于真正人脸,如果该传统嵌入特征或近似嵌入特征对应的候选人脸属于真正人脸,则记录人脸信息;如果该传统嵌入特征或近似嵌入特征对应的候选人脸不属于真正人脸,则结束。4) Through the verification module, the traditional embedded feature or approximate embedded feature is verified, and it is judged whether the candidate face corresponding to each traditional embedded feature or approximate embedded feature belongs to a real face, if the candidate face corresponding to the traditional embedded feature or approximate embedded feature If the face belongs to the real face, record the face information; if the candidate face corresponding to the traditional embedded feature or the approximate embedded feature does not belong to the real face, then end.

请参考图4,验证模块由一个四层的全连接卷积神经网络(称为验证卷积神经网络,简称CNN-V)组成,以用于进行特征验证,即判别该传统嵌入特征或近似嵌入特征对应的候选人脸是否属于真正人脸并修正对应的候选人脸位置与尺度。如果不属于真正人脸,则忽略该传统嵌入特征或近似嵌入特征对应的候选人脸;如果属于真正人脸,则将该传统嵌入特征或近似嵌入特征对应的修正后的候选人脸位置与尺度加入检测结果中。Please refer to Figure 4. The verification module consists of a four-layer fully connected convolutional neural network (called a verification convolutional neural network, or CNN-V for short), which is used for feature verification, that is, to distinguish the traditional embedded features or approximate embedded features. Whether the candidate face corresponding to the feature belongs to the real face and correct the position and scale of the corresponding candidate face. If it does not belong to a real face, ignore the candidate face corresponding to the traditional embedded feature or the approximate embedded feature; if it belongs to the real face, then the corrected candidate face position and scale corresponding to the traditional embedded feature or the approximate embedded feature added to the test results.

通过验证模块对传统嵌入特征或近似嵌入特征进行分类与回归,从而判别出候选属于真正的人脸或非人脸,并对人脸位置与尺度进行修正,从而得到精度更高的人脸检测性能。Through the verification module, the traditional embedded features or approximate embedded features are classified and regressed, so as to determine whether the candidate is a real face or a non-face, and correct the position and scale of the face, so as to obtain a face detection performance with higher accuracy .

因此,本发明提出的一种人脸检测方法及装置联合了候选模块的候选卷积神经网络CNN-P、候选模块的特征卷积神经网络CNN-F、嵌入模块和验证模块的验证卷积神经网络CNN-V,来达到本发明的目的。Therefore, a face detection method and device proposed by the present invention combine the candidate convolutional neural network CNN-P of the candidate module, the feature convolutional neural network CNN-F of the candidate module, the verification convolutional neural network of the embedding module and the verification module. Network CNN-V, to achieve the purpose of the present invention.

下面具体描述嵌入模块的嵌入变换所采用的方法。The method adopted for the embedding transformation of the embedding module is described in detail below.

1、传统的LLE方法。1. The traditional LLE method.

请参考图3,通过传统的LLE方法,将蒙面遮挡的候选特征xi在预先构造好的传统外部特征空间中进行投影变换,得到嵌入特征vi,该嵌入特征vi可以有效消除由于蒙面遮挡带来的特征不完整和不精确问题,具有很好的抗遮挡能力。其中xi的下标i用于标记不同的候选特征;vi的下标i用于标记不同的嵌入特征。嵌入特征vi称为传统嵌入特征。Please refer to Figure 3. Through the traditional LLE method, the masked candidate features x i are projected and transformed in the pre-constructed traditional external feature space to obtain embedded features v i , which can effectively eliminate the The incomplete and inaccurate features caused by surface occlusion have good anti-occlusion capabilities. Among them, the subscript i of xi is used to mark different candidate features; the subscript i of v i is used to mark different embedded features. Embedded features v i are called traditional embedded features.

所述传统外部特征空间由参考人脸特征和参考非人脸特征组成,其表示成特征字典的形式,即D=[D + ,D-],这里D+是参考人脸特征字典,D-是参考非人脸特征字典,通常来说D+和D-规模都有上百万。The traditional external feature space is composed of reference face features and reference non-face features, which are expressed in the form of a feature dictionary, that is, D=[D + , D- ], where D + is a reference face feature dictionary, and D- It is a reference to the non-face feature dictionary. Generally speaking, the D + and D- scales have millions.

所述参考人脸特征和参考非人脸特征,通过构建参考候选特征集实现。具体地,对标注好的大型无遮挡的参考人脸数据集Sn,使用候选模块进行候选人脸检测及候选特征提取。判断候选特征属于人脸特征还是非人脸特征,将这些候选特征分成参考人脸特征和参考非人脸特征,分别存入参考人脸特征字典D+和参考非人脸特征字典D-。其中判断候选特征属于人脸特征还是非人脸特征,是通过计算该候选特征对应的候选人脸位置与标注好的人脸位置之间的重叠度来确定,其重叠度用交并比(Intersection-over-Union,IoU)来度量。通常传统方法中交并比大于0.5被判断为人脸,小于0.5则被判断为非人脸。与传统方法相比,本发明中交并比大于0.7被判断为参考人脸,交并比小于0.3被判断为参考非人脸,使得到的参考人脸与参考非人脸具有更好的区分性,可以保证参考候选特征具有更好的辨识能力。The reference face features and reference non-face features are realized by constructing a reference candidate feature set. Specifically, for the labeled large-scale unoccluded reference face dataset S n , the candidate module is used to detect candidate faces and extract candidate features. Determine whether the candidate features belong to face features or non-face features, divide these candidate features into reference face features and reference non-face features, and store them in the reference face feature dictionary D + and the reference non-face feature dictionary D respectively. Among them, judging whether the candidate feature is a face feature or a non-face feature is determined by calculating the degree of overlap between the candidate face position corresponding to the candidate feature and the marked face position, and the overlap degree is determined by the intersection ratio (Intersection -over-Union, IoU) to measure. Usually, in the traditional method, if the intersection ratio is greater than 0.5, it is judged as a human face, and if it is less than 0.5, it is judged as a non-human face. Compared with the traditional method, in the present invention, the intersection and union ratio greater than 0.7 is judged as a reference face, and the intersection and union ratio is less than 0.3 is judged as a reference non-human face, so that the obtained reference face and the reference non-human face have a better distinction It can ensure that the reference candidate features have better recognition ability.

对于每一个带噪声的候选特征xi,都从D+和D-中选择距离xi最邻近的特征集构成特征子字典Di(Di的下标i用于标记不同的候选特征对应的特征子字典),然后利用LLE算法进行投影变换,获得一个新的特征表达即传统嵌入特征vi,该过程的求解公式如下:For each noisy candidate feature x i , select the feature set closest to x i from D + and D - to form a feature subdictionary D i (the subscript i of D i is used to mark the corresponding feature sub-dictionary), and then use the LLE algorithm for projection transformation to obtain a new feature expression, that is, the traditional embedded feature v i , the solution formula of this process is as follows:

满足vi≥0(1) satisfy v i ≥ 0(1)

2、快速近似LLE方法。2. Fast approximate LLE method.

本发明提出一种快速近似LLE方法,对于每一个带噪声的候选特征xi,利用快速近似LLE方法进行投影变换,获得一个近似嵌入特征该方法求解公式如下:The present invention proposes a fast approximate LLE method. For each noisy candidate feature x i , use the fast approximate LLE method to perform projection transformation to obtain an approximate embedded feature The solution formula of this method is as follows:

满足 Satisfy

上述公式(2)中,是近似外部特征空间,该空间是从参考人脸特征字典D+和非人脸特征字典D-中选择具有代表性的特征组成的字典。对每个候选特征xi不再需要构造其对应的特征子字典Di,每个候选特征xi都使用固定的近似外部特征空间进行投影变换,得到近似嵌入特征 In the above formula (2), is an approximate external feature space, which is a dictionary composed of representative features selected from the reference face feature dictionary D + and non-face feature dictionary D . For each candidate feature xi, it is no longer necessary to construct its corresponding feature sub-dictionary D i , and each candidate feature xi uses a fixed approximate external feature space Perform projection transformation to obtain approximate embedded features

下面具体描述快速近似LLE方法中近似外部特征空间的构造。The construction of the approximate external feature space in the fast approximate LLE method is described in detail below.

所述近似外部特征空间的构造方法是通过从外部的数据库中寻找最相似的参考人脸或差异最大的参考非人脸,进行近似外部特征空间构造。The construction method of the approximate external feature space is to construct the approximate external feature space by finding the most similar reference human face or the most different reference non-human face from an external database.

请参考图5,该图是近似外部特征空间构造的流程图,是从D+和D-中选择最具代表性的特征组成,其包括具有代表性的参考人脸特征字典和具有代表性的参考非人脸特征字典表示为本发明提出的近似外部特征空间构造方法具体分成以下几步:Please refer to Figure 5, which is an approximate external feature space Constructed flow chart, is to select the most representative feature composition from D + and D- , which includes a representative reference face feature dictionary and a representative reference non-face feature dictionary Expressed as The approximate external feature space proposed by the present invention The construction method is divided into the following steps:

1)构建参考人脸和参考非人脸特征字典:其与上述传统的LLE方法中相同,对标注好的大型无遮挡的参考人脸数据集Sn,使用候选模块进行候选人脸检测及候选特征提取。根据候选特征属于人脸特征还是非人脸特征,将这些候选特征分别存入参考人脸特征字典D+和参考非人脸特征字典D-。判断候选特征属于人脸特征还是非人脸特征,是通过计算该候选特征对应的候选人脸位置与标注好的人脸位置之间的重叠度来确定,其重叠度用交并比IoU来度量。通常传统方法中交并比大于0.5被判断为人脸,小于0.5则被判断为非人脸。与传统方法相比,本发明中交并比大于0.7被判断为参考人脸,交并比小于0.3被判断为参考非人脸,使得到的参考人脸与参考非人脸具有更好的区分性,可以保证参考候选特征具有更好的辨识能力。1) Construct reference face and reference non-face feature dictionaries: it is the same as in the above-mentioned traditional LLE method. For the marked large-scale unoccluded reference face dataset S n , use the candidate module to detect candidate faces and candidate faces. feature extraction. According to whether the candidate features belong to face features or non-face features, these candidate features are respectively stored in the reference face feature dictionary D + and the reference non-face feature dictionary D . Judging whether a candidate feature belongs to a face feature or a non-face feature is determined by calculating the degree of overlap between the candidate face position corresponding to the candidate feature and the marked face position, and the degree of overlap is measured by the intersection-over-union ratio (IoU) . Usually, in the traditional method, if the intersection ratio is greater than 0.5, it is judged as a human face, and if it is less than 0.5, it is judged as a non-human face. Compared with the traditional method, in the present invention, the intersection and union ratio greater than 0.7 is judged as a reference face, and the intersection and union ratio is less than 0.3 is judged as a reference non-human face, so that the obtained reference face and the reference non-human face have a better distinction It can ensure that the reference candidate features have better recognition ability.

2)构建蒙面人脸和蒙面非人脸特征字典:类似上述步骤1),对标注好的大型蒙面人脸数据集Sm,使用候选模块进行候选人脸检测及候选特征提取。根据候选特征属于蒙面人脸特征还是蒙面非人脸特征,将这些候选特征分成蒙面人脸特征字典和蒙面非人脸特征字典由于蒙面人脸检测的定位精度通常会低于无遮挡的人脸检测,本发明中交并比大于0.6被判断为蒙面人脸,交并比小于0.4被判断为蒙面非人脸,以选择质量更好的蒙面人脸候选特征。2) Construct masked face and masked non-face feature dictionaries: similar to the above step 1), for the marked large-scale masked face dataset S m , use the candidate module to detect candidate faces and extract candidate features. According to whether the candidate features belong to masked face features or masked non-face features, these candidate features are divided into masked face feature dictionary and a dictionary of masked non-face features Since the positioning accuracy of masked face detection is usually lower than that of unoccluded face detection, in the present invention, an intersection ratio greater than 0.6 is judged as a masked face, and an intersection ratio less than 0.4 is judged as a masked non-human face. To select masked face candidate features with better quality.

3)选择具有代表性的参考人脸特征字典 从参考人脸特征字典D+中选择,是D+的一个子集即 的代表性表明它在代表蒙面人脸时具有好的表征能力同时在代表蒙面非人脸时具有区分能力。从而,在稀疏地代表蒙面人脸特征字典时应有最小的错误,同时在稀疏地代表蒙面非人脸特征字典应有最大的错误。因此,通过求解下列公式(3)得到:3) Select a representative reference face feature dictionary Select from the reference face feature dictionary D + , which is a subset of D + namely The representativeness of σ indicates that it has good representational power when representing masked faces and discriminative power when representing masked non-human faces. thereby, In a sparsely represented masked face feature dictionary should have minimal error while sparsely representing masked non-face feature dictionaries There should be maximum error. therefore, By solving the following formula (3), we get:

满足 Satisfy

上述公式(3)属于稀疏编码处理,公式中α1和α2分别是利用代表某个蒙面人脸特征x1和某个蒙面非人脸特征x2需要的稀疏系数向量。稀疏系数向量中仅有一个元素是1,其它元素是0。利用稀疏系数向量的约束条件,稀疏编码处理等价于从中寻找最近邻。由于中的各个特征来自于参考人脸特征字典D+,公式(3)的优化问题与经典的稀疏编码方式不一样,用经典的优化算法难以进行求解。所以,本发明提出一种贪婪方法有效地从参考人脸特征字典D+中构建在提出的贪婪方法中,本发明首先计算参考人脸特征字典D+中每个参考人脸特征的损失该损失表示为与蒙面人脸特征字典的最近邻特征的距离和与蒙面非人脸特征字典的最近邻特征的距离之差,其通过以下公式(4)实现:The above formula (3) belongs to sparse coding processing, and α 1 and α 2 in the formula are respectively Represents the sparse coefficient vector required for a certain masked face feature x1 and a certain masked non-face feature x2 . Only one element in the sparse coefficient vector is 1, and the other elements are 0. Using the constraints of sparse coefficient vectors, the sparse coding process is equivalent to starting from Find the nearest neighbor in . because Each feature in comes from the reference face feature dictionary D + . The optimization problem of formula (3) is different from the classical sparse coding method, and it is difficult to solve it with the classic optimization algorithm. Therefore, the present invention proposes a greedy method to construct effectively from the reference face feature dictionary D + In the proposed greedy method, the present invention firstly calculates each reference face feature in the reference face feature dictionary D + Loss This loss is expressed as Dictionary with masked face features The distance of the nearest neighbor feature and Dictionary with masked non-face features The distance difference of the nearest neighbor feature of , which is realized by the following formula (4):

满足 Satisfy

上述公式(4)中,ρ1和ρ2是两个平衡系数,用于平衡特征之间的距离,实际处理中通常取1以加速计算,每个参考人脸特征很少被用于代表中的蒙面人脸特征和中的蒙面非人脸特征。通过计算损失获得按照损失由小到大升序排列的参考人脸特征列表,列表中排在最前面的参考人脸特征在代表蒙面人脸特征方面的能力最强,而代表蒙面非人脸特征的能力最弱。采用这种方式,能够通过迭代的方式,不断地将列表前M个参考人脸特征加入到一个特征池P+中,构造出最终的优选的,M大于等于1且小于等于50。具体地,令初始特征池为空即然后在第t步采用来选择前M个候选,得到接着,中的特征用于更新然后用于求解公式(3)中的目标函数。In the above formula (4), ρ 1 and ρ 2 are two balance coefficients, which are used to balance the distance between features. In actual processing, 1 is usually used to speed up the calculation. Each reference face feature rarely used to represent Masked face features in and Masked non-face features in . Calculate the loss by Obtain a list of reference face features arranged in ascending order of loss. The reference face features at the top of the list have the strongest ability to represent masked face features, while the ability to represent masked non-face features the weakest. In this way, the top M reference face features of the list can be continuously added to a feature pool P + in an iterative manner to construct the final Preferably, M is greater than or equal to 1 and less than or equal to 50. Specifically, let the initial feature pool be empty or Then at step t use To select the top M candidates, get then, The features in are used to update Then used to solve the objective function in formula (3).

4)选择具有代表性的参考非人脸特征字典 从参考非人脸特征字典D-中选择,是D-的一个子集即的代表性表明它在代表蒙面非人脸时具有好的表征能力同时在代表蒙面人脸时具有区分能力。从而,在稀疏地代表蒙面非人脸特征字典时应有最小的错误,同时在稀疏地代表蒙面人脸特征字典时应有最大的错误。因此,能够通过求解下列公式(5)得到:4) Select a representative reference non-face feature dictionary Select from the reference non - face feature dictionary D-, which is a subset of D- The representativeness of σ indicates that it has good representational power when representing masked non-human faces and discriminative power when representing masked human faces. thereby, Sparsely represent masked non-face feature dictionaries should have minimal error while sparsely representing masked face feature dictionaries should have the greatest error. therefore, can be obtained by solving the following formula (5):

满足 Satisfy

上述公式(5)属于稀疏编码处理,公式中α1和α2分别是利用代表某个蒙面人脸特征x1和某个蒙面非人脸特征x2需要的稀疏系数向量。稀疏系数向量中仅有一个元素是1,其它元素是0。利用稀疏系数向量的约束条件,稀疏编码处理等价于从中寻找最近邻。由于中的各个特征来自于参考非人脸特征字典D-,公式(5)的优化问题与经典的稀疏编码方式不一样,用经典的优化算法难以进行求解。所以,本发明提出一种贪婪方法有效地从参考非人脸特征字典D-中构建在提出的贪婪方法中,本发明首先计算参考非人脸特征字典D-中每个参考非人脸特征的损失该损失表示为与蒙面非人脸特征字典的最近邻特征的距离和与蒙面人脸特征字典的最近邻特征的距离之差,通过以下公式(6)实现:满足 The above formula (5) belongs to sparse coding processing, and α 1 and α 2 in the formula are respectively Represents the sparse coefficient vector required for a certain masked face feature x1 and a certain masked non-face feature x2 . Only one element in the sparse coefficient vector is 1, and the other elements are 0. Using the constraints of sparse coefficient vectors, the sparse coding process is equivalent to starting from Find the nearest neighbor in . because Each feature in comes from the reference non-face feature dictionary D - , the optimization problem of formula (5) is different from the classical sparse coding method, and it is difficult to solve it with the classical optimization algorithm. Therefore, the present invention proposes a greedy method to construct effectively from the reference non - face feature dictionary D- In the proposed greedy method, the present invention first calculates each reference non-face feature in the reference non - face feature dictionary D- Loss This loss is expressed as Dictionary with masked non-face features The distance of the nearest neighbor feature and Dictionary with masked face features The distance difference between the nearest neighbor features of is realized by the following formula (6): Satisfy

上述公式(6)中,ρ1和ρ2是两个平衡系数,用于平衡特征之间的距离,实际处理中通常取1以加速计算,每个参考非人脸特征很少被用于代表中的蒙面人脸特征和中的蒙面非人脸特征。通过计算损失获得按照损失由小到大升序排列的参考非人脸特征列表,列表中排在最前面的参考非人脸特征在代表蒙面非人脸特征方面的能力最强,而代表蒙面人脸特征的能力最弱。采用这种方式,能够通过迭代的方式,不断地将列表前M个参考非人脸特征加入到一个特征池P-中,构造出最终的优选的,M大于等于1且小于等于50。具体地,令初始特征池为空即然后在第t步采用来选择前M个候选,得到接着,中的特征用于更新然后用于求解公式(5)中的目标函数。In the above formula (6), ρ 1 and ρ 2 are two balance coefficients, which are used to balance the distance between features. In actual processing, they usually take 1 to speed up the calculation. Each reference non-face feature rarely used to represent Masked face features in and Masked non-face features in . Calculate the loss by Obtain a list of reference non-face features arranged in ascending order of loss. The reference non-face features at the top of the list have the strongest ability to represent masked non-face features, while the representative masked face features the weakest ability. In this way, the top M reference non-face features of the list can be continuously added to a feature pool P - in an iterative manner to construct the final Preferably, M is greater than or equal to 1 and less than or equal to 50. Specifically, let the initial feature pool be empty or Then at step t use To select the top M candidates, get then, The features in are used to update Then used to solve the objective function in formula (5).

5)合并字典,得到近似外部特征空间 5) Merge dictionaries to obtain approximate external feature space

上述步骤中,步骤1)和2)没有严格先后顺序,可以先后或并行进行;步骤3)和4)没有严格先后顺序,可以先后或并行进行。通过上述步骤,构造出近似外部特征空间该近似外部特征空间是从大量参考人脸特征和参考非人脸特征中选择最有代表性的特征组成,其选择策略是通过与大量蒙面人脸特征和蒙面非人脸特征进行比较得到,包含的特征能够很好地代表蒙面人脸特征同时也能区分蒙面非人脸特征,因此利用近似外部特征空间对候选特征进行嵌入投影得到的嵌入特征对蒙面人脸具有很好的表征能力。另一方面,与传统的LLE方法相比,本发明提出的快速近似LLE方法构造的近似外部特征空间比每个候选特征xi对应的局部特征空间Di要大,对每个候选特征xi,进行投影变换后,得到的近似嵌入特征比传统的LLE方法得到的传统嵌入特征vi维度要高,一定程度上弥补了快速近似带来的特征表征损失,所以本发明提出的快速近似LLE方法构造的近似外部特征空间用于蒙面人脸检测中,对检测精度几乎没有影响。Among the above steps, steps 1) and 2) are not in strict order and can be performed sequentially or in parallel; steps 3) and 4) are not in strict order and can be performed successively or in parallel. Through the above steps, an approximate external feature space is constructed The approximate external feature space is composed of selecting the most representative features from a large number of reference face features and reference non-face features. The selection strategy is obtained by comparing with a large number of masked face features and masked non-face features. , the included features can well represent masked face features and can also distinguish masked non-face features, so using the approximate external feature space The embedded features obtained by embedding projection of candidate features have good representation ability for masked faces. On the other hand, compared with the traditional LLE method, the approximate external feature space constructed by the fast approximate LLE method proposed by the present invention is larger than the local feature space D i corresponding to each candidate feature x i , for each candidate feature x i , after projective transformation, the obtained approximate embedded features Compared with the traditional LLE method, the dimension of the traditional embedding feature v i is higher, and to a certain extent, it compensates for the loss of feature representation caused by the fast approximation. Therefore, the approximate external feature space constructed by the fast approximate LLE method proposed by the present invention It is used in masked face detection and has almost no impact on detection accuracy.

通过比较近似外部特征空间对应的具有代表性的参考人脸图像和具有代表性的参考非人脸图像的示例,可以发现,选择的具有代表性的参考人脸图像包含不同外观、佩戴、肤色、表情等,因此能够很好地代表蒙面人脸并同时很好地区分蒙面非人脸;选择的具有代表性的参考非人脸图像则是纹理区域、不完整人脸、含较多背景的人脸,因此能够很好地代表蒙面非人脸并同时很好地区分蒙面人脸。By comparing examples of representative reference face images and representative reference non-face images corresponding to the approximate external feature space, it can be found that the selected representative reference face images contain different appearance, wearing, skin color, Expressions, etc., so it can well represent masked faces and at the same time distinguish masked non-faces well; the selected representative reference non-face images are textured areas, incomplete faces, and more backgrounds Therefore, it can well represent masked non-human faces and distinguish masked faces well at the same time.

以上实施仅用以说明本发明的技术方案而非对其进行限制,本领域的普通技术人员可以对本发明的技术方案进行修改或者等同替换,而不脱离本发明的精神和范围,本发明的保护范围应以权利要求书所述为准。The above implementation is only used to illustrate the technical solution of the present invention and not to limit it. Those skilled in the art can modify or equivalently replace the technical solution of the present invention without departing from the spirit and scope of the present invention. Protection of the present invention The scope should be defined by the claims.

Claims (7)

1. a kind of method for detecting human face, step include:
1) candidate face detection is carried out to image to be detected, obtains candidate face image;
2) candidate feature extraction is carried out to the candidate face image, obtains candidate feature;
3) insertion transformation is carried out to the candidate feature, obtains tradition insertion feature or approximate insertion feature;
4) it to the tradition insertion feature or approximate insertion feature, is verified by classifying with regression algorithm, obtains detection knot Fruit;
Wherein, in step 3), after the candidate feature carries out insertion transformation by a surface space built in advance, Obtain tradition insertion feature or approximate insertion feature;The surface space is that conventional external feature space or proximate exterior are special Insertion transformation described in sign space is locally linear embedding into method or quick approximation is locally linear embedding into method realization using traditional;It passes The method that is locally linear embedding into of system carries out insertion transformation to the candidate feature with noise using conventional external feature space, is passed System insertion feature;Quick approximation, which is locally linear embedding into, is embedded in the candidate feature with noise using proximate exterior feature space Transformation obtains approximate insertion feature;
The quick approximation is locally linear embedding into the building method of proximate exterior feature space in method, comprising the following steps:
A) candidate face detection is carried out to the reference face data set marked and candidate feature is extracted, judge that candidate feature belongs to These candidate features are stored in reference to face characteristics dictionary respectively and refer to non-face feature by face characteristic or non-face feature Dictionary;
B) candidate face detection is carried out to the masked face data set marked and candidate feature is extracted, judge that candidate feature belongs to These candidate features are stored in masked face characteristics dictionary and masked non-by masked face characteristic or masked non-face feature respectively Face characteristic dictionary;
C) from above-mentioned with reference to selecting representative to represent above-mentioned masked face characteristics dictionary in face characteristics dictionary With reference to face characteristics dictionary;
D) from above-mentioned with reference to selecting representative to represent above-mentioned masked non-face tagged word in non-face characteristics dictionary The non-face characteristics dictionary of the reference of allusion quotation;
E) merging is above-mentioned representative with reference to face characteristics dictionary and representative with reference to non-face characteristics dictionary, obtains To proximate exterior feature space.
2. the method as described in claim 1, which is characterized in that in step a), by calculating the corresponding candidate of the candidate feature Degree of overlapping between face location and the face location marked determines that degree of overlapping is handed over and ratio is to measure, wherein friendship is simultaneously Candidate feature is judged for the feature with reference to face than being greater than 0.7, is handed over and than judging candidate feature for reference to inhuman less than 0.3 The feature of face.
3. the method as described in claim 1, which is characterized in that step b), by calculating the corresponding candidate of the candidate feature Degree of overlapping between face position and the face location marked determines that degree of overlapping is handed over and ratio is to measure, wherein hand over and compare Candidate feature is judged greater than 0.6 for the feature of masked face, handed over and than judging candidate feature to be masked non-face less than 0.4 Feature.
4. the method as described in claim 1, which is characterized in that use greedy algorithm from reference face characteristics dictionary in step c) The representative reference face characteristics dictionary of middle selection;The greedy algorithm refers to calculating with reference to each in face characteristics dictionary With reference to the loss of face characteristic, obtain by the reference face feature list for losing ascending ascending order arrangement, before taking the list most The reference face characteristic in face represents masked face characteristic;Wherein the loss refers to each reference face characteristic and masked face The distance of the arest neighbors feature of characteristics dictionary and each arest neighbors feature with reference to face characteristic and masked non-face characteristics dictionary Distance difference.
5. the method as described in claim 1, which is characterized in that using greedy algorithm from reference to non-face tagged word in step d) It is selected in allusion quotation representative with reference to non-face characteristics dictionary;The greedy algorithm refers to that calculating refers to non-face characteristics dictionary In each loss with reference to non-face feature, obtain taking by the non-face feature list of reference for losing the arrangement of ascending ascending order The non-face feature of the reference of the list foremost represents masked non-face feature;Wherein the loss refers to each with reference to inhuman Face feature is at a distance from the arest neighbors feature of masked non-face characteristics dictionary and each special with reference to non-face feature and masked face Levy the difference of the distance of the arest neighbors feature of dictionary.
6. a kind of human face detection device, including candidate block, insertion module and authentication module;
The candidate block is used to carry out candidate face detection to image to be detected and extracts candidate feature;
The insertion module obtains tradition insertion feature or approximate insertion is special for carrying out insertion transformation to the candidate feature Sign;
The authentication module is used to test above-mentioned tradition insertion feature or approximate insertion feature with regression algorithm by classifying Card, to obtain testing result to the end;
Wherein, the insertion module carries out insertion change to the candidate feature by a surface space built in advance After changing, tradition insertion feature or approximate insertion feature are obtained;The surface space is conventional external feature space or approximation Insertion transformation described in surface space is locally linear embedding into method or quick approximation is locally linear embedding into method using traditional It realizes;Traditional method that is locally linear embedding into carries out insertion change to the candidate feature with noise using conventional external feature space It changes, obtains tradition insertion feature;Quick approximation is locally linear embedding into special to the candidate with noise using proximate exterior feature space Sign carries out insertion transformation, obtains approximate insertion feature;
The quick approximation is locally linear embedding into the building method of proximate exterior feature space in method, comprising the following steps:
A) candidate face detection is carried out to the reference face data set marked and candidate feature is extracted, judge that candidate feature belongs to These candidate features are stored in reference to face characteristics dictionary respectively and refer to non-face feature by face characteristic or non-face feature Dictionary;
B) candidate face detection is carried out to the masked face data set marked and candidate feature is extracted, judge that candidate feature belongs to These candidate features are stored in masked face characteristics dictionary and masked non-by masked face characteristic or masked non-face feature respectively Face characteristic dictionary;
C) from above-mentioned with reference to selecting representative to represent above-mentioned masked face characteristics dictionary in face characteristics dictionary With reference to face characteristics dictionary;
D) from above-mentioned with reference to selecting representative to represent above-mentioned masked non-face tagged word in non-face characteristics dictionary The non-face characteristics dictionary of the reference of allusion quotation;
E) merging is above-mentioned representative with reference to face characteristics dictionary and representative with reference to non-face characteristics dictionary, obtains To proximate exterior feature space.
7. device as claimed in claim 6, which is characterized in that the candidate block obtains multiple candidate features, then embedding Enter after carrying out insertion transformation by a surface space built in advance in module, obtains tradition insertion feature or approximation It is embedded in feature;The surface space is conventional external feature space or proximate exterior feature space;The insertion transformation is adopted Method is locally linear embedding into or quick approximation is locally linear embedding into method and realizes with traditional.
CN201710127367.3A 2017-03-06 2017-03-06 A face detection method and device Active CN107066943B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201710127367.3A CN107066943B (en) 2017-03-06 2017-03-06 A face detection method and device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710127367.3A CN107066943B (en) 2017-03-06 2017-03-06 A face detection method and device

Publications (2)

Publication Number Publication Date
CN107066943A CN107066943A (en) 2017-08-18
CN107066943B true CN107066943B (en) 2019-10-25

Family

ID=59622056

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710127367.3A Active CN107066943B (en) 2017-03-06 2017-03-06 A face detection method and device

Country Status (1)

Country Link
CN (1) CN107066943B (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108563982B (en) * 2018-01-05 2020-01-17 百度在线网络技术(北京)有限公司 Method and apparatus for detecting image
CN110363126A (en) * 2019-07-04 2019-10-22 杭州视洞科技有限公司 A kind of plurality of human faces real-time tracking and out of kilter method

Family Cites Families (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101202845B (en) * 2007-11-14 2011-05-18 北京大学 Method for changing infrared image into visible light image and device
WO2010032297A1 (en) * 2008-09-17 2010-03-25 富士通株式会社 Image processing device, image processing method, and image processing program
CN101393608A (en) * 2008-11-04 2009-03-25 清华大学 A visual object recognition method and device based on manifold distance analysis
CN101493885B (en) * 2009-02-27 2012-01-04 中国人民解放军空军工程大学 Embedded human face characteristic extracting method based on nuclear neighborhood protection
CN104978549B (en) * 2014-04-03 2019-04-02 北京邮电大学 Three-dimensional face images feature extracting method and system
US10410749B2 (en) * 2014-10-21 2019-09-10 uBiome, Inc. Method and system for microbiome-derived characterization, diagnostics and therapeutics for cutaneous conditions
CN105469063B (en) * 2015-12-04 2019-03-05 苏州大学 The facial image principal component feature extracting method and identification device of robust
CN106056553B (en) * 2016-05-31 2021-02-26 李炎然 Image restoration method based on tight frame feature dictionary

Also Published As

Publication number Publication date
CN107066943A (en) 2017-08-18

Similar Documents

Publication Publication Date Title
CN110110642B (en) Pedestrian re-identification method based on multi-channel attention features
WO2019120115A1 (en) Facial recognition method, apparatus, and computer apparatus
CN103605972B (en) Non-restricted environment face verification method based on block depth neural network
CN107330397B (en) A Pedestrian Re-identification Method Based on Large-Interval Relative Distance Metric Learning
CN110796026A (en) A Pedestrian Re-identification Method Based on Global Feature Splicing
CN111126240B (en) Three-channel feature fusion face recognition method
CN110717411A (en) A Pedestrian Re-identification Method Based on Deep Feature Fusion
WO2010092952A1 (en) Pattern recognition device
CN108960211A (en) A kind of multiple target human body attitude detection method and system
Jammalamadaka et al. Has my algorithm succeeded? an evaluator for human pose estimators
CN110674874A (en) A Fine-Grained Image Recognition Method Based on Object Fine Component Detection
CN110991321B (en) Video pedestrian re-identification method based on tag correction and weighting feature fusion
CN112862860B (en) Object perception image fusion method for multi-mode target tracking
CN109977895A (en) A kind of wild animal video object detection method based on multi-characteristic fusion
CN106778686A (en) A kind of copy video detecting method and system based on deep learning and graph theory
CN105574475A (en) Common vector dictionary based sparse representation classification method
CN114677633B (en) Multi-component feature fusion-based pedestrian detection multi-target tracking system and method
CN108960047A (en) Face De-weight method in video monitoring based on the secondary tree of depth
CN110414299A (en) A computer vision-based method for kinship analysis of monkey faces
CN106096517A (en) A kind of face identification method based on low-rank matrix Yu eigenface
CN111401113A (en) A Pedestrian Re-identification Method Based on Human Pose Estimation
CN113537119B (en) Transmission line connecting part detection method based on improved Yolov4-tiny
CN108460336A (en) A kind of pedestrian detection method based on deep learning
Mao et al. Training a scene-specific pedestrian detector using tracklets
CN107066943B (en) A face detection method and device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant