[go: up one dir, main page]

WO2002007096A1 - Device for tracking feature point on face - Google Patents

Device for tracking feature point on face Download PDF

Info

Publication number
WO2002007096A1
WO2002007096A1 PCT/JP2000/004798 JP0004798W WO0207096A1 WO 2002007096 A1 WO2002007096 A1 WO 2002007096A1 JP 0004798 W JP0004798 W JP 0004798W WO 0207096 A1 WO0207096 A1 WO 0207096A1
Authority
WO
WIPO (PCT)
Prior art keywords
feature point
face
specific pattern
person
image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Ceased
Application number
PCT/JP2000/004798
Other languages
French (fr)
Japanese (ja)
Inventor
Kentaro Hayashi
Kazuhiko Sumi
Manabu Hashimoto
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Mitsubishi Electric Corp
Original Assignee
Mitsubishi Electric Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Mitsubishi Electric Corp filed Critical Mitsubishi Electric Corp
Priority to PCT/JP2000/004798 priority Critical patent/WO2002007096A1/en
Publication of WO2002007096A1 publication Critical patent/WO2002007096A1/en
Anticipated expiration legal-status Critical
Ceased legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions

Definitions

  • the present invention relates to a technology for tracking a feature point of a face in an input image, and relates to a face feature point tracking device for detecting and tracking a feature portion of a person's face such as eyes and nose. is there. Background art
  • Japanese Patent Application Laid-Open No. 9-251534 as shown in a block diagram in FIG. Some were disclosed.
  • This device consists of an image input unit 111, a face region extraction unit 112, a feature point extraction unit 113, a feature point set candidate selection unit 114, a pattern evaluation unit 115, a normalization generation unit 111 6.
  • an image of a person to be recognized is input by the image input unit 111.
  • the face region extraction unit 112 extracts the person's face region from the input image
  • the feature point extraction unit 113 extracts the eyeball (black eye) from the extracted face regions by using the degree of separation filter.
  • Feature point candidates for the face such as the face and nostrils are extracted.
  • the feature point set candidate selection unit 114 narrows down the feature point set candidates from among the feature point candidates extracted by the feature point extraction unit 113 using structural constraints of the face.
  • the 'similarity calculation unit 115a generates a feature extracted based on each feature point for each feature point set selected by the feature point set candidate selection unit 114.
  • Partial templates such as eye, nose, mouth area etc
  • the degree of similarity to the pattern of port 1 15b is calculated, the degree of consistency of the weighted sum is obtained, and the feature point set having the highest degree of consistency is selected as the correct feature point set.
  • the normalization generation unit 1 16 generates a normalized image using the feature point set selected as correct.
  • the similarity calculating unit 117a compares the normalized image obtained in the normalization generation unit 116 with the dictionary image 111b of each registrant registered in advance. The similarity is calculated, and the person representing the dictionary image having a high similarity is identified as the person.
  • feature points of a face are extracted from an input image as described above.
  • an eyeball Each feature point candidate of the face such as the iris and nostrils is extracted, and each feature point set (; :) extracted by the pattern evaluation unit 115 using the pattern matching method is used. Since the point set is selected, if the eyes cannot be extracted because the eyes are closed and the nostrils cannot be extracted due to the angle of the face, the pattern evaluation at the next stage cannot be performed accurately or at all. There is a problem that the face feature points cannot be tracked stably and accurately because they may not exist.
  • the present invention has been made in order to solve the above-described problems, and it is not possible to detect a feature point such as a closed eye or a nostril due to a positional relationship between a face and an image input unit (camera). It is an object of the present invention to provide a facial feature point tracking device that can stably track feature points even in a case. Disclosure of the invention
  • a first feature point tracking device for a face sequentially detects the position of a characteristic portion of the face of a person from an image of the face of the person captured in time series, Device that traces,
  • An image pattern including a characteristic portion of the person's face stored in advance is defined as a specific pattern, and the specific pattern or an image pattern close to the specific pattern exists at any position in the captured image.
  • Invariant feature point position detecting means for detecting a position of an invariant feature point included in a characteristic portion of the human face
  • An output position integrating means for storing an image of a characteristic portion of the detected person's face as a new specific pattern is provided.
  • a change in the position of a characteristic portion of the face can be detected if detected by at least one of the specific pattern position detecting means and the invariant feature point position detecting means, so that more stable detection is possible.
  • the specific pattern is updated each time the detection is performed, so the specific pattern position detection means should detect the characteristic part of the face even if it changes over time. Can be done, and mouth-bust tracking becomes possible.
  • a second facial feature point tracking device is the first facial feature point tracking device, wherein the nose position detecting means for detecting the nose position from the captured image of the face of the person;
  • the apparatus further includes feature point detection means for detecting the position of a characteristic portion of the face of a person other than the nose from the nose position detected by the nose position detection means.
  • the position of the nose which is relatively easy to detect, is first detected, and based on the relative positional relationship between the nose and the characteristic part of the face other than the nose, the characteristic of the face of the person other than the nose is detected.
  • the detection is easier than in the case where the position of the feature point is directly detected from the captured image.
  • a third facial feature point tracking device is characterized in that in the first facial feature point tracking device, a change in the position of a characteristic portion of the person's face obtained from the output position integrating means is examined.
  • a movement state detecting means for accumulating the time required for the position change if the change in the position is smaller than a predetermined threshold value; and, if the accumulated time is longer than a preset time, accumulating the preset time.
  • setting means for updating the predetermined threshold value to the minimum value of the amount of change in the position changed for each of the accumulated times.
  • FIG. 1 is a block diagram showing the configuration of a face feature point tracking device according to the first embodiment of the present invention
  • FIG. 2 is a flowchart for explaining the operation of the face feature point tracking device according to the first embodiment of the present invention.
  • FIG. 3 is a block diagram showing the configuration of a face feature point tracking device according to the second embodiment of the present invention.
  • FIG. 4 is a diagram for explaining the operation of the face feature point tracking device according to the second embodiment of the present invention.
  • FIG. 5 is a block diagram showing the configuration of the face feature point tracking device according to the third embodiment of the present invention.
  • FIG. 6 is a diagram showing the operation of the face feature point tracking device according to the third embodiment of the present invention.
  • FIG. 7 is a block diagram showing the configuration of a conventional person authentication device. BEST MODE FOR CARRYING OUT THE INVENTION
  • FIG. 1 is a block diagram showing the configuration of a face feature point tracking device according to the first embodiment of the present invention
  • FIG. 2 is a flowchart for explaining the operation of the face feature point tracking device according to the first embodiment of the present invention.
  • reference numeral 1 denotes an imaging means for inputting an image of a person's face, which is, for example, a CCD camera.
  • Numeral 2 is included when the face is imaged from the input image obtained by the imaging means 1 such as a pupil of the eye (black eye) or a nostril of the nose, and the shape thereof has a relatively small variation with time.
  • Invariant feature points and are invariant feature point position detecting means for detecting this invariant feature point and detecting its position. After detecting the invariant feature points (pupils, nostrils), the invariant feature point detection means 2 calculates and outputs the position of the center point (center of gravity, etc.).
  • Specific pattern position detecting means for detecting a position of a specific pattern on a face such as an eye or a nose.
  • the specific pattern is, for example, a region (such as a rectangle) surrounding the invariant feature points.
  • the position of the specific pattern for example, the position of the center point (such as the center of gravity) of this area is used.
  • the invariant feature point position detecting means 2, the specific pattern position detecting means 3, and the output position integrating means 4 are realized by, for example, a computer.
  • FIG. 2 is a diagram for explaining the operation of the facial feature point tracking apparatus according to the first embodiment.
  • invariant feature points are pupils and nostrils
  • specific patterns are eyes and nose
  • eyes and nose of a face are tracked.
  • step 1 it is checked whether the feature point position was detected in the previous calculation, and if so, tracking from the input image based on the feature point position (eye and nose position) obtained in the previous calculation. Set the search area to search for the target feature point.
  • a rectangular area having a certain size surrounding the eyes and a rectangular area having a certain size surrounding the nose are set.
  • the rectangular area to be searched (referred to as a search rectangular area) is usually a rectangular area of a certain size centered on the feature point position obtained by the previous calculation.
  • the size of this rectangular area may be determined as appropriate according to the time interval for capturing the input image, the time interval for detecting the feature points, the speed at which the person moves the face, etc. Also, when the feature point position is not detected in the previous calculation In other words, when detecting feature points for the first time, for example, binarizing the input image using a means (not shown) for binarizing the input image, and detecting the position of the nostril using general knowledge of the face structure Set the eye search rectangular area from.
  • step 2 a search area for searching for invariant feature points is set in the search rectangular area set in step 1.
  • a rectangular area centered on the feature point used in step 1 and smaller than the rectangular area set in step 1 is set.
  • the positions of the invariant feature points usually exist within a specific pattern, and the size of the region is smaller than that of the specific pattern.
  • the entire search area set in step 1 is invariant feature
  • the position of the invariant feature point can be calculated in a shorter time than in the case where the search area is the position of the point.
  • the invariant feature point is a pupil
  • a shape (circle) corresponding to the pupil is set, and a rectangular area including the shape around the feature point is set as a search area for searching for the invariant feature point. It is.
  • the size of the rectangular area for searching for the invariant feature points may be appropriately determined according to the time interval for capturing the input image, the time interval for detecting the invariant feature points, the speed at which the pupil moves, and the like.
  • the search area is set based on the relative positional relationship between the invariant feature point position and the specific pattern position obtained by the previous calculation, and the area with the highest probability of the invariant feature point being present, such as statistics, is calculated.
  • the minimum rectangular area which is estimated using a method and includes an area where the probability that an invariant feature point exists is sufficiently high, may be used as a search area for searching for an invariant feature point.
  • invariant feature points are detected from the search area for searching for the invariant feature points set in step 2, and the positions of the invariant feature points are output.
  • invariant feature points for example, the shape of invariant feature points stored in advance (pupil shape for a pupil, nostril shape for a nostril). More specifically, for example, by examining where in the search area set in step 2 a shape that matches the shape corresponding to the invariant feature point stored in advance, the invariant feature point is detected, and the detected shape Calculate the center position
  • the separation degree filter (not shown) is used.
  • the invariant feature points can be detected with higher accuracy if the image in the search area set in step 2 is used.
  • the invariant feature point is the pupil, part or all of the pupil will be hidden during the movement while closing the eyes, such as blinking. Therefore, since the shape of the invariant feature point stored in advance is different from the shape of the pupil in the actual input image, the invariant feature point described above is not detected.
  • an invariant feature point cannot be detected in the search area set in step 2, it is determined that blinking is in progress, and processing such as not detecting the position of the invariant feature point is performed.
  • the additional detection improves the reliability of detection in step 3.
  • a position where an invariant feature point such as a pupil or a nostril exists can be detected from an image.
  • Step 2 and Step 3 Immediately before performing Step 2 and Step 3, or Step 2 and Step Perform step 4 while performing step 3.
  • Step 4 specifically detects a specific pattern from the search rectangular area set in step 1 and outputs the center position of the specific pattern.
  • the specific pattern obtained by the previous calculation is used as a template image pattern
  • a specific pattern is detected from the search area using a method such as pattern matching, and the center position of the detected specific pattern is input this time. Output as the position of the specific pattern included in the image.
  • Pattern matching is a method of detecting a specific pattern existing in a search area of an input image by finding a position where the sum of differences between a specific pattern and a corresponding pixel value of the input image is small.
  • the position where the specific pattern of the eyes and the nose exists can be detected from the image.
  • the position of the invariant feature point obtained in step 3 and the position of the specific pattern obtained in step 4 should ideally coincide with each other, but in reality this may not be the case.
  • step 5 for example, the midpoint of the position obtained by steps 3 and 4 is adopted.
  • step 5 the midpoint of the two feature point positions was adopted, but it was not the midpoint, but an internally divided point with a weight that is closer to the feature point position obtained from one of the steps. You may.
  • the feature point position obtained in step 5 is stored as feature point position information in step 6, and is prepared for feature point position detection performed immediately after. In this way, the positions of the feature points on the face can be stably tracked.
  • step 2 if no invariant feature point is detected in the search area set in step 2, it is determined that blinking is in progress, and processing such as not detecting the position of the invariant feature point is performed. If this output is made, the position of the specific nodal output in step 4 is set as the position of the feature point, so that the position of the invariant score point cannot be detected due to blinking or the like. Even feature points can be tracked.
  • step 5 The output of step 5 is used in step 6, step 7, and step 9.
  • step 6 the information on the position of the feature point output in step 5 is retained and updated, and is used as the position of the feature point in step 1 in the next detection.
  • the search area in step 1 can be set to an appropriate position.
  • a specific pattern (template) is obtained from the surrounding image information.
  • step 8 the acquired specific pattern is used as a new specific pattern, and the updated specific pattern is used as a specific pattern used for detecting the next specific pattern. However, this can be detected appropriately.
  • step 9 the feature point position obtained in step 5 is output, and the process ends.
  • the face feature point tracking operation as described above is performed at every time interval at which the image capturing means 1 takes in images in time series, for example, at every 1/30 second.
  • the flowchart shown in FIG. 2 is an example, and another flowchart may be used as long as the input / output relationship of each step is appropriate.
  • the position of the center point of the pupil or the nostril based on the position of the invariant feature point detected by the invariant feature point position detection means 2 and the pattern matching by the specific pattern position detection means 3
  • the output integration means 4 adjusts to the value taking into account the detection values from both detection means 2 and 3.
  • feature points far from the true position are not output, and feature points on the face, such as eyes and nose, can be tracked in the mouth past.
  • the invariant feature point position detecting means 2 and the specific pattern position detecting means 3 respectively detect the invariant feature point position and the specific pattern position independently, either one of the detecting means 2 (or 3) is used. Even when detection by the other means is impossible, the detection result by the other detection means 3 (or 2) can be obtained. Can be tracked stably.
  • the characteristic point position obtained by the output position integrating means 4 and the specific pattern acquired based on the characteristic point position are updated, and the next invariable characteristic point position detecting means 2 or the specific pattern position detecting means 3 are updated. Therefore, the specific pattern position detecting means 4 can detect the characteristic portion of the person's face following the change even when the characteristic portion changes with time. Therefore, robust tracking becomes possible.
  • the detection time of the specific pattern is shorter than when the entire image of the person's face is set as the search area. Can also be shortened. Example 2.
  • FIG. 3 is a block diagram showing the configuration of a face feature point tracking device according to the second embodiment of the present invention
  • FIG. 4 is a flowchart for explaining the operation of the face feature point tracking device according to the second embodiment of the present invention.
  • reference numeral 5 denotes a means for detecting the position of the nose from the image of the face obtained by the imaging means 1
  • reference numeral 6 denotes a position of the nose detected by the nose position detection means 5, and the imaging means 1 is used.
  • This is means for detecting the position of the eyes from the image of the face obtained by the above.
  • the nose position detecting means 5 and the eye position detecting means 6 are both realized by a computer.
  • the facial feature point tracking apparatus configured as described above can be operated, for example, in the order shown in the flowchart of FIG.
  • the image capturing means 1, the invariant feature point position detecting means 2, and the identification are used in the same manner as in the first embodiment.
  • the feature points of the face are tracked by the pattern position detecting means 3 and the output position integrating means 4.
  • the positions of both eyes and the nose are detected by the nose position detecting means 5 and the eye position detecting means 6, and the feature point positions are calculated based on the detection results.
  • FIG. 4 is a flowchart for explaining the operation of the facial feature point tracking apparatus according to the second embodiment.
  • step 11 it is determined whether or not the feature point position has been obtained last time.
  • the procedure for detecting a feature point when the feature point position has been previously obtained is the same as that described in the first embodiment with reference to FIG. 2, and a description thereof will be omitted.
  • the nostrils appear in black areas in the image. Therefore, in order to detect the nose region, it is sufficient to detect these two black regions, and the detection can be performed with relatively high accuracy.
  • a specific candidate region where the nose is likely to be extracted is extracted in step 12 and the image is binarized with a specific threshold value for the extracted region.
  • step 12 above in order to cope with a change in brightness of the image, binarization may be performed using a value obtained by adding a specific value to the darkest pixel value based on the minimum pixel value in the image as a threshold value.
  • step 13 a region having an area within a specific range is extracted from the image binarized in step 1 2, and a certain range Detects two regions that are located on the left and right at a distance.
  • These two regions are nostrils, and the position of the nose can be represented by, for example, the midpoint of each center of gravity of the two regions.
  • step 14 an area where an eye is likely to be seen relatively is set from the position of the nose detected in step 13.
  • step 15 the region where the eyes set in step 14 are likely to exist is binarized.
  • step 16 a region having an area within a specific range is extracted from the image binarized in step 15, and an arbitrary part of the region is extracted.
  • two regions that are arranged on the left and right within a certain range are detected as pupils, and the center position of each pupil is set as a feature point position.
  • the positions of the nose and both eyes are detected, and these positions are output as feature point positions in step 9.
  • the feature point positions are retained in step 6 to prepare for the next tracking.
  • a predetermined range is cut out around the feature point position, and this is obtained as a specific pattern.
  • the position of the nose that is relatively easy to detect is detected first.
  • other feature points for example, the positions of eyes are detected, so that stable and mouth-to-mouth tracking can be performed.
  • the flowchart shown in FIG. 4 is an example, and another flowchart may be used as long as the input / output relationship of each step is appropriate.
  • the positions of the eyes and the nose are tracked as the feature point positions.
  • the present invention is not limited to this.
  • the positions of the mouth may be used.
  • the position of the nose which is relatively easy to detect is first detected, and the position of the mouth is detected based on this. .
  • the position of the nose which is relatively easy to detect, is first detected, and the characteristic portion of the face of the person other than the nose is determined based on the relative positional relationship between the nose and the characteristic portion of the face other than the nose.
  • the position of the feature point of a person's face other than the nose is easier to detect than if it is detected directly from the captured image.
  • FIG. 5 is a block diagram showing a configuration of a facial feature point tracking apparatus according to Embodiment 3 of the present invention.
  • FIG. 6 is a flowchart for explaining the operation of the facial feature point tracking device according to the third embodiment of the present invention.
  • 7 is a movement width detecting means corresponding to the movement state detecting means
  • 8 is a reference movement width
  • 9 is a reference tracking time
  • 10 is a stable tracking determination means
  • 11 is a reference movement width setting means
  • 1 2 Are reference tracking time setting means, which are both realized by a computer and constitute a stability tracking means for tracking the stability of the face.
  • the reference movement width setting means 11 and the reference tracking time setting means 12 constitute a setting means.
  • the stability tracking determination means 10 uses the output of the output position integrating means 4, the reference movement width 8 and the reference tracking time 9 to detect the movement width of the feature point position by the movement width detection means 7, and based on the detection result, The tracking determination means 10 determines whether stable tracking is performed and outputs the result.
  • the reference movement width setting means 11 updates the reference movement width 8
  • the reference tracking time setting means 12 updates the reference tracking time 9.
  • step 21 it is determined whether or not this stability tracking method is to be used for the first time, and if so, in step 22, the reference tracking time is set to a predetermined specific fixed value. Then, the reference motion width is set to a predetermined specific fixed value. Then, the change of the position of the feature point is examined for a time corresponding to the reference tracking time.
  • the movement width of the position of the feature point is set to a value smaller than, for example, the reference movement width.
  • step 23 the positions of the feature points obtained by the previous calculation and the current calculation are used.
  • the movement width of the feature point position is calculated from the obtained feature point position. At this time, for example, if the movement width is defined as the relative distance of the feature point position from the previous feature point position, the movement width of the feature point position becomes the length of the vector representing the difference between the positions.
  • step 24 it is checked whether or not the motion width of the feature point position set in step 22 or step 23 is smaller than the reference motion width.
  • step 25 the tracking time is set to 0 in step 25, the fact that the tracking is not stable (not stable) is output in step 26, and the process returns to step 21.
  • the motion width is smaller than the reference motion width in step 24, the time required to change the position of the feature point in step 24 in step 27, that is, the time when the previous feature point position was calculated and the current time The difference from the time when the feature point position was calculated is accumulated in the tracking time.
  • step 28 it is checked whether this tracking time is longer than the reference tracking time.
  • step 26 if it is determined that the tracking time is equal to or shorter than the reference tracking time, it is output in step 26 that the tracking is not stable tracking, and the process returns to step 21.
  • step 28 If it is determined in step 28 that the tracking time is longer than the reference tracking time, the tracking time used in step 29 is set as a new reference tracking time, and the updated reference tracking time is held in step 30. .
  • step 32 the updated reference motion width is held, and in step 33, the fact that the tracking is stable is output.
  • it is used for tracking feature points of a face to track the 3D direction of the face, and enables stable and robust tracking of the 3D direction of the face. It is also used for feature point detection and tracking for face individual identification, and enables stable and mouth-bust personal identification of faces.

Landscapes

  • Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Theoretical Computer Science (AREA)
  • Collating Specific Patterns (AREA)

Abstract

A device for tracking a feature point on a face comprising specific pattern position measuring means (3) for measuring the position of either a specific pattern being an image pattern including a feature portion of the face of a person stored in advance or an image pattern similar to the specific pattern in a captured image, constant feature point position measuring means (2) for measuring the position of a constant feature point included in the feature portion of the face of the person, and output position integrating means (4) for measuring the position of the feature portion in the captured image on the basis of the output from the contestant feature point position measuring means and the output from the specific pattern position measuring means and storing therein the image of the feature portion as a new specific pattern.

Description

明 細 顔の特徴点追跡装置 技術分野  Detail Facial feature point tracking device Technical field

本発明は、 入力画像中の顔の特徴点を追跡する技術に関するものであ り、 目、 鼻等の人物の顔の特徴的な部分を検出し追跡する顔の特徴点追 跡装置に関するものである。 背景技術  The present invention relates to a technology for tracking a feature point of a face in an input image, and relates to a face feature point tracking device for detecting and tracking a feature portion of a person's face such as eyes and nose. is there. Background art

従来、 入力画像から顔特徴点を抽出し、 その特徴点を用いて人物認証 を行う人物認証装置として、 第 7図にプロック図で示すような特開平 9 - 2 5 1 5 3 4号公報に開示されているものがあった。  Conventionally, as a person authentication device that extracts a facial feature point from an input image and performs personal authentication using the feature point, Japanese Patent Application Laid-Open No. 9-251534, as shown in a block diagram in FIG. Some were disclosed.

本装置は、 画像入力部 1 1 1、 顔領域抽出部 1 1 2、 特徴点抽出部 1 1 3、 特徴点セヅト候補選択部 1 1 4、 パターン評価部 1 1 5、 正規化 生成部 1 1 6、 認識部 1 1 7からなる。  This device consists of an image input unit 111, a face region extraction unit 112, a feature point extraction unit 113, a feature point set candidate selection unit 114, a pattern evaluation unit 115, a normalization generation unit 111 6. Recognition unit 1 17

次に動作について説明する。 まず、 画像入力部 1 1 1により、 認識対 象となる人物の画像を入力する。 次に、 顔領域抽出部 1 1 2により入力 画像から当該人物の顏領域を抽出し、 特徴点抽出部 1 1 3は抽出された 顔領域の中から分離度フィル夕を用いて眼球 (黒目) や鼻穴などの顔の 各特徴点候補を抽出する。 特徴点セット候補選択部 1 1 4は特徴点抽出 部 1 1 3によって抽出された各特徴点候補の中から顔の構造的な制約を 用いて各特徴点のセット候補を絞り込む。 パターン評価部 1 1 5におい ては、 '類似度計算部 1 1 5 aが特徴点セット候補選択部 1 1 4によって 選択された各特徴点セットに対して、 各特徴点を基準に切り出した特徴 点近傍パターンと予め登録してある目、 鼻、 口領域などの部分テンプレ ート 1 1 5 bのパターンとの類似度を計算してその加重和の整合度を求 め、 最も高い整合度を持つ特徴点セットを正しい特徴点セットとして選 択する。 正規化生成部 1 1 6はその正しいと選択された特徴点セットを 用いて正規化画像を生成する。 認識部 1 1 7においては、 類似度計算部 1 1 7 aが正規ィ匕生成部 1 1 6で得られた規化画像と予め登録されてい る各登録者の辞書画像 1 1 7 bとの類似度を計算し、 類似度が高い辞書 画像を表す人物を当人と識別する。 Next, the operation will be described. First, an image of a person to be recognized is input by the image input unit 111. Next, the face region extraction unit 112 extracts the person's face region from the input image, and the feature point extraction unit 113 extracts the eyeball (black eye) from the extracted face regions by using the degree of separation filter. Feature point candidates for the face such as the face and nostrils are extracted. The feature point set candidate selection unit 114 narrows down the feature point set candidates from among the feature point candidates extracted by the feature point extraction unit 113 using structural constraints of the face. In the pattern evaluation unit 115, the 'similarity calculation unit 115a generates a feature extracted based on each feature point for each feature point set selected by the feature point set candidate selection unit 114. Partial templates such as eye, nose, mouth area etc The degree of similarity to the pattern of port 1 15b is calculated, the degree of consistency of the weighted sum is obtained, and the feature point set having the highest degree of consistency is selected as the correct feature point set. The normalization generation unit 1 16 generates a normalized image using the feature point set selected as correct. In the recognizing unit 117, the similarity calculating unit 117a compares the normalized image obtained in the normalization generation unit 116 with the dictionary image 111b of each registrant registered in advance. The similarity is calculated, and the person representing the dictionary image having a high similarity is identified as the person.

従来の入力画像からの顔の特徴点抽出は以上のように行われており、 この手法を用いて顔特徴点を追跡した場合、 特徴点抽出部 1 1 3により 分離度フィルタを用いて眼球 (黒目) や鼻穴などの顔の各特徴点候補を 抽出し、 このようにして抽出された各特徴点セット (;:対してパターン評 価部 1 1 5によりパターンマッチの手法を用いて正しい特徴点セットと して選択するので、 例えば目をつぶつていて眼球が抽出できない場合や 顔の角度により鼻穴が抽出できない場合などには、 次段階のパターン評 価も正確に行えなかったり全く行えなかったりする場合があるので、 顏 特徴点を追跡を安定にしかも正確に行うことができないという問題点が ある。  Conventionally, feature points of a face are extracted from an input image as described above. When a face feature point is tracked using this method, an eyeball ( Each feature point candidate of the face such as the iris and nostrils is extracted, and each feature point set (; :) extracted by the pattern evaluation unit 115 using the pattern matching method is used. Since the point set is selected, if the eyes cannot be extracted because the eyes are closed and the nostrils cannot be extracted due to the angle of the face, the pattern evaluation at the next stage cannot be performed accurately or at all. There is a problem that the face feature points cannot be tracked stably and accurately because they may not exist.

本発明は上記のような問題点を解消するためになされたもので、 目を つぶっていたり、 顔と画像入力部 (カメラ) との位置関係により鼻孔が 検出できないといったように特徴点が検出できない場合であっても安定 かつ口バストに特徴点を追跡することができる顔の特徴点追跡装置を提 供することを目的としている。 発明の開示  The present invention has been made in order to solve the above-described problems, and it is not possible to detect a feature point such as a closed eye or a nostril due to a positional relationship between a face and an image input unit (camera). It is an object of the present invention to provide a facial feature point tracking device that can stably track feature points even in a case. Disclosure of the invention

本発明に係る第 1の顔の特徴点追跡装置は、 時系列に取り込まれる人 物の顔の画像から前記人物の顔の特徴的な部分の位置を逐次検出し、 追 跡する装置であって、 A first feature point tracking device for a face according to the present invention sequentially detects the position of a characteristic portion of the face of a person from an image of the face of the person captured in time series, Device that traces,

予め記憶された前記人物の顔の特徴的な部分を含む画像パタンを特定 パ夕ンとし、 前記特定パタンもしくは前記特定パタンに近い画像パ夕ン が前記取り込まれた画像中のどの位置に存在するかを検出する特定パ夕 ン位置検出手段、  An image pattern including a characteristic portion of the person's face stored in advance is defined as a specific pattern, and the specific pattern or an image pattern close to the specific pattern exists at any position in the captured image. Specific panel position detecting means for detecting

前記人物の顔の特徴的な部分に含まれる不変特徴点の位置を検出する 不変特徴点位置検出手段、 および  Invariant feature point position detecting means for detecting a position of an invariant feature point included in a characteristic portion of the human face; and

前記不変特徴点位置検出手段からの出力と前記特定パ夕ン位置検出手 段からの出力とにより前記取り込まれた画像中にある前記人物の顔の特 徴的な部分の位置を検出するとともに、 前記検出した人物の顔の特徴的 な部分の画像を新たな特定パ夕ンとして記憶する出力位置統合手段を備 えたものである。  Detecting a position of a characteristic portion of the person's face in the captured image based on an output from the invariant feature point position detecting means and an output from the specific pattern position detecting means; An output position integrating means for storing an image of a characteristic portion of the detected person's face as a new specific pattern is provided.

これによれば、 特定パタン位置検出手段および不変特徴点位置検出手 段の少なくともいずれか一方により検出されれば顔の特徴的な部分の位 置の変化を検出できるため、 より安定した検出が可能となるばかりか、 検出が行われるごとに特定パタンを更新していくので、 特定パ夕ン位置 検出手段は顔の特徴的な部分が時間的に変化してもこれに追従して検出 することができるようになり、 口バストな追跡が可能となる。  According to this, a change in the position of a characteristic portion of the face can be detected if detected by at least one of the specific pattern position detecting means and the invariant feature point position detecting means, so that more stable detection is possible. In addition, the specific pattern is updated each time the detection is performed, so the specific pattern position detection means should detect the characteristic part of the face even if it changes over time. Can be done, and mouth-bust tracking becomes possible.

本発明に係る第 2の顔の特徴点追跡装置は、 第 1の顔の特徴点追跡装 置において、 取り込まれた人物の顔の画像から鼻の位置を検出する鼻の 位置検出手段、 および前記鼻の位置検出手段により検出された鼻の位置 から前記鼻以外の人物の顔の特徴的な部分の位置を検出する特徴点検出 手段を備えたものである。  A second facial feature point tracking device according to the present invention is the first facial feature point tracking device, wherein the nose position detecting means for detecting the nose position from the captured image of the face of the person; The apparatus further includes feature point detection means for detecting the position of a characteristic portion of the face of a person other than the nose from the nose position detected by the nose position detection means.

これによれば、 比較的検出が容易な鼻の位置をまず検出し、 鼻と鼻以 外の顔の特徴的な部分との相対的な位置関係に基き、 鼻以外の人物の顔 の特徴的な部分の位置を検出するようにしたので、 鼻以外の人物の顔の 特徴点の位置を取り込まれた画像から直接検出する場合に比較して検出 が容易になる。 According to this, the position of the nose, which is relatively easy to detect, is first detected, and based on the relative positional relationship between the nose and the characteristic part of the face other than the nose, the characteristic of the face of the person other than the nose is detected. Of the face of the person other than the nose The detection is easier than in the case where the position of the feature point is directly detected from the captured image.

本発明に係る第 3の顔の特徴点追跡装置は、 第 1の顔の特徴点追跡装 置において、 出力位置統合手段から得られた人物の顔の特徴的な部分の 位置の変化を調べるとともに、 上記位置の変化が所定の閾値よりも小さ ければ位置の変化に要した時間を累積する動き状態検出手段、 および 累積した時間が予め設定した時間よりも大きくなると、 予め設定した 時間を前記累積した時間に更新するとともに、 前記所定の閾値を前記累 積した時間毎に変化した位置の変化量うちの最小値に更新する設定手段 を備えたものである。  A third facial feature point tracking device according to the present invention is characterized in that in the first facial feature point tracking device, a change in the position of a characteristic portion of the person's face obtained from the output position integrating means is examined. A movement state detecting means for accumulating the time required for the position change if the change in the position is smaller than a predetermined threshold value; and, if the accumulated time is longer than a preset time, accumulating the preset time. And setting means for updating the predetermined threshold value to the minimum value of the amount of change in the position changed for each of the accumulated times.

これによれば、 人物の顔の動的な変化を検出するだけではなく、 動か ない状態を検出することができるようになるばかりか、 設定手段を備え たので、 動かないと判断するために用いる所定の閾値として予め設定し た値は個人の特性に応じた値に収束するため、 個人の特性に応じた判定 が可能となる。 図面の簡単な説明  According to this, it is possible to not only detect a dynamic change of a person's face, but also to detect a state in which the person does not move. Since the value set in advance as the predetermined threshold value converges to a value according to the characteristics of the individual, it is possible to make a determination according to the characteristics of the individual. BRIEF DESCRIPTION OF THE FIGURES

第 1図は本発明の実施例 1による顔の特徴点追跡装置の構成を示すプ ロック図、 第 2図は本発明の実施例 1による顔の特徴点追跡装置の動作 を説明するためのフローチャート図、 第 3図は本発明の実施例 2による 顔の特徴点追跡装置の構成を示すプロック図、 第 4図は本発明の実施例 2による顔の特徴点追跡装置の動作を説明するためのフローチヤ一ト図 、 第 5図は本発明に実施例 3による顔の特徴点追跡装置の構成を示すブ ロック図、 第 6図は本発明の実施例 3による顔の特徴点追跡装置の動作 を説明するためのフローチャート図、 第 7図は従来の人物認証装置の構 成を示すプロック図である。 発明を実施するための最良の形態 FIG. 1 is a block diagram showing the configuration of a face feature point tracking device according to the first embodiment of the present invention, and FIG. 2 is a flowchart for explaining the operation of the face feature point tracking device according to the first embodiment of the present invention. FIG. 3 is a block diagram showing the configuration of a face feature point tracking device according to the second embodiment of the present invention. FIG. 4 is a diagram for explaining the operation of the face feature point tracking device according to the second embodiment of the present invention. FIG. 5 is a block diagram showing the configuration of the face feature point tracking device according to the third embodiment of the present invention. FIG. 6 is a diagram showing the operation of the face feature point tracking device according to the third embodiment of the present invention. FIG. 7 is a block diagram showing the configuration of a conventional person authentication device. BEST MODE FOR CARRYING OUT THE INVENTION

実施例 1 .  Example 1

以下、 この発明の実施例を図に基づいて説明する。  Hereinafter, embodiments of the present invention will be described with reference to the drawings.

第 1図は本発明の実施例 1による顔の特徴点追跡装置の構成を示すブ ロック図、 第 2図は本発明の実施例 1による顔の特徴点追跡装置の動作 を説明するためのフローチャート図である。  FIG. 1 is a block diagram showing the configuration of a face feature point tracking device according to the first embodiment of the present invention, and FIG. 2 is a flowchart for explaining the operation of the face feature point tracking device according to the first embodiment of the present invention. FIG.

第 1図において、 1は人物の顏の画像を入力するための撮像手段であ り、 例えば C C Dカメラからなる。  In FIG. 1, reference numeral 1 denotes an imaging means for inputting an image of a person's face, which is, for example, a CCD camera.

2は撮像手段 1で得られた入力画像から、 例えば目の瞳孔 (黒目) や 鼻の鼻孔といったように顔を撮像したときに含まれるもので、 その形状 の時間的な変動が比較的少ないものを不変特徴点とし、 この不変特徴点 を検出するとともにその位置を検出するための不変特徴点位置検出手段 である。 不変特徴点検出手段 2は不変特徴点 (瞳孔、 鼻孔) を検出した 後、 その中心点 (重心など) の位置を算出し出力する。  Numeral 2 is included when the face is imaged from the input image obtained by the imaging means 1 such as a pupil of the eye (black eye) or a nostril of the nose, and the shape thereof has a relatively small variation with time. Are invariant feature points, and are invariant feature point position detecting means for detecting this invariant feature point and detecting its position. After detecting the invariant feature points (pupils, nostrils), the invariant feature point detection means 2 calculates and outputs the position of the center point (center of gravity, etc.).

3は予め記憶された人物の顔の特徴的な部分を含む画像パタンを特定 パタンとし、 特定パタンもしくは特定パタンに近い画像パタンが、 取り 込まれた画像中のどの位置に存在するかを検出する特定パタン位置検出 手段であり、 例えば目、 鼻などの顔上の特定のパタンの位置を検出する 。 特定パタンとは例えば不変特徴点を囲む領域 (四角形など) である。 特定パタンの位置としては、 例えばこの領域の中心点 (重心など) の位 置等が用いられる。  3 designates an image pattern including a characteristic portion of a person's face stored in advance as a specific pattern, and detects at which position in the captured image the specific pattern or an image pattern close to the specific pattern exists. Specific pattern position detecting means for detecting a position of a specific pattern on a face such as an eye or a nose. The specific pattern is, for example, a region (such as a rectangle) surrounding the invariant feature points. As the position of the specific pattern, for example, the position of the center point (such as the center of gravity) of this area is used.

4は不変特徴点位置検出手段 2の出力と特定パ夕ン位置検出手段 3と の出力とにより撮像手段 1から取り込まれた画像中にある人物の顔の特 徴的な部分の位置を検出するとともに、 前記検出した人物の顔の特徴的 な部分の画像を新たな特定パ夕ンとして記憶する出力位置統合手段であ る o 4 detects the position of the characteristic portion of the face of the person in the image captured from the imaging means 1 by the output of the invariant feature point position detecting means 2 and the output of the specific pattern position detecting means 3 Output position integrating means for storing the image of the characteristic portion of the detected face of the person as a new specific pattern. O

不変特徴点位置検出手段 2、 特定パ夕ン位置検出手段 3および出力位 置統合手段 4は例えばコンピュータにより実現される。  The invariant feature point position detecting means 2, the specific pattern position detecting means 3, and the output position integrating means 4 are realized by, for example, a computer.

次に実施例 1の顔の特徴点追跡装置の動作を説明する。  Next, the operation of the facial feature point tracking apparatus according to the first embodiment will be described.

第 2図は実施例 1の顔の特徴点追跡装置の動作を説明するための図で ある。  FIG. 2 is a diagram for explaining the operation of the facial feature point tracking apparatus according to the first embodiment.

この実施例では、 不変特徴点を瞳孔、 鼻孔とし、 特定パタンを目、 鼻 とし、 顔の目、 鼻を追跡する場合について説明する。  In this embodiment, a case will be described in which invariant feature points are pupils and nostrils, specific patterns are eyes and nose, and eyes and nose of a face are tracked.

まず、 ステップ 1において、 前回の計算において特徴点位置が検出さ れたかどうかを調べ、 あれば前回の計算により得られた特徴点位置 (目 と鼻の位置) を基に、 入力画像中から追跡対象となる特徴点を探索すベ き探索領域を設定する。  First, in step 1, it is checked whether the feature point position was detected in the previous calculation, and if so, tracking from the input image based on the feature point position (eye and nose position) obtained in the previous calculation. Set the search area to search for the target feature point.

この実施例では目、 鼻を追跡するため、 探索領域としては、 例えば目 の周りを囲う一定の大きさの矩形領域、 鼻の周りを囲う一定の大きさの 矩形領域を設定する。  In this embodiment, in order to track the eyes and the nose, as the search area, for example, a rectangular area having a certain size surrounding the eyes and a rectangular area having a certain size surrounding the nose are set.

この探索の対象となる矩形領域 (探索矩形領域と称す) は通常、 前回 の計算により得られた特徴点位置を中心とする一定の大きさの矩形領域 である。  The rectangular area to be searched (referred to as a search rectangular area) is usually a rectangular area of a certain size centered on the feature point position obtained by the previous calculation.

この矩形領域の大きさは入力画像を取り込む時間間隔、 特徴点の検出 を行う時間間隔、 人物が顔を動かす速度などによって適宜決めればよい また、 前回の計算において特徴点位置が検出されていない場合、 つま り特徴点の検出を初めて行う場合、 例えば入力画像を二値化する手段 ( 図示せず) を用いて画像を二値化し、 顔構造の一般的知識を用いて鼻孔 位置を検出してから目の探索矩形領域を設定する。  The size of this rectangular area may be determined as appropriate according to the time interval for capturing the input image, the time interval for detecting the feature points, the speed at which the person moves the face, etc. Also, when the feature point position is not detected in the previous calculation In other words, when detecting feature points for the first time, for example, binarizing the input image using a means (not shown) for binarizing the input image, and detecting the position of the nostril using general knowledge of the face structure Set the eye search rectangular area from.

この具体的な設定方法ついては後出の実施例 2で詳細に説明する。 このように入力画像中から探索矩形領域を設定することにより、 入力 画像全体を特徴点を探索するための探索領域とするのに比べ探索に要す る時間をより少なくすることができる。 The specific setting method will be described in detail in a second embodiment described later. By setting the search rectangular area in the input image in this way, the time required for the search can be reduced as compared to setting the entire input image as a search area for searching for a feature point.

次に、 ステップ 2において、 ステップ 1で設定した探索矩形領域内に 不変特徴点を探索するための探索領域を設定する。設定においては例え ばステップ 1において用いた特徴点を中心位置とする矩形領域であって ステップ 1で設定した矩形領域よりも小さな矩形領域を設定するもので ある。  Next, in step 2, a search area for searching for invariant feature points is set in the search rectangular area set in step 1. In the setting, for example, a rectangular area centered on the feature point used in step 1 and smaller than the rectangular area set in step 1 is set.

不変特徴点の位置は、 通常、 特定パタン内に存在するものであり、 そ の領域の大きさも特定ノ タンのそれに比べ小さい。  The positions of the invariant feature points usually exist within a specific pattern, and the size of the region is smaller than that of the specific pattern.

'ステップ 2により、 ステヅプ 1によつて設定された探索矩形領域内の さらに内側に不変特徴点を探索するための矩形領域を設定することによ り、 ステップ 1において設定した探索領域全体を不変特徴点の位置の探 索領域とする場合に比べ、 不変特徴点の位置をより短い時間で算出する ことができる。  'By setting a rectangular area for searching for invariant feature points further inside the search rectangular area set in step 1 in step 2, the entire search area set in step 1 is invariant feature The position of the invariant feature point can be calculated in a shorter time than in the case where the search area is the position of the point.

この実施例では、 不変特徴点を瞳孔としているため、 瞳孔に対応する 形状 (円形) を設定し、 特徴点を中心としてこの形状を含む矩形領域を 不変特徴点を探索する探索領域と設定するものである。  In this embodiment, since the invariant feature point is a pupil, a shape (circle) corresponding to the pupil is set, and a rectangular area including the shape around the feature point is set as a search area for searching for the invariant feature point. It is.

この不変特徴点を探索するための矩形領域の大きさは入力画像を取り 込む時間間隔、 不変特徴点の検出を行う時間間隔、 瞳孔が動く速度など によって適宜決めればよい。  The size of the rectangular area for searching for the invariant feature points may be appropriately determined according to the time interval for capturing the input image, the time interval for detecting the invariant feature points, the speed at which the pupil moves, and the like.

あるいはステツプ 2における探索領域の設定は、 前回の計算により得 た不変特徴点位置と特定パ夕ン位置との相対位置関係を基に、 最も不変 特徴点が存在する確率の高い領域を統計などの手法を用いて見積もり、 不変特徴点が存在する確率が十分高い領域を含む最小矩形領域を不変特 徴点を探索するための探索領域としてもよい。 次に、 ステップ 3において、 ステップ 2で設定した不変特徴点を探索 するための探索領域の中から不変特徴点を検出するとともに不変特徴点 の位置を出力する。 Alternatively, in step 2, the search area is set based on the relative positional relationship between the invariant feature point position and the specific pattern position obtained by the previous calculation, and the area with the highest probability of the invariant feature point being present, such as statistics, is calculated. The minimum rectangular area, which is estimated using a method and includes an area where the probability that an invariant feature point exists is sufficiently high, may be used as a search area for searching for an invariant feature point. Next, in step 3, invariant feature points are detected from the search area for searching for the invariant feature points set in step 2, and the positions of the invariant feature points are output.

これは、 不変特徴点に関する知識、 例えば予め記憶した不変特徴点の 形状 (瞳孔であれば瞳孔の形状、 鼻孔であれば鼻孔の形状) に基き行う 。 より具体的には例えば予め記憶した不変特徴点に対応する形状と一致 するものがステップ 2で設定した探索領域のどこに存在するのかを調べ ることにより、 不変特徴点を検出し、 検出した形状の中心位置を算出し This is performed based on knowledge about invariant feature points, for example, the shape of invariant feature points stored in advance (pupil shape for a pupil, nostril shape for a nostril). More specifically, for example, by examining where in the search area set in step 2 a shape that matches the shape corresponding to the invariant feature point stored in advance, the invariant feature point is detected, and the detected shape Calculate the center position

、 これを不変特徴点の位置として出力するものである。 This is output as the position of the invariant feature point.

またこのとき、 不変特 j数点の形状がわかっており、 その内外部の輝度 分布が異なっている個所を不変特徴点として選択しているのであれば、 分離度フィル夕 (図示せず) をステップ 2において設定した探索領域内 の画像にかければより精度高く不変特徴点を検出することが可能となる  Also, at this time, if the shape of the invariant feature j points is known, and a portion where the luminance distribution inside and outside is different is selected as the invariant feature point, the separation degree filter (not shown) is used. The invariant feature points can be detected with higher accuracy if the image in the search area set in step 2 is used.

不変特徴点を瞳孔とした場合、 瞬きといったように目をつぶろうとす る間の動作において、 瞳孔の一部、 または全部が隠されてしまう。 このため、 予め記憶した不変特徴点の形状と、 実際の入力画像内の瞳 孔の形状とが異なるため上記で述べた不変特徴点が検出がなされない。If the invariant feature point is the pupil, part or all of the pupil will be hidden during the movement while closing the eyes, such as blinking. Therefore, since the shape of the invariant feature point stored in advance is different from the shape of the pupil in the actual input image, the invariant feature point described above is not detected.

.従って、 ステップ 2において設定し探索領域内に不変特徴点が検出で きなかったような場合、 瞬きの最中であると判断し、 不変特徴点の位置 を検出しないようにするなどの処理を追カ卩することでステップ 3におけ る検出の信頼性を向上することができる。 Therefore, if an invariant feature point cannot be detected in the search area set in step 2, it is determined that blinking is in progress, and processing such as not detecting the position of the invariant feature point is performed. The additional detection improves the reliability of detection in step 3.

以上により瞳孔や鼻孔などの不変特徴点が存在する位置を画像中から 検出することができる。  As described above, a position where an invariant feature point such as a pupil or a nostril exists can be detected from an image.

次に、 ステヅプ 2およびステップ 3を実行した直後、 またはステップ Then, immediately after performing Step 2 and Step 3, or

2およびステップ 3を実行する直前、 またはステツプ 2およびステヅプ 3を実行中に、 ステップ 4を行う。 Immediately before performing Step 2 and Step 3, or Step 2 and Step Perform step 4 while performing step 3.

ステップ 4は具体的には、 ステップ 1で設定した探索矩形領域の中か ら特定パ夕ンの検出を行うとともに、 特定パ夕ンの中心位置を出力する ものである。 つまり前回の計算により得られた特定パタンをテンプレー ト画像パタンとし、 パタンマッチングなどの手法を用いて探索領域の中 から特定パタンを検出するとともに、 検出した特定パ夕ンの中心位置を 今回の入力画像中に含まれる特定パタンの位置として出力する。  Step 4 specifically detects a specific pattern from the search rectangular area set in step 1 and outputs the center position of the specific pattern. In other words, the specific pattern obtained by the previous calculation is used as a template image pattern, a specific pattern is detected from the search area using a method such as pattern matching, and the center position of the detected specific pattern is input this time. Output as the position of the specific pattern included in the image.

パタンマッチとは、 例えば特定ノ タンと入力画像の対応する画素値の 差の総和が小さくなる位置を見つけることにより、 入力画像の探索領域 中に存在する特定パタンを検出する手法である。  Pattern matching is a method of detecting a specific pattern existing in a search area of an input image by finding a position where the sum of differences between a specific pattern and a corresponding pixel value of the input image is small.

以上により、 目と鼻の特定パタンが存在する位置を画像中から検出す ることができる。  As described above, the position where the specific pattern of the eyes and the nose exists can be detected from the image.

さて、 ステップ 3により得られる不変特徴点の位置とステップ 4によ り得られる特定パタンの位置は、 理想的には一致するはずであるが、 現 実にはそうならない場合がある。  By the way, the position of the invariant feature point obtained in step 3 and the position of the specific pattern obtained in step 4 should ideally coincide with each other, but in reality this may not be the case.

したがって、 これらの出力を用い統合することにより、 両者の位置か ら特徴点の真の位置を求める必要がある。  Therefore, it is necessary to obtain the true position of the feature point from both positions by integrating using these outputs.

ここではステップ 5において、 例えばステップ 3とステップ 4により それそれ得られる位置の中点を採用する。  Here, in step 5, for example, the midpoint of the position obtained by steps 3 and 4 is adopted.

こうすれば一方のステップ 3 (または 4 ) より得られた一方の位置の 誤差が大きくても他方のステップ 4 (または 3 ) より得られた他方の位 置が真値に近ければ、 正しい位置からあまりかけ離れていない値を得る ことができる。  In this way, even if the error of one position obtained from one step 3 (or 4) is large, if the other position obtained from the other step 4 (or 3) is close to the true value, the correct position It is possible to obtain values that are not far apart.

これにより、 両方のステップ 3 , 4より得られた位置が共に誤差が大 きいという場合を除いて安定かつロバストに特徴点の位置を得ることが できる。 なお、 ステップ 5では 2つの特徴点位置の中点を採用したが、 中点で はなくどちらか一方のステップより得られた特徴点位置に近い位置とす る重みづけを有する内分点であってもよい。 As a result, the positions of the feature points can be obtained stably and robustly unless the positions obtained from both steps 3 and 4 have large errors. Note that in step 5, the midpoint of the two feature point positions was adopted, but it was not the midpoint, but an internally divided point with a weight that is closer to the feature point position obtained from one of the steps. You may.

ステップ 5で得られた特徴点位置をステツプ 6で特徴点位置情報とし て保持し、 直後の時刻で行う特徴点位置検出に備える。 このようにして 、 顔上の特徴点位置を安定に追跡することができる。  The feature point position obtained in step 5 is stored as feature point position information in step 6, and is prepared for feature point position detection performed immediately after. In this way, the positions of the feature points on the face can be stably tracked.

更には、 ステップ 2において設定した探索領域内に不変特徴点が検出 できなかったような場合、 瞬きの最中であると判断し、 不変特徴点の位 置を検出しないようにするなどの処理を追加した場合、 この出力がなさ れると、 スッテツプ 4により出力される特定ノ 夕ンの位置を特徴点の位 置とすることにより、 瞬き等により不変得徴点の位置が検出できない場 合であっても、 特徴点を追跡することができる。  Furthermore, if no invariant feature point is detected in the search area set in step 2, it is determined that blinking is in progress, and processing such as not detecting the position of the invariant feature point is performed. If this output is made, the position of the specific nodal output in step 4 is set as the position of the feature point, so that the position of the invariant score point cannot be detected due to blinking or the like. Even feature points can be tracked.

ステップ 5の出力はステップ 6、 ステップ 7、 ステップ 9で用いる。 ステップ 6ではステップ 5で出力される特徴点の位置の情報を保持、 更新し、 次回の検出におけるステップ 1の特徴点の位置とする。  The output of step 5 is used in step 6, step 7, and step 9. In step 6, the information on the position of the feature point output in step 5 is retained and updated, and is used as the position of the feature point in step 1 in the next detection.

このようにすることにより、 顔が動いたとしてもステップ 1における 探索領域を適切な位置に設定することができるのである。  In this way, even if the face moves, the search area in step 1 can be set to an appropriate position.

ステヅプアでは、 ステップ 5で得られた特徴点の位置を基にその周辺 の画像情報から特定パタン (テンプレート) を取得する。  In the step, based on the positions of the feature points obtained in step 5, a specific pattern (template) is obtained from the surrounding image information.

ステップ 8では取得した特定パタンを新たな特定パ夕ンとし、 この更 新した特定パタンを次回の特定パタンの検出に用いる特定パタンとする このようにすることにより特定パタンが時間的に変化する場合であつ ても、 適切にこれを検出することができる。  In step 8, the acquired specific pattern is used as a new specific pattern, and the updated specific pattern is used as a specific pattern used for detecting the next specific pattern. However, this can be detected appropriately.

ステップ 9において、 ステップ 5で得られた特徴点位置を出力して処 理を終了する。 以上説明したような顔の特徴点追跡動作は撮像手段 1が画像を時系列 に取り込む時間間隔毎、 例えば 1 / 3 0秒毎に行われる。 In step 9, the feature point position obtained in step 5 is output, and the process ends. The face feature point tracking operation as described above is performed at every time interval at which the image capturing means 1 takes in images in time series, for example, at every 1/30 second.

なお、 第 2図に示したフローチャート図は一例であり、 各ステップの 入出力関係が適正でさえあれば別のフローチャート図であってもよい。 以上説明したように、 本実施例では、 不変特徴点位置検出手段 2によ り検出された不変特徴点の位置に基づく瞳孔や鼻孔の中心点の位置と、 特定パタン位置検出手段 3でパタンマッチにより検出された特定パタン の位置から導き出される目や鼻の中心点の位置とが異なる場合にも、 出 力統合手段 4により両検出手段 2、 3からの検出値を加味した値に調整 するので、 真の位置とかけ離れた特徴点位置が出力されることはなく、 目や鼻などの顏上の特徴点を口パストに追跡することができる。  Note that the flowchart shown in FIG. 2 is an example, and another flowchart may be used as long as the input / output relationship of each step is appropriate. As described above, in the present embodiment, the position of the center point of the pupil or the nostril based on the position of the invariant feature point detected by the invariant feature point position detection means 2 and the pattern matching by the specific pattern position detection means 3 Even if the position of the center point of the eye or nose derived from the position of the specific pattern detected by the above is different from the position of the center point of the eye or nose, the output integration means 4 adjusts to the value taking into account the detection values from both detection means 2 and 3. However, feature points far from the true position are not output, and feature points on the face, such as eyes and nose, can be tracked in the mouth past.

また、 不変特徴点位置検出手段 2と特定パタン位置検出手段 3とはそ れそれ独立して不変特徴点位置と特定パタン位置とを検出するので、 ど ちらか一方の検出手段 2 (または 3 ) による検出が不可能である場合に も他方の検出手段 3 (または 2 ) による検出結果が得られるので、 これ を出力位置統合手段 4で特徴点位置として出力することにより目や鼻な どの顔上の特徴点を安定に追跡することができる。  In addition, since the invariant feature point position detecting means 2 and the specific pattern position detecting means 3 respectively detect the invariant feature point position and the specific pattern position independently, either one of the detecting means 2 (or 3) is used. Even when detection by the other means is impossible, the detection result by the other detection means 3 (or 2) can be obtained. Can be tracked stably.

また、 本実施例では出力位置統合手段 4によって得られた特徴点位置 、 この特徴点位置に基づいて取得した特定パタンを更新し、 次回の不変 特徴点位置検出手段 2または特定パタン位置検出手段 3による検出に用 いるので、 特定パタン位置検出手段 4は、 人物の顏の特徴となる部分が 時間に応じて変化するような場合であっても、 この変化に追従して検出 することが可能となるため、 ロバストな追跡が可能となる。  In this embodiment, the characteristic point position obtained by the output position integrating means 4 and the specific pattern acquired based on the characteristic point position are updated, and the next invariable characteristic point position detecting means 2 or the specific pattern position detecting means 3 are updated. Therefore, the specific pattern position detecting means 4 can detect the characteristic portion of the person's face following the change even when the characteristic portion changes with time. Therefore, robust tracking becomes possible.

更に、 人物の顔の画像から探索領域を設定し、 設定した探索領域中か ら特定パタンの位置を検出するため、 人物の顔の画像全てを探索領域と する場合に比べて特定パタンの検出時間を短縮することも可能である。 実施例 2 . , Furthermore, since the search area is set from the image of the person's face and the position of the specific pattern is detected from the set search area, the detection time of the specific pattern is shorter than when the entire image of the person's face is set as the search area. Can also be shortened. Example 2.

第 3図は本発明の実施例 2による顔の特徴点追跡装置の構成を示すプ ロック図、 第 4図は本発明の実施例 2による顔の特徴点追跡装置の動作 を説明するためのフローチャート図である。  FIG. 3 is a block diagram showing the configuration of a face feature point tracking device according to the second embodiment of the present invention, and FIG. 4 is a flowchart for explaining the operation of the face feature point tracking device according to the second embodiment of the present invention. FIG.

第 3図において、 5は撮像手段 1によって得られた顔の画像から鼻の 位置を検出する手段、 6は鼻の位置検出手段 5によつて検出された鼻の 位置を用いて、 撮像手段 1によって得られた顔の画像から目の位置を検 出する手段であり、 これら鼻の位置検出手段 5および目の位置検出手段 6は共にコンピュータにより実現される。  In FIG. 3, reference numeral 5 denotes a means for detecting the position of the nose from the image of the face obtained by the imaging means 1, and reference numeral 6 denotes a position of the nose detected by the nose position detection means 5, and the imaging means 1 is used. This is means for detecting the position of the eyes from the image of the face obtained by the above. The nose position detecting means 5 and the eye position detecting means 6 are both realized by a computer.

上記のように構成された顔の特徴点追跡装置は、 例えば第 4図のフロ —チャート図に示す順序で動作させることができる。  The facial feature point tracking apparatus configured as described above can be operated, for example, in the order shown in the flowchart of FIG.

以下、 第 4図のフローチャートに従って、 特徴点位置として目と鼻の 位置を追跡する場合について説明する。  Hereinafter, a case where the positions of the eyes and the nose are tracked as feature point positions will be described with reference to the flowchart of FIG.

もし、 前回本装置を利用し、 前回に特徴点位置を得ているのであれば 、 その特徴点位置を用いて、 実施例 1と同様に撮像手段 1、 不変特徴手 点位置検出手段 2、 特定パタン位置検出手段 3、 出力位置統合手段 4に より顏の特徴点を追跡する。  If the feature point position has been previously obtained by using the present apparatus, the image capturing means 1, the invariant feature point position detecting means 2, and the identification are used in the same manner as in the first embodiment. The feature points of the face are tracked by the pattern position detecting means 3 and the output position integrating means 4.

また、 前回に特徴点位置を得ていなければ、 鼻の位置検出手段 5、 目 の位置検出手段 6により両目と鼻の位置を検出し、 この検出結果に基づ き特徴点位置を算出する。  If the feature point positions have not been obtained previously, the positions of both eyes and the nose are detected by the nose position detecting means 5 and the eye position detecting means 6, and the feature point positions are calculated based on the detection results.

第 4図は実施例 2の顔の特徴点追跡装置の動作を説明するためのフロ —チャート図である。  FIG. 4 is a flowchart for explaining the operation of the facial feature point tracking apparatus according to the second embodiment.

第 4図において、 まず、 ステップ 1 1において、 前回に特徴点位置を 得ているかどうかの判断を行う。  In FIG. 4, first, in step 11, it is determined whether or not the feature point position has been obtained last time.

これは例えば出力位置統合手段 4が特徴点の位置を保持しているかど うかを調べればよい。 This is, for example, whether the output position integration means 4 holds the position of the feature point. You just have to check.

前回に特徴点位置を得ている場合の特徴点の検出の手順は、 実施例 1 で第 2図を用いて説明したのと同じであるので説明を省略する。  The procedure for detecting a feature point when the feature point position has been previously obtained is the same as that described in the first embodiment with reference to FIG. 2, and a description thereof will be omitted.

以下では、 前回に特徴点位置を得ていない場合の特徴点追跡の手順に ついて説明する。  The following describes the procedure for feature point tracking when the feature point position has not been obtained previously.

顏を斜め下から撮像する場合、 鼻孔は画像中で黒い領域に映る。 した がって、 鼻の領域を検出するためには、 これら 2つの黒領域を検出すれ ばよく、 比較的精度良く検出することができる。  When the face is imaged diagonally below, the nostrils appear in black areas in the image. Therefore, in order to detect the nose region, it is sufficient to detect these two black regions, and the detection can be performed with relatively high accuracy.

鼻の位置を検出するために、 ステップ 1 2により、 鼻が存在するであ ろう特定候補領域を抽出し、 この抽出した領域に対し、 特定の閾値で画 像を二値化する。  In order to detect the position of the nose, a specific candidate region where the nose is likely to be extracted is extracted in step 12 and the image is binarized with a specific threshold value for the extracted region.

なお、 上記ステップ 1 2では、 画像の明度変化に対応するために、 画 像中の最喑ピクセル値を基準にして最暗ピクセル値に特定値を加算した 値を閾値として二値化してもよい。  In step 12 above, in order to cope with a change in brightness of the image, binarization may be performed using a value obtained by adding a specific value to the darkest pixel value based on the minimum pixel value in the image as a threshold value. .

次にステップ 1 3においては、 上記ステップ 1 2によって二値化され た画像中から、 ある特定範囲内の面積を持つ領域を抽出し、 それら領域 中の任意の組み合わせの中から、 一定範囲内の距離にある左右に並んだ 2つの領域を検出する。  Next, in step 13, a region having an area within a specific range is extracted from the image binarized in step 1 2, and a certain range Detects two regions that are located on the left and right at a distance.

この 2つの領域が鼻孔であり、 鼻の位置は例えば 2つの領域の各重心 の中点で代表することができる。  These two regions are nostrils, and the position of the nose can be represented by, for example, the midpoint of each center of gravity of the two regions.

ステップ 1 4においては、 上記ステップ 1 3によって検出された鼻の 位置から、 相対的に見て目が存在する可能性のある領域を設定する。 ステップ 1 5においては、 上記ステップ 1 4によって設定された目が 存在する可能性がある領域を二値化する。  In step 14, an area where an eye is likely to be seen relatively is set from the position of the nose detected in step 13. In step 15, the region where the eyes set in step 14 are likely to exist is binarized.

ステヅプ 1 6においては、 上記ステップ 1 5によって二値化された画 像から、 ある特定範囲内の面積を持つ領域を抽出し、 それら領域中の任 意の組み合わせの中から、 一定範囲内の距離にある左右に並んだ 2つの 領域を瞳孔として検出するとともにそれそれの瞳孔の中心位置を特徴点 の位置とする。 In step 16, a region having an area within a specific range is extracted from the image binarized in step 15, and an arbitrary part of the region is extracted. Among the desired combinations, two regions that are arranged on the left and right within a certain range are detected as pupils, and the center position of each pupil is set as a feature point position.

以上により、 鼻と両目の位置が検出され、 これらの位置を特徴点位置 としてステップ 9により出力する。 また、 ステップ 6により特徴点位置 を保持し、 次回の追跡に備える。 さらに、 ステップ 7により特徴点位置 を中心に所定範囲を切り出し、 これを特定パ夕ンとして取得する。 以上説明したように、 本実施例では、 追跡開始時や追跡に失敗した場 合などのように、 前回に特徴点位置を得ていない場合、 比較的検出しや すい鼻の位置をまず検出し、 これに基づいて他の特徴点である例えば目 の位置を検出するので、 安定で口バストな追跡が可能となる。  As described above, the positions of the nose and both eyes are detected, and these positions are output as feature point positions in step 9. In addition, the feature point positions are retained in step 6 to prepare for the next tracking. Further, in step 7, a predetermined range is cut out around the feature point position, and this is obtained as a specific pattern. As described above, in this embodiment, when the feature point position has not been obtained previously, such as when tracking is started or when tracking fails, the position of the nose that is relatively easy to detect is detected first. However, based on this, other feature points, for example, the positions of eyes are detected, so that stable and mouth-to-mouth tracking can be performed.

なお、 第 4図に示したフローチャート図は一例であり、 各ステップの 入出力関係が適正でさえあれば別のフローチャート図であってもよい。 なお、 上記実施例 1および実施例 2では特徴点位置として目と鼻の位 置を追跡する場合について説明したが、 これに限るものではなく、 例え ば口の位置などであってもよい。 この場合にも、 実施例 2においては、 比較的検出しやすい鼻の位置をまず検出し、 これに基づいて口の位置を 検出する。 .  Note that the flowchart shown in FIG. 4 is an example, and another flowchart may be used as long as the input / output relationship of each step is appropriate. In the first and second embodiments, the case where the positions of the eyes and the nose are tracked as the feature point positions has been described. However, the present invention is not limited to this. For example, the positions of the mouth may be used. Also in this case, in the second embodiment, the position of the nose which is relatively easy to detect is first detected, and the position of the mouth is detected based on this. .

このように比較的検出が容易な鼻の位置をまず検出し、 鼻と鼻以外の 顔の特徴的な部分との相対的な位置関係に基き、 鼻以外の人物の顏の特 徴的な部分の位置を検出するようにしたので、 鼻以外の人物の顔の特徴 点の位置を取り込まれた画像から直接検出する場合に比較して検出が容 易になる o 実施例 3 .  In this way, the position of the nose, which is relatively easy to detect, is first detected, and the characteristic portion of the face of the person other than the nose is determined based on the relative positional relationship between the nose and the characteristic portion of the face other than the nose. O The position of the feature point of a person's face other than the nose is easier to detect than if it is detected directly from the captured image.o Example 3.

第 5図は本発明の実施例 3による顔の特徴点追跡装置の構成を示すブ ロック図、 第 6図は本発明の実施例 3による顔の特徴点追跡装置の動作 を説明するためのフローチャート図である。 FIG. 5 is a block diagram showing a configuration of a facial feature point tracking apparatus according to Embodiment 3 of the present invention. FIG. 6 is a flowchart for explaining the operation of the facial feature point tracking device according to the third embodiment of the present invention.

第 5図において、 7は動き状態検出手段に相当する動き幅検出手段、 8は参照動き幅、 9は参照追跡時間、 1 0は安定追跡判定手段、 1 1は 参照動き幅設定手段、 1 2は参照追跡時間設定手段であり、 これらは共 にコンピュータにより実現され、 これらにより顔の安定度を追跡する安 定度追跡手段を構成している。 また、 参照動き幅設定手段 1 1と参照追 跡時間設定手段 1 2により設定手段を構成している。  In FIG. 5, 7 is a movement width detecting means corresponding to the movement state detecting means, 8 is a reference movement width, 9 is a reference tracking time, 10 is a stable tracking determination means, 11 is a reference movement width setting means, and 1 2 Are reference tracking time setting means, which are both realized by a computer and constitute a stability tracking means for tracking the stability of the face. The reference movement width setting means 11 and the reference tracking time setting means 12 constitute a setting means.

安定度追跡判定手段 1 0は、 出力位置統合手段 4の出力と、 参照動き 幅 8および参照追跡時間 9を用いて動き幅検出手段 7により特徴点位置 の動き幅を検出し、 検出結果から安定追跡判定手段 1 0により安定追跡 かどうかを判断し、 結果を出力する。  The stability tracking determination means 10 uses the output of the output position integrating means 4, the reference movement width 8 and the reference tracking time 9 to detect the movement width of the feature point position by the movement width detection means 7, and based on the detection result, The tracking determination means 10 determines whether stable tracking is performed and outputs the result.

安定追跡判定手段 1 0の結果を受けて、 参照動き幅設定手段 1 1が参 照動き幅 8を更新し、 参照追跡時間設定手段 1 2が参照追跡時間 9を更 新する。  In response to the result of the stable tracking determination means 10, the reference movement width setting means 11 updates the reference movement width 8, and the reference tracking time setting means 12 updates the reference tracking time 9.

以下では、 例えば図 2に示したフローチャート図を用い特徴点の位置 が検出されたとして、 安定度追跡手段における処理の流れを説明する。 まず最初にステップ 2 1において、 本安定度追跡手段を初めて利用す るかどうかを判断し、 もし初めてであれば、 ステップ 2 2において、 参 照追跡時間を予め定められた特定の固 値に設定し、 参照動き幅を予め 定められた特定の固定値に設定する。 この後、 参照追跡時間に相当する 時間の間、 特徴点の位置の変化を調べる。  In the following, the flow of processing in the stability tracking means will be described assuming that the position of a feature point has been detected, for example, using the flowchart shown in FIG. First, in step 21, it is determined whether or not this stability tracking method is to be used for the first time, and if so, in step 22, the reference tracking time is set to a predetermined specific fixed value. Then, the reference motion width is set to a predetermined specific fixed value. Then, the change of the position of the feature point is examined for a time corresponding to the reference tracking time.

また、 特徴点の位置の動き幅をたとえば参照動き幅よりも小さな値に 設定しておく。  Also, the movement width of the position of the feature point is set to a value smaller than, for example, the reference movement width.

ステヅプ 2 1によって初めてでないと判断された場合には、 ステップ 2 3により、 前回の計算により得られた特徴点の位置と今回の計算によ り得られた特徴点の位置とから特徴点位置の動き幅を計算する。 このと き、 たとえば動き幅を前回特徴点位置からの特徴点位置の相対距離であ ると定義すれば、 特徴点位置の動き幅は位置の差分を表すべクトルの長 さとなる。 If it is determined in step 21 that this is not the first time, in step 23, the positions of the feature points obtained by the previous calculation and the current calculation are used. The movement width of the feature point position is calculated from the obtained feature point position. At this time, for example, if the movement width is defined as the relative distance of the feature point position from the previous feature point position, the movement width of the feature point position becomes the length of the vector representing the difference between the positions.

次にステップ 2 4において、 上記ステップ 2 2またはステップ 2 3に より設定された特徴点位置の動き幅が参照動き幅よりも小さいかどうか を調べる。  Next, in step 24, it is checked whether or not the motion width of the feature point position set in step 22 or step 23 is smaller than the reference motion width.

その結果、 動き幅が参照動き幅と同じかまたはそれよりも大きい場合 には、 人物が顔を動かしており、 一定の状態にないことを意味する。 こ の場合ステップ 2 5に進み、 ステップ 2 5によって追跡時間を 0にし、 ステップ 2 6により安定追跡ではない (安定した状態でない) 旨を出力 し、 ステップ 2 1に戻る。  As a result, if the motion width is equal to or larger than the reference motion width, it means that the person is moving the face and is not in a constant state. In this case, the process proceeds to step 25, the tracking time is set to 0 in step 25, the fact that the tracking is not stable (not stable) is output in step 26, and the process returns to step 21.

ステップ 2 4において動き幅が参照動き幅よりも小さい場合にはステ ップ 2 7によってステップ 2 4における特徴点の位置の変化に要した時 間、 すなわち前回の特徴点位置を算出した時刻と今回の特徴点位置を算 出した時刻との差を追跡時間に累積する。  If the motion width is smaller than the reference motion width in step 24, the time required to change the position of the feature point in step 24 in step 27, that is, the time when the previous feature point position was calculated and the current time The difference from the time when the feature point position was calculated is accumulated in the tracking time.

次に、 ステップ 2 8によりこの追跡時間が参照追跡時間よりも長いか どうかを調べる。  Next, in step 28, it is checked whether this tracking time is longer than the reference tracking time.

その結果、 追跡時間が参照追跡時間と同じかまたはそれよりも短いと 判断された場合には、 ステップ 2 6により安定追跡ではないと出力し、 ステップ 2 1に戻る。  As a result, if it is determined that the tracking time is equal to or shorter than the reference tracking time, it is output in step 26 that the tracking is not stable tracking, and the process returns to step 21.

ステップ 2 8において参照追跡時間よりも追跡時間が長いと判断され た場合にはステップ 2 9で用いた追跡時間を新たな参照追跡時間とし、 ステップ 3 0においてこの更新された参照追跡時間を保持する。  If it is determined in step 28 that the tracking time is longer than the reference tracking time, the tracking time used in step 29 is set as a new reference tracking time, and the updated reference tracking time is held in step 30. .

次に、 ステヅプ 3 1において追跡時間として累積された各時間間隔に 対応する特徴点の位置の変化 (動き幅) のうちの最小のものを新たな参 照動き幅とし、 ステップ 3 2においてこの更新された参照動き幅を保持 し、 ステップ 3 3において安定追跡である旨を出力し、 ステップ 2 1に 民る。 Next, the smallest one of the change (movement width) of the position of the feature point corresponding to each time interval accumulated as the tracking time in step 31 is newly added. In step 32, the updated reference motion width is held, and in step 33, the fact that the tracking is stable is output.

このように構成することにより、 人物の顔の動的な変化を検出するだ けではなく、 動かない状態を検出することができるようになるばかりか 、 安定追跡である (動かない) と判断するために用いる所定の閾値に対 応する参照動き幅、 予め設定した値に対応する参照追跡時間は個人の特 性に応じた値に収束するため、 使用するほど個人の特性に応じた判定が 可能となる。 産業上の利用可能性  With this configuration, it is possible not only to detect dynamic changes in the face of a person, but also to detect an immovable state, and to judge that tracking is stable (does not move). The reference movement width corresponding to the predetermined threshold value used for reference and the reference tracking time corresponding to the preset value converge to a value according to the characteristics of the individual, so that the more it is used, the more the determination according to the characteristics of the individual is possible Becomes Industrial applicability

例えば顏の 3次元方向を追跡するための顔の特徴点追跡に用いられ、 安定かつロバストな顔の 3次元方向追跡が可能となる。 また、 顔の個人 識別のための特徴点検出および追跡に用いられ、 安定かつ口バストな顏 の個人識別が可能となる。  For example, it is used for tracking feature points of a face to track the 3D direction of the face, and enables stable and robust tracking of the 3D direction of the face. It is also used for feature point detection and tracking for face individual identification, and enables stable and mouth-bust personal identification of faces.

Claims

請 求 の 範 囲 The scope of the claims 1 . 時系列に取り込まれる人物の顔の画像から前記人物の顔の特徴的な 部分の位置を逐次検出し、 追跡する装置であって、 1. A device that sequentially detects and tracks the position of a characteristic portion of a person's face from an image of the person's face captured in time series, 予め記憶された前記人物の顔の特徴的な部分を含む画像パタンを特定 パ夕ンとし、 前記特定パ夕ンもしくは前記特定パタンに近い画像パ夕ン が前記取り込まれた画像中のどの位置に存在するかを検出する特定パ夕 ン位置検出手段、  An image pattern including a characteristic portion of the person's face stored in advance is specified as a specific pattern, and the specific pattern or an image pattern close to the specific pattern is located at any position in the captured image. Specific pattern position detecting means for detecting the presence 前記人物の顔の特徴的な部分に含まれる不変特徴点の位置を検出する 不変特徴点位置検出手段、 および  Invariant feature point position detecting means for detecting a position of an invariant feature point included in a characteristic portion of the human face; and 前記不変特徴点位置検出手段からの出力と前記特定パ夕ン位置検出手 段からの出力とにより前記取り込まれた画像中にある前記人物の顔の特 徴的な部分の位置を検出するとともに、 前記検出した人物の顔の特徴的 な部分の画像を新たな特定パ夕ンとして記憶する出力位置統合手段を備 えたことを特徴とする顔の特徴点追跡装置。  Detecting a position of a characteristic portion of the person's face in the captured image based on an output from the invariant feature point position detecting means and an output from the specific pattern position detecting means; A feature point tracking device for a face, comprising output position integration means for storing an image of a characteristic portion of the detected face of a person as a new specific pattern. 2 . 請求の範囲第 1項に記載の顔の特徴点追跡装置において、  2. The facial feature point tracking device according to claim 1, 取り込まれた人物の顔の画像から鼻の位置を検出する鼻の位置検出手 段、 および  A nose position detecting means for detecting the nose position from the captured image of the person's face, and 前記鼻の位置検出手段により検出された鼻の位置から前記鼻以外の人 物の顔の特徴的な部分の位置を検出する特徴点検出手段を備えたことを 特徴とする顔の特徴点追跡装置。  A feature point tracking device comprising: feature point detection means for detecting the position of a characteristic portion of a human face other than the nose from the nose position detected by the nose position detection means. . 3 . 請求の範囲第 1項に記載の顔の特徴点追跡装置において、  3. The facial feature point tracking device according to claim 1, 出力位置統合手段から得られた人物の顔の特徴的な部分の位置の変化 を調べるとともに、 上記位置の変化が所定の閾値よりも小さければ位置 の変化に要した時間を累積する動き状態検出手段、 および  In addition to examining the change in the position of the characteristic portion of the person's face obtained from the output position integrating means, if the change in the position is smaller than a predetermined threshold, the movement state detecting means accumulating the time required for the change in position , and 累積した時間が予め設定した時間よりも大きくなると、 予め設定した 時間を前記累積した時間に更新するとともに、 前記所定の閾値を前記累 積した時間毎に変化した位置の変化量うちの最小値に更新する設定手段 を備えたことを特徴とする顔の特徴点追跡装置。 If the accumulated time is longer than the preset time, the preset Setting means for updating the time to the accumulated time and updating the predetermined threshold value to the minimum value of the amount of change in the position changed for each accumulated time. Tracking device.
PCT/JP2000/004798 2000-07-17 2000-07-17 Device for tracking feature point on face Ceased WO2002007096A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
PCT/JP2000/004798 WO2002007096A1 (en) 2000-07-17 2000-07-17 Device for tracking feature point on face

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/JP2000/004798 WO2002007096A1 (en) 2000-07-17 2000-07-17 Device for tracking feature point on face

Publications (1)

Publication Number Publication Date
WO2002007096A1 true WO2002007096A1 (en) 2002-01-24

Family

ID=11736270

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2000/004798 Ceased WO2002007096A1 (en) 2000-07-17 2000-07-17 Device for tracking feature point on face

Country Status (1)

Country Link
WO (1) WO2002007096A1 (en)

Cited By (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2005078311A (en) * 2003-08-29 2005-03-24 Fujitsu Ltd Facial part tracking device, eye state determination device, and computer program
JP2007212430A (en) * 2006-08-07 2007-08-23 Kurabo Ind Ltd Photogrammetry apparatus and photogrammetry system
JP2008501172A (en) * 2004-05-28 2008-01-17 ソニー・ユナイテッド・キングダム・リミテッド Image comparison method
JP2010026804A (en) * 2008-07-18 2010-02-04 Olympus Corp Image processing device, image processing program and image processing method
WO2012032747A1 (en) * 2010-09-06 2012-03-15 日本電気株式会社 Feature point selecting system, feature point selecting method, feature point selecting program
JP2012254221A (en) * 2011-06-09 2012-12-27 Canon Inc Image processing apparatus, method for controlling the same, and program
JP2013179687A (en) * 2013-05-22 2013-09-09 Yoichiro Ito Authentication method and system by using connection line generated from optical information difference
JP2013225316A (en) * 2013-05-22 2013-10-31 Yoichiro Ito Method and system for extracting optical information difference for individual difference authentication
US9008360B2 (en) 2010-02-12 2015-04-14 Yoichiro Ito Authentication system, and method for registering and matching authentication information
JP2017027602A (en) * 2015-07-24 2017-02-02 株式会社リコー Object tracking method and device
KR20210060554A (en) * 2018-12-10 2021-05-26 텐센트 테크놀로지(센젠) 컴퍼니 리미티드 Face landmark detection method and device, computer device, and storage medium

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH04174309A (en) * 1990-11-07 1992-06-22 Nissan Motor Co Ltd Driver's eye position detecting apparatus and condition detecting apparatus
JPH06337998A (en) * 1993-05-31 1994-12-06 Nec Corp Vehicle detector, vehicle tracking device and vehicle monitoring device
JPH08287216A (en) * 1995-04-18 1996-11-01 Sanyo Electric Co Ltd In-face position recognizing method
JPH1139469A (en) * 1997-07-24 1999-02-12 Mitsubishi Electric Corp Face image processing device
US5926251A (en) * 1997-08-12 1999-07-20 Mitsubishi Denki Kabushiki Kaisha Eye image tracking apparatus
JP2000148977A (en) * 1998-11-10 2000-05-30 Meidensha Corp Vehicle detecting method, recording medium with software realizing the same recorded therein and device for realizing method

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH04174309A (en) * 1990-11-07 1992-06-22 Nissan Motor Co Ltd Driver's eye position detecting apparatus and condition detecting apparatus
JPH06337998A (en) * 1993-05-31 1994-12-06 Nec Corp Vehicle detector, vehicle tracking device and vehicle monitoring device
JPH08287216A (en) * 1995-04-18 1996-11-01 Sanyo Electric Co Ltd In-face position recognizing method
JPH1139469A (en) * 1997-07-24 1999-02-12 Mitsubishi Electric Corp Face image processing device
US5926251A (en) * 1997-08-12 1999-07-20 Mitsubishi Denki Kabushiki Kaisha Eye image tracking apparatus
JP2000148977A (en) * 1998-11-10 2000-05-30 Meidensha Corp Vehicle detecting method, recording medium with software realizing the same recorded therein and device for realizing method

Cited By (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2005078311A (en) * 2003-08-29 2005-03-24 Fujitsu Ltd Facial part tracking device, eye state determination device, and computer program
JP2008501172A (en) * 2004-05-28 2008-01-17 ソニー・ユナイテッド・キングダム・リミテッド Image comparison method
JP2007212430A (en) * 2006-08-07 2007-08-23 Kurabo Ind Ltd Photogrammetry apparatus and photogrammetry system
JP2010026804A (en) * 2008-07-18 2010-02-04 Olympus Corp Image processing device, image processing program and image processing method
US9008360B2 (en) 2010-02-12 2015-04-14 Yoichiro Ito Authentication system, and method for registering and matching authentication information
US9111150B2 (en) 2010-02-12 2015-08-18 Yoichiro Ito Authentication system, and method for registering and matching authentication information
WO2012032747A1 (en) * 2010-09-06 2012-03-15 日本電気株式会社 Feature point selecting system, feature point selecting method, feature point selecting program
JP2012254221A (en) * 2011-06-09 2012-12-27 Canon Inc Image processing apparatus, method for controlling the same, and program
JP2013225316A (en) * 2013-05-22 2013-10-31 Yoichiro Ito Method and system for extracting optical information difference for individual difference authentication
JP2013179687A (en) * 2013-05-22 2013-09-09 Yoichiro Ito Authentication method and system by using connection line generated from optical information difference
JP2017027602A (en) * 2015-07-24 2017-02-02 株式会社リコー Object tracking method and device
KR20210060554A (en) * 2018-12-10 2021-05-26 텐센트 테크놀로지(센젠) 컴퍼니 리미티드 Face landmark detection method and device, computer device, and storage medium
JP2022502751A (en) * 2018-12-10 2022-01-11 ▲騰▼▲訊▼科技(深▲セン▼)有限公司 Face keypoint detection method, device, computer equipment and computer program
KR102592270B1 (en) * 2018-12-10 2023-10-19 텐센트 테크놀로지(센젠) 컴퍼니 리미티드 Facial landmark detection method and apparatus, computer device, and storage medium
US11915514B2 (en) 2018-12-10 2024-02-27 Tencent Technology (Shenzhen) Company Limited Method and apparatus for detecting facial key points, computer device, and storage medium
JP7512262B2 (en) 2018-12-10 2024-07-08 ▲騰▼▲訊▼科技(深▲セン▼)有限公司 Facial keypoint detection method, device, computer device and computer program

Similar Documents

Publication Publication Date Title
Kawaguchi et al. Detection of eyes from human faces by Hough transform and separability filter
CN107438854B (en) System and method for performing fingerprint-based user authentication using images captured by a mobile device
US7127086B2 (en) Image processing apparatus and method
JP4863423B2 (en) IDENTIFICATION SYSTEM AND METHOD USING IRIS AND COMPUTER-READABLE RECORDING MEDIUM CONTAINING IDENTIFICATION PROGRAM FOR PERFORMING THE METHOD
US8515136B2 (en) Image processing device, image device, image processing method
KR102554391B1 (en) Iris recognition based user authentication apparatus and method thereof
US12014578B2 (en) Authentication device, authentication method, and recording medium
JP2006133946A (en) Moving object recognition device
KR101626837B1 (en) Method and apparatus for convergence biometric authentication based on finger joint and finger vein
WO2002071316A1 (en) Non-contact type human iris recognition method by correction of rotated iris image
WO2002007096A1 (en) Device for tracking feature point on face
KR20020022295A (en) Device And Method For Face Recognition Using 3 Dimensional Shape Information
US11380133B2 (en) Domain adaptation-based object recognition apparatus and method
JP3823760B2 (en) Robot equipment
CN112766065A (en) Mobile terminal examinee identity authentication method, device, terminal and storage medium
KR20000035050A (en) Method for photographing and recognizing a face
JP2007140823A (en) Face matching device, face matching method and program
KR100862526B1 (en) Method and system for building face video database and face recognition method and system using same.
CN117037272A (en) Method and system for monitoring fall of old people
JP4860289B2 (en) Robot equipment
CN114821731A (en) Dynamic face recognition optimization method based on domestic CPU and OS
WO2022130616A1 (en) Authentication method, information processing device, and authentication program
JP7605218B2 (en) Image processing device, image processing method, and program
KR101718244B1 (en) Apparatus and method of processing wide angle image for recognizing face
JP2000067237A (en) Person identifying device

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A1

Designated state(s): JP US

AL Designated countries for regional patents

Kind code of ref document: A1

Designated state(s): AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE

121 Ep: the epo has been informed by wipo that ep was designated in this application
122 Ep: pct application non-entry in european phase