[go: up one dir, main page]

CN110443884B - Method and device for hand motion reconstruction - Google Patents

Method and device for hand motion reconstruction Download PDF

Info

Publication number
CN110443884B
CN110443884B CN201910646553.7A CN201910646553A CN110443884B CN 110443884 B CN110443884 B CN 110443884B CN 201910646553 A CN201910646553 A CN 201910646553A CN 110443884 B CN110443884 B CN 110443884B
Authority
CN
China
Prior art keywords
hand
depth data
depth
information
preset
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201910646553.7A
Other languages
Chinese (zh)
Other versions
CN110443884A (en
Inventor
刘烨斌
李梦成
戴琼海
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hangzhou Xinchangyuan Technology Co ltd
Original Assignee
Tsinghua University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tsinghua University filed Critical Tsinghua University
Priority to CN201910646553.7A priority Critical patent/CN110443884B/en
Publication of CN110443884A publication Critical patent/CN110443884A/en
Application granted granted Critical
Publication of CN110443884B publication Critical patent/CN110443884B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T17/00Three dimensional [3D] modelling, e.g. data description of 3D objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/50Depth or shape recovery
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/80Analysis of captured images to determine intrinsic or extrinsic camera parameters, i.e. camera calibration
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence
    • G06T2207/10021Stereoscopic video; Stereoscopic image sequence

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Computer Graphics (AREA)
  • Geometry (AREA)
  • Software Systems (AREA)
  • User Interface Of Digital Computer (AREA)
  • Processing Or Creating Images (AREA)

Abstract

本发明提出一种手部运动重建方法和装置,其中,方法包括:获取手部深度数据集,其中,手部深度数据集中的每个手部深度数据中包含手部深度图片信息及其对应的手部骨架坐标;控制预设手部模型分别根据手部深度数据集中每个手部深度数据调整姿态,获取姿态调整后的拟合手部模型的姿态参数;根据与每个手部深度数据对应的拟合手部模型的姿态参数高斯分布函数,以便于根据高斯分布函数进行手部运动重建。本发明解决了现有技术中人手的活动较为灵活复杂,与物体交互过程中常常伴随着严重遮挡的技术问题,通过同构单视点的RGB‑D数据,可以获得更加精确的三维重建结果,在较简单的硬件条件下可以获得较好的物体与手部交互重建结果。

Figure 201910646553

The present invention provides a hand motion reconstruction method and device, wherein the method includes: acquiring a hand depth data set, wherein each hand depth data in the hand depth data set includes hand depth picture information and its corresponding Hand skeleton coordinates; control the preset hand model to adjust the posture according to each hand depth data in the hand depth data set, and obtain the posture parameters of the fitted hand model after the posture adjustment; according to the corresponding hand depth data The Gaussian distribution function of the pose parameters of the fitted hand model is used to reconstruct the hand motion according to the Gaussian distribution function. The present invention solves the technical problem that the activities of the human hand are relatively flexible and complex in the prior art, and the interaction with objects is often accompanied by severe occlusion. Through the isomorphic single-view RGB-D data, more accurate three-dimensional reconstruction results can be obtained. Under simpler hardware conditions, better object-hand interaction reconstruction results can be obtained.

Figure 201910646553

Description

Hand motion reconstruction method and device
Technical Field
The invention relates to the technical field of computer vision, in particular to a hand motion reconstruction method and device.
Background
In the field of three-dimensional dynamic reconstruction, three-dimensional reconstruction of human hands is a very hot problem. Accurate dynamic hand reconstruction is very challenging due to the complex and flexible movements of human hands and often accompanied by the problem of self-occlusion. The hand is used as a main way for the physical contact between the human body and the outside, and the hand reconstruction has great application in the fields of virtual reality, man-machine interaction and the like.
Generally, a person is constantly interacting with the outside world, and the most frequent is physical interaction of the hand with the outside world. If the interaction process of the hand and the outside can be well reconstructed, the method has great help for understanding deep meaning of interaction of people and the outside, and has great effect on the fields of artificial intelligence, computer interaction, VR games and the like in industry. However, the movement of the human hand is flexible and complex, a very serious occlusion problem is often accompanied in the process of interacting with an object, and how to reconstruct the interaction between the hand and the outside from a single viewpoint is a challenging problem.
Disclosure of Invention
The present invention is directed to solving, at least to some extent, one of the technical problems in the related art.
Therefore, a first objective of the present invention is to provide a hand motion reconstruction method, which can obtain a more accurate three-dimensional reconstruction result and obtain a better object-hand interaction reconstruction result under a simpler hardware condition.
The second purpose of the invention is to provide a hand motion reconstruction device.
A third object of the invention is to propose a computer device.
A fourth object of the invention is to propose a non-transitory computer-readable storage medium.
To achieve the above object, a first aspect of the present invention provides a hand motion reconstruction method, including: acquiring a hand depth data set, wherein each hand depth data in the hand depth data set comprises hand depth picture information and hand skeleton coordinates corresponding to the hand depth picture information; controlling a preset hand model to adjust the posture according to each hand depth data in the hand depth data set respectively, and obtaining posture parameters of the fitting hand model after posture adjustment; and according to the gesture parameter Gaussian distribution function of the fitted hand model corresponding to each hand depth data, reconstructing hand motion according to the Gaussian distribution function.
The hand motion reconstruction method solves the technical problems that in the prior art, the movement of a human hand is flexible and complex, and serious shielding is often caused in the process of interacting with an object, can obtain a more accurate three-dimensional reconstruction result through isomorphic single-viewpoint RGB-D data, and can obtain a better object and hand interaction reconstruction result under a simpler hardware condition.
In an embodiment of the present invention, the controlling the preset hand model to adjust the posture according to each hand depth data in the hand depth data set, and obtain posture parameters of the fitting hand model after posture adjustment includes: acquiring first hand depth data meeting preset conditions in the hand depth data set; determining a first posture parameter of the preset hand model according to a point cloud matching algorithm and hand depth information in the first hand depth data; constructing a regression matrix of the first posture parameter through a gradient descent iterative algorithm and hand skeleton coordinates in the first hand depth data; determining second hand depth data in the hand depth data set except the first hand depth data, and calculating fitting skeleton coordinates of the preset hand model according to the regression matrix and hand skeleton coordinates corresponding to the second hand depth data; and calculating a second posture parameter of the preset hand model according to the fitting skeleton coordinate.
In an embodiment of the present invention, the acquiring first hand depth data in the hand depth data set that meets a preset condition includes: determining a reference attitude parameter corresponding to each hand depth data; calculating a difference value between the reference attitude parameter and a preset initial attitude parameter; and determining hand depth data corresponding to the difference value smaller than a preset threshold value as the first hand depth data.
In an embodiment of the present invention, the hand motion reconstruction method further includes: acquiring continuous multi-frame images of interaction between a user hand and an object based on a preset RGB-D camera; extracting first color information and first depth information of the hand of the user and second color information and second depth information of the object according to the continuous multi-frame images; acquiring the motion state information of the object according to the second color information and the second depth information; extracting depth information of a first key point of the hand of the user according to the first depth information; estimating the depth information of a second key point of the hand of the user according to the depth information of the first key point and the Gaussian distribution function; and simulating the interactive animation of the object and the hand of the user according to the depth information of the first key point, the depth information of the second key point, the first color information and the motion state information.
In an embodiment of the present invention, the estimating depth information of a second keypoint of the hand of the user according to the depth information of the first keypoint and the gaussian distribution function includes: determining estimated depth information of the second key point according to the depth information of the first key point and a preset algorithm; calculating the confidence of the estimated depth information according to the Gaussian distribution function; and detecting whether the confidence coefficient is greater than a preset threshold value, if not, modifying the estimated depth information until the confidence coefficient is greater than the preset threshold value, and taking the modified estimated depth information as the depth information of the second key point.
In an embodiment of the present invention, before the acquiring, based on the preset RGB-D camera, a continuous multi-frame image of a user's hand interacting with an object, the method includes: acquiring internal parameters and external parameters of an RGB module in the RGB-D camera; acquiring human body action images shot by the RGB module and the depth module in the RGB-D camera simultaneously; and correcting the internal parameters and the external parameters according to a preset function and the human motion images shot at the same time.
In order to achieve the above object, a second embodiment of the present invention provides a hand motion reconstruction device, including: the system comprises a first acquisition module, a second acquisition module and a third acquisition module, wherein the first acquisition module is used for acquiring a hand depth data set, and each hand depth data in the hand depth data set comprises hand depth picture information and hand skeleton coordinates corresponding to the hand depth picture information; the control module is used for controlling a preset hand model to adjust the posture according to each hand depth data in the hand depth data set; the second acquisition module is used for acquiring the posture parameters of the fitting hand model after the posture adjustment; and the reconstruction module is used for reconstructing hand motion according to the Gaussian distribution function of the attitude parameters of the fitted hand model corresponding to each hand depth data.
In an embodiment of the present invention, the second obtaining module includes: the determining unit is used for determining a first posture parameter of the preset hand model according to a point cloud matching algorithm and hand depth information in the first hand depth data; and the calculation unit is used for calculating a second posture parameter of the preset hand model according to the fitting skeleton coordinate.
According to the hand motion reconstruction device, the technical problems that in the prior art, movement of a human hand is flexible and complex and serious shielding is often caused in an interaction process with an object are solved through the first acquisition module, the control module, the second acquisition module and the reconstruction module, a more accurate three-dimensional reconstruction result can be obtained through isomorphic single-viewpoint RGB-D data, and a better object and hand interaction reconstruction result can be obtained under a simpler hardware condition.
To achieve the above object, a third aspect of the present invention provides a computer device, which includes a memory, a processor, and a computer program stored in the memory and executable on the processor, and when the processor executes the computer program, the hand motion reconstruction method according to the first aspect of the present invention is implemented.
In order to achieve the above object, a fourth aspect of the present invention provides a non-transitory computer-readable storage medium, wherein the computer program, when executed by a processor, implements the hand motion reconstruction method according to the first aspect of the above embodiments.
Additional aspects and advantages of the invention will be set forth in part in the description which follows and, in part, will be obvious from the description, or may be learned by practice of the invention.
Drawings
The foregoing and/or additional aspects and advantages of the present invention will become apparent and readily appreciated from the following description of the embodiments, taken in conjunction with the accompanying drawings of which:
fig. 1 is a schematic flow chart of a hand motion reconstruction method according to an embodiment of the present invention;
FIG. 2 is a schematic flow chart illustrating another hand motion reconstruction method according to an embodiment of the present invention;
fig. 3 is a schematic structural diagram of a hand movement reconstruction apparatus according to an embodiment of the present invention;
fig. 4 is a schematic structural diagram of another hand motion reconstruction device according to an embodiment of the present invention.
Detailed Description
Reference will now be made in detail to embodiments of the present invention, examples of which are illustrated in the accompanying drawings, wherein like or similar reference numerals refer to the same or similar elements or elements having the same or similar function throughout. The embodiments described below with reference to the drawings are illustrative and intended to be illustrative of the invention and are not to be construed as limiting the invention.
The hand motion reconstruction method and apparatus of the embodiments of the present invention are described below with reference to the drawings.
Fig. 1 is a schematic flow chart of a hand motion reconstruction method according to an embodiment of the present invention.
In view of the foregoing embodiments, an embodiment of the present invention provides a hand motion reconstruction method, as shown in fig. 1, the hand motion reconstruction method includes the following steps:
step 101, a hand depth data set is obtained, wherein each hand depth data in the hand depth data set comprises hand depth picture information and hand skeleton coordinates corresponding to the hand depth picture information.
Specifically, disclosed hand depth data sets are obtained, the data sets are defined differently, but each hand data includes a large amount of depth picture information and corresponding hand skeleton coordinates thereof, which can be a depth map of a human hand and matched hand skeleton three-dimensional coordinate marks, and in order to unify data, a MANO hand model needs to be fitted to various data sets.
And 102, controlling the preset hand model to adjust the posture according to each hand depth data in the hand depth data set respectively, and obtaining posture parameters of the fitting hand model after posture adjustment.
Specifically, control is predetermine the hand model and is concentrated every hand depth data adjustment gesture according to hand depth data respectively, and wherein the gesture can include the length, thickness etc. of finger, obtains the gesture parameter of the fitting hand model after the gesture adjustment, and wherein the gesture parameter can be the rotation angle of every finger joint, incline direction etc. and the process of specifically fitting is: acquiring first hand depth data meeting preset conditions in a hand depth data set; determining a first posture parameter of a preset hand model according to a point cloud matching algorithm and hand depth information in the first hand depth data; constructing a regression matrix of the first attitude parameter through a gradient descent iterative algorithm and hand skeleton coordinates in the first hand depth data; determining second hand depth data except the first hand depth data in the hand depth data set, and calculating fitting skeleton coordinates of a preset hand model according to the regression matrix and hand skeleton coordinates corresponding to the second hand depth data; and calculating a second posture parameter of the preset hand model according to the fitting skeleton coordinates.
Further, in obtaining first hand depth data meeting preset conditions in the hand depth data set, firstly, a reference attitude parameter corresponding to each hand depth data is determined, wherein whether the data volume of the depth data is large needs to be judged, such as the number of pixel points and the depth smoothness among the pixel points, then, a difference value between the reference attitude parameter and a preset initial attitude parameter is calculated, and then, the hand depth data corresponding to the difference value smaller than a preset threshold value is determined to be the first hand depth data.
It can be understood that as a possible implementation manner of the embodiment of the present invention, first hand depth data meeting preset conditions, i.e., several sets of data with simple actions, are selected, and an ICP point cloud matching method is used to fit a MANO hand model to the depth; secondly, learning regression matrixes of two frameworks among a few groups of data, because the data volume is small, directly solving the regression matrixes possibly underdetermined, but because some consistency exists among different hand framework definitions, 1 norm constraint can be added, the regression matrixes between the hand frameworks of the MANO model and the hand frameworks of the data set can be estimated through a gradient descent iterative algorithm, finally, the frameworks defined by the MANO of the remaining data are obtained through the regression matrixes, and the posture parameters are obtained through combining the depth images.
And 103, according to the gesture parameter Gaussian distribution function of the fitted hand model corresponding to each hand depth data, hand motion reconstruction is conducted according to the Gaussian distribution function.
Specifically, according to the gesture parameter gaussian distribution function of the fitting hand model corresponding to each hand depth data, the gesture parameters of the hand model can be analyzed by using a statistical method to obtain mixed gaussian distribution of the gesture parameters, and the mixed gaussian distribution is used as gesture prior distribution of the hand, so that hand motion reconstruction is facilitated.
After obtaining a gaussian distribution function of the pose parameters of the fitted hand model corresponding to each hand depth data, hand motion reconstruction is performed according to the gaussian distribution function, specifically, an embodiment of the present invention provides a hand motion reconstruction method, as shown in fig. 2, the method includes the following steps:
step 201, acquiring continuous multiframe images of the interaction between the hand of the user and the object based on a preset RGB-D camera.
Specifically, after calibration is completed, a preset RGB-D camera collects a continuous multi-frame image of interaction between a human hand and an object, which may be an RGB-D sequence in this example.
It should be noted that before acquiring a continuous multi-frame image of a user hand interacting with an object based on a preset RGB-D camera, internal and external parameters of an RGB module in the RGB-D camera need to be acquired; acquiring human body action images shot by a depth module in an RGB module and an RGB-D camera simultaneously; and correcting the internal parameters and the external parameters according to the preset function and the human motion image shot at the same time.
Specifically, a color (RGB) picture and a depth (depth) picture of the depth camera have a certain viewing angle difference, and the camera needs to be calibrated by using a checkerboard method or other methods. Because the Kinect camera has the function of human body identification, can utilize and carry out camera demarcation:
let the rgb camera internally refer to the following equation (1) and externally refer to the following equation (2):
Figure BDA0002133780640000051
Figure BDA0002133780640000052
the projection matrix from the three-dimensional space to the two-dimensional plane of the rgb image is shown in equation (3):
Figure BDA0002133780640000053
the Depth camera and the color camera shoot human body actions synchronously, and the corresponding relation between the RGB picture and the Depth picture can be obtained by utilizing a map function carried in the SDK of the Kinect camera, so that the camera can be calibrated through the relation.
Step 202, extracting first color information and first depth information of the hand of the user and second color information and second depth information of the object according to the continuous multi-frame images.
Specifically, image information and depth information of the hand of the user and image information and depth information of the object are extracted according to the RGB-D sequence, wherein RGB-D pictures of different viewpoints of the object are obtained, and a three-dimensional model of the object can be reconstructed by using an existing multi-viewpoint reconstruction algorithm, such as a kinect fusion algorithm carried by a computer vision open source library OpenCv 4.0 or commercial software Agisoft Metashape Pro.
And step 203, acquiring the motion state information of the object according to the second color information and the second depth information.
Specifically, the motion state of the object is identified from the RGB-D data through a depth learning method according to the image information and the depth information of the object. The motion state information of the object includes a motion position of the object, a shape of the object, and the like.
And step 204, extracting the depth information of the first key point of the hand of the user according to the first depth information.
Specifically, according to the depth information of the hand of the user, the depth information of the collected or non-shielded sparse key points of the hand is identified from the RGB-D data through a depth learning method such as OpenPose.
And step 205, estimating the depth information of the second key point of the hand of the user according to the depth information of the first key point and the Gaussian distribution function.
Specifically, estimated depth information of a second key point is determined according to depth information of a first key point and a preset algorithm, confidence of the estimated depth information is calculated according to a Gaussian distribution function, whether the confidence is greater than a preset threshold is detected, if not, the estimated depth information is modified until the confidence is greater than the preset threshold, and the modified estimated depth information is used as the depth information of the second key point. In this example, the second keypoint may be a point that is not acquired or occluded, and since the point that is not acquired or occluded is estimated by a gaussian function, the confidence level needs to be determined, and the estimated depth information is modified until the confidence level is greater than a preset threshold, and the estimated depth information is used as the depth information of the occluded point.
And step 206, simulating the interactive animation of the object and the hand of the user according to the depth information of the first key point, the depth information of the second key point, the first color information and the motion state information.
Specifically, the interactive animation of the object and the hand of the user is simulated according to the collected depth information of the point and the depth information of the shielded point, the image information of the hand of the user and the motion state of the object estimated by a preset algorithm.
The hand motion reconstruction method provided by the embodiment of the invention solves the technical problems that the movement of a human hand is flexible and complex and severe shielding is often accompanied in the process of interacting with an object in the prior art, can obtain a more accurate three-dimensional reconstruction result through isomorphic single-viewpoint RGB-D data, and can obtain a better object and hand interaction reconstruction result under a simpler hardware condition.
In order to realize the embodiment, the invention further provides a hand motion reconstruction device.
Fig. 3 is a schematic structural diagram of a hand motion reconstruction device according to an embodiment of the present invention.
As shown in fig. 3, the hand motion reconstruction apparatus includes: the hand depth data processing method comprises a first obtaining module 10, a control module 20, a second obtaining module 30 and a reconstruction module 40, wherein the first obtaining module 10 is used for obtaining a hand depth data set, each hand depth data in the hand depth data set comprises hand depth picture information and hand skeleton coordinates corresponding to the hand depth picture information, then the control module 20 controls a preset hand model to adjust the posture according to each hand depth data in the hand depth data set, and then the second obtaining module 30 obtains the posture parameters of a fitting hand model after the posture is adjusted; as shown in fig. 4, on the basis of fig. 3, the method further includes: the determining unit 31 is configured to determine a first pose parameter of a preset hand model according to a point cloud matching algorithm and hand depth information in the first hand depth data; and the calculating unit 32 is used for calculating a second attitude parameter of the preset hand model according to the fitted skeleton coordinates, and finally, the reconstructing module 40 is used for reconstructing hand motion according to the Gaussian distribution function of the attitude parameter of the fitted hand model corresponding to each hand depth data so as to reconstruct hand motion according to the Gaussian distribution function.
It should be noted that the explanation of the hand motion reconstruction method embodiment is also applicable to the hand motion reconstruction device of this embodiment, and is not repeated here.
In order to implement the above embodiments, the present invention further provides a computer device, which includes a memory, a processor, and a computer program stored in the memory and executable on the processor, and when the processor executes the computer program, the hand motion reconstruction method as described in the above embodiments is implemented.
In order to implement the above embodiments, the present invention further proposes a non-transitory computer readable storage medium, wherein when being executed by a processor, the computer program implements the hand motion reconstruction method as described in the above embodiments.
In the description herein, references to the description of the term "one embodiment," "some embodiments," "an example," "a specific example," or "some examples," etc., mean that a particular feature, structure, material, or characteristic described in connection with the embodiment or example is included in at least one embodiment or example of the invention. In this specification, the schematic representations of the terms used above are not necessarily intended to refer to the same embodiment or example. Furthermore, the particular features, structures, materials, or characteristics described may be combined in any suitable manner in any one or more embodiments or examples. Furthermore, various embodiments or examples and features of different embodiments or examples described in this specification can be combined and combined by one skilled in the art without contradiction.
Furthermore, the terms "first", "second" and "first" are used for descriptive purposes only and are not to be construed as indicating or implying relative importance or implicitly indicating the number of technical features indicated. Thus, a feature defined as "first" or "second" may explicitly or implicitly include at least one such feature. In the description of the present invention, "a plurality" means at least two, e.g., two, three, etc., unless specifically limited otherwise.
Any process or method descriptions in flow charts or otherwise described herein may be understood as representing modules, segments, or portions of code which include one or more executable instructions for implementing steps of a custom logic function or process, and alternate implementations are included within the scope of the preferred embodiment of the present invention in which functions may be executed out of order from that shown or discussed, including substantially concurrently or in reverse order, depending on the functionality involved, as would be understood by those reasonably skilled in the art of the present invention.
The logic and/or steps represented in the flowcharts or otherwise described herein, e.g., an ordered listing of executable instructions that can be considered to implement logical functions, can be embodied in any computer-readable medium for use by or in connection with an instruction execution system, apparatus, or device, such as a computer-based system, processor-containing system, or other system that can fetch the instructions from the instruction execution system, apparatus, or device and execute the instructions. For the purposes of this description, a "computer-readable medium" can be any means that can contain, store, communicate, propagate, or transport the program for use by or in connection with the instruction execution system, apparatus, or device. More specific examples (a non-exhaustive list) of the computer-readable medium would include the following: an electrical connection (electronic device) having one or more wires, a portable computer diskette (magnetic device), a Random Access Memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber device, and a portable compact disc read-only memory (CDROM). Additionally, the computer-readable medium could even be paper or another suitable medium upon which the program is printed, as the program can be electronically captured, via for instance optical scanning of the paper or other medium, then compiled, interpreted or otherwise processed in a suitable manner if necessary, and then stored in a computer memory.
It should be understood that portions of the present invention may be implemented in hardware, software, firmware, or a combination thereof. In the above embodiments, the various steps or methods may be implemented in software or firmware stored in memory and executed by a suitable instruction execution system. If implemented in hardware, as in another embodiment, any one or combination of the following techniques, which are known in the art, may be used: a discrete logic circuit having a logic gate circuit for implementing a logic function on a data signal, an application specific integrated circuit having an appropriate combinational logic gate circuit, a Programmable Gate Array (PGA), a Field Programmable Gate Array (FPGA), or the like.
It will be understood by those skilled in the art that all or part of the steps carried by the method for implementing the above embodiments may be implemented by hardware related to instructions of a program, which may be stored in a computer readable storage medium, and when the program is executed, the program includes one or a combination of the steps of the method embodiments.
In addition, functional units in the embodiments of the present invention may be integrated into one processing module, or each unit may exist alone physically, or two or more units are integrated into one module. The integrated module can be realized in a hardware mode, and can also be realized in a software functional module mode. The integrated module, if implemented in the form of a software functional module and sold or used as a stand-alone product, may also be stored in a computer readable storage medium.
The storage medium mentioned above may be a read-only memory, a magnetic or optical disk, etc. Although embodiments of the present invention have been shown and described above, it is understood that the above embodiments are exemplary and should not be construed as limiting the present invention, and that variations, modifications, substitutions and alterations can be made to the above embodiments by those of ordinary skill in the art within the scope of the present invention.

Claims (6)

1.一种手部运动重建方法,其特征在于,包括以下步骤:1. a hand motion reconstruction method, is characterized in that, comprises the following steps: 获取手部深度数据集,其中,所述手部深度数据集中的每个手部深度数据中包含手部深度图片信息及其对应的手部骨架坐标;Obtaining a hand depth data set, wherein each hand depth data in the hand depth data set includes hand depth picture information and its corresponding hand skeleton coordinates; 控制预设手部模型分别根据所述手部深度数据集中每个手部深度数据调整姿态,获取姿态调整后的拟合手部模型的姿态参数;Control the preset hand model to adjust the posture according to each hand depth data in the hand depth data set, and obtain the posture parameters of the fitted hand model after the posture adjustment; 根据与所述每个手部深度数据对应的所述拟合手部模型的姿态参数高斯分布函数,以便于根据所述高斯分布函数进行手部运动重建,其中,所述控制预设手部模型分别根据所述手部深度数据集中每个手部深度数据调整姿态,获取姿态调整后的拟合手部模型的姿态参数,包括:According to the posture parameter Gaussian distribution function of the fitted hand model corresponding to the depth data of each hand, so as to perform hand motion reconstruction according to the Gaussian distribution function, wherein the control presets the hand model Adjust the posture according to each hand depth data in the hand depth data set, and obtain the posture parameters of the fitted hand model after the posture adjustment, including: 获取所述手部深度数据集中每个手部深度数据对应的参考姿态参数;Obtain the reference posture parameter corresponding to each hand depth data in the hand depth data set; 计算所述参考姿态参数与预设初始姿态参数的差值,然后确定小于预设阈值的差值对应的手部深度数据为第一手部深度数据;Calculate the difference between the reference attitude parameter and the preset initial attitude parameter, and then determine the hand depth data corresponding to the difference smaller than the preset threshold as the first hand depth data; 根据点云匹配算法和所述第一手部深度数据中的手部深度信息,确定所述预设手部模型的第一姿态参数;Determine the first posture parameter of the preset hand model according to the point cloud matching algorithm and the hand depth information in the first hand depth data; 通过梯度下降迭代算法和所述第一手部深度数据中的手部骨架坐标,构建所述第一姿态参数的回归矩阵;Construct the regression matrix of the first posture parameter by using the gradient descent iterative algorithm and the hand skeleton coordinates in the first hand depth data; 确定所述手部深度数据集中除所述第一手部深度数据之外的第二手部深度数据,根据所述回归矩阵和所述第二手部深度数据对应的手部骨架坐标计算所述预设手部模型的拟合骨架坐标;Determine the second hand depth data in the hand depth data set except the first hand depth data, and calculate the hand skeleton coordinates according to the regression matrix and the hand skeleton coordinates corresponding to the second hand depth data The fitted skeleton coordinates of the preset hand model; 根据所述拟合骨架坐标计算所述预设手部模型的第二姿态参数;Calculate the second posture parameter of the preset hand model according to the fitted skeleton coordinates; 基于预设的RGB-D相机获取用户手部与物体交互的连续多帧图像;Based on the preset RGB-D camera to obtain continuous multi-frame images of the interaction between the user's hand and the object; 根据所述连续多帧图像提取所述用户手部的第一彩色信息和第一深度信息,以及所述物体的第二彩色信息和第二深度信息;extracting first color information and first depth information of the user's hand, and second color information and second depth information of the object according to the consecutive multi-frame images; 根据所述第二彩色信息和第二深度信息获取所述物体的运动状态信息;Acquiring motion state information of the object according to the second color information and the second depth information; 根据所述第一深度信息提取所述用户手部的第一关键点的深度信息;extracting depth information of the first key point of the user's hand according to the first depth information; 根据所述第一关键点的深度信息和所述高斯分布函数估算所述用户手部的第二关键点的深度信息;Estimating the depth information of the second key point of the user's hand according to the depth information of the first key point and the Gaussian distribution function; 根据所述第一关键点的深度信息、所述第二关键点的深度信息、所述第一彩色信息和所述运动状态信息模拟所述物体与所述用户手部的交互动画。The interaction animation between the object and the user's hand is simulated according to the depth information of the first key point, the depth information of the second key point, the first color information and the motion state information. 2.如权利要求1所述的方法,其特征在于,所述根据所述第一关键点的深度信息和所述高斯分布函数估算所述用户手部的第二关键点的深度信息,包括:2. The method according to claim 1, wherein the estimating the depth information of the second key point of the user's hand according to the depth information of the first key point and the Gaussian distribution function comprises: 根据所述第一关键点的深度信息和预设算法确定所述第二关键点的估算深度信息;Determine the estimated depth information of the second key point according to the depth information of the first key point and a preset algorithm; 根据所述高斯分布函数计算所述估算深度信息的置信度;calculating the confidence of the estimated depth information according to the Gaussian distribution function; 检测所述置信度是否大于预设阈值,若不大于所述预设阈值,则修改所述估算深度信息,直至所述置信度大于所述预设阈值,将修改后的所述估算深度信息作为所述第二关键点的深度信息。Detecting whether the confidence level is greater than a preset threshold, and if it is not greater than the preset threshold, modify the estimated depth information until the confidence level is greater than the preset threshold, and use the modified estimated depth information as depth information of the second keypoint. 3.如权利要求1所述的方法,其特征在于,在所述基于预设的RGB-D相机获取用户手部与物体交互的连续多帧图像之前,包括:3. The method according to claim 1, wherein before the preset RGB-D camera acquires the continuous multi-frame images of the interaction between the user's hand and the object, the method comprises: 获取所述RGB-D相机中RGB模组的内参和外参;Obtain the internal parameters and external parameters of the RGB module in the RGB-D camera; 获取所述RGB模组和所述RGB-D相机中深度模组同时拍摄的人体动作图像;Acquiring the human action images simultaneously shot by the RGB module and the depth module in the RGB-D camera; 根据预设函数和所述同时拍摄的人体动作图像修正所述内参和外参。The internal parameter and the external parameter are corrected according to the preset function and the simultaneously captured human action image. 4.一种手部运动重建装置,其特征在于,所述装置包括:4. A hand motion reconstruction device, wherein the device comprises: 第一获取模块,用于获取手部深度数据集,其中,所述手部深度数据集中的每个手部深度数据中包含手部深度图片信息及其对应的手部骨架坐标;a first acquisition module, configured to acquire a hand depth data set, wherein each hand depth data in the hand depth data set includes hand depth picture information and its corresponding hand skeleton coordinates; 控制模块,用于控制预设手部模型分别根据所述手部深度数据集中每个手部深度数据调整姿态;a control module for controlling the preset hand model to adjust the posture according to each hand depth data in the hand depth data set; 第二获取模块,用于获取姿态调整后的拟合手部模型的姿态参数,其中,所述第二获取模块,具体用于:获取所述手部深度数据集中每个手部深度数据对应的参考姿态参数;The second acquisition module is used to acquire the posture parameters of the fitted hand model after posture adjustment, wherein the second acquisition module is specifically used for: acquiring the corresponding hand depth data in the hand depth data set reference attitude parameters; 计算所述参考姿态参数与预设初始姿态参数的差值,然后确定小于预设阈值的差值对应的手部深度数据为第一手部深度数据;Calculate the difference between the reference attitude parameter and the preset initial attitude parameter, and then determine the hand depth data corresponding to the difference smaller than the preset threshold as the first hand depth data; 根据点云匹配算法和所述第一手部深度数据中的手部深度信息,确定所述预设手部模型的第一姿态参数;Determine the first posture parameter of the preset hand model according to the point cloud matching algorithm and the hand depth information in the first hand depth data; 通过梯度下降迭代算法和所述第一手部深度数据中的手部骨架坐标,构建所述第一姿态参数的回归矩阵;Construct the regression matrix of the first posture parameter by using the gradient descent iterative algorithm and the hand skeleton coordinates in the first hand depth data; 确定所述手部深度数据集中除所述第一手部深度数据之外的第二手部深度数据,根据所述回归矩阵和所述第二手部深度数据对应的手部骨架坐标计算所述预设手部模型的拟合骨架坐标;Determine the second hand depth data in the hand depth data set except the first hand depth data, and calculate the hand skeleton coordinates according to the regression matrix and the hand skeleton coordinates corresponding to the second hand depth data The fitted skeleton coordinates of the preset hand model; 根据所述拟合骨架坐标计算所述预设手部模型的第二姿态参数;Calculate the second posture parameter of the preset hand model according to the fitted skeleton coordinates; 重建模块,用于根据与所述每个手部深度数据对应的所述拟合手部模型的姿态参数高斯分布函数,以便于根据所述高斯分布函数进行手部运动重建;a reconstruction module, configured to perform hand motion reconstruction according to the posture parameter Gaussian distribution function of the fitted hand model corresponding to the depth data of each hand, so as to perform hand motion reconstruction according to the Gaussian distribution function; 动画处理模块,用于基于预设的RGB-D相机获取用户手部与物体交互的连续多帧图像,根据所述连续多帧图像提取所述用户手部的第一彩色信息和第一深度信息,以及所述物体的第二彩色信息和第二深度信息,根据所述第二彩色信息和第二深度信息获取所述物体的运动状态信息,根据所述第一深度信息提取所述用户手部的第一关键点的深度信息,根据所述第一关键点的深度信息和所述高斯分布函数估算所述用户手部的第二关键点的深度信息,根据所述第一关键点的深度信息、所述第二关键点的深度信息、所述第一彩色信息和所述运动状态信息模拟所述物体与所述用户手部的交互动画。The animation processing module is used to obtain continuous multi-frame images of the interaction between the user's hand and the object based on the preset RGB-D camera, and extract the first color information and the first depth information of the user's hand according to the continuous multi-frame images , and the second color information and second depth information of the object, obtain the motion state information of the object according to the second color information and the second depth information, and extract the user's hand according to the first depth information The depth information of the first key point, the depth information of the second key point of the user's hand is estimated according to the depth information of the first key point and the Gaussian distribution function, and the depth information of the first key point is estimated according to the depth information of the first key point. , the depth information of the second key point, the first color information and the motion state information simulate the interactive animation of the object and the user's hand. 5.一种计算机设备,其特征在于,包括存储器、处理器及存储在所述存储器上并可在所述处理器上运行的计算机程序,所述处理器执行所述计算机程序时,实现如权利要求1-3任一所述的手部运动重建方法。5. A computer device, characterized in that it comprises a memory, a processor and a computer program stored on the memory and running on the processor, and when the processor executes the computer program, the computer program as claimed in the claim is realized. The hand motion reconstruction method described in any one of requirements 1-3. 6.一种非临时性计算机可读存储介质,其上存储有计算机程序,其特征在于,所述计算机程序被处理器执行时实现如权利要求1-3任一所述的手部运动重建方法。6. A non-transitory computer-readable storage medium on which a computer program is stored, characterized in that, when the computer program is executed by a processor, the hand motion reconstruction method according to any one of claims 1-3 is implemented .
CN201910646553.7A 2019-07-17 2019-07-17 Method and device for hand motion reconstruction Active CN110443884B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910646553.7A CN110443884B (en) 2019-07-17 2019-07-17 Method and device for hand motion reconstruction

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910646553.7A CN110443884B (en) 2019-07-17 2019-07-17 Method and device for hand motion reconstruction

Publications (2)

Publication Number Publication Date
CN110443884A CN110443884A (en) 2019-11-12
CN110443884B true CN110443884B (en) 2021-07-30

Family

ID=68430670

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910646553.7A Active CN110443884B (en) 2019-07-17 2019-07-17 Method and device for hand motion reconstruction

Country Status (1)

Country Link
CN (1) CN110443884B (en)

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111290577B (en) * 2020-01-22 2024-03-22 北京明略软件系统有限公司 Non-contact input method and device
CN111429499B (en) * 2020-02-24 2023-03-10 中山大学 High-precision three-dimensional reconstruction method for hand skeleton based on single depth camera
CN111596767B (en) * 2020-05-27 2023-05-30 广州市大湾区虚拟现实研究院 Gesture capturing method and device based on virtual reality
CN113674395B (en) * 2021-07-19 2023-04-18 广州紫为云科技有限公司 3D hand lightweight real-time capturing and reconstructing system based on monocular RGB camera

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102262783A (en) * 2011-08-16 2011-11-30 清华大学 Method and system for restructuring motion of three-dimensional gesture
CN108564618A (en) * 2018-04-11 2018-09-21 清华大学 Hand geometry motion method for reconstructing and device based on multi-voxel proton block
CN109636831A (en) * 2018-12-19 2019-04-16 安徽大学 A method of estimation 3 D human body posture and hand information

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10559119B2 (en) * 2017-08-31 2020-02-11 Intel Corporation Method and apparatus for natural hand visualization

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102262783A (en) * 2011-08-16 2011-11-30 清华大学 Method and system for restructuring motion of three-dimensional gesture
CN108564618A (en) * 2018-04-11 2018-09-21 清华大学 Hand geometry motion method for reconstructing and device based on multi-voxel proton block
CN109636831A (en) * 2018-12-19 2019-04-16 安徽大学 A method of estimation 3 D human body posture and hand information

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
《3D Hand Shape and Pose from Images in the Wild》;Adnane Boukhayma等;《arXiv》;20190209;第1-12页 *
《Embodied Hands: Modeling and Capturing Hands and Bodies Together》;JAVIER ROMERO等;《ACM Transactions on Graphics》;20171130;第245:1-17页 *
《Single Image 3D Hand Reconstruction with Mesh Convolutions》;Dominik Kulon等;《arXiv》;20190513;第1-13页 *

Also Published As

Publication number Publication date
CN110443884A (en) 2019-11-12

Similar Documents

Publication Publication Date Title
CN110443884B (en) Method and device for hand motion reconstruction
KR101560508B1 (en) Method and arrangement for 3-dimensional image model adaptation
Achenbach et al. Fast generation of realistic virtual humans
Stoll et al. Fast articulated motion tracking using a sums of gaussians body model
Weiss et al. Home 3D body scans from noisy image and range data
KR101183000B1 (en) A system and method for 3D space-dimension based image processing
JP7005622B2 (en) Recognition processing device, recognition processing method and program
CN109242961A (en) A kind of face modeling method, apparatus, electronic equipment and computer-readable medium
JP6676562B2 (en) Image synthesizing apparatus, image synthesizing method, and computer program
KR20150117553A (en) Method, apparatus and computer readable recording medium for eye gaze tracking
US10789784B2 (en) Image display method, electronic device, and non-transitory computer readable recording medium for quickly providing simulated two-dimensional head portrait as reference after plastic operation
JP2016071645A (en) Object three-dimensional model restoration method, apparatus and program
CN108734772A (en) High accuracy depth image acquisition methods based on Kinect fusion
CN109407824B (en) Method and device for synchronous motion of human body model
KR101454780B1 (en) Apparatus and method for generating texture for three dimensional model
CN115862124B (en) Line-of-sight estimation method and device, readable storage medium and electronic equipment
CN111742352B (en) Method for modeling three-dimensional object and electronic equipment
JP6770208B2 (en) Information processing device
JP6799468B2 (en) Image processing equipment, image processing methods and computer programs
Rendle et al. Volumetric avatar reconstruction with spatio-temporally offset rgbd cameras
Lim et al. 3-D reconstruction using the kinect sensor and its application to a visualization system
CN111105489A (en) Data synthesis method and apparatus, storage medium, and electronic apparatus
Amirkhanov et al. WithTeeth: Denture Preview in Augmented Reality.
Shen et al. Virtual mirror by fusing multiple RGB-D cameras
US20250118027A1 (en) Face micro detail recovery via patch scanning, interpolation, and style transfer

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
TR01 Transfer of patent right
TR01 Transfer of patent right

Effective date of registration: 20221230

Address after: Room 3346, Floor 3, International Innovation Expo Center, No. 267, Kejiyuan Road, Baiyang Street, Qiantang District, Hangzhou, Zhejiang 310020

Patentee after: Hangzhou Xinchangyuan Technology Co.,Ltd.

Address before: 100084 Tsinghua Yuan, Beijing, Haidian District

Patentee before: TSINGHUA University