CN115937373B - Avatar driving method, apparatus, device and storage medium - Google Patents
Avatar driving method, apparatus, device and storage medium Download PDFInfo
- Publication number
- CN115937373B CN115937373B CN202211670321.3A CN202211670321A CN115937373B CN 115937373 B CN115937373 B CN 115937373B CN 202211670321 A CN202211670321 A CN 202211670321A CN 115937373 B CN115937373 B CN 115937373B
- Authority
- CN
- China
- Prior art keywords
- posture
- pose
- transformation coefficient
- difference
- vertex
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Classifications
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02D—CLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
- Y02D10/00—Energy efficient computing, e.g. low power processors, power management or thermal management
Landscapes
- Processing Or Creating Images (AREA)
Abstract
本公开提供了一种虚拟形象驱动方法、装置、设备以及储介质,涉及人工智能技术领域,尤其涉及虚拟人、元宇宙、增强现实、虚拟现实、混合现实、扩展现实等技术领域。具体实现方案为:接收输入数据流,输入数据流包括多个时序帧,任意一个时序帧与面部姿态相关联,面部姿态包括第一部位姿态和第二部位姿态,针对多个时序帧中的任意一个目标时序帧,根据初始基准姿态以及第一部位姿态,确定第一姿态变换系数;根据第一姿态变换系数更新初始基准姿态,得到更新基准姿态;根据更新基准姿态以及第二部位姿态,确定第二姿态变换系数;根据第二姿态变换系数驱动虚拟形象。
The present disclosure provides a virtual image driving method, device, equipment and storage medium, and relates to the technical field of artificial intelligence, especially to the technical fields of virtual human, metaverse, augmented reality, virtual reality, mixed reality, extended reality and other technical fields. The specific implementation plan is: receiving an input data stream. The input data stream includes multiple timing frames. Any timing frame is associated with a facial posture. The facial posture includes a first part posture and a second part posture. For any of the multiple timing frames, In a target timing frame, the first posture transformation coefficient is determined based on the initial reference posture and the first part posture; the initial reference posture is updated according to the first posture transformation coefficient to obtain the updated reference posture; and the third posture is determined based on the updated reference posture and the second part posture. Two posture transformation coefficients; driving the virtual image according to the second posture transformation coefficient.
Description
技术领域Technical field
本公开涉及人工智能技术领域,尤其涉及虚拟人、元宇宙、增强现实、虚拟现实、混合现实、扩展现实等技术领域,具体涉及一种虚拟形象驱动方法、装置、设备以及存储介质。The present disclosure relates to the technical field of artificial intelligence, especially to the technical fields of virtual human, metaverse, augmented reality, virtual reality, mixed reality, extended reality, etc., and specifically to a virtual image driving method, device, equipment and storage medium.
背景技术Background technique
随着计算机技术和互联网技术的发展,可以通过虚拟形象提供生活、娱乐等方面的各项功能服务。例如,一些虚拟形象提供视觉显示服务,如何使得虚拟形象作出的面部表情更真实是一个亟需解决的问题。With the development of computer technology and Internet technology, various functional services in life, entertainment and other aspects can be provided through virtual images. For example, some avatars provide visual display services. How to make the facial expressions made by avatars more realistic is an urgent problem that needs to be solved.
发明内容Contents of the invention
本公开提供了一种虚拟形象驱动方法、装置、设备以及存储介质。The present disclosure provides a virtual image driving method, device, equipment and storage medium.
根据本公开的一方面,提供了一种虚拟形象驱动方法,包括:接收输入数据流,其中,输入数据流包括多个时序帧,任意一个时序帧与面部姿态相关联,面部姿态包括第一部位姿态和第二部位姿态,第二部位姿态包括多个第二子部位姿态,任意多个第一部位姿态之间的相关性小于与多个第二子部位姿态之间的相关性,针对多个时序帧中的任意一个目标时序帧,根据初始基准姿态以及第一部位姿态,确定第一姿态变换系数;根据第一姿态变换系数更新初始基准姿态,得到更新基准姿态;根据更新基准姿态以及第二部位姿态,确定第二姿态变换系数;以及根据第二姿态变换系数驱动虚拟形象。According to an aspect of the present disclosure, an avatar driving method is provided, including: receiving an input data stream, wherein the input data stream includes a plurality of time series frames, any one of the time series frames is associated with a facial gesture, and the facial gesture includes a first part The posture and the second part posture, the second part posture includes multiple second sub-part postures, the correlation between any multiple first part postures is smaller than the correlation with the multiple second sub-part postures, for multiple For any target timing frame in the timing frame, determine the first posture transformation coefficient according to the initial reference posture and the first part posture; update the initial reference posture according to the first posture transformation coefficient to obtain the updated reference posture; based on the updated reference posture and the second posture Part posture, determine the second posture transformation coefficient; and drive the virtual image according to the second posture transformation coefficient.
根据本公开的另一方面,提供了一种虚拟形象驱动装置,包括:输入数据流接收模块,用于接收输入数据流,其中,输入数据流包括多个时序帧,任意一个时序帧与面部姿态相关联,面部姿态包括第一部位姿态和第二部位姿态,第二部位姿态包括多个第二子部位姿态,任意多个第一部位姿态之间的相关性小于多个第二子部位姿态之间的相关性;第一姿态变换系数确定模块,用于针对多个时序帧中的任意一个目标时序帧,根据初始基准姿态以及第一部位姿态,确定第一姿态变换系数;更新基准姿态确定模块,用于根据第一姿态变换系数更新初始基准姿态,得到更新基准姿态;第二姿态变换系数确定模块,用于根据更新基准姿态以及第二部位姿态,确定第二姿态变换系数;虚拟形象驱动模块,用于根据第二姿态变换系数驱动虚拟形象。According to another aspect of the present disclosure, an avatar driving device is provided, including: an input data stream receiving module, configured to receive an input data stream, wherein the input data stream includes a plurality of time series frames, any one of the time series frames and facial gestures Correlation, the facial posture includes a first part posture and a second part posture, the second part posture includes multiple second sub-part postures, and the correlation between any multiple first part postures is smaller than the multiple second sub-part postures. The correlation between , used to update the initial reference posture according to the first posture transformation coefficient to obtain the updated reference posture; the second posture transformation coefficient determination module is used to determine the second posture transformation coefficient according to the updated reference posture and the second part posture; the avatar driving module , used to drive the virtual image according to the second posture transformation coefficient.
根据本公开的另一方面,提供了一种电子设备,包括:至少一个处理器和与至少一个处理器通信连接的存储器。其中,存储器存储有可被至少一个处理器执行的指令,指令被至少一个处理器执行,以使至少一个处理器能够执行本公开实施例的方法。According to another aspect of the present disclosure, an electronic device is provided, including: at least one processor and a memory communicatively connected with the at least one processor. The memory stores instructions that can be executed by at least one processor, and the instructions are executed by at least one processor, so that at least one processor can execute the method of the embodiment of the present disclosure.
根据本公开的另一方面,提供了一种存储有计算机指令的非瞬时计算机可读存储介质,计算机指令用于使计算机执行本公开实施例的方法。According to another aspect of the present disclosure, there is provided a non-transitory computer-readable storage medium storing computer instructions for causing a computer to perform methods of embodiments of the present disclosure.
应当理解,本部分所描述的内容并非旨在标识本公开的实施例的关键或重要特征,也不用于限制本公开的范围。本公开的其它特征将通过以下的说明书而变得容易理解。It should be understood that what is described in this section is not intended to identify key or important features of the embodiments of the disclosure, nor is it intended to limit the scope of the disclosure. Other features of the present disclosure will become readily understood from the following description.
附图说明Description of the drawings
附图用于更好地理解本方案,不构成对本公开的限定。其中:The accompanying drawings are used to better understand the present solution and do not constitute a limitation of the present disclosure. in:
图1示意性示出了根据本公开实施例的虚拟形象驱动方法和装置的系统架构图;Figure 1 schematically shows a system architecture diagram of an avatar driving method and device according to an embodiment of the present disclosure;
图2A示意性示出了根据本公开实施例的虚拟形象驱动方法的流程图;Figure 2A schematically shows a flow chart of an avatar driving method according to an embodiment of the present disclosure;
图2B示意性示出了下巴-嘴唇区域姿态的示意图;Figure 2B schematically shows a schematic diagram of the posture of the chin-lip region;
图3示意性示出了根据本公开另一实施例的虚拟形象驱动方法的示意图;Figure 3 schematically shows a schematic diagram of an avatar driving method according to another embodiment of the present disclosure;
图4示意性示出了根据本公开又一实施例的虚拟形象驱动方法的确定第三姿态变换系数示意图;Figure 4 schematically shows a schematic diagram of determining the third posture transformation coefficient of the avatar driving method according to yet another embodiment of the present disclosure;
图5示意性示出了根据本公开又一实施例的虚拟形象驱动装置的框图;以及Figure 5 schematically shows a block diagram of an avatar driving device according to yet another embodiment of the present disclosure; and
图6示意性示出了可以实现本公开实施例的虚拟形象驱动方法的电子设备的框图。FIG. 6 schematically shows a block diagram of an electronic device that can implement the avatar driving method of an embodiment of the present disclosure.
具体实施方式Detailed ways
以下结合附图对本公开的示范性实施例做出说明,其中包括本公开实施例的各种细节以助于理解,应当将它们认为仅仅是示范性的。因此,本领域普通技术人员应当认识到,可以对这里描述的实施例做出各种改变和修改,而不会背离本公开的范围和精神。同样,为了清楚和简明,以下的描述中省略了对公知功能和结构的描述。Exemplary embodiments of the present disclosure are described below with reference to the accompanying drawings, in which various details of the embodiments of the present disclosure are included to facilitate understanding and should be considered to be exemplary only. Accordingly, those of ordinary skill in the art will recognize that various changes and modifications can be made to the embodiments described herein without departing from the scope and spirit of the disclosure. Also, descriptions of well-known functions and constructions are omitted from the following description for clarity and conciseness.
在此使用的术语仅仅是为了描述具体实施例,而并非意在限制本公开。在此使用的术语“包括”、“包含”等表明了所述特征、步骤、操作和/或部件的存在,但是并不排除存在或添加一个或多个其他特征、步骤、操作或部件。The terminology used herein is for the purpose of describing particular embodiments only and is not intended to limit the disclosure. The terms "comprising," "comprising," and the like, as used herein, indicate the presence of stated features, steps, operations, and/or components but do not exclude the presence or addition of one or more other features, steps, operations, or components.
在此使用的所有术语(包括技术和科学术语)具有本领域技术人员通常所理解的含义,除非另外定义。应注意,这里使用的术语应解释为具有与本说明书的上下文相一致的含义,而不应以理想化或过于刻板的方式来解释。All terms (including technical and scientific terms) used herein have the same meaning as commonly understood by one of ordinary skill in the art, unless otherwise defined. It should be noted that the terms used here should be interpreted to have meanings consistent with the context of this specification and should not be interpreted in an idealized or overly rigid manner.
在使用类似于“A、B和C等中至少一个”这样的表述的情况下,一般来说应该按照本领域技术人员通常理解该表述的含义来予以解释(例如,“具有A、B和C中至少一个的系统”应包括但不限于单独具有A、单独具有B、单独具有C、具有A和B、具有A和C、具有B和C、和/或具有A、B、C的系统等)。Where an expression similar to "at least one of A, B, C, etc." is used, it should generally be interpreted in accordance with the meaning that a person skilled in the art generally understands the expression to mean (e.g., "having A, B and C "A system with at least one of" shall include, but is not limited to, systems with A alone, B alone, C alone, A and B, A and C, B and C, and/or systems with A, B, C, etc. ).
图1示意性示出了根据本公开一实施例的虚拟形象驱动方法和装置的系统架构。需要注意的是,图1所示仅为可以应用本公开实施例的系统架构的示例,以帮助本领域技术人员理解本公开的技术内容,但并不意味着本公开实施例不可以用于其他设备、系统、环境或场景。Figure 1 schematically shows the system architecture of an avatar driving method and device according to an embodiment of the present disclosure. It should be noted that Figure 1 is only an example of a system architecture to which embodiments of the present disclosure can be applied, to help those skilled in the art understand the technical content of the present disclosure, but does not mean that the embodiments of the present disclosure cannot be used in other applications. Device, system, environment or scenario.
如图1所示,根据该实施例的系统架构100可以包括客户端101、102、103,网络104和服务器105。网络104用以在客户端101、102、103和服务器105之间提供通信链路的介质。网络104可以包括各种连接类型,例如有线、无线通信链路或者光纤电缆等等。As shown in Figure 1, the system architecture 100 according to this embodiment may include clients 101, 102, 103, a network 104 and a server 105. Network 104 is the medium used to provide communication links between clients 101, 102, 103 and server 105. Network 104 may include various connection types, such as wired, wireless communication links, or fiber optic cables, among others.
用户可以使用客户端101、102、103通过网络104与服务器105交互,以接收或发送消息等。客户端101、102、103上可以安装有各种通讯客户端应用,例如购物类应用、网页浏览器应用、搜索类应用、即时通信工具、邮箱客户端、社交平台软件等(仅为示例)。Users may use clients 101, 102, 103 to interact with server 105 over network 104 to receive or send messages, etc. Various communication client applications can be installed on the clients 101, 102, and 103, such as shopping applications, web browser applications, search applications, instant messaging tools, email clients, social platform software, etc. (only examples).
客户端101、102、103可以是具有显示屏并且支持网页浏览的各种电子设备,包括但不限于智能手机、平板电脑、膝上型便携计算机和台式计算机等等。本公开实施例的客户端101、102、103例如可以运行应用程序。The clients 101, 102, and 103 may be various electronic devices having a display screen and supporting web browsing, including but not limited to smartphones, tablet computers, laptop computers, desktop computers, and the like. The clients 101, 102, and 103 of the embodiments of the present disclosure may, for example, run applications.
服务器105可以是提供各种服务的服务器,例如对用户利用客户端101、102、103所浏览的网站提供支持的后台管理服务器(仅为示例)。后台管理服务器可以对接收到的用户请求等数据进行分析等处理,并将处理结果(例如根据用户请求获取或生成的网页、信息、或数据等)反馈给客户端。另外,服务器105还可以是云服务器,即服务器105具有云计算功能。The server 105 may be a server that provides various services, such as a backend management server that provides support for websites browsed by users using the clients 101, 102, and 103 (example only). The background management server can analyze and process the received user request and other data, and feed back the processing results (such as web pages, information, or data obtained or generated according to the user request) to the client. In addition, the server 105 can also be a cloud server, that is, the server 105 has a cloud computing function.
需要说明的是,本公开实施例所提供的虚拟形象驱动方法可以由服务器105执行。相应地,本公开实施例所提供的虚拟形象驱动装置可以设置于服务器105中。本公开实施例所提供的虚拟形象驱动方法也可以由不同于服务器105且能够与客户端101、102、103和/或服务器105通信的服务器或服务器集群执行。相应地,本公开实施例所提供的虚拟形象驱动装置也可以设置于不同于服务器105且能够与客户端101、102、103和/或服务器105通信的服务器或服务器集群中。It should be noted that the avatar driving method provided by the embodiment of the present disclosure can be executed by the server 105 . Correspondingly, the avatar driving device provided by the embodiment of the present disclosure can be provided in the server 105. The virtual image driving method provided by the embodiment of the present disclosure can also be executed by a server or server cluster that is different from the server 105 and can communicate with the clients 101, 102, 103 and/or the server 105. Correspondingly, the avatar driving device provided by the embodiment of the present disclosure can also be provided in a server or server cluster that is different from the server 105 and can communicate with the clients 101, 102, 103 and/or the server 105.
在一种示例中,服务器105可以通过网络104获取来自客户端101、102、103的输入数据流,并基于输入数据流进行虚拟形象驱动。In one example, the server 105 can obtain input data streams from the clients 101, 102, and 103 through the network 104, and drive the avatar based on the input data streams.
应该理解,图1中的客户端、网络和服务器的数目仅仅是示意性的。根据实现需要,可以具有任意数目的客户端、网络和服务器。It should be understood that the number of clients, networks and servers in Figure 1 is only illustrative. You can have any number of clients, networks, and servers depending on your implementation needs.
应注意,本公开的技术方案中,所涉及的用户个人信息的收集、存储、使用、加工、传输、提供和公开等处理,均符合相关法律法规的规定,且不违背公序良俗。It should be noted that in the technical solution of this disclosure, the collection, storage, use, processing, transmission, provision and disclosure of user personal information are in compliance with relevant laws and regulations and do not violate public order and good customs.
在本公开的技术方案中,在获取或采集用户个人信息之前,均获取了用户的授权或同意。In the technical solution of the present disclosure, the user's authorization or consent is obtained before obtaining or collecting the user's personal information.
本公开实施例提供了一种虚拟形象驱动方法,下面结合图1的系统架构,参考图2A~图5来描述根据本公开示例性实施方式的虚拟形象驱动方法。本公开实施例的虚拟形象驱动方法例如可以由图1所示的服务器105来执行。Embodiments of the present disclosure provide an avatar driving method. The avatar driving method according to an exemplary embodiment of the present disclosure will be described below in conjunction with the system architecture of FIG. 1 and with reference to FIGS. 2A to 5 . The avatar driving method according to the embodiment of the present disclosure may be executed by the server 105 shown in FIG. 1 , for example.
图2A示意性示出了根据本公开一实施例的虚拟形象驱动方法的流程图。FIG. 2A schematically shows a flow chart of an avatar driving method according to an embodiment of the present disclosure.
如图2A所示,本公开实施例的虚拟形象驱动方法200例如可以包括操作S210~操作S250。As shown in FIG. 2A , the avatar driving method 200 according to the embodiment of the present disclosure may include operations S210 to S250, for example.
在操作S210,接收输入数据流。In operation S210, an input data stream is received.
输入数据流包括多个时序帧,任意一个时序帧与面部姿态相关联,面部姿态包括第一部位姿态和第二部位姿态,第二部位姿态包括多个第二子部位姿态,任意多个第一部位姿态之间的相关性小于多个第二子部位姿态之间的相关性。The input data stream includes multiple timing frames, and any timing frame is associated with a facial posture. The facial posture includes a first part posture and a second part posture. The second part posture includes a plurality of second sub-part postures, and any multiple first part postures. The correlation between part poses is smaller than the correlation between multiple second sub-part poses.
任意多个第一部位姿态之间的相关性小于多个第二子部位姿态之间的相关性可以理解为任意多个第一部位姿态之间的相关性小于与任意一个第二部位姿态关联的多个第二子部位姿态之间的相关性。The correlation between any multiple first part poses is smaller than the correlation between multiple second sub-part poses. It can be understood that the correlation between any multiple first part poses is smaller than the correlation associated with any second part pose. Correlation between poses of multiple second subparts.
示例性地,输入数据流例如可以基于相机阵列对某一个真人对象扫描时间维度上连续的三维面部表情,对扫描的三维面部区域进行规整化处理得到,使得输入数据流可以体现时间维度的时序帧,针对任意一个目标时序帧,还可以体现该目标时序帧显示的图元顶点位置,图元顶点位置表征图元顶点的位置,图元顶点可以理解为图元(图形元素)的顶点,可以由几何顶点组合为图元,图元例如可以包括点、线段、多边形等,每一个图像显示的目标对象可以由多个图元组成,图元顶点位置例如是包括x坐标、y坐标以及z坐标的三维数据。For example, the input data stream can be obtained by scanning a real person object's continuous three-dimensional facial expressions in the time dimension based on a camera array, and regularizing the scanned three-dimensional facial areas, so that the input data stream can reflect the time-series frames in the time dimension. , for any target timing frame, it can also reflect the vertex position of the primitive displayed in the target timing frame. The vertex position of the primitive represents the position of the vertex of the primitive. The vertex of the primitive can be understood as the vertex of the primitive (graphical element), which can be expressed by Geometric vertices are combined into primitives. The primitives can include points, line segments, polygons, etc., for example. The target object displayed by each image can be composed of multiple primitives. The vertex position of the primitive includes, for example, x coordinates, y coordinates, and z coordinates. 3D data.
示例性地,第一部位姿态和第二部位姿态例如可以预先对面部进行部位划分得到。For example, the first part pose and the second part pose may be obtained by dividing the face into parts in advance.
第一部位姿态也可以包括一个,此时,可以认为第一部位姿态相比于第二部位姿态的多个第二子部位姿态更独立。The first part posture may also include one. In this case, the first part posture may be considered to be more independent than the plurality of second sub-part postures of the second part posture.
示例性地,可以将面部的眼睛区域、鼻子区域以及下巴区域作为第一部位区域,针对任意一个目标时序帧显示的面部姿态,第一部位姿态对应于第一部位区域的姿态。第一部位姿态例如可以包括眼睛区域姿态、鼻子区域姿态以及下巴区域姿态。For example, the eye area, nose area and chin area of the face can be used as the first part area, and for the facial posture displayed in any target timing frame, the first part posture corresponds to the posture of the first part area. The first part posture may include, for example, an eye area posture, a nose area posture, and a chin area posture.
示例性地,还可以将例如下巴-牙齿区域、下巴-嘴唇区域以及眼睛-眉毛区域作为第二区域,针对任意一个目标时序帧显示的面部姿态,第二部位姿态对应于第二部位区域的姿态。第二部位姿态例如可以包括下巴-牙齿区域姿态、下巴-嘴唇区域姿态以及眼睛-眉毛区域姿态。For example, the chin-tooth area, the chin-lip area, and the eye-eyebrow area can also be used as the second area. For the facial posture displayed in any target timing frame, the second part posture corresponds to the posture of the second part area. . The second part posture may include, for example, a chin-tooth region posture, a chin-lip region posture, and an eye-brow region posture.
针对下巴-牙齿区域姿态这一第二部位姿态,相应的第二子部位姿态包括下巴子区域姿态和牙齿子区域姿态。针对下巴-嘴唇区域姿态这一第二部位姿态,相应的第二子部位姿态包括下巴子区域姿态和嘴唇子区域姿态。针对眼睛-眉毛区域姿态这一第二部位姿态,相应的第二子部位姿态包括眼睛子区域姿态和眉毛子区域姿态。For the second part posture of the chin-tooth region posture, the corresponding second sub-part posture includes the chin sub-region posture and the tooth sub-region posture. For the second part posture of the chin-lip region posture, the corresponding second sub-part posture includes the chin sub-region posture and the lip sub-region posture. For the second part posture of the eye-eyebrow region posture, the corresponding second sub-part posture includes the eye sub-region posture and the eyebrow sub-region posture.
如图2B所示,示意性示出了下巴-嘴唇区域姿态的多个具体示例。As shown in FIG. 2B , multiple specific examples of chin-lip region poses are schematically shown.
需要说明的是,任意多个第一部位姿态之间的相关性小于与任意一个第二部位姿态关联的第二子部位姿态之间的相关性可以理解为在作出相应表情时,第一部位姿态之间的关联程度较低,但是与任意一个第二部位姿态关联的第二子部位姿态之间的关联程度较高。例如,在上述示例中,针对眼睛区域和下巴区域对应的两个第一部位姿态,由于眼睛运动不会关联到下巴同步运动,因此由于眼睛运动引起的表情变化应该体现在眼睛区域姿态变化,下巴区域姿态应该不发生变化。因此,眼睛区域姿态与下巴区域姿态之间应该是具有低相关性或者低耦合性的,例如可以将各个第一部位姿态解耦。It should be noted that if the correlation between any multiple first part postures is less than the correlation between the second sub-part postures associated with any second part posture, it can be understood that when making the corresponding expression, the first part posture The degree of correlation between them is low, but the degree of correlation between the second sub-part postures associated with any second part posture is high. For example, in the above example, for the two first part postures corresponding to the eye area and chin area, since the eye movement will not be associated with the synchronous movement of the chin, the expression changes caused by the eye movement should be reflected in the posture changes in the eye area and chin The regional attitude should not change. Therefore, the eye region posture and the chin region posture should have low correlation or low coupling, for example, each first part posture can be decoupled.
例如,在上述示例中,下巴-牙齿区域姿态这一第二区域姿态涉及下巴区域和牙齿区域。实际的人体结构中下巴和牙齿是一体连接的,所以由于下巴运动引起的表情变化,实际上会关联到牙齿的同步运动。因此,针对下巴-牙齿区域姿态这一第二部位姿态,相关联的下巴区域姿态与牙齿区域姿态应该是具有强相关性或者强耦合性的。For example, in the above example, the second region posture, the chin-tooth region posture, involves the chin region and the tooth region. In the actual human body structure, the jaw and teeth are integrally connected, so changes in expression caused by jaw movement are actually related to the synchronous movement of the teeth. Therefore, for the second part posture of the chin-tooth region posture, the associated chin region posture and tooth region posture should have strong correlation or strong coupling.
在操作S220,针对多个时序帧中的任意一个目标时序帧,根据初始基准姿态以及第一部位姿态,确定第一姿态变换系数。In operation S220, for any target timing frame among the plurality of timing frames, the first posture transformation coefficient is determined based on the initial reference posture and the first part posture.
示例性地,初始基准姿态例如可以是输入数据流的某一个时序帧的面部姿态。For example, the initial reference pose may be the facial pose of a certain timing frame of the input data stream.
姿态变换系数可以理解为可重建为相应的面部姿态的系数数值。本公开实施例中,第一姿态变换系数、第二姿态变换系数等仅用于区分不同的姿态变换系数。The pose transformation coefficient can be understood as the coefficient value that can be reconstructed into the corresponding facial pose. In the embodiment of the present disclosure, the first posture transformation coefficient, the second posture transformation coefficient, etc. are only used to distinguish different posture transformation coefficients.
示例性地,姿态变换系数例如可以包括blendshape,即混合变形系数。For example, the posture transformation coefficient may include blendshape, that is, a blend deformation coefficient.
示例性地,初始基准姿态例如可以是“无表情”的面部姿态。By way of example, the initial reference pose may be a "no expression" facial pose, for example.
在操作S230,根据第一姿态变换系数更新初始基准姿态,得到更新基准姿态。In operation S230, the initial reference posture is updated according to the first posture transformation coefficient to obtain an updated reference posture.
在操作S240,根据更新基准姿态以及第二部位姿态,确定第二姿态变换系数。In operation S240, the second posture transformation coefficient is determined according to the updated reference posture and the second part posture.
在操作S250,根据第二姿态变换系数驱动虚拟形象。In operation S250, the avatar is driven according to the second posture transformation coefficient.
根据本公开实施例的虚拟形象驱动方法,通过接收输入数据流以及针对多个时序帧中的任意一个目标时序帧,根据初始基准姿态以及第一部位姿态,确定的第一姿态变换系数,可以实现对初始基准姿态中对应第一部位区域的姿态进行更新,通过根据第一姿态变换系数更新初始基准姿态得到更新基准姿态,可以实现将第一部位区域的更新的姿态作为新的基准姿态,通过根据更新基准姿态以及第二部位姿态确定的第二姿态变换系数可以在第一部位姿态更新的基础上对第二部位姿态进行更新,由此驱动的虚拟形象的面部表情更加真实。According to the avatar driving method according to the embodiment of the present disclosure, by receiving the input data stream and the first posture transformation coefficient determined based on the initial reference posture and the first part posture for any one of the target timing frames among the plurality of timing frames, it can be achieved The posture corresponding to the first part region in the initial reference posture is updated, and the updated reference posture is obtained by updating the initial reference posture according to the first posture transformation coefficient. The updated posture of the first part area can be used as the new reference posture. Updating the reference posture and the second posture transformation coefficient determined by the second part's posture can update the second part's posture based on the first part's posture update, so that the facial expression of the virtual image driven by this is more realistic.
根据本公开实施例的虚拟形象驱动方法,可以对面部姿态基于相关性进行细分,由于不同的部位之间的相关性会影响到作出面部表情的真实性(例如,一些实施方式根据相应的姿态变换系数重建的面部姿态会显示出下巴运动使得嘴巴张开,但是牙齿未运动的情况,与真实的人体结构做出的表情冲突),因此,根据本公开实施例的虚拟形象驱动方法可以基于低相关性的第一部位姿态和高相关性的第二子部位姿态,可以降低虚拟形象驱动出现的面部表情混乱等情况,使得驱动虚拟形象作出的表情更真实,真实性例如体现在低相关性的第一部位区域不会互相关联作出面部表情、高相关性的第二部位区域不会独立作出面部表情。According to the avatar driving method of the embodiment of the present disclosure, facial gestures can be subdivided based on correlation, because the correlation between different parts will affect the authenticity of facial expressions (for example, some embodiments can use The facial posture reconstructed by the transformation coefficient will show that the jaw moves to open the mouth, but the teeth do not move, which conflicts with the expression made by the real human body structure). Therefore, the avatar driving method according to the embodiment of the present disclosure can be based on low The first part posture of correlation and the second sub-part posture of high correlation can reduce the confusion of facial expressions caused by the avatar driver, making the expressions driven by the avatar more realistic. The authenticity is reflected in the low correlation, for example. The first area will not be correlated with each other to make facial expressions, and the second area with high correlation will not be able to make facial expressions independently.
图3示意性示出了根据本公开另一实施例的虚拟形象驱动方法的示意图。FIG. 3 schematically shows a schematic diagram of an avatar driving method according to another embodiment of the present disclosure.
如图3所示,根据本公开又一实施例的虚拟形象驱动方法300例如可以利用以下实施例实现操作S350的根据第二姿态变换系数驱动虚拟形象的具体示例。As shown in FIG. 3 , the avatar driving method 300 according to another embodiment of the present disclosure can, for example, use the following embodiments to implement a specific example of driving the avatar according to the second posture transformation coefficient of operation S350.
输入数据流包括顶点流数据,顶点流数据的每一个时序帧由多个图元顶点位置表征面部姿态。在图3的示例中,示意性示出了输入数据流Input为顶点流数据,例如顶点流数据的目标时序帧Pi由多个图元顶点位置307表征对应的面部姿态fp-i。The input data stream includes vertex stream data, and each temporal frame of the vertex stream data represents the facial posture by multiple primitive vertex positions. In the example of FIG. 3 , it is schematically shown that the input data stream Input is vertex stream data. For example, the target timing frame Pi of the vertex stream data represents the corresponding facial pose fp-i by multiple primitive vertex positions 307 .
在操作S351,针对目标时序帧Pi,根据第二姿态变换系数306以及目标时序帧Pi对应的图元顶点位置307,确定第二姿态变换差异数据308。In operation S351, for the target temporal frame Pi, the second posture transformation difference data 308 is determined according to the second posture transformation coefficient 306 and the primitive vertex position 307 corresponding to the target temporal frame Pi.
针对任意一个目标时序帧Pi,对应的图元顶点位置307是相机拍摄的真实的面部姿态,第二姿态变换系数是用于重建面部姿态的,根据第二姿态变换系数重建的面部姿态与真实的面部姿态(真实的面部姿态可以利用相应的图元顶点位置表征)之间具有误差,可以利用第二姿态变换差异数据表征第二姿态变换系数重建的面部姿态与真实面部姿态之间的差异程度。For any target timing frame Pi, the corresponding primitive vertex position 307 is the real facial posture captured by the camera. The second posture transformation coefficient is used to reconstruct the facial posture. The facial posture reconstructed according to the second posture transformation coefficient is consistent with the real facial posture. If there is an error between facial poses (the real facial pose can be represented by the corresponding primitive vertex position), the second pose transformation difference data can be used to characterize the degree of difference between the facial pose reconstructed by the second pose transformation coefficient and the real facial pose.
在操作S352,根据第二姿态变换差异数据308以及第二姿态变换系数306,确定第三姿态变换系数309。In operation S352, a third posture transformation coefficient 309 is determined based on the second posture transformation difference data 308 and the second posture transformation coefficient 306.
在操作S353,根据第三姿态变换系数309驱动虚拟形象310。In operation S353, the avatar 310 is driven according to the third posture transformation coefficient 309.
在上述第二姿态变换差异数据表征第二姿态变换系数重建面部姿态与真实面部姿态之间的差异程度的情况下,根据第二变换差异数据以及第二姿态变换系数可以确定第三姿态变换系数。第三姿态变换系数相比于第二姿态变换系数用于重建面部姿态的误差更小。In the case where the second posture transformation difference data represents the degree of difference between the facial posture reconstructed by the second posture transformation coefficient and the real facial posture, the third posture transformation coefficient can be determined according to the second transformation difference data and the second posture transformation coefficient. The error of the third posture transformation coefficient for reconstructing the facial posture is smaller than that of the second posture transformation coefficient.
根据本公开实施例的虚拟形象驱动方法,通过针对目标时序帧,根据第二姿态变换系数以及目标时序帧对应的图元顶点位置,确定的第二姿态变换差异数据可以衡量第二姿态变换系数重建的面部姿态的误差。通过根据第二姿态变换差异数据以及第二姿态变换系数,可以基于误差对第二姿态变换系数进行调整,得到的第三姿态变换系数的误差更小,根据第三姿态变换系数驱动的虚拟形象可以作出更准确的面部表情。According to the avatar driving method of the embodiment of the present disclosure, by aiming at the target timing frame, the determined second posture transformation difference data can measure the reconstruction of the second posture transformation coefficient according to the second posture transformation coefficient and the primitive vertex position corresponding to the target timing frame. facial pose error. By using the second posture transformation difference data and the second posture transformation coefficient, the second posture transformation coefficient can be adjusted based on the error, and the error of the obtained third posture transformation coefficient is smaller. The avatar driven according to the third posture transformation coefficient can Make more accurate facial expressions.
在图3的示例中,还示意性示出了操作S310~操作S340,操作S310~操作S340分别与上述实施例的操作S210~操作S240类似。In the example of FIG. 3 , operations S310 to S340 are also schematically shown. Operations S310 to S340 are respectively similar to operations S210 to S240 of the above embodiment.
例如,在图3的示例中,示意性示出了操作S310的接收输入数据流Input。输入数据流Input例如包括时序帧P1至时序帧PN的共计N个时序帧,每一个时序帧与相应的面部姿态相关联。例如,时序帧P1与面部姿态fp-1关联。以目标时序帧Pi为例,关联的面部姿态fp-i包括第一部位姿态301和第二部位姿态302。For example, in the example of FIG. 3 , the receiving input data stream Input of operation S310 is schematically shown. The input data stream Input includes, for example, a total of N timing frames from timing frame P 1 to timing frame PN, and each timing frame is associated with a corresponding facial gesture. For example, temporal frame P 1 is associated with facial pose fp-1. Taking the target temporal frame Pi as an example, the associated facial pose fp-i includes a first part pose 301 and a second part pose 302.
在图3的示例中,示意性示出了操作S320的根据初始基准姿态303以及第一部位姿态301,确定第一姿态变换系数304。还示意性示出了S330的根据第一姿态变换系数304更新初始基准姿态303,得到更新基准姿态305。还示意性示出了S340的根据更新基准姿态305以及第二部位姿态302,确定第二姿态变换系数306。In the example of FIG. 3 , it is schematically shown that operation S320 determines the first posture transformation coefficient 304 according to the initial reference posture 303 and the first part posture 301 . It is also schematically shown that S330 updates the initial reference posture 303 according to the first posture transformation coefficient 304 to obtain the updated reference posture 305. Determining the second posture transformation coefficient 306 according to the updated reference posture 305 and the second part posture 302 in S340 is also schematically shown.
图4示意性示出了根据本公开又一实施例的虚拟形象驱动方法的确定第三姿态变换系数的示意图。FIG. 4 schematically shows a schematic diagram of determining the third posture transformation coefficient of the avatar driving method according to yet another embodiment of the present disclosure.
如图4所示,根据本公开又一实施例的虚拟形象驱动方法,例如可以利用以下实施例实现根据第二姿态变换差异数据以及第二姿态变换系数,确定第三姿态变换系数的具体示例。As shown in FIG. 4 , according to the avatar driving method according to another embodiment of the present disclosure, for example, the following embodiment can be used to implement a specific example of determining the third posture transformation coefficient according to the second posture transformation difference data and the second posture transformation coefficient.
在操作S461,根据第二姿态变换差异数据401和初始基准姿态402,确定第一差异姿态403,第一差异姿态由相应的图元顶点的位置404表征。In operation S461, a first difference posture 403 is determined based on the second posture transformation difference data 401 and the initial reference posture 402, and the first difference posture is represented by the position 404 of the corresponding primitive vertex.
第二姿态变换差异数据作为误差类型的数据,根据第二姿态变换差异数据和初始基准姿态确定的第一差异姿态可以以面部姿态的方式呈现误差类型的第二姿态变换差异数据。The second posture transformation difference data serves as the error type data, and the first difference posture determined based on the second posture transformation difference data and the initial reference posture may present the second posture transformation difference data of the error type in the form of a facial posture.
在操作S462,利用第一骨骼节点数据405对第一差异姿态403进行骨骼分解,得到骨骼-顶点第一关联数据406。In operation S462, the first differential posture 403 is decomposed using the first skeleton node data 405 to obtain the first skeleton-vertex associated data 406.
第一骨骼节点数据405包括第一骨骼节点的数量405-M和第一骨骼节点的位姿405-Pos,骨骼-顶点第一关联数据表征任意一个第一骨骼节点与第一差异姿态相应的图元顶点之间的关联权重。The first skeletal node data 405 includes the number of first skeletal nodes 405-M and the pose 405-Pos of the first skeletal node. The first bone-vertex associated data represents a graph corresponding to any first skeletal node and the first difference pose. The association weight between meta-vertices.
第一差异姿态相应的图元顶点可以理解为利用图元顶点位置表征第一差异姿态的情况下,相应的图元顶点。The corresponding primitive vertex of the first difference posture can be understood as the corresponding primitive vertex when the position of the primitive vertex is used to represent the first difference posture.
每一个骨骼节点可以关联第一差异姿态相应的多个图元顶点,每一个被关联的第一差异姿态相应的图元顶点与骨骼节点均具有相应的关联权重,即可以实现第一骨骼节点与第一差异姿态相应的图元顶点之间一对多的关联,并且第一骨骼节点的位姿具有x轴、y轴、z轴的3个移动自由度和3个转动自由度的共计6个自由度,而图元顶点位置具有x轴、y轴、z轴的3个移动自由度,相比于第一差异姿态相应的图元顶点的位置,利用第一骨骼节点对第一差异姿态进行骨骼分解,得到骨骼-顶点第一关联数据具有更优的表达性,可以拟合更复杂、更细微的面部姿态,适应于第一差异姿态的细微的形变(第一差异姿态是基于误差类型的第二姿态差异数据得到的,所以第一差异姿态是比较细微的形变)。另外,一些复杂的面部姿态是真实人类难以作出的,通过本公开实施例的虚拟形象驱动方法,例如可以驱动虚拟形象作出真实人类难以作出的面部姿态。Each bone node can be associated with multiple primitive vertices corresponding to the first difference posture. Each associated primitive vertex and bone node corresponding to the first difference posture have corresponding association weights, that is, the first bone node and There is a one-to-many association between the primitive vertices corresponding to the first difference posture, and the posture of the first skeletal node has 3 degrees of freedom of movement on the x-axis, y-axis, and z-axis and 3 degrees of freedom of rotation, a total of 6 Degrees of freedom, and the vertex position of the primitive has three degrees of freedom of movement in the x-axis, y-axis, and z-axis. Compared with the position of the primitive vertex corresponding to the first difference posture, the first skeletal node is used to perform the first difference posture Bone decomposition to obtain the first bone-vertex associated data has better expressivity, can fit more complex and subtle facial postures, and is adapted to the subtle deformation of the first difference posture (the first difference posture is based on the error type The second posture difference data is obtained, so the first difference posture is a relatively subtle deformation). In addition, some complex facial gestures are difficult for real humans to make. Through the virtual image driving method of the embodiment of the present disclosure, for example, the virtual image can be driven to make facial gestures that are difficult for real humans to make.
示例性地,第一骨骼节点的数量和位姿例如可以由相关人员人工设置,或者可以根据第一差异姿态对应的姿态(形变)基于骨骼分解原理自动推理得到。For example, the number and posture of the first skeletal nodes can be manually set by relevant personnel, or can be automatically inferred based on the principle of skeletal decomposition according to the posture (deformation) corresponding to the first difference posture.
在第一差异姿态以面部姿态的方式呈现误差类型的第二姿态变换差异数据的情况下,例如可以由相关人员在重建的可视化的第一差异姿态的基础上人工设置第一骨骼节点的数量和位姿。In the case where the first difference pose presents second pose transformation difference data of the error type in the form of a facial pose, for example, the number of first skeletal nodes and Posture.
示例性地,例如可以基于线性蒙皮分解算法对第一差异姿态进行骨骼分解,线性蒙皮分解算法简称SSDR(Smooth Skinning Decomposition With Rigid Bones)。SSDR模型可以从一系列动作中分解出线性蒙皮数据,从而通过一定数量的骨骼和顶点权重图来近似拟合出该动作对应的形变。For example, the first difference posture may be decomposed into bones based on a linear skinning decomposition algorithm, which is referred to as SSDR (Smooth Skinning Decomposition With Rigid Bones). The SSDR model can decompose linear skinning data from a series of actions, and then approximately fit the deformation corresponding to the action through a certain number of bone and vertex weight maps.
在操作S463,根据骨骼-顶点第一关联数据406,确定与骨骼-节点第一关联数据对应的第二差异姿态407。In operation S463, according to the first bone-vertex correlation data 406, the second difference posture 407 corresponding to the first bone-node correlation data is determined.
在操作S464,根据第二姿态变换系数408和第二差异姿态407,确定第三姿态变换系数409。In operation S464, a third posture transformation coefficient 409 is determined based on the second posture transformation coefficient 408 and the second difference posture 407.
需要说明的是,第二差异姿态与第一差异姿态在重建结果方面是相同的,两者的区别在于第一差异姿态是利用相应的图元顶点位置表征的,第二差异姿态是利用骨骼-节点第一关联数据表征的。It should be noted that the reconstruction results of the second difference pose and the first difference pose are the same. The difference between the two is that the first difference pose is represented by the corresponding primitive vertex position, and the second difference pose is represented by the skeleton- The node is represented by the first associated data.
在需要确定用于重建面部姿态的第三姿态变换系数的时候,利用第二差异姿态可以引入骨骼-节点第一关联数据,骨骼-节点第一关联数据的表达能力更强且数据量相比于图元顶点位置具有更少的数据量,可以适应更细微以及更大量的面部姿态以及形变。When it is necessary to determine the third posture transformation coefficient for reconstructing the facial posture, the second difference posture can be used to introduce the first bone-node associated data. The first bone-node associated data has stronger expressive ability and the data volume is compared to Primitive vertex positions have less data and can accommodate more subtle and larger facial poses and deformations.
示例性地,根据本公开又一实施例的虚拟形象驱动方法,例如还可以利用以下实施例实现根据第三姿态变换系数驱动虚拟形象的具体示例:针对目标时序帧,根据第三姿态变换系数以及目标时序帧对应的图元顶点位置,确定第三姿态变换差异数据。在第三姿态变换差异数据不满足姿态变换差异条件的情况下,根据第三姿态变换差异数据和第三姿态变换系数,确定第四姿态变换系数,其中,第四姿态变换系数对应的面部姿态与目标时序帧对应的图元顶点位置之间的差异数值满足姿态变换差异条件。根据第四姿态变换系数驱动虚拟形象。Illustratively, according to the avatar driving method according to another embodiment of the present disclosure, for example, the following embodiment can also be used to implement a specific example of driving the avatar according to the third posture transformation coefficient: for the target timing frame, according to the third posture transformation coefficient and The vertex position of the primitive corresponding to the target timing frame determines the third attitude transformation difference data. When the third posture transformation difference data does not satisfy the posture transformation difference condition, the fourth posture transformation coefficient is determined based on the third posture transformation difference data and the third posture transformation coefficient, where the facial posture corresponding to the fourth posture transformation coefficient is the same as the facial posture corresponding to the fourth posture transformation coefficient. The numerical difference between the primitive vertex positions corresponding to the target timing frame satisfies the attitude transformation difference condition. The virtual image is driven according to the fourth posture transformation coefficient.
根据本公开实施例的虚拟形象驱动方法,为了获得更加准确的姿态变换系数,可以通过姿态变换差异条件将姿态变化系数控制在指定的误差范围内。在上述实施例确定的第三姿态变换系数不满足姿态变换差异条件的情况下,可以基于第三姿态变换系数确定的第三姿态变换差异数据继续调整第三姿态变换系数,直到得到的第四姿态变换系数满足姿态变换差异条件,使得根据第四姿态变换系数驱动虚拟形象具有更高的准确性。According to the avatar driving method of the embodiment of the present disclosure, in order to obtain a more accurate posture transformation coefficient, the posture change coefficient can be controlled within a specified error range through the posture transformation difference condition. When the third posture transformation coefficient determined in the above embodiment does not satisfy the posture transformation difference condition, the third posture transformation coefficient can be continuously adjusted based on the third posture transformation difference data determined by the third posture transformation coefficient until the fourth posture is obtained. The transformation coefficient satisfies the posture transformation difference condition, so that driving the virtual image according to the fourth posture transformation coefficient has higher accuracy.
示例性地,根据本公开又一实施例的虚拟形象驱动方法,例如可以利用以下实施例实现根据第三姿态变换差异数据和第三姿态变换系数,确定第四姿态变换系数的具体示例:根据第三姿态变换差异数据和初始基准姿态,确定第三差异姿态。利用第二骨骼节点数据对第三差异姿态进行骨骼分解,得到骨骼-顶点第二关联数据。根据骨骼-顶点第二关联数据,确定与骨骼-节点第二关联数据对应的第四差异姿态。根据第三姿态变换系数和第四差异姿态,确定第四姿态变换系数。Illustratively, according to the avatar driving method according to another embodiment of the present disclosure, for example, the following embodiment can be used to implement a specific example of determining the fourth posture transformation coefficient according to the third posture transformation difference data and the third posture transformation coefficient: according to the third posture transformation difference data and the third posture transformation coefficient. The three-posture transformation difference data and the initial reference posture are determined to determine the third difference posture. Use the second skeletal node data to perform skeletal decomposition on the third difference posture to obtain the second bone-vertex associated data. According to the second bone-vertex related data, a fourth difference posture corresponding to the second bone-node related data is determined. The fourth posture transformation coefficient is determined based on the third posture transformation coefficient and the fourth difference posture.
第三差异姿态由相应的图元顶点的位置表征。The third difference pose is characterized by the position of the corresponding primitive vertex.
第二骨骼节点数据包括第二骨骼节点的数量和位姿,骨骼-顶点第二关联数据表征任意一个第二骨骼节点与第三差异姿态相应的图元顶点之间的关联权重。The second skeletal node data includes the number and posture of the second skeletal node, and the second bone-vertex association data represents the association weight between any second skeletal node and the primitive vertex corresponding to the third difference posture.
需要说明的是,在上述实施例先确定了第二姿态差异数据的基础上,之后进一步确定的第三姿态变换差异数据相比于第二姿态差异数据,对应更加细微和更大量的面部姿态和形变。根据本公开实施例的虚拟形象驱动方法,通过骨骼分解进行迭代,可以在提高准确性的基础上,应对细微、大量的面部姿态和形变,具体的原理如上述实施例的说明,在此不再赘述。It should be noted that, on the basis of first determining the second posture difference data in the above embodiment, the third posture transformation difference data further determined corresponds to more subtle and larger amounts of facial postures and facial gestures than the second posture difference data. deformation. According to the avatar driving method of the embodiment of the present disclosure, iteration through bone decomposition can cope with subtle and large-scale facial postures and deformations on the basis of improving accuracy. The specific principle is as explained in the above embodiment and will not be discussed here. Repeat.
示例性地,根据本公开又一实施例的虚拟形象驱动方法,姿态变换系数是根据姿态变换函数得到的,姿态变换函数与图元顶点的位置、基准姿态相关,基准姿态包括初始基准姿态、更新基准姿态中的至少一个。姿态变换系数包括第一姿态变换系数、第二姿态变换系数、第三姿态变换系数以及第四姿态变换系数中的至少一个。第三姿态变换系数为根据第二姿态变换系数得到,第四姿态变换系数为根据第三姿态变换系数得到。Exemplarily, according to the avatar driving method according to another embodiment of the present disclosure, the posture transformation coefficient is obtained according to the posture transformation function. The posture transformation function is related to the position of the primitive vertex and the reference posture. The reference posture includes the initial reference posture, the updated At least one of the base poses. The posture transformation coefficient includes at least one of a first posture transformation coefficient, a second posture transformation coefficient, a third posture transformation coefficient and a fourth posture transformation coefficient. The third posture transformation coefficient is obtained based on the second posture transformation coefficient, and the fourth posture transformation coefficient is obtained based on the third posture transformation coefficient.
需要说明的是,这里的“第一姿态变换系数”、“第二姿态变换系数”、“第三姿态变换系数”以及“第四姿态变换系数”分别与上述实施例的“第一姿态变换系数”、“第二姿态变换系数”、“第三姿态变换系数”以及“第四姿态变换系数”指代相同。It should be noted that the “first posture transformation coefficient”, “second posture transformation coefficient”, “third posture transformation coefficient” and “fourth posture transformation coefficient” here are respectively the same as the “first posture transformation coefficient” in the above embodiment. ”, “second posture transformation coefficient”, “third posture transformation coefficient” and “fourth posture transformation coefficient” refer to the same.
示例性地,例如可以利用以下公式(1)表征姿态变换函数。For example, the posture transformation function can be characterized by the following formula (1).
C(x)=basic+B·x (1)C(x)=basic+B·x (1)
x表征对应于图元顶点的姿态系数的数值,x的数值例如可以在0至1之间。basic表征基准姿态,例如上述实施例的初始基准姿态或者更新基准姿态。B表征相应的图元顶点的位置。x represents the value of the attitude coefficient corresponding to the vertex of the primitive, and the value of x can be between 0 and 1, for example. basic represents the base posture, such as the initial base posture or the updated base posture in the above embodiment. B represents the position of the corresponding primitive vertex.
例如,可以将目标时序帧对应的图元顶点位置作为B,初始基准姿态作为basic,利用公式(1),可以得到针对目标时序帧的第一姿态变换系数x以及第一姿态变换系数对应的面部姿态C(x)。For example, the vertex position of the primitive corresponding to the target timing frame can be regarded as B, and the initial reference posture can be regarded as basic. Using formula (1), the first posture transformation coefficient x for the target timing frame and the face corresponding to the first posture transformation coefficient can be obtained Attitude C(x).
例如,可以将第一姿态变换系数对应的面部姿态C(x)的图元顶点位置作为B,更新基准姿态作为basic,利用公式(1),可以得到针对目标时序帧的第二姿态变换系数x以及第二姿态变换系数对应的面部姿态C(x)。For example, the primitive vertex position of the facial pose C(x) corresponding to the first pose transformation coefficient can be used as B, and the base pose can be updated as basic. Using formula (1), the second pose transformation coefficient x for the target timing frame can be obtained and the facial pose C(x) corresponding to the second pose transformation coefficient.
示例性地,根据本公开又一实施例的虚拟形象驱动方法,姿态变换差异数据是根据姿态变换差异函数得到的,姿态变换差异函数与姿态变换系数和关键顶点的位置相关;关键顶点是根据表征性从全量的图元顶点中确定的;姿态变换差异数据包括第二姿态变换差异数据以及第三姿态变换差异数据中的至少一个。第三姿态变换差异数据为根据第三姿态变换系数得到。Exemplarily, according to the avatar driving method of another embodiment of the present disclosure, the posture transformation difference data is obtained according to the posture transformation difference function, and the posture transformation difference function is related to the posture transformation coefficient and the position of the key vertex; the key vertex is obtained according to the representation The properties are determined from all primitive vertices; the posture transformation difference data includes at least one of the second posture transformation difference data and the third posture transformation difference data. The third posture transformation difference data is obtained based on the third posture transformation coefficient.
需要说明的是,这里的“姿态变换系数”、“第三姿态变换差异数据”分别与上述实施例的“姿态变换系数”、“第三姿态变换差异数据”指代相同。It should be noted that the "posture transformation coefficient" and "third posture transformation difference data" here are respectively the same as the "posture transformation coefficient" and "third posture transformation difference data" in the above embodiment.
示例性地,例如可以利用以下公式(2)表征姿态变换差异函数。For example, the attitude transformation difference function can be characterized by the following formula (2).
需要说明的是,公式(1)中x的数值可以是变化的,即x的取值可以是变化的,x的每一个具体取值可以对应相关的面部姿态,可以利用公式(2)以误差条件对x的取值进行限定,可以确定x的一个准确的具体数值。It should be noted that the value of x in formula (1) can change, that is, the value of x can change. Each specific value of x can correspond to the relevant facial posture. Formula (2) can be used to calculate the error The condition limits the value of x and can determine an accurate specific value of x.
公式(2)中,i表征任意一个目标时序帧,bi表征关键顶点,Xpre表征由当前帧的上一帧计算出的姿态变换系数,公式(2)可以针对输入数据流整体,αi、bi、β1以及β2均表征系数。In formula (2), i represents any target timing frame, b i represents the key vertex, and Xpre represents the attitude transformation coefficient calculated from the previous frame of the current frame. Formula (2) can be used for the entire input data stream, α i b i , β 1 and β 2 all represent coefficients.
综上,根据本公开实施例的虚拟形象驱动方法,通过例如基于第一部位姿态、第二部位姿态、第二姿态变换差异数据、第三姿态变换差异数据等阶段性的多次顶点流解算,可以将例如第一部位姿态、第二部位姿态等相互解耦,解耦可以理解为将原有的一次执行面部姿态重建过程分为多次,每一次得到阶段性的姿态变换系数,在提高虚拟形象驱动准确性的同时,也便于在不同人像之间迁移,本公开实施例的虚拟形象驱动方法具有更高的虚拟形象驱动效率,可以应用于广泛的场景,例如虚拟主播、虚拟客服、虚拟教师等场景。To sum up, according to the avatar driving method according to the embodiment of the present disclosure, through, for example, staged multiple vertex flow calculations based on the first part posture, the second part posture, the second posture transformation difference data, the third posture transformation difference data, etc. , for example, the posture of the first part, the posture of the second part, etc. can be decoupled from each other. Decoupling can be understood as dividing the original once-executed facial posture reconstruction process into multiple times, and each time the phased posture transformation coefficient is obtained. In improving While avatar driving is accurate, it also facilitates migration between different portraits. The avatar driving method of the embodiment of the present disclosure has higher avatar driving efficiency and can be applied to a wide range of scenarios, such as virtual anchors, virtual customer service, virtual Scenes such as teachers.
图5示意性示出了根据本公开一实施例的虚拟形象驱动装置的框图。FIG. 5 schematically shows a block diagram of an avatar driving device according to an embodiment of the present disclosure.
如图5所示,本公开实施例的虚拟形象驱动装置500例如包括输入数据流接收模块510、第一姿态变换系数确定模块520、更新基准姿态确定模块530、第二姿态变换系数确定模块540以及虚拟形象驱动模块540。As shown in Figure 5, the avatar driving device 500 in the embodiment of the present disclosure includes, for example, an input data stream receiving module 510, a first posture transformation coefficient determination module 520, an updated reference posture determination module 530, a second posture transformation coefficient determination module 540, and Avatar driver module 540.
输入数据流接收模块510,用于接收输入数据流,其中,输入数据流包括多个时序帧,任意一个时序帧与面部姿态相关联,面部姿态包括第一部位姿态和第二部位姿态,第二部位姿态包括多个第二子部位姿态,任意多个第一部位姿态之间的相关性小于多个第二子部位姿态之间的相关性。The input data stream receiving module 510 is configured to receive an input data stream, where the input data stream includes a plurality of time series frames, any one of the time series frames is associated with a facial posture, the facial posture includes a first part posture and a second part posture, and a second part posture. The part posture includes a plurality of second sub-part postures, and the correlation between any plurality of first part postures is smaller than the correlation between the plurality of second sub-part postures.
第一姿态变换系数确定模块520,用于针对多个时序帧中的任意一个目标时序帧,根据初始基准姿态以及第一部位姿态,确定第一姿态变换系数。The first posture transformation coefficient determination module 520 is configured to determine the first posture transformation coefficient according to the initial reference posture and the first part posture for any target timing frame among the plurality of timing frames.
更新基准姿态确定模块530,用于根据第一姿态变换系数更新初始基准姿态,得到更新基准姿态。The updated reference posture determination module 530 is used to update the initial reference posture according to the first posture transformation coefficient to obtain the updated reference posture.
第二姿态变换系数确定模块540,用于根据更新基准姿态以及第二部位姿态,确定第二姿态变换系数。The second posture transformation coefficient determination module 540 is used to determine the second posture transformation coefficient according to the updated reference posture and the second part posture.
虚拟形象驱动模块550,用于根据第二姿态变换系数驱动虚拟形象。The virtual image driving module 550 is used to drive the virtual image according to the second posture transformation coefficient.
根据本公开实施例,输入数据流包括顶点流数据,顶点流数据的每一个时序帧由多个图元顶点位置表征面部姿态;虚拟形象驱动模块包括:第二姿态变换差异数据确定子模块,用于针对目标时序帧,根据第二姿态变换系数以及目标时序帧对应的图元顶点位置,确定第二姿态变换差异数据;第三姿态变换系数确定子模块,用于根据第二姿态变换差异数据以及第二姿态变换系数,确定第三姿态变换系数;虚拟形象驱动子模块,用于根据第三姿态变换系数驱动虚拟形象。According to the embodiment of the present disclosure, the input data stream includes vertex stream data, and each temporal frame of the vertex stream data represents the facial posture by multiple primitive vertex positions; the avatar driving module includes: a second posture transformation difference data determination sub-module, using For the target timing frame, determine the second posture transformation difference data according to the second posture transformation coefficient and the primitive vertex position corresponding to the target timing frame; the third posture transformation coefficient determination submodule is used to transform the difference data according to the second posture and The second posture transformation coefficient determines the third posture transformation coefficient; the virtual image driving submodule is used to drive the virtual image according to the third posture transformation coefficient.
根据本公开实施例,第三姿态变换系数确定子模块包括:第一差异姿态确定单元,用于根据第二姿态变换差异数据和初始基准姿态,确定第一差异姿态,第一差异姿态由相应的图元顶点的位置表征;骨骼-顶点第一关联数据确定单元,用于利用第一骨骼节点数据对第一差异姿态进行骨骼分解,得到骨骼-顶点第一关联数据,其中,第一骨骼节点数据包括第一骨骼节点的数量和位姿,骨骼-顶点第一关联数据表征任意一个第一骨骼节点与第一差异姿态相应的图元顶点之间的关联权重;第二差异姿态确定单元,用于根据骨骼-顶点第一关联数据,确定与骨骼-节点第一关联数据对应的第二差异姿态;第三姿态变换系数确定单元,用于根据第二姿态变换系数和第二差异姿态,确定第三姿态变换系数。According to an embodiment of the present disclosure, the third posture transformation coefficient determination sub-module includes: a first difference posture determination unit, configured to determine the first difference posture according to the second posture transformation difference data and the initial reference posture, and the first difference posture is determined by the corresponding The position representation of the vertex of the primitive; the bone-vertex first associated data determination unit is used to perform skeletal decomposition of the first difference posture using the first bone node data to obtain the first bone-vertex associated data, wherein the first bone node data Including the number and posture of the first skeletal nodes, the first bone-vertex associated data represents the association weight between any first skeletal node and the primitive vertex corresponding to the first difference posture; the second difference posture determination unit is used for Determine the second difference posture corresponding to the first bone-node correlation data according to the first bone-vertex correlation data; the third posture transformation coefficient determination unit is used to determine the third posture transformation coefficient according to the second posture transformation coefficient and the second difference posture. Attitude transformation coefficient.
根据本公开实施例,虚拟形象驱动子模块包括:第三姿态变换差异数据确定单元,用于针对目标时序帧,根据第三姿态变换系数以及目标时序帧对应的图元顶点位置,确定第三姿态变换差异数据;第四姿态变换系数确定单元,用于在第三姿态变换差异数据不满足姿态变换差异条件的情况下,根据第三姿态变换差异数据和第三姿态变换系数,确定第四姿态变换系数,其中,第四姿态变换系数对应的面部姿态与目标时序帧对应的图元顶点位置之间的差异数值满足姿态变换差异条件;虚拟形象驱动单元,用于根据第四姿态变换系数驱动虚拟形象。According to an embodiment of the present disclosure, the avatar driving submodule includes: a third posture transformation difference data determination unit, configured to determine the third posture for the target timing frame according to the third posture transformation coefficient and the primitive vertex position corresponding to the target timing frame. transformation difference data; a fourth posture transformation coefficient determination unit, configured to determine the fourth posture transformation based on the third posture transformation difference data and the third posture transformation coefficient when the third posture transformation difference data does not satisfy the posture transformation difference condition. coefficient, wherein the difference value between the facial posture corresponding to the fourth posture transformation coefficient and the primitive vertex position corresponding to the target timing frame satisfies the posture transformation difference condition; the avatar driving unit is used to drive the avatar according to the fourth posture transformation coefficient .
根据本公开实施例,第四姿态变换系数确定单元包括:第三差异姿态确定子单元,用于根据第三姿态变换差异数据和初始基准姿态,确定第三差异姿态,第三差异姿态由相应的图元顶点的位置表征;骨骼-顶点第二关联数据确定子单元,用于利用第二骨骼节点数据对第三差异姿态进行骨骼分解,得到骨骼-顶点第二关联数据,其中,第二骨骼节点数据包括第二骨骼节点的数量和位姿,骨骼-顶点第二关联数据表征任意一个第二骨骼节点与第三差异姿态相应的图元顶点之间的关联权重;第四差异姿态确定子单元,用于根据骨骼-顶点第二关联数据,确定与骨骼-节点第二关联数据对应的第四差异姿态;第四姿态变换系数确定子单元,用于根据第三姿态变换系数和第四差异姿态,确定第四姿态变换系数。According to an embodiment of the present disclosure, the fourth posture transformation coefficient determination unit includes: a third difference posture determination subunit, configured to determine the third difference posture according to the third posture transformation difference data and the initial reference posture, and the third difference posture is determined by the corresponding The position representation of the primitive vertices; the bone-vertex second associated data determination subunit is used to use the second bone node data to perform bone decomposition on the third difference posture to obtain the bone-vertex second associated data, where the second bone node The data includes the number and posture of the second skeletal node. The second bone-vertex association data represents the association weight between any second skeletal node and the primitive vertex corresponding to the third difference posture; the fourth difference posture determination subunit, Used to determine the fourth difference posture corresponding to the second bone-node related data according to the second bone-vertex related data; the fourth posture transformation coefficient determination subunit is used to determine the fourth difference posture based on the third posture transformation coefficient and the fourth difference posture. Determine the fourth posture transformation coefficient.
根据本公开实施例,第一部位姿态包括以下中的至少一个:眼睛区域姿态、鼻子区域姿态以及下巴区域姿态;第二部位姿态包括以下中的至少一个:下巴-牙齿区域姿态、下巴-嘴唇区域姿态、眼睛-眉毛区域姿态,下巴-牙齿区域姿态包括下巴子区域姿态和牙齿子区域姿态,下巴-嘴唇区域姿态包括下巴子区域姿态和嘴唇子区域姿态,眼睛-眉毛区域姿态包括眼睛子区域姿态和眉毛子区域姿态。According to an embodiment of the present disclosure, the first part posture includes at least one of the following: eye area posture, nose area posture, and chin area posture; the second part posture includes at least one of the following: chin-tooth area posture, chin-lip area Posture, eye-eyebrow area posture, chin-tooth area posture includes chin sub-region posture and tooth sub-region posture, chin-lip area posture includes chin sub-region posture and lip sub-region posture, eye-eyebrow area posture includes eye sub-region posture and eyebrow area posture.
根据本公开实施例,姿态变换系数是根据姿态变换函数得到的,姿态变换函数与基准姿态、图元顶点位置相关,基准姿态包括初始基准姿态、更新基准姿态;姿态变换系数包括第一姿态变换系数、第二姿态变换系数、第三姿态变换系数以及第四姿态变换系数中的至少一个。第三姿态变换系数为根据第二姿态变换系数得到,第四姿态变换系数为根据第三姿态变换系数得到。According to the embodiment of the present disclosure, the posture transformation coefficient is obtained according to the posture transformation function. The posture transformation function is related to the reference posture and the vertex position of the primitive. The reference posture includes the initial reference posture and the updated reference posture; the posture transformation coefficient includes the first posture transformation coefficient. , at least one of the second posture transformation coefficient, the third posture transformation coefficient and the fourth posture transformation coefficient. The third posture transformation coefficient is obtained based on the second posture transformation coefficient, and the fourth posture transformation coefficient is obtained based on the third posture transformation coefficient.
根据本公开实施例,姿态变换差异数据是根据姿态变换差异函数得到的,姿态变换差异函数与姿态变换系数和关键顶点的位置相关;关键顶点是根据表征性从全量的图元顶点中确定的;姿态变换差异数据包括第二姿态变换差异数据以及第三姿态变换差异数据中的至少一个。第三姿态变换差异数据为根据第三姿态变换系数得到。According to embodiments of the present disclosure, the posture transformation difference data is obtained based on the posture transformation difference function, which is related to the posture transformation coefficient and the position of the key vertex; the key vertex is determined from the full amount of primitive vertices based on representation; The posture transformation difference data includes at least one of the second posture transformation difference data and the third posture transformation difference data. The third posture transformation difference data is obtained based on the third posture transformation coefficient.
应该理解,本公开装置部分的实施例与本公开方法部分的实施例对应相同或类似,所解决的技术问题和所达到的技术效果也对应相同或类似,本公开在此不再赘述。It should be understood that the embodiments of the device part of the disclosure are the same or similar to the embodiments of the method part of the disclosure, and the technical problems solved and the technical effects achieved are also the same or similar, and the disclosure will not be repeated here.
根据本公开的实施例,本公开还提供了一种电子设备、一种可读存储介质和一种计算机程序产品。According to embodiments of the present disclosure, the present disclosure also provides an electronic device, a readable storage medium, and a computer program product.
图6示出了可以用来实施本公开的实施例的示例电子设备600的示意性框图。电子设备旨在表示各种形式的数字计算机,诸如,膝上型计算机、台式计算机、工作台、个人数字助理、服务器、刀片式服务器、大型计算机、和其它适合的计算机。电子设备还可以表示各种形式的移动装置,诸如,个人数字处理、蜂窝电话、智能电话、可穿戴设备和其它类似的计算装置。本文所示的部件、它们的连接和关系、以及它们的功能仅仅作为示例,并且不意在限制本文中描述的和/或者要求的本公开的实现。Figure 6 shows a schematic block diagram of an example electronic device 600 that may be used to implement embodiments of the present disclosure. Electronic devices are intended to refer to various forms of digital computers, such as laptop computers, desktop computers, workstations, personal digital assistants, servers, blade servers, mainframe computers, and other suitable computers. Electronic devices may also represent various forms of mobile devices, such as personal digital assistants, cellular phones, smart phones, wearable devices, and other similar computing devices. The components shown herein, their connections and relationships, and their functions are examples only and are not intended to limit implementations of the disclosure described and/or claimed herein.
如图6所示,设备600包括计算单元601,其可以根据存储在只读存储器(ROM)602中的计算机程序或者从存储单元608加载到随机访问存储器(RAM)603中的计算机程序,来执行各种适当的动作和处理。在RAM 603中,还可存储设备600操作所需的各种程序和数据。计算单元601、ROM 602以及RAM 603通过总线604彼此相连。输入/输出(I/O)接口605也连接至总线604。As shown in FIG. 6 , the device 600 includes a computing unit 601 that can execute according to a computer program stored in a read-only memory (ROM) 602 or loaded from a storage unit 608 into a random access memory (RAM) 603 Various appropriate actions and treatments. In the RAM 603, various programs and data required for the operation of the device 600 can also be stored. Computing unit 601, ROM 602 and RAM 603 are connected to each other via bus 604. An input/output (I/O) interface 605 is also connected to bus 604.
设备600中的多个部件连接至I/O接口605,包括:输入单元606,例如键盘、鼠标等;输出单元607,例如各种类型的显示器、扬声器等;存储单元608,例如磁盘、光盘等;以及通信单元609,例如网卡、调制解调器、无线通信收发机等。通信单元609允许设备600通过诸如因特网的计算机网络和/或各种电信网络与其他设备交换信息/数据。Multiple components in device 600 are connected to I/O interface 605, including: input unit 606, such as keyboard, mouse, etc.; output unit 607, such as various types of displays, speakers, etc.; storage unit 608, such as magnetic disk, optical disk, etc. ; and communication unit 609, such as a network card, modem, wireless communication transceiver, etc. The communication unit 609 allows the device 600 to exchange information/data with other devices through computer networks such as the Internet and/or various telecommunications networks.
计算单元601可以是各种具有处理和计算能力的通用和/或专用处理组件。计算单元601的一些示例包括但不限于中央处理单元(CPU)、图形处理单元(GPU)、各种专用的人工智能(AI)计算芯片、各种运行机器学习模型算法的计算单元、数字信号处理器(DSP)、以及任何适当的处理器、控制器、微控制器等。计算单元601执行上文所描述的各个方法和处理,例如虚拟形象驱动方法。例如,在一些实施例中,虚拟形象驱动方法可被实现为计算机软件程序,其被有形地包含于机器可读介质,例如存储单元608。在一些实施例中,计算机程序的部分或者全部可以经由ROM 602和/或通信单元609而被载入和/或安装到设备600上。当计算机程序加载到RAM 603并由计算单元601执行时,可以执行上文描述的虚拟形象驱动方法的一个或多个步骤。备选地,在其他实施例中,计算单元601可以通过其他任何适当的方式(例如,借助于固件)而被配置为执行虚拟形象驱动方法。Computing unit 601 may be a variety of general and/or special purpose processing components having processing and computing capabilities. Some examples of the computing unit 601 include, but are not limited to, a central processing unit (CPU), a graphics processing unit (GPU), various dedicated artificial intelligence (AI) computing chips, various computing units that run machine learning model algorithms, digital signal processing processor (DSP), and any appropriate processor, controller, microcontroller, etc. The computing unit 601 performs various methods and processes described above, such as the avatar driving method. For example, in some embodiments, the avatar driving method may be implemented as a computer software program that is tangibly included in a machine-readable medium, such as the storage unit 608. In some embodiments, part or all of the computer program may be loaded and/or installed onto device 600 via ROM 602 and/or communication unit 609. When the computer program is loaded into the RAM 603 and executed by the computing unit 601, one or more steps of the avatar driving method described above may be performed. Alternatively, in other embodiments, the computing unit 601 may be configured to perform the avatar driving method in any other suitable manner (eg, by means of firmware).
本文中以上描述的系统和技术的各种实施方式可以在数字电子电路系统、集成电路系统、场可编程门阵列(FPGA)、专用集成电路(ASIC)、专用标准产品(ASSP)、芯片上系统的系统(SOC)、复杂可编程逻辑设备(CPLD)、计算机硬件、固件、软件、和/或它们的组合中实现。这些各种实施方式可以包括:实施在一个或者多个计算机程序中,该一个或者多个计算机程序可在包括至少一个可编程处理器的可编程系统上执行和/或解释,该可编程处理器可以是专用或者通用可编程处理器,可以从存储系统、至少一个输入装置、和至少一个输出装置接收数据和指令,并且将数据和指令传输至该存储系统、该至少一个输入装置、和该至少一个输出装置。Various implementations of the systems and techniques described above may be implemented in digital electronic circuit systems, integrated circuit systems, field programmable gate arrays (FPGAs), application specific integrated circuits (ASICs), application specific standard products (ASSPs), systems on a chip implemented in a system (SOC), complex programmable logic device (CPLD), computer hardware, firmware, software, and/or combinations thereof. These various embodiments may include implementation in one or more computer programs executable and/or interpreted on a programmable system including at least one programmable processor, the programmable processor The processor, which may be a special purpose or general purpose programmable processor, may receive data and instructions from a storage system, at least one input device, and at least one output device, and transmit data and instructions to the storage system, the at least one input device, and the at least one output device. An output device.
用于实施本公开的方法的程序代码可以采用一个或多个编程语言的任何组合来编写。这些程序代码可以提供给通用计算机、专用计算机或其他可编程数据处理装置的处理器或控制器,使得程序代码当由处理器或控制器执行时使流程图和/或框图中所规定的功能/操作被实施。程序代码可以完全在机器上执行、部分地在机器上执行,作为独立软件包部分地在机器上执行且部分地在远程机器上执行或完全在远程机器或服务器上执行。Program code for implementing the methods of the present disclosure may be written in any combination of one or more programming languages. These program codes may be provided to a processor or controller of a general-purpose computer, special-purpose computer, or other programmable data processing device, such that the program codes, when executed by the processor or controller, cause the functions specified in the flowcharts and/or block diagrams/ The operation is implemented. The program code may execute entirely on the machine, partly on the machine, as a stand-alone software package, partly on the machine and partly on a remote machine or entirely on the remote machine or server.
在本公开的上下文中,机器可读介质可以是有形的介质,其可以包含或存储以供指令执行系统、装置或设备使用或与指令执行系统、装置或设备结合地使用的程序。机器可读介质可以是机器可读信号介质或机器可读储存介质。机器可读介质可以包括但不限于电子的、磁性的、光学的、电磁的、红外的、或半导体系统、装置或设备,或者上述内容的任何合适组合。机器可读存储介质的更具体示例会包括基于一个或多个线的电气连接、便携式计算机盘、硬盘、随机存取存储器(RAM)、只读存储器(ROM)、可擦除可编程只读存储器(EPROM或快闪存储器)、光纤、便捷式紧凑盘只读存储器(CD-ROM)、光学储存设备、磁储存设备、或上述内容的任何合适组合。In the context of this disclosure, a machine-readable medium may be a tangible medium that may contain or store a program for use by or in connection with an instruction execution system, apparatus, or device. The machine-readable medium may be a machine-readable signal medium or a machine-readable storage medium. Machine-readable media may include, but are not limited to, electronic, magnetic, optical, electromagnetic, infrared, or semiconductor systems, devices or devices, or any suitable combination of the foregoing. More specific examples of machine-readable storage media would include one or more wire-based electrical connections, laptop disks, hard drives, random access memory (RAM), read only memory (ROM), erasable programmable read only memory (EPROM or flash memory), optical fiber, portable compact disk read-only memory (CD-ROM), optical storage device, magnetic storage device, or any suitable combination of the above.
为了提供与用户的交互,可以在计算机上实施此处描述的系统和技术,该计算机具有:用于向用户显示信息的显示装置(例如,CRT(阴极射线管)或者LCD(液晶显示器)监视器);以及键盘和指向装置(例如,鼠标或者轨迹球),用户可以通过该键盘和该指向装置来将输入提供给计算机。其它种类的装置还可以用于提供与用户的交互;例如,提供给用户的反馈可以是任何形式的传感反馈(例如,视觉反馈、听觉反馈、或者触觉反馈);并且可以用任何形式(包括声输入、语音输入或者、触觉输入)来接收来自用户的输入。To provide interaction with a user, the systems and techniques described herein may be implemented on a computer having a display device (eg, a CRT (cathode ray tube) or LCD (liquid crystal display) monitor) for displaying information to the user ); and a keyboard and pointing device (eg, a mouse or a trackball) through which a user can provide input to the computer. Other kinds of devices may also be used to provide interaction with the user; for example, the feedback provided to the user may be any form of sensory feedback (e.g., visual feedback, auditory feedback, or tactile feedback); and may be provided in any form, including Acoustic input, voice input or tactile input) to receive input from the user.
可以将此处描述的系统和技术实施在包括后台部件的计算系统(例如,作为数据服务器)、或者包括中间件部件的计算系统(例如,应用服务器)、或者包括前端部件的计算系统(例如,具有图形用户界面或者网络浏览器的用户计算机,用户可以通过该图形用户界面或者该网络浏览器来与此处描述的系统和技术的实施方式交互)、或者包括这种后台部件、中间件部件、或者前端部件的任何组合的计算系统中。可以通过任何形式或者介质的数字数据通信(例如,通信网络)来将系统的部件相互连接。通信网络的示例包括:局域网(LAN)、广域网(WAN)和互联网。The systems and techniques described herein may be implemented in a computing system that includes back-end components (e.g., as a data server), or a computing system that includes middleware components (e.g., an application server), or a computing system that includes front-end components (e.g., A user's computer having a graphical user interface or web browser through which the user can interact with implementations of the systems and technologies described herein), or including such backend components, middleware components, or any combination of front-end components in a computing system. The components of the system may be interconnected by any form or medium of digital data communication (eg, a communications network). Examples of communication networks include: local area network (LAN), wide area network (WAN), and the Internet.
计算机系统可以包括客户端和服务器。客户端和服务器一般远离彼此并且通常通过通信网络进行交互。通过在相应的计算机上运行并且彼此具有客户端-服务器关系的计算机程序来产生客户端和服务器的关系。Computer systems may include clients and servers. Clients and servers are generally remote from each other and typically interact over a communications network. The relationship of client and server is created by computer programs running on corresponding computers and having a client-server relationship with each other.
应该理解,可以使用上面所示的各种形式的流程,重新排序、增加或删除步骤。例如,本发公开中记载的各步骤可以并行地执行也可以顺序地执行也可以不同的次序执行,只要能够实现本公开公开的技术方案所期望的结果,本文在此不进行限制。It should be understood that various forms of the process shown above may be used, with steps reordered, added or deleted. For example, each step described in the present disclosure can be executed in parallel, sequentially, or in a different order. As long as the desired results of the technical solution disclosed in the present disclosure can be achieved, there is no limitation here.
上述具体实施方式,并不构成对本公开保护范围的限制。本领域技术人员应该明白的是,根据设计要求和其他因素,可以进行各种修改、组合、子组合和替代。任何在本公开的精神和原则之内所作的修改、等同替换和改进等,均应包含在本公开保护范围之内。The above-mentioned specific embodiments do not constitute a limitation on the scope of the present disclosure. It will be understood by those skilled in the art that various modifications, combinations, sub-combinations and substitutions are possible depending on design requirements and other factors. Any modifications, equivalent substitutions, improvements, etc. made within the spirit and principles of this disclosure shall be included in the protection scope of this disclosure.
Claims (16)
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN202211670321.3A CN115937373B (en) | 2022-12-23 | 2022-12-23 | Avatar driving method, apparatus, device and storage medium |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN202211670321.3A CN115937373B (en) | 2022-12-23 | 2022-12-23 | Avatar driving method, apparatus, device and storage medium |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| CN115937373A CN115937373A (en) | 2023-04-07 |
| CN115937373B true CN115937373B (en) | 2023-10-03 |
Family
ID=86697564
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| CN202211670321.3A Active CN115937373B (en) | 2022-12-23 | 2022-12-23 | Avatar driving method, apparatus, device and storage medium |
Country Status (1)
| Country | Link |
|---|---|
| CN (1) | CN115937373B (en) |
Citations (7)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN110991327A (en) * | 2019-11-29 | 2020-04-10 | 深圳市商汤科技有限公司 | Interactive method and apparatus, electronic device and storage medium |
| CN112509099A (en) * | 2020-11-30 | 2021-03-16 | 北京百度网讯科技有限公司 | Avatar driving method, apparatus, device and storage medium |
| US10949648B1 (en) * | 2018-01-23 | 2021-03-16 | Snap Inc. | Region-based stabilized face tracking |
| CN112581573A (en) * | 2020-12-15 | 2021-03-30 | 北京百度网讯科技有限公司 | Avatar driving method, apparatus, device, medium, and program product |
| WO2022021686A1 (en) * | 2020-07-28 | 2022-02-03 | 完美世界(北京)软件科技发展有限公司 | Method and apparatus for controlling virtual object, and storage medium and electronic apparatus |
| CN115147523A (en) * | 2022-07-07 | 2022-10-04 | 北京百度网讯科技有限公司 | Virtual image driving method and device, device, medium and program product |
| WO2022222572A1 (en) * | 2021-04-19 | 2022-10-27 | 上海商汤智能科技有限公司 | Method and apparatus for driving interaction object, device, and storage medium |
Family Cites Families (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN110111247B (en) * | 2019-05-15 | 2022-06-24 | 浙江商汤科技开发有限公司 | Face deformation processing method, device and equipment |
| CN111223171B (en) * | 2020-01-14 | 2025-06-27 | 腾讯科技(深圳)有限公司 | Image processing method, device, electronic device and storage medium |
| CN111694429B (en) * | 2020-06-08 | 2023-06-02 | 北京百度网讯科技有限公司 | Virtual object driving method, device, electronic device and readable storage |
| CN112150638B (en) * | 2020-09-14 | 2024-01-26 | 北京百度网讯科技有限公司 | Virtual object image synthesis method, device, electronic equipment and storage medium |
-
2022
- 2022-12-23 CN CN202211670321.3A patent/CN115937373B/en active Active
Patent Citations (7)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US10949648B1 (en) * | 2018-01-23 | 2021-03-16 | Snap Inc. | Region-based stabilized face tracking |
| CN110991327A (en) * | 2019-11-29 | 2020-04-10 | 深圳市商汤科技有限公司 | Interactive method and apparatus, electronic device and storage medium |
| WO2022021686A1 (en) * | 2020-07-28 | 2022-02-03 | 完美世界(北京)软件科技发展有限公司 | Method and apparatus for controlling virtual object, and storage medium and electronic apparatus |
| CN112509099A (en) * | 2020-11-30 | 2021-03-16 | 北京百度网讯科技有限公司 | Avatar driving method, apparatus, device and storage medium |
| CN112581573A (en) * | 2020-12-15 | 2021-03-30 | 北京百度网讯科技有限公司 | Avatar driving method, apparatus, device, medium, and program product |
| WO2022222572A1 (en) * | 2021-04-19 | 2022-10-27 | 上海商汤智能科技有限公司 | Method and apparatus for driving interaction object, device, and storage medium |
| CN115147523A (en) * | 2022-07-07 | 2022-10-04 | 北京百度网讯科技有限公司 | Virtual image driving method and device, device, medium and program product |
Non-Patent Citations (3)
| Title |
|---|
| Jian Luo ; Jin Tang ; Xiaoming Xiao.Robust arbitrary view gait recognition based on parametric 3D human body reconstruction and virtual posture synthesis.《Pattern Recognition》.2016,361–377. * |
| 基于OPENGL的人体姿态数据仿真;刘凯;柴毅;冯文武;;计算机仿真(04);267-270、278 * |
| 基于骨骼数据的三维人体行走姿态模拟;李锦;童立靖;英祥;杨金秋;;数字技术与应用(10);58-61 * |
Also Published As
| Publication number | Publication date |
|---|---|
| CN115937373A (en) | 2023-04-07 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| CN113658309B (en) | Three-dimensional reconstruction method, device, equipment and storage medium | |
| CN112862933B (en) | Method, apparatus, device and storage medium for optimizing model | |
| CN114820905B (en) | Virtual image generation method and device, electronic equipment and readable storage medium | |
| CN112819971B (en) | Method, device, device and medium for generating virtual image | |
| CN115147265B (en) | Avatar generation method, apparatus, electronic device, and storage medium | |
| CN113052962B (en) | Model training method, information output method, device, equipment and storage medium | |
| CN114792355B (en) | Virtual image generation method and device, electronic equipment and storage medium | |
| US20180276870A1 (en) | System and method for mass-animating characters in animated sequences | |
| CN112820408B (en) | Surgical risk determination method, related device and computer program product | |
| CN114708374A (en) | Virtual image generation method, device, electronic device and storage medium | |
| CN116188742A (en) | Virtual object control method, device, equipment and storage medium | |
| CN116843807B (en) | Virtual image generation method, virtual image model training method, virtual image generation device, virtual image model training device and electronic equipment | |
| CN116524165B (en) | Migration method, device, equipment and storage medium for three-dimensional expression model | |
| CN111311712B (en) | Video frame processing method and device | |
| CN113380269B (en) | Video image generation method, apparatus, device, medium, and computer program product | |
| CN114529649A (en) | Image processing method and device | |
| CN114972587A (en) | Expression-driven method, device, electronic device, and readable storage medium | |
| CN115937373B (en) | Avatar driving method, apparatus, device and storage medium | |
| CN112562043A (en) | Image processing method and device and electronic equipment | |
| CN116843808B (en) | Rendering, model training and virtual image generating method and device based on point cloud | |
| CN114820908B (en) | Virtual image generation method, device, electronic device and storage medium | |
| CN117788653A (en) | Animation generation method, animation generation system and computer-readable medium | |
| CN111834021A (en) | Data interaction method, apparatus, device and storage medium | |
| CN116320223A (en) | Photographing method, photographing device, photographing equipment and storage medium | |
| CN116363331B (en) | Image generation method, device, equipment and storage medium |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| PB01 | Publication | ||
| PB01 | Publication | ||
| SE01 | Entry into force of request for substantive examination | ||
| SE01 | Entry into force of request for substantive examination | ||
| GR01 | Patent grant | ||
| GR01 | Patent grant |