[go: up one dir, main page]

CN110378372A - Diagram data recognition methods, device, computer equipment and storage medium - Google Patents

Diagram data recognition methods, device, computer equipment and storage medium Download PDF

Info

Publication number
CN110378372A
CN110378372A CN201910503194.XA CN201910503194A CN110378372A CN 110378372 A CN110378372 A CN 110378372A CN 201910503194 A CN201910503194 A CN 201910503194A CN 110378372 A CN110378372 A CN 110378372A
Authority
CN
China
Prior art keywords
matrix
feature map
target
convolutional neural
trained
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201910503194.XA
Other languages
Chinese (zh)
Inventor
张一帆
史磊
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nanjing Artificial Intelligence Chip Innovation Institute Institute Of Automation Chinese Academy Of Sciences
Institute of Automation of Chinese Academy of Science
Original Assignee
Nanjing Artificial Intelligence Chip Innovation Institute Institute Of Automation Chinese Academy Of Sciences
Institute of Automation of Chinese Academy of Science
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nanjing Artificial Intelligence Chip Innovation Institute Institute Of Automation Chinese Academy Of Sciences, Institute of Automation of Chinese Academy of Science filed Critical Nanjing Artificial Intelligence Chip Innovation Institute Institute Of Automation Chinese Academy Of Sciences
Priority to CN201910503194.XA priority Critical patent/CN110378372A/en
Publication of CN110378372A publication Critical patent/CN110378372A/en
Priority to PCT/CN2019/129708 priority patent/WO2020248581A1/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting

Landscapes

  • Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Image Analysis (AREA)

Abstract

本申请涉及一种图数据识别方法、装置、计算机设备和存储介质。所述方法包括:获取输入已训练的卷积神经网络的当前卷积层的输入特征图,输入特征图为根据图像数据生成的特征图,获取当前卷积层的第一偏置矩阵,其中第一偏置矩阵为生成已训练的卷积神经网络时生成的矩阵,根据输入特征图生成第二偏置矩阵,获取参考邻接矩阵,计算参考邻接矩阵、第一偏置矩阵和第二偏置矩阵的和,得到目标邻接矩阵,获取当前卷积层的卷积核,根据当前卷积层的卷积核、目标邻接矩阵和输入特征图生成目标输出特征图,根据目标输出特征图,识别出图数据的识别结果。对现有的固定的邻接矩阵基础上增加可调整的偏置矩阵,提高已训练的卷积神经网络的识别准确率。

The present application relates to a graph data recognition method, device, computer equipment and storage medium. The method includes: obtaining an input feature map of the current convolutional layer of the trained convolutional neural network, the input feature map is a feature map generated according to image data, and obtaining a first bias matrix of the current convolutional layer, wherein the first A bias matrix is the matrix generated when the trained convolutional neural network is generated, a second bias matrix is generated according to the input feature map, a reference adjacency matrix is obtained, and a reference adjacency matrix, the first bias matrix and the second bias matrix are calculated The sum of the target adjacency matrix is obtained, the convolution kernel of the current convolution layer is obtained, the target output feature map is generated according to the convolution kernel of the current convolution layer, the target adjacency matrix and the input feature map, and the image is identified according to the target output feature map Data recognition results. An adjustable bias matrix is added to the existing fixed adjacency matrix to improve the recognition accuracy of the trained convolutional neural network.

Description

图数据识别方法、装置、计算机设备和存储介质Image data recognition method, device, computer equipment and storage medium

技术领域technical field

本申请涉及计算机技术领域,尤其涉及一种图数据识别方法、装置、计算机设备和存储介质。The present application relates to the field of computer technology, and in particular to a method, device, computer equipment and storage medium for identifying image data.

背景技术Background technique

在骨骼点数据中,人体是由若干预先定义好的关键关节点在相机坐标系中的坐标来表示的。它可以很方便地通过深度摄像头(例如Kinect)以及各种姿态估计算法(例如OpenPose)获得。图1为Kinect深度摄像机所定义的人体的关键关节点。它将人体定义为25个关键关节点的三维坐标。由于行为往往是以视频的形式存在的,所以一个长度为T帧的行为可以用Tx25x3的张量来表示。In the skeleton point data, the human body is represented by the coordinates of several predefined key joint points in the camera coordinate system. It can be easily obtained by depth cameras (such as Kinect) and various pose estimation algorithms (such as OpenPose). Figure 1 shows the key joint points of the human body defined by the Kinect depth camera. It defines the human body as the three-dimensional coordinates of 25 key joint points. Since behaviors often exist in the form of video, a behavior with a length of T frames can be represented by a Tx25x3 tensor.

参照图2,图2为一个实施例中的时空图。每个关节点定义为图的节点,关节点之间的物理连接定义为图的边,并且在相邻帧的同一个节点间加上时间维度的边,得到一张可以描述人体行为的时空图。Referring to FIG. 2, FIG. 2 is a space-time diagram in one embodiment. Each joint point is defined as a node of the graph, and the physical connection between the joint points is defined as the edge of the graph, and the edge of the time dimension is added between the same nodes of adjacent frames to obtain a space-time graph that can describe human behavior .

目前常见的基于骨骼点的行为识别方法为图卷积。图卷积和普通卷积操作不同,在图上做卷积时,每一个节点的邻节点数是不固定的,而卷积操作的参数是固定的,为了将固定数量的参数和不定数量的临节点数对应起来,需要定义映射函数,通过映射函数实现参数和节点的对应。如定义卷积核大小为三,如图3所示,三个参数分别对应于远离人体中心的点001,靠近人体中心点000的点002和卷积点本身003。则卷积操作可以用公式(1)表示:At present, the common behavior recognition method based on skeleton points is graph convolution. Graph convolution is different from ordinary convolution operations. When convolution is performed on a graph, the number of neighbor nodes of each node is not fixed, and the parameters of the convolution operation are fixed. In order to combine a fixed number of parameters with an indeterminate number To correspond to the number of adjacent nodes, it is necessary to define a mapping function, and realize the correspondence between parameters and nodes through the mapping function. If the size of the convolution kernel is defined as three, as shown in Figure 3, the three parameters correspond to the point 001 away from the center of the human body, the point 002 close to the center of the human body 000, and the convolution point itself 003. Then the convolution operation can be expressed by formula (1):

其中f是输入输出特征张量,w是卷积参数,v是图中节点,l代表节点与参数间的映射函数,Z是归一化函数。在具体实现时,映射函数可以通过图的邻接矩阵来实现,通过邻接矩阵表示的卷积操作如公式(2)所示:Among them, f is the input and output feature tensor, w is the convolution parameter, v is the node in the graph, l represents the mapping function between the node and the parameter, and Z is the normalization function. In the specific implementation, the mapping function can be realized through the adjacency matrix of the graph, and the convolution operation represented by the adjacency matrix is shown in formula (2):

其中A代表图的邻接矩阵,K是卷积核大小,Λ用于对A进行归一化处理。通过与邻接矩阵A相乘,从特征张量中“筛选”出所需要的节点并与对应的参数相乘。Where A represents the adjacency matrix of the graph, K is the size of the convolution kernel, and Λ is used to normalize A. By multiplying with the adjacency matrix A, the required nodes are "filtered" from the feature tensor and multiplied with the corresponding parameters.

上述通过邻接矩阵表示的卷积操作时,邻接矩阵定义了用于图卷积网络中的人体图的拓扑结构。人体姿态多种多样,固定的拓扑结构无法准确地描述人体的每一种姿态,从而导致识别准确率低下。When the above convolution operation is represented by the adjacency matrix, the adjacency matrix defines the topology of the human body graph used in the graph convolutional network. There are various postures of the human body, and the fixed topological structure cannot accurately describe each posture of the human body, resulting in low recognition accuracy.

发明内容Contents of the invention

为了解决上述技术问题,本申请提供了一种图数据识别方法、装置、计算机设备和存储介质。In order to solve the above-mentioned technical problems, the present application provides a graph data recognition method, device, computer equipment and storage medium.

第一方面,本申请提供了一种图数据识别方法,包括:In the first aspect, the present application provides a graph data recognition method, including:

获取输入已训练的卷积神经网络的当前卷积层的输入特征图,输入特征图为根据图像数据生成的特征图;Obtain the input feature map of the current convolution layer of the trained convolutional neural network, the input feature map is a feature map generated according to the image data;

获取当前卷积层的第一偏置矩阵,其中第一偏置矩阵为生成已训练的卷积神经网络时生成的矩阵;Obtain the first bias matrix of the current convolutional layer, where the first bias matrix is the matrix generated when the trained convolutional neural network is generated;

根据输入特征图生成第二偏置矩阵;Generate a second bias matrix based on the input feature map;

获取参考邻接矩阵,计算参考邻接矩阵、第一偏置矩阵和第二偏置矩阵的和,得到目标邻接矩阵;Obtain the reference adjacency matrix, calculate the sum of the reference adjacency matrix, the first bias matrix and the second bias matrix, and obtain the target adjacency matrix;

获取当前卷积层的卷积核;Get the convolution kernel of the current convolution layer;

根据当前卷积层的卷积核、目标邻接矩阵和输入特征图生成目标输出特征图;Generate the target output feature map according to the convolution kernel of the current convolution layer, the target adjacency matrix and the input feature map;

根据目标输出特征图,识别出图数据的识别结果。According to the target output feature map, the recognition result of the map data is recognized.

第二方面,本申请提供了一种特图生成装置,包括:In a second aspect, the present application provides a special map generation device, including:

数据获取模块,用于获取输入已训练的卷积神经网络的当前卷积层的输入特征图,输入特征图为根据图像数据生成的特征图,获取当前卷积层的第一偏置矩阵,其中第一偏置矩阵为生成已训练的卷积神经网络时生成的矩阵;The data acquisition module is used to obtain the input feature map of the current convolutional layer of the trained convolutional neural network, the input feature map is a feature map generated according to the image data, and obtain the first bias matrix of the current convolutional layer, wherein The first bias matrix is a matrix generated when the trained convolutional neural network is generated;

第二偏置矩阵生成模块,用于根据输入特征图生成第二偏置矩阵;The second offset matrix generation module is used to generate the second offset matrix according to the input feature map;

目标邻接矩阵生成模块,用于获取参考邻接矩阵,计算参考邻接矩阵、第一偏置矩阵和第二偏置矩阵的和,得到目标邻接矩阵;The target adjacency matrix generation module is used to obtain the reference adjacency matrix, calculate the sum of the reference adjacency matrix, the first offset matrix and the second offset matrix, and obtain the target adjacency matrix;

目标输出特征图生成模块,用于获取当前卷积层的卷积核,根据当前卷积层的卷积核、目标邻接矩阵和输入特征图生成目标输出特征图。The target output feature map generation module is used to obtain the convolution kernel of the current convolution layer, and generate the target output feature map according to the convolution kernel of the current convolution layer, the target adjacency matrix and the input feature map.

识别模块,用于根据目标输出特征图,识别出图数据的识别结果。The recognition module is configured to output a feature map according to the target, and recognize a recognition result of the map data.

一种计算机设备,包括存储器、处理器及存储在存储器上并可在处理器上运行的计算机程序,所述处理器执行所述计算机程序时实现以下步骤:A computer device, comprising a memory, a processor, and a computer program stored on the memory and operable on the processor, and the processor implements the following steps when executing the computer program:

获取输入已训练的卷积神经网络的当前卷积层的输入特征图,输入特征图为根据图像数据生成的特征图;Obtain the input feature map of the current convolution layer of the trained convolutional neural network, the input feature map is a feature map generated according to the image data;

获取当前卷积层的第一偏置矩阵,其中第一偏置矩阵为生成已训练的卷积神经网络时生成的矩阵;Obtain the first bias matrix of the current convolutional layer, where the first bias matrix is the matrix generated when the trained convolutional neural network is generated;

根据输入特征图生成第二偏置矩阵;Generate a second bias matrix based on the input feature map;

获取参考邻接矩阵,计算参考邻接矩阵、第一偏置矩阵和第二偏置矩阵的和,得到目标邻接矩阵;Obtain the reference adjacency matrix, calculate the sum of the reference adjacency matrix, the first bias matrix and the second bias matrix, and obtain the target adjacency matrix;

获取当前卷积层的卷积核;Get the convolution kernel of the current convolution layer;

根据当前卷积层的卷积核、目标邻接矩阵和输入特征图生成目标输出特征图;Generate the target output feature map according to the convolution kernel of the current convolution layer, the target adjacency matrix and the input feature map;

根据目标输出特征图,识别出图数据的识别结果。According to the target output feature map, the recognition result of the map data is recognized.

一种计算机可读存储介质,其上存储有计算机程序,所述计算机程序被处理器执行时实现以下步骤:A computer-readable storage medium, on which a computer program is stored, and when the computer program is executed by a processor, the following steps are implemented:

获取输入已训练的卷积神经网络的当前卷积层的输入特征图,输入特征图为根据图像数据生成的特征图;Obtain the input feature map of the current convolution layer of the trained convolutional neural network, the input feature map is a feature map generated according to the image data;

获取当前卷积层的第一偏置矩阵,其中第一偏置矩阵为生成已训练的卷积神经网络时生成的矩阵;Obtain the first bias matrix of the current convolutional layer, where the first bias matrix is the matrix generated when the trained convolutional neural network is generated;

根据输入特征图生成第二偏置矩阵;Generate a second bias matrix based on the input feature map;

获取参考邻接矩阵,计算参考邻接矩阵、第一偏置矩阵和第二偏置矩阵的和,得到目标邻接矩阵;Obtain the reference adjacency matrix, calculate the sum of the reference adjacency matrix, the first bias matrix and the second bias matrix, and obtain the target adjacency matrix;

获取当前卷积层的卷积核;Get the convolution kernel of the current convolution layer;

根据当前卷积层的卷积核、目标邻接矩阵和输入特征图生成目标输出特征图;Generate the target output feature map according to the convolution kernel of the current convolution layer, the target adjacency matrix and the input feature map;

根据目标输出特征图,识别出图数据的识别结果。According to the target output feature map, the recognition result of the map data is recognized.

上述图数据识别方法、装置、计算机设备和存储介质,所述方法包括:获取输入已训练的卷积神经网络的当前卷积层的输入特征图,输入特征图为根据图像数据生成的特征图,获取当前卷积层的第一偏置矩阵,其中第一偏置矩阵为生成已训练的卷积神经网络时生成的矩阵,根据输入特征图生成第二偏置矩阵,获取参考邻接矩阵,计算参考邻接矩阵、第一偏置矩阵和第二偏置矩阵的和,得到目标邻接矩阵,获取当前卷积层的卷积核,根据当前卷积层的卷积核、目标邻接矩阵和输入特征图生成目标输出特征图,根据目标输出特征图,识别出图数据的识别结果。对已训练的卷积神经网络中各个卷积层中的邻接矩阵增加偏置矩阵,偏置矩阵中的第一偏置矩阵为根据需求确定的已矩阵,和根据输入特征图生成的第二偏置矩阵为根据输入数据生成的矩阵,增加偏置矩阵能够表征需求所需的特征和输入数据的样本特征,提高生成特征图的准确性,进而提高已训练的卷积神经网络的识别准确率。The above image data recognition method, device, computer equipment and storage medium, the method includes: obtaining the input feature map of the current convolution layer of the trained convolutional neural network, the input feature map is a feature map generated according to the image data, Obtain the first bias matrix of the current convolutional layer, where the first bias matrix is the matrix generated when the trained convolutional neural network is generated, generate the second bias matrix according to the input feature map, obtain the reference adjacency matrix, and calculate the reference The sum of the adjacency matrix, the first bias matrix and the second bias matrix is obtained to obtain the target adjacency matrix, and the convolution kernel of the current convolution layer is obtained, which is generated according to the convolution kernel of the current convolution layer, the target adjacency matrix and the input feature map The target output feature map is used to identify the recognition result of the map data according to the target output feature map. Add a bias matrix to the adjacency matrix in each convolutional layer in the trained convolutional neural network, the first bias matrix in the bias matrix is the matrix determined according to the requirements, and the second bias matrix generated according to the input feature map The offset matrix is a matrix generated according to the input data. Adding the offset matrix can represent the required features and the sample features of the input data, improve the accuracy of the generated feature map, and then improve the recognition accuracy of the trained convolutional neural network.

附图说明Description of drawings

此处的附图被并入说明书中并构成本说明书的一部分,示出了符合本发明的实施例,并与说明书一起用于解释本发明的原理。The accompanying drawings, which are incorporated in and constitute a part of this specification, illustrate embodiments consistent with the invention and together with the description serve to explain the principles of the invention.

为了更清楚地说明本发明实施例或现有技术中的技术方案,下面将对实施例或现有技术描述中所需要使用的附图作简单地介绍,显而易见地,对于本领域普通技术人员而言,在不付出创造性劳动性的前提下,还可以根据这些附图获得其他的附图。In order to more clearly illustrate the technical solutions in the embodiments of the present invention or the prior art, the following will briefly introduce the drawings that need to be used in the description of the embodiments or the prior art. Obviously, for those of ordinary skill in the art, In other words, other drawings can also be obtained from these drawings without paying creative labor.

图1为一个实施例中Kinect深度摄像机所定义的人体的关键关节点的示意图;Fig. 1 is the schematic diagram of the key joint point of the human body defined by Kinect depth camera in an embodiment;

图2为一个实施例中描述人体行为的时空图;Fig. 2 is a space-time diagram describing human behavior in one embodiment;

图3为一个实施例中图卷积中定义的节点示意图;Fig. 3 is a schematic diagram of nodes defined in graph convolution in an embodiment;

图4为一个实施例中一个实施例中图数据识别方法的应用环境图;Fig. 4 is an application environment diagram of a graph data identification method in an embodiment in an embodiment;

图5为一个实施例中特征图生成方法的流程示意图;Fig. 5 is a schematic flow chart of a method for generating a feature map in an embodiment;

图6为一个实施例中卷积层的数据处理流程示意图;FIG. 6 is a schematic diagram of a data processing flow of a convolutional layer in an embodiment;

图7为一个实施例中特征图生成装置的结构框图;Fig. 7 is a structural block diagram of a feature map generation device in an embodiment;

图8为一个实施例中计算机设备的内部结构图。Figure 8 is a diagram of the internal structure of a computer device in one embodiment.

具体实施方式Detailed ways

为使本申请实施例的目的、技术方案和优点更加清楚,下面将结合本申请实施例中的附图,对本申请实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例是本申请的一部分实施例,而不是全部的实施例。基于本申请中的实施例,本领域普通技术人员在没有做出创造性劳动的前提下所获得的所有其他实施例,都属于本申请保护的范围。In order to make the purposes, technical solutions and advantages of the embodiments of the present application clearer, the technical solutions in the embodiments of the present application will be clearly and completely described below in conjunction with the drawings in the embodiments of the present application. Obviously, the described embodiments It is a part of the embodiments of this application, but not all of them. Based on the embodiments in the present application, all other embodiments obtained by persons of ordinary skill in the art without making creative efforts belong to the protection scope of the present application.

图4为一个实施例中图数据识别方法的应用环境图。参照图4,该图数据识别方法应用于特征图生成系统。该特征图生成系统包括终端110和服务器120。终端110和服务器120通过网络连接。终端或服务器获取输入已训练的卷积神经网络的当前卷积层的输入特征图,输入特征图为根据图像数据生成的特征图,输入特征图为是通过提取图数据得到的特征图,获取当前卷积层的第一偏置矩阵,其中第一偏置矩阵为生成已训练的卷积神经网络时生成的矩阵,根据输入特征图生成第二偏置矩阵,获取参考邻接矩阵,计算参考邻接矩阵、第一偏置矩阵和第二偏置矩阵的和,得到目标邻接矩阵,获取当前卷积层的卷积核,根据当前卷积层的卷积核、目标邻接矩阵和输入特征图生成目标输出特征图,根据目标输出特征图,识别出图数据的识别结果。终端110具体可以是台式终端或移动终端,移动终端具体可以手机、平板电脑、笔记本电脑等中的至少一种。服务器120可以用独立的服务器或者是多个服务器组成的服务器集群来实现。Fig. 4 is an application environment diagram of the image data identification method in an embodiment. Referring to Fig. 4, the graph data recognition method is applied to a feature graph generation system. The feature map generation system includes a terminal 110 and a server 120 . Terminal 110 and server 120 are connected via a network. The terminal or server obtains the input feature map of the current convolutional layer of the trained convolutional neural network. The input feature map is a feature map generated based on image data, and the input feature map is a feature map obtained by extracting image data. Obtain the current The first bias matrix of the convolutional layer, where the first bias matrix is the matrix generated when the trained convolutional neural network is generated, the second bias matrix is generated according to the input feature map, the reference adjacency matrix is obtained, and the reference adjacency matrix is calculated , the sum of the first bias matrix and the second bias matrix to obtain the target adjacency matrix, obtain the convolution kernel of the current convolution layer, and generate the target output according to the convolution kernel of the current convolution layer, the target adjacency matrix and the input feature map The feature map, according to the target output feature map, recognizes the recognition result of the map data. The terminal 110 may specifically be a desktop terminal or a mobile terminal, and the mobile terminal may specifically be at least one of a mobile phone, a tablet computer, a notebook computer, and the like. The server 120 can be implemented by an independent server or a server cluster composed of multiple servers.

如图5所示,在一个实施例中,提供了一种图数据识别方法。本实施例主要以该方法应用于上述图4中的终端110(或服务器120)来举例说明。参照图5,该图数据识别方法具体包括如下步骤:As shown in FIG. 5 , in one embodiment, a method for identifying graph data is provided. This embodiment is mainly described by taking the method applied to the terminal 110 (or server 120) in FIG. 4 above as an example. Referring to Fig. 5, the data identification method in this figure specifically includes the following steps:

步骤S201,获取输入已训练的卷积神经网络的当前卷积层的输入特征图。Step S201, obtaining the input feature map of the current convolutional layer of the trained convolutional neural network.

在本具体实施例中,输入特征图为根据图像数据生成的特征图。In this specific embodiment, the input feature map is a feature map generated according to image data.

具体地,已训练的卷积神经网络是指通过大量的携带标签的图数据训练得到的,其中图数据为人体行为的时空图,时空图如图2所示。图数据携带的标签包括人体的行为,如拍手、跳跃、拉手和打架等个人行为或多人行为。已训练的卷积神经网络包含多个卷积层,当前卷积层可以为卷积神经网络中的任意一个卷积层。上一个卷积层的输出数据为当前卷积层的输入数据,获取上一个卷积层中的输出数据作为当前卷积层的输入数据,输入数据为输入特征图。Specifically, the trained convolutional neural network is obtained by training a large amount of labeled graph data, where the graph data is a spatio-temporal graph of human behavior, and the spatio-temporal graph is shown in Figure 2. The labels carried by the graph data include human behaviors, such as individual or multi-person behaviors such as clapping, jumping, pulling hands, and fighting. The trained convolutional neural network contains multiple convolutional layers, and the current convolutional layer can be any convolutional layer in the convolutional neural network. The output data of the previous convolutional layer is the input data of the current convolutional layer, and the output data of the previous convolutional layer is obtained as the input data of the current convolutional layer, and the input data is the input feature map.

步骤S202,获取当前卷积层的第一偏置矩阵。Step S202, acquiring the first bias matrix of the current convolutional layer.

具体地,第一偏置矩阵为生成已训练的卷积神经网络时生成的矩阵。第一偏置矩阵是根据训练需求得到的偏置矩阵,不同的训练需求,是指卷积神经网络训练后用于做什么的需求,如用于识别拍手和用于识别打架时,得到的第一偏置矩阵不相同。Specifically, the first bias matrix is a matrix generated when the trained convolutional neural network is generated. The first offset matrix is the offset matrix obtained according to the training requirements. Different training requirements refer to the requirements for what to do after the convolutional neural network is trained. A bias matrix is not the same.

步骤S203,根据输入特征图生成第二偏置矩阵。Step S203, generating a second bias matrix according to the input feature map.

具体地,第二偏置矩阵为根据输入特征图生成的矩阵。通过对输入矩阵进行降维、归一化等操作得到第二偏置矩阵。第二偏置矩阵与输入特征图相关,根据当前卷积层中的函数对输入特征图进行映射,得到映射后的矩阵,将对映射后的矩阵进行乘积运算,对乘积运算结果进行归一化等处理,得到第二偏置矩阵。Specifically, the second bias matrix is a matrix generated according to the input feature map. The second bias matrix is obtained by performing operations such as dimensionality reduction and normalization on the input matrix. The second bias matrix is related to the input feature map. The input feature map is mapped according to the function in the current convolution layer to obtain the mapped matrix. The mapped matrix will be multiplied and the result of the multiplication operation will be normalized. and so on to obtain the second bias matrix.

在一个实施例中,步骤S203,包括:采用已训练的卷积神经网络中的降维函数对输入特征图进行降维,得到降维矩阵,归一化降维矩阵,得到归一化矩阵,归一化矩阵为第二偏置矩阵。In one embodiment, step S203 includes: using a dimensionality reduction function in the trained convolutional neural network to perform dimensionality reduction on the input feature map to obtain a dimensionality reduction matrix, and normalize the dimensionality reduction matrix to obtain a normalized matrix, The normalization matrix is the second bias matrix.

具体地,降维函数包括第一降维函数和第二降维函数,根据第一降维函数对输入特征图对进行降维得到第一特征图,根据第二降维函数对输入特征图进行降维得到第二特征图。计算第一特征图和第二特征图的乘积,得到第一乘积矩阵。对输入特征图采用不同的降维函数进行降维,得到不同的降维矩阵,其中降维函数的参数是根据需求进行训练得到的。计算两个降维后的矩阵的乘积,即第一乘积矩阵,第一乘积矩阵中的各个点表示横坐标对应的点和纵坐标对应的点之间的特征相似度。如1表示关节点1,2表示关节点2,则第一乘积矩阵中的坐标(1,2)中的矩阵元素,表示关节点1和关节点2之间的特征相似度。对第一乘积矩阵进行归一化操作得到归一化矩阵,归一化操作,对数据进行归一化操作可以提高数据计算的精度,加快数据的收敛速度。将归一化后的第一乘积矩阵作为第二偏置矩阵。Specifically, the dimensionality reduction function includes a first dimensionality reduction function and a second dimensionality reduction function. According to the first dimensionality reduction function, the input feature map pair is dimensionally reduced to obtain the first feature map, and according to the second dimensionality reduction function, the input feature map is Dimensionality reduction obtains the second feature map. Calculate the product of the first feature map and the second feature map to obtain a first product matrix. Different dimensionality reduction functions are used to reduce the dimensionality of the input feature map, and different dimensionality reduction matrices are obtained. The parameters of the dimensionality reduction functions are obtained by training according to the requirements. Calculate the product of the two dimension-reduced matrices, that is, the first product matrix, and each point in the first product matrix represents the feature similarity between the point corresponding to the abscissa and the point corresponding to the ordinate. For example, 1 represents joint point 1, and 2 represents joint point 2, then the matrix elements in the coordinates (1, 2) in the first product matrix represent the feature similarity between joint point 1 and joint point 2. Performing a normalization operation on the first product matrix to obtain a normalized matrix, the normalization operation, and performing the normalization operation on the data can improve the accuracy of data calculation and speed up the convergence speed of the data. The normalized first product matrix is used as the second bias matrix.

步骤S204,获取参考邻接矩阵,计算参考邻接矩阵、第一偏置矩阵和第二偏置矩阵的和,得到目标邻接矩阵。Step S204, obtaining the reference adjacency matrix, calculating the sum of the reference adjacency matrix, the first offset matrix and the second offset matrix, to obtain the target adjacency matrix.

具体地,第一偏置矩阵、第二偏置矩阵和参考邻接矩阵具有相同的维度信息,计算参考邻接矩阵、第一偏置矩阵和第二偏置矩阵的和,即对相同位置的矩阵元素直接进行相加,得到目标邻接矩阵。Specifically, the first offset matrix, the second offset matrix, and the reference adjacency matrix have the same dimension information, and the sum of the reference adjacency matrix, the first offset matrix, and the second offset matrix is calculated, that is, for the matrix elements at the same position Add directly to get the target adjacency matrix.

步骤S205,获取当前卷积层的卷积核。Step S205, obtaining the convolution kernel of the current convolution layer.

具体地,每个卷积层包含多个卷积核,各个卷积层对应的卷积核数量可以相同或不同,各个卷积核可以相同也可以不相同。卷积核是用于对图像进行卷积运算,不同的卷积核可以提取不同的图像特征。采用目标邻接矩阵对输入特征图进行特征提取,得到特征图。Specifically, each convolution layer includes multiple convolution kernels, the number of convolution kernels corresponding to each convolution layer may be the same or different, and each convolution kernel may be the same or different. Convolution kernels are used to perform convolution operations on images, and different convolution kernels can extract different image features. The target adjacency matrix is used to perform feature extraction on the input feature map to obtain the feature map.

步骤S205,根据当前卷积层的卷积核、目标邻接矩阵和输入特征图生成目标输出特征图。Step S205, generating a target output feature map according to the convolution kernel of the current convolution layer, the target adjacency matrix and the input feature map.

具体地,采用目标邻接矩阵对应输入特征图进行特征提取,得到对应的特征图,采用卷积核对根据目标路径矩阵提取的特征图进行卷积运算,将卷积运算得到的特征图作为目标输出特征图。采用目标邻接矩阵对输入特征图进行特征提取,能够提取到更为准确的特征。Specifically, the target adjacency matrix is used to perform feature extraction corresponding to the input feature map, and the corresponding feature map is obtained, and the convolution kernel is used to perform convolution operation on the feature map extracted according to the target path matrix, and the feature map obtained by the convolution operation is used as the target output feature picture. Using the target adjacency matrix to extract features from the input feature map, more accurate features can be extracted.

在一个实施例中,输入特征图至少包括三个维度,目标邻接矩阵包括至少三个维度,步骤S205,包括:In one embodiment, the input feature map includes at least three dimensions, and the target adjacency matrix includes at least three dimensions. Step S205 includes:

步骤S2051,重塑输入特征图,得到重塑特征图。Step S2051, reshaping the input feature map to obtain the reshaping feature map.

在本具体实施例中,重塑特征图的第一维度为输入特征图的第一维度和第二维度的乘积。In this specific embodiment, the first dimension of the reshaped feature map is the product of the first dimension and the second dimension of the input feature map.

具体地,重塑是指对输入特征图进行调整,使得第一维度和第二维度的乘积为重塑特征图的第一维度,如将包含三个维度的输入特征图调整成2个维度的重塑特征图,假设输入特征图为C×M×N,其中第一维度为C,第二维度为M,第三维度为N,则可以重塑图为C×M×N,第一维度为C和M的乘积CM,第二维度与输入特征图的第三维图相同,保持整个输入特征图的元素不变,总元素为C、M和N的乘积CMN。其中第一维度C为通道素,第二维度M为输入特征图的行数,第三维度N为输入特征图的列数。其中N代表的是人体关节点的数量,在Kinect中N定义为25。对矩阵进行重塑是为了方便运算。Specifically, reshaping refers to adjusting the input feature map so that the product of the first dimension and the second dimension is the first dimension of the reshaping feature map, such as adjusting the input feature map containing three dimensions into two dimensions. Reshape the feature map, assuming that the input feature map is C×M×N, where the first dimension is C, the second dimension is M, and the third dimension is N, then the reshaping map can be C×M×N, the first dimension It is the product CM of C and M, the second dimension is the same as the third dimension of the input feature map, keeping the elements of the entire input feature map unchanged, and the total element is the product CMN of C, M and N. The first dimension C is the channel pixel, the second dimension M is the number of rows of the input feature map, and the third dimension N is the number of columns of the input feature map. Among them, N represents the number of joint points of the human body, and N is defined as 25 in Kinect. The reshaping of the matrix is for convenience of operation.

步骤S2052,计算重塑特征图和目标邻接矩阵的各个通道的矩阵的乘积,得到各个通道的第二乘积矩阵。Step S2052, calculating the product of the reshaped feature map and the matrix of each channel of the target adjacency matrix to obtain a second product matrix of each channel.

具体地,重塑特征图的第二维度与目标邻接矩阵的各个通道的矩阵的第一维度相同,如重塑特征图为C×M×N,目标邻接矩阵为C×N×N,各个通道矩阵为N×N,则重塑特征图与各个通道的矩阵的乘积矩阵为CMN。Specifically, the second dimension of the reshaped feature map is the same as the first dimension of the matrix of each channel of the target adjacency matrix. For example, the reshaped feature map is C×M×N, the target adjacency matrix is C×N×N, and each channel The matrix is N×N, then the product matrix of the reshaped feature map and the matrix of each channel is CMN.

步骤S2053,反重塑各个通道的第二乘积矩阵,得到各个通道的反重塑特征图。Step S2053, dereshaping the second product matrix of each channel to obtain the deremodeling feature map of each channel.

具体地,反重塑是重塑的逆过程,如重塑是将三维矩阵转换为二位矩阵,则反重塑为将二维矩阵转换为三维矩阵,如各个通道的乘积矩阵为CM×N,则反重塑后得到的反重塑特征图为C×M×N的矩阵。Specifically, anti-reshaping is the inverse process of reshaping. For example, reshaping is to convert a three-dimensional matrix into a two-dimensional matrix, then anti-reshaping is to convert a two-dimensional matrix into a three-dimensional matrix. For example, the product matrix of each channel is CM×N , then the anti-reshaping feature map obtained after anti-reshaping is a C×M×N matrix.

步骤S2054,根据各个通道的卷积核,根据各个通道的卷积核对反重塑特征图进行卷积运算,得到当前卷积层各个通道的目标特征图。Step S2054, according to the convolution kernel of each channel, perform convolution operation on the anti-reshaping feature map according to the convolution kernel of each channel, and obtain the target feature map of each channel of the current convolution layer.

步骤S2055,对各个通道的目标特征图进行求和,得到当前卷积层的输出特征图,将当前卷积层的输出特征图作为目标输出特征图。In step S2055, the target feature map of each channel is summed to obtain the output feature map of the current convolution layer, and the output feature map of the current convolution layer is used as the target output feature map.

具体地,通过各个通道对应的卷积核,对反重塑特征图进行特征提取,得到各个卷积核对应的特征,由各个卷积核提取到的特征组成各个通道的目标特征图。计算各个通道的目标特征图的和,即对应位置的矩阵元素相加,得到输出特征图,该输出特征图为当前卷积层的输出特征图。Specifically, through the convolution kernels corresponding to each channel, feature extraction is performed on the anti-reshaping feature map to obtain the features corresponding to each convolution kernel, and the features extracted by each convolution kernel form the target feature map of each channel. Calculate the sum of the target feature maps of each channel, that is, add the matrix elements at the corresponding positions to obtain the output feature map, which is the output feature map of the current convolutional layer.

在一个实施例中,步骤S205,还包括:In one embodiment, step S205 further includes:

步骤S2056,判断输出特征图的通道数是否与输入特征图的通道数一致。Step S2056, judging whether the number of channels of the output feature map is consistent with the number of channels of the input feature map.

步骤S2057,当一致时,将输入特征图对输出特征图的和,作为当前卷积层的目标输出特征图。Step S2057, when consistent, use the sum of the input feature map and the output feature map as the target output feature map of the current convolutional layer.

步骤S2058,当不一致时,对输入特征图进行卷积运算,得到与输出特征图的通道数一致的卷积特征图,将卷积特征图与输出特征图的和,作为目标输出特征图。Step S2058, if they are inconsistent, perform a convolution operation on the input feature map to obtain a convolution feature map with the same number of channels as the output feature map, and use the sum of the convolution feature map and the output feature map as the target output feature map.

具体地,根据各个通道的卷积核和对应的额反重塑矩阵生成的输出特征图的各个通道矩阵,判断输出特征图是通道是否与输入特征图的通道数一致,当一致时,输入特征图与输出特征对应的位置的元素进行相加,得到目标输出特征图。当不一致时,对输入特征图进行卷积运算,得到与输出特征图具有相同通道的卷积特征图,计算卷积特征图与输出特征图相同位置的元素的和,得到目标输出特征图。通过判断输入特征图与输出特征图的通道数,根据通道数确定目标输出特征图,提高了目标输出特征图的准确性。Specifically, according to each channel matrix of the output feature map generated by the convolution kernel of each channel and the corresponding frontal inverse reshaping matrix, it is judged whether the channel of the output feature map is consistent with the number of channels of the input feature map. When they are consistent, the input The feature map is added to the element at the position corresponding to the output feature to obtain the target output feature map. When inconsistent, perform convolution operation on the input feature map to obtain a convolution feature map with the same channel as the output feature map, calculate the sum of the elements at the same position in the convolution feature map and the output feature map, and obtain the target output feature map. By judging the channel numbers of the input feature map and the output feature map, the target output feature map is determined according to the channel number, which improves the accuracy of the target output feature map.

步骤S206,根据目标输出特征图,识别出图数据的识别结果。Step S206, according to the target output feature map, identify the recognition result of the map data.

具体地,将目标输出特征图输入已训练的卷积神经网络中的识别层,通过识别层识别目标输出特征图对应的候选识别结果,从候选识别结果中选择识别概率最大的候选识别结果,作为目标识别结果,将目标识别结果作为图数据对应的识别结果。如识别类型包括拍手、跳跃、牵手三个类型时,其中拍手对应的识别概率为0.89,跳跃对应的识别概率为0.01,牵手对应的识别概率为0.1时,则图数据对应的识别结果为拍手。Specifically, the target output feature map is input into the recognition layer in the trained convolutional neural network, the candidate recognition results corresponding to the target output feature map are identified through the recognition layer, and the candidate recognition result with the highest recognition probability is selected from the candidate recognition results as Target recognition result, the target recognition result is used as the recognition result corresponding to the graph data. For example, when the recognition types include clapping, jumping, and holding hands, the recognition probability corresponding to clapping is 0.89, the recognition probability corresponding to jumping is 0.01, and the recognition probability corresponding to holding hands is 0.1, then the recognition result corresponding to the graph data is clapping.

在一个实施例中,当已训练的卷积神经网络中包括当前卷积层,和当前卷积层的下一个卷积层时,将目标输出特征图作为下一个卷积层的输入特征图,将下一卷积层作为当前卷积层,进入获取输入已训练的卷积神经网络的当前卷积层的输入特征图,直至已训练的卷积神经网络中的各个卷积层都完成时,输出最后一个卷积层的目标输出特征图,将最后一个卷积层的目标输出特征图输入识别层,得到图书对应的识别结果。具有相同网络结构的卷积层的数据处理流程相同。In one embodiment, when the trained convolutional neural network includes the current convolutional layer and the next convolutional layer of the current convolutional layer, the target output feature map is used as the input feature map of the next convolutional layer, Use the next convolutional layer as the current convolutional layer, and enter to obtain the input feature map of the current convolutional layer of the trained convolutional neural network until all convolutional layers in the trained convolutional neural network are completed. Output the target output feature map of the last convolutional layer, input the target output feature map of the last convolutional layer into the recognition layer, and obtain the corresponding recognition result of the book. The data processing flow of convolutional layers with the same network structure is the same.

上述图数据识别方法,获取输入已训练的卷积神经网络的当前卷积层的输入特征图,输入特征图为根据图像数据生成的特征图,获取当前卷积层的第一偏置矩阵,其中第一偏置矩阵为生成已训练的卷积神经网络时生成的矩阵,根据输入特征图生成第二偏置矩阵,获取参考邻接矩阵,计算参考邻接矩阵、第一偏置矩阵和第二偏置矩阵的和,得到目标邻接矩阵,获取当前卷积层的卷积核,根据当前卷积层的卷积核、目标邻接矩阵和输入特征图生成目标输出特征图,根据目标输出特征图,识别出图数据的识别结果。对现有的固定的邻接矩阵基础上增加可调整的偏置矩阵,偏置矩阵中的第一偏置矩阵为根据训练需求得到的,故能够好的表征需求所对应的特征,第二偏置矩阵根据任务需求确定参数,且根据输入的图数据生成,能够表征各个图数据的特征,从图数据本身和需求同时确定的偏置矩阵,能够更好的标准各个图数据的本身的特征,从而提高已训练的卷积神经网络的识别准确率。在一个实施例中,步骤S206,包括:The above image data recognition method obtains the input feature map of the current convolutional layer of the trained convolutional neural network, the input feature map is a feature map generated according to the image data, and obtains the first bias matrix of the current convolutional layer, where The first bias matrix is the matrix generated when the trained convolutional neural network is generated, the second bias matrix is generated according to the input feature map, the reference adjacency matrix is obtained, and the reference adjacency matrix, the first bias matrix and the second bias are calculated The sum of the matrices to obtain the target adjacency matrix, obtain the convolution kernel of the current convolution layer, generate the target output feature map according to the convolution kernel of the current convolution layer, the target adjacency matrix and the input feature map, and identify the The recognition result of graph data. An adjustable offset matrix is added to the existing fixed adjacency matrix. The first offset matrix in the offset matrix is obtained according to the training requirements, so it can represent the features corresponding to the requirements well. The second offset matrix The matrix determines the parameters according to the task requirements, and is generated according to the input graph data, which can represent the characteristics of each graph data. The offset matrix determined from the graph data itself and the requirements can better standardize the characteristics of each graph data itself, so that Improve the recognition accuracy of the trained convolutional neural network. In one embodiment, step S206 includes:

步骤S2061,当当前卷积层为已训练的卷积神经网络中的最后一个卷积层时,判断多个目标输出特征图中是否存在需要合并的目标特征图。Step S2061, when the current convolutional layer is the last convolutional layer in the trained convolutional neural network, determine whether there is a target feature map that needs to be merged in multiple target output feature maps.

步骤S2062,当存在时,合并需要合并的目标输出特征图,得到合并特征图。Step S2062, if it exists, merge the target output feature maps that need to be merged to obtain the merged feature map.

步骤S2063,当合并特征图包含全部目标输出特征图时,对合并特征图进行识别,得到合并特征图对应的识别结果。Step S2063, when the merged feature map contains all target output feature maps, identify the merged feature map to obtain a recognition result corresponding to the merged feature map.

步骤S2064,当合并特征图包含全部目标输出特征图时,对合并特征图进行识别,得到合并特征图对应的识别结果,对未合并的目标输出特征图进行识别,得到未合并的目标输出特征图对应的识别结果。Step S2064, when the merged feature map contains all target output feature maps, identify the merged feature map to obtain the recognition result corresponding to the merged feature map, and identify the unmerged target output feature map to obtain the unmerged target output feature map corresponding recognition results.

步骤S2065,当不存在时,对各个目标输出特征图进行识别,得到各个目标输出特征图对应的识别结果。Step S2065, if it does not exist, identify each target output feature map, and obtain the recognition result corresponding to each target output feature map.

具体地,当当前卷积层为为已训练的卷积神经网络中的最后一个卷积层时,表示卷积运算已经解释,提取到了图数据最终的目标输出特征图,根据最终的目标输出特征图进行识别之前,判断各个最终的目标输出特征图是否需要合并。需要进行合并是因为,有的行为需要多人才能完成,如牵手、打架等行为至少需要两个人才能够完成。对于个人可以完成的行为,则无需合并目标输出特征图,直接对目标输出特征图进行识别,将目标输出特征图的识别结果,作为图数据的识别结果。图数据的识别结果可以为一个或多个子识别结果,当图数据的输入包含个行为时,得到的识别结果包含多个行为,各个行为作为一个子识别结果。Specifically, when the current convolutional layer is the last convolutional layer in the trained convolutional neural network, it means that the convolution operation has been explained, and the final target output feature map of the graph data has been extracted, and the final target output feature Before the map is recognized, it is judged whether each final target output feature map needs to be merged. The need for merging is because some behaviors require multiple people to complete, such as holding hands, fighting and other behaviors require at least two people to complete. For behaviors that individuals can complete, there is no need to combine the target output feature map, and the target output feature map is directly recognized, and the recognition result of the target output feature map is used as the recognition result of the graph data. The recognition result of graph data can be one or more sub-recognition results. When the input of graph data includes a behavior, the obtained recognition result includes multiple behaviors, and each behavior is regarded as a sub-recognition result.

当存在需要合并的目标输出特征图时,确定哪些目标输出特征图为需要合并的,如何合并的,将需要合并的目标输出特征图进行合并,直接对合并特征图,其中合并特征图可以为一个或多个,具体的合并结果与输入数据中的人体行为相关,如图数据中包括拍手、牵手、打架等多个类型的行为时,牵手和打架需要进行合并,则合并特征图包括两个。直接对合并特征图进行识别,得到合并特征图的识别结果。当存在部分合并,部分不合并的情况时,不合并的目标输出特征图直接进行识别,得到不合并的目标输出特征图对应的识别结果。When there are target output feature maps that need to be merged, determine which target output feature maps need to be merged and how to merge them, merge the target output feature maps that need to be merged, and directly merge the feature maps, where the merged feature map can be one or more, the specific merging result is related to the human behavior in the input data, as shown in the figure, when the data includes multiple types of behaviors such as clapping, holding hands, fighting, etc., holding hands and fighting need to be merged, and the merged feature map includes two. The merged feature map is directly recognized, and the recognition result of the merged feature map is obtained. When there is a situation of partly merging and partly not merging, the non-merging target output feature map is directly recognized, and the recognition result corresponding to the non-merging target output feature map is obtained.

在一个具体的实施例中,参照图6,图6为一个实施例中卷积层的数据处理流程示意图。图中的fin为当前卷积层的输入特征图,fout为当前卷积层的输出特征图。输出特征图采用输入特征图的具体表示方式如公式(3)所示:In a specific embodiment, refer to FIG. 6 , which is a schematic diagram of a data processing flow of a convolutional layer in an embodiment. In the figure, fin is the input feature map of the current convolutional layer, and f out is the output feature map of the current convolutional layer. The output feature map adopts the specific representation of the input feature map as shown in formula (3):

式中,Ak为参考邻接矩阵中的第k个邻接矩阵,Bk为偏置矩阵中的第k个邻接矩阵,Ck为偏置矩阵中的第k个邻接矩阵,softmax(S)表示对矩阵S进行归一化一算,Wk为卷积核的第k个参数,Kv为卷积核的大小,卷积核的大小可以自定义,如可以设置Kv=3或5。为第一降维函数,为第二降维函数。假设输入特征图的尺寸信息为Cin×T×N,其中Cin代表输入通道数,T代表图数据的帧数,N代表kinect定义的关节节点数,其中N=25。对输入特征图进行重塑得到CinT×N的重塑特征图,偏置矩阵Bk为训练卷积神经网络后得到的矩阵,Bk与Ak具有相同的尺寸信息,即为N×N,计算计算第一偏置矩阵Bk、第二篇值矩阵CK与参考邻接矩阵Ak的和,得到目标邻接矩阵的各个通道的矩阵,计算目标邻接矩阵各个通道的矩阵与重塑特征图的乘积,得到各个通道的第二乘积矩阵,并对各个通道的第二乘积矩阵进行反重塑,得到反重塑矩阵,获取个卷积核Wk,其,通过各个通道的卷积核对反重塑矩阵进行卷积运算,得到各个通道对应的输出特征图。判断输出特征图是否与输入特征图的通道数是否一致,当不一致时,通过残差网络res,其中残差网络中的卷积核大小为1×1。将输入特征图调整致与输出特征图的通道数一致的矩阵,计算调整后的输入特征图与输出特征图的和,得到目标输出特征图。当一致时,计算输入特征图与输出特征图的和,得到目标输出特征图。根据目标输出特征图对各个图数据的行为进行识别,得到对应的识别结果。In the formula, A k is the kth adjacency matrix in the reference adjacency matrix, B k is the kth adjacency matrix in the bias matrix, C k is the kth adjacency matrix in the bias matrix, softmax(S) means Perform normalization calculation on the matrix S, W k is the kth parameter of the convolution kernel, Kv is the size of the convolution kernel, the size of the convolution kernel can be customized, for example, K v =3 or 5 can be set. is the first dimensionality reduction function, is the second dimensionality reduction function. Assume that the size information of the input feature map is C in × T × N, where C in represents the number of input channels, T represents the number of frames of graph data, and N represents the number of joint nodes defined by kinect, where N=25. Reshape the input feature map to obtain the reshaped feature map of C in T×N. The bias matrix B k is the matrix obtained after training the convolutional neural network. B k and A k have the same size information, which is N× N, calculate and calculate the sum of the first bias matrix B k , the second value matrix C K and the reference adjacency matrix A k to obtain the matrix of each channel of the target adjacency matrix, and calculate the matrix and reshape features of each channel of the target adjacency matrix The product of the graphs is obtained to obtain the second product matrix of each channel, and the second product matrix of each channel is reversely reshaped to obtain the reverse remodeling matrix, and a convolution kernel W k is obtained, which, through the convolution kernel of each channel The anti-reshaping matrix performs convolution operation to obtain the output feature map corresponding to each channel. Determine whether the output feature map is consistent with the number of channels of the input feature map. If not, pass the residual network res, where the convolution kernel size in the residual network is 1×1. Adjust the input feature map to a matrix with the same number of channels as the output feature map, calculate the sum of the adjusted input feature map and output feature map, and obtain the target output feature map. When consistent, calculate the sum of the input feature map and the output feature map to obtain the target output feature map. According to the target output feature map, the behavior of each graph data is recognized, and the corresponding recognition result is obtained.

当上述特征图的生过程为训练卷积神经网络的数据处理过程,且当各个图数据对应的识别与图数据的标签中的类别不一致时,根据预设损失函数各个图数据对应的损失值,根据梯度回传算法回传损失值,得到卷积层的回传值,根据回传值更新对应的卷积层的各个通道的卷积核w的参数、第一偏置矩阵B和用于计算第二偏置矩阵C的映射函数的参数。When the generation process of the above feature map is the data processing process of training the convolutional neural network, and when the identification corresponding to each graph data is inconsistent with the category in the label of the graph data, according to the loss value corresponding to each graph data according to the preset loss function, Return the loss value according to the gradient return algorithm to obtain the return value of the convolution layer, update the parameters of the convolution kernel w of each channel of the corresponding convolution layer according to the return value, the first bias matrix B and the calculation Parameters of the mapping function for the second bias matrix C.

当上述特征图的生过程为训练后的已训练的卷积神经网络的数据处理过程时,根据目标输出特征图识别的结果作为图数据对应的识别结果。将图数据输入已训练的卷积神经网路,卷积神经网络包含多个卷积层和识别层,各个卷积层包括卷积核和目标邻接矩阵,通过各个卷积层的目标邻接矩阵对图数据进行特征提取,得到对应的图像特征图集合,通过卷积核对图像特征集合进行卷积运算,得到各个卷积层的目标输出特征图,识别层的上一个卷积层的目标输出特征图作为识别层的输入数据,根据各个图数据的目标输出特征图,识别出对应的人体行为类型。When the above feature map generation process is the data processing process of the trained convolutional neural network, the recognition result of the feature map is output according to the target as the recognition result corresponding to the map data. Input the graph data into the trained convolutional neural network. The convolutional neural network includes multiple convolutional layers and recognition layers. Each convolutional layer includes a convolution kernel and a target adjacency matrix. The target adjacency matrix of each convolutional layer is used to Feature extraction is performed on the image data to obtain the corresponding image feature map set, and the convolution operation is performed on the image feature set through the convolution kernel to obtain the target output feature map of each convolution layer, and the target output feature map of the previous convolution layer of the recognition layer As the input data of the identification layer, according to the target output feature map of each graph data, the corresponding human behavior type is identified.

图5为一个实施例中图数据识别方法的流程示意图。应该理解的是,虽然图5的流程图中的各个步骤按照箭头的指示依次显示,但是这些步骤并不是必然按照箭头指示的顺序依次执行。除非本文中有明确的说明,这些步骤的执行并没有严格的顺序限制,这些步骤可以以其它的顺序执行。而且,图5中的至少一部分步骤可以包括多个子步骤或者多个阶段,这些子步骤或者阶段并不必然是在同一时刻执行完成,而是可以在不同的时刻执行,这些子步骤或者阶段的执行顺序也不必然是依次进行,而是可以与其它步骤或者其它步骤的子步骤或者阶段的至少一部分轮流或者交替地执行。Fig. 5 is a schematic flowchart of a method for identifying image data in an embodiment. It should be understood that although the various steps in the flow chart of FIG. 5 are shown sequentially as indicated by the arrows, these steps are not necessarily executed sequentially in the order indicated by the arrows. Unless otherwise specified herein, there is no strict order restriction on the execution of these steps, and these steps can be executed in other orders. Moreover, at least some of the steps in Fig. 5 may include multiple sub-steps or multiple stages, these sub-steps or stages are not necessarily executed at the same moment, but may be executed at different moments, the execution of these sub-steps or stages The order is not necessarily performed sequentially, but may be performed alternately or alternately with at least a part of other steps or sub-steps or stages of other steps.

在一个实施例中,如图7所示,提供了一种特图生成装置200,包括:In one embodiment, as shown in FIG. 7 , a special map generation device 200 is provided, including:

数据获取模块201,用于获取输入已训练的卷积神经网络的当前卷积层的输入特征图,输入特征图为根据图像数据生成的特征图,获取当前卷积层的第一偏置矩阵,其中第一偏置矩阵为生成已训练的卷积神经网络时生成的矩阵。The data acquisition module 201 is used to obtain the input feature map of the current convolutional layer of the trained convolutional neural network, the input feature map is a feature map generated according to the image data, and obtain the first bias matrix of the current convolutional layer, Wherein the first bias matrix is a matrix generated when the trained convolutional neural network is generated.

第二偏置矩阵生成模块202,用于根据输入特征图生成第二偏置矩阵。The second offset matrix generation module 202 is configured to generate a second offset matrix according to the input feature map.

目标邻接矩阵生模块203,用于获取参考邻接矩阵,计算参考邻接矩阵、第一偏置矩阵和第二偏置矩阵的和,得到目标邻接矩阵。The target adjacency matrix generating module 203 is configured to obtain a reference adjacency matrix, calculate the sum of the reference adjacency matrix, the first offset matrix and the second offset matrix, and obtain the target adjacency matrix.

目标输出特征图生成模块204,用于获取当前卷积层的卷积核,根据当前卷积层的卷积核、目标邻接矩阵和输入特征图生成目标输出特征图。The target output feature map generating module 204 is configured to obtain the convolution kernel of the current convolution layer, and generate the target output feature map according to the convolution kernel of the current convolution layer, the target adjacency matrix and the input feature map.

识别模块205,用于根据目标输出特征图,识别出图数据的识别结果。The recognition module 205 is configured to recognize the recognition result of the map data according to the target output feature map.

在一个实施例中,第二偏置矩阵生成模块具体用于采用已训练的卷积神经网络中的降维函数对输入特征图进行降维,得到降维矩阵,归一化降维矩阵,得到归一化矩阵,归一化矩阵为第二偏置矩阵。In one embodiment, the second offset matrix generation module is specifically used to reduce the dimensionality of the input feature map by using the dimensionality reduction function in the trained convolutional neural network to obtain the dimensionality reduction matrix, and normalize the dimensionality reduction matrix to obtain Normalization matrix, the normalization matrix is the second bias matrix.

在一个实施例中,第二偏置矩阵生成模块具体用于根据降维函数中的第一降维函数对输入特征图的各个通道的矩阵进行降维,得到各个通道的第一降维矩阵,根据降维函数中的第二降维函数对输入特征图的各个通道的矩阵进行降维,得到各个通道的第二降维矩阵,其中,降维函数包括两个,输入特征图至少包括三个维度,第一维度为通道数,计算各个通道的第一降维矩阵和第二降维矩阵的乘积,得到各个通道的第一乘积矩阵,归一化各个通道的第一乘积矩阵,得到归一化矩阵对应的通道的矩阵。In one embodiment, the second offset matrix generation module is specifically configured to perform dimensionality reduction on the matrix of each channel of the input feature map according to the first dimensionality reduction function in the dimensionality reduction function, to obtain the first dimensionality reduction matrix of each channel, According to the second dimensionality reduction function in the dimensionality reduction function, the matrix of each channel of the input feature map is reduced in dimension to obtain the second dimensionality reduction matrix of each channel, wherein the dimensionality reduction function includes two, and the input feature map includes at least three dimension, the first dimension is the number of channels, calculate the product of the first dimensionality reduction matrix and the second dimensionality reduction matrix of each channel, obtain the first product matrix of each channel, normalize the first product matrix of each channel, and obtain normalization The matrix of channels corresponding to the transformation matrix.

在一个实施例中,上述图数据识别装置,还包括:In one embodiment, the above image data identification device further includes:

网络生成模块,用于生成已训练的卷积神经网络。其中网络生成模块,包括:Network generation module for generating trained convolutional neural networks. Among them, the network generation module includes:

数据获取单元,用于获取包含多个训练图数据的训练集合,训练图数据携带标签信息。The data acquisition unit is configured to acquire a training set including a plurality of training graph data, and the training graph data carries label information.

特征提取单元,用于将训练图数据和标签信息输入初始卷积神经网络,通过初始卷积神经网络提取各个训练图数据的特征。The feature extraction unit is used to input the training graph data and label information into the initial convolutional neural network, and extract the features of each training graph data through the initial convolutional neural network.

识别单元,用于根据各个训练图数据的特征,识别出各个训练图数据对应的识别结果。The recognition unit is configured to recognize the recognition result corresponding to each training graph data according to the characteristics of each training graph data.

损失值计算单元,用于按照预设损失函数计算各个训练图数据的识别结果和标签的损失值。The loss value calculation unit is used to calculate the recognition result of each training image data and the loss value of the label according to a preset loss function.

网络确定单元,用于当损失值小于或等于预设损失值时,得到已训练的卷积神经网络。The network determination unit is used to obtain the trained convolutional neural network when the loss value is less than or equal to the preset loss value.

在一个实施例中,网络确定单元,还包括:In one embodiment, the network determining unit further includes:

参数更新子单元,用于当损失值大于预设损失值时,根据损失值通过梯度回传算法更新初始卷积神经网络的网络参数。The parameter update subunit is used to update the network parameters of the initial convolutional neural network through the gradient backpropagation algorithm according to the loss value when the loss value is greater than the preset loss value.

网络确定子单元,用于采用更新了网络参数的初始卷积神经网络作为初始卷积神经网络,进入将训练图数据和标签信息输入初始卷积神经网络,直至按照预设损失函数计算各个训练图数据的识别结果和标签的损失值,小于或等于预设损失值时,得到已训练的卷积神经网络。The network determination subunit is used to adopt the initial convolutional neural network with updated network parameters as the initial convolutional neural network, enter the training graph data and label information into the initial convolutional neural network, and calculate each training graph according to the preset loss function When the recognition result of the data and the loss value of the label are less than or equal to the preset loss value, a trained convolutional neural network is obtained.

在一个实施例中,网络确定子单元具体通过梯度回传算法将损失值回传到任意一个卷积层时,得到各个卷积层的回传值,根据各个卷积层的回传值更新卷积层的网络参数,其中,初始卷积神经网络模型包括至少一个卷积层,网络参数包括初始降维函数的参数和初始偏置矩阵的参数。In one embodiment, when the network determination subunit specifically returns the loss value to any convolution layer through the gradient return algorithm, it obtains the return value of each convolution layer, and updates the volume according to the return value of each convolution layer. The network parameters of the product layer, wherein the initial convolutional neural network model includes at least one convolutional layer, and the network parameters include the parameters of the initial dimensionality reduction function and the parameters of the initial bias matrix.

图8示出了一个实施例中计算机设备的内部结构图。该计算机设备具体可以是图4中的终端110(或服务器120)。如图8所示,该计算机设备包括该计算机设备包括通过系统总线连接的处理器、存储器、网络接口、输入装置和显示屏。其中,存储器包括非易失性存储介质和内存储器。该计算机设备的非易失性存储介质存储有操作系统,还可存储有计算机程序,该计算机程序被处理器执行时,可使得处理器实现图数据识别方法。该内存储器中也可储存有计算机程序,该计算机程序被处理器执行时,可使得处理器执行图数据识别方法。计算机设备的显示屏可以是液晶显示屏或者电子墨水显示屏,计算机设备的输入装置可以是显示屏上覆盖的触摸层,也可以是计算机设备外壳上设置的按键、轨迹球或触控板,还可以是外接的键盘、触控板或鼠标等。Figure 8 shows a diagram of the internal structure of a computer device in one embodiment. The computer device may specifically be the terminal 110 (or server 120) in FIG. 4 . As shown in FIG. 8 , the computer equipment includes a processor, a memory, a network interface, an input device, and a display screen connected through a system bus. Wherein, the memory includes a non-volatile storage medium and an internal memory. The non-volatile storage medium of the computer device stores an operating system, and may also store a computer program, and when the computer program is executed by the processor, the processor may realize the image data identification method. A computer program may also be stored in the internal memory, and when the computer program is executed by the processor, the processor may execute the image data identification method. The display screen of the computer equipment may be a liquid crystal display screen or an electronic ink display screen, and the input device of the computer equipment may be a touch layer covered on the display screen, or a button, a trackball or a touch pad provided on the casing of the computer equipment, or It can be an external keyboard, touchpad or mouse.

本领域技术人员可以理解,图8中示出的结构,仅仅是与本申请方案相关的部分结构的框图,并不构成对本申请方案所应用于其上的计算机设备的限定,具体的计算机设备可以包括比图中所示更多或更少的部件,或者组合某些部件,或者具有不同的部件布置。Those skilled in the art can understand that the structure shown in FIG. 8 is only a block diagram of a partial structure related to the solution of this application, and does not constitute a limitation on the computer equipment to which the solution of this application is applied. The specific computer equipment can be More or fewer components than shown in the figures may be included, or some components may be combined, or have a different arrangement of components.

在一个实施例中,本申请提供的特图生成装置可以实现为一种计算机程序的形式,计算机程序可在如图8所示的计算机设备上运行。计算机设备的存储器中可存储组成该特图生成装置的各个程序模块,比如,图7所示的数据获取模块201、第二偏置矩阵生成模块202、目标邻接矩阵生模块203、目标输出特征图生成模块204和识别模块205。各个程序模块构成的计算机程序使得处理器执行本说明书中描述的本申请各个实施例的图数据识别方法中的步骤。In an embodiment, the device for generating a special map provided in the present application may be implemented in the form of a computer program, and the computer program may be run on a computer device as shown in FIG. 8 . The memory of the computer equipment can store the various program modules that make up the special map generation device, such as the data acquisition module 201 shown in Figure 7, the second bias matrix generation module 202, the target adjacency matrix generation module 203, and the target output feature map A generation module 204 and an identification module 205 . The computer program constituted by each program module enables the processor to execute the steps in the image data identification method of each embodiment of the application described in this specification.

例如,图8所示的计算机设备可以通过如图7所示的特图生成装置中的数据获取模块201执行获取输入已训练的卷积神经网络的当前卷积层的输入特征图,输入特征图为根据图像数据生成的特征图,获取当前卷积层的第一偏置矩阵,其中第一偏置矩阵为生成已训练的卷积神经网络时生成的矩阵。计算机设备可以通过第二偏置矩阵生成模块202执行根据输入特征图生成第二偏置矩阵。计算机设备可以通过目标邻接矩阵生模块203执行获取参考邻接矩阵,计算参考邻接矩阵、第一偏置矩阵和第二偏置矩阵的和,得到目标邻接矩阵。计算机设备可以通过目标输出特征图生成模块204执行获取当前卷积层的卷积核,根据当前卷积层的卷积核、目标邻接矩阵和输入特征图生成目标输出特征图。计算机设备可以通过识别模块205执行根据目标输出特征图,识别出图数据的识别结果。For example, the computer equipment shown in FIG. 8 can perform the acquisition of the input feature map of the current convolutional layer of the trained convolutional neural network through the data acquisition module 201 in the special map generation device as shown in FIG. Obtain the first bias matrix of the current convolutional layer for the feature map generated according to the image data, where the first bias matrix is a matrix generated when the trained convolutional neural network is generated. The computer device can generate the second offset matrix according to the input feature map through the second offset matrix generation module 202 . The computer device can obtain the reference adjacency matrix through the target adjacency matrix generation module 203, calculate the sum of the reference adjacency matrix, the first bias matrix and the second bias matrix, and obtain the target adjacency matrix. The computer device can obtain the convolution kernel of the current convolution layer through the target output feature map generation module 204, and generate the target output feature map according to the convolution kernel of the current convolution layer, the target adjacency matrix and the input feature map. The computer device can use the recognition module 205 to execute the recognition result of identifying the graph data according to the target output feature graph.

在一个实施例中,提供了一种计算机设备,包括存储器、处理器及存储在存储器上并可在处理器上运行的计算机程序,处理器执行计算机程序时实现以下步骤:获取输入已训练的卷积神经网络的当前卷积层的输入特征图,输入特征图为根据图像数据生成的特征图,获取当前卷积层的第一偏置矩阵,其中第一偏置矩阵为生成已训练的卷积神经网络时生成的矩阵,根据输入特征图生成第二偏置矩阵,获取参考邻接矩阵,计算参考邻接矩阵、第一偏置矩阵和第二偏置矩阵的和,得到目标邻接矩阵,获取当前卷积层的卷积核,根据当前卷积层的卷积核、目标邻接矩阵和输入特征图生成目标输出特征图,根据目标输出特征图,识别出图数据的识别结果。In one embodiment, a computer device is provided, including a memory, a processor, and a computer program stored on the memory and operable on the processor. When the processor executes the computer program, the following steps are implemented: obtaining the input trained volume The input feature map of the current convolutional layer of the product neural network, the input feature map is the feature map generated according to the image data, and the first bias matrix of the current convolutional layer is obtained, wherein the first bias matrix is the generated convolution of the trained The matrix generated by the neural network generates the second bias matrix according to the input feature map, obtains the reference adjacency matrix, calculates the sum of the reference adjacency matrix, the first bias matrix and the second bias matrix, obtains the target adjacency matrix, and obtains the current volume The convolution kernel of the product layer generates the target output feature map according to the convolution kernel of the current convolution layer, the target adjacency matrix and the input feature map, and recognizes the recognition result of the graph data according to the target output feature map.

在一个实施例中,根据输入特征图生成的第二偏置矩阵,包括:采用已训练的卷积神经网络中的降维函数对输入特征图进行降维,得到降维矩阵,归一化降维矩阵,得到归一化矩阵,归一化矩阵为第二偏置矩阵。In one embodiment, the second bias matrix generated according to the input feature map includes: using the dimensionality reduction function in the trained convolutional neural network to perform dimensionality reduction on the input feature map to obtain a dimensionality reduction matrix, normalized reduction dimension matrix to obtain a normalized matrix, and the normalized matrix is the second bias matrix.

在一个实施例中,降维函数包括两个,输入特征图至少包括三个维度,其中,第一维度为通道数,包括:根据降维函数中的第一降维函数对输入特征图的各个通道的矩阵进行降维,得到各个通道的第一降维矩阵,根据降维函数中的第二降维函数对输入特征图的各个通道的矩阵进行降维,得到各个通道的第二降维矩阵,计算各个通道的第一降维矩阵和第二降维矩阵的乘积,得到各个通道的第一乘积矩阵,归一化各个通道的第一乘积矩阵,得到归一化矩阵对应的通道的矩阵。In one embodiment, the dimension reduction function includes two, and the input feature map includes at least three dimensions, wherein the first dimension is the number of channels, including: according to the first dimension reduction function in the dimension reduction function, each of the input feature maps The matrix of the channel is dimensionally reduced to obtain the first dimensionality reduction matrix of each channel, and the matrix of each channel of the input feature map is reduced according to the second dimensionality reduction function in the dimensionality reduction function to obtain the second dimensionality reduction matrix of each channel , calculate the product of the first dimensionality reduction matrix and the second dimensionality reduction matrix of each channel, obtain the first product matrix of each channel, normalize the first product matrix of each channel, and obtain the matrix of the channel corresponding to the normalization matrix.

在一个实施例中,处理器执行计算机程序时还实现以下步骤:生成已训练的卷积神经网络的步骤,包括:获取包含多个训练图数据的训练集合,训练图数据携带标签信息,将训练图数据和标签信息输入初始卷积神经网络,通过初始卷积神经网络提取各个训练图数据的特征,根据各个训练图数据的特征,识别出各个训练图数据对应的识别结果,按照预设损失函数计算各个训练图数据的识别结果和标签的损失值,当损失值小于或等于预设损失值时,得到已训练的卷积神经网络。In one embodiment, the following steps are also implemented when the processor executes the computer program: the step of generating a trained convolutional neural network includes: obtaining a training set comprising a plurality of training graph data, the training graph data carrying label information, and training The graph data and label information are input into the initial convolutional neural network, and the characteristics of each training graph data are extracted through the initial convolutional neural network. According to the characteristics of each training graph data, the recognition results corresponding to each training graph data are identified, and the preset loss function is used. Calculate the recognition result of each training image data and the loss value of the label. When the loss value is less than or equal to the preset loss value, a trained convolutional neural network is obtained.

在一个实施例中,处理器执行计算机程序时还实现以下步骤:当损失值大于预设损失值时,根据损失值通过梯度回传算法更新初始卷积神经网络的网络参数,采用更新了网络参数的初始卷积神经网络作为初始卷积神经网络,进入将训练图数据和标签信息输入初始卷积神经网络,直至按照预设损失函数计算各个训练图数据的识别结果和标签的损失值,小于或等于预设损失值时,得到已训练的卷积神经网络。In one embodiment, when the processor executes the computer program, the following steps are also implemented: when the loss value is greater than the preset loss value, update the network parameters of the initial convolutional neural network through the gradient return algorithm according to the loss value, and use the updated network parameters The initial convolutional neural network is used as the initial convolutional neural network, and the training image data and label information are input into the initial convolutional neural network until the recognition result of each training image data and the loss value of the label are calculated according to the preset loss function, which is less than or When equal to the preset loss value, the trained convolutional neural network is obtained.

在一个实施例中,初始卷积神经网络模型包括至少一个卷积层,卷积层中包括初始偏置矩阵和初始降维函数,根据损失值通过梯度回传算法更新初始卷积神经网络的网络参数,包括:通过梯度回传算法将损失值回传到任意一个卷积层时,得到各个卷积层的回传值,根据各个卷积层的回传值更新初始降维函数的参数和初始偏置矩阵的参数。In one embodiment, the initial convolutional neural network model includes at least one convolutional layer, the convolutional layer includes an initial bias matrix and an initial dimensionality reduction function, and the network of the initial convolutional neural network is updated through a gradient backpropagation algorithm according to the loss value Parameters, including: when the loss value is returned to any convolution layer through the gradient return algorithm, the return value of each convolution layer is obtained, and the parameters of the initial dimensionality reduction function and the initial dimensionality reduction function are updated according to the return value of each convolution layer. Parameters for the bias matrix.

在一个实施例中,根据目标输出特征图,识别出图数据对应的识别结果,包括:当当前卷积层为已训练的卷积神经网络中的最后一个卷积层时,判断多个目标输出特征图中是否存在需要合并的目标特征图,当存在时,合并需要合并的目标输出特征图,得到合并特征图,当合并特征图包含全部目标输出特征图时,对合并特征图进行识别,得到合并特征图对应的识别结果,当合并特征图包含全部目标输出特征图时,对合并特征图进行识别,得到合并特征图对应的识别结果,对未合并的目标输出特征图进行识别,得到未合并的目标输出特征图对应的识别结果,当不存在时,对各个目标输出特征图进行识别,得到各个目标输出特征图对应的识别结果。In one embodiment, according to the target output feature map, identifying the recognition result corresponding to the graph data includes: when the current convolutional layer is the last convolutional layer in the trained convolutional neural network, judging multiple target output Whether there is a target feature map that needs to be merged in the feature map, if it exists, merge the target output feature map that needs to be merged to obtain the merged feature map, when the merged feature map contains all target output feature maps, identify the merged feature map to get The recognition result corresponding to the merged feature map. When the merged feature map contains all target output feature maps, the merged feature map is recognized to obtain the recognition result corresponding to the merged feature map. The unmerged target output feature map is recognized to obtain the unmerged feature map. If there is no recognition result corresponding to the target output feature map, each target output feature map is recognized, and the recognition result corresponding to each target output feature map is obtained.

在一个实施例中,提供了一种计算机可读存储介质,其上存储有计算机程序,计算机程序被处理器执行时实现以下步骤:获取输入已训练的卷积神经网络的当前卷积层的输入特征图,输入特征图为根据图像数据生成的特征图,获取当前卷积层的第一偏置矩阵,其中第一偏置矩阵为生成已训练的卷积神经网络时生成的矩阵,根据输入特征图生成第二偏置矩阵,获取参考邻接矩阵,计算参考邻接矩阵、第一偏置矩阵和第二偏置矩阵的和,得到目标邻接矩阵,获取当前卷积层的卷积核,根据当前卷积层的卷积核、目标邻接矩阵和输入特征图生成目标输出特征图,根据目标输出特征图,识别出图数据的识别结果。In one embodiment, a computer-readable storage medium is provided, on which a computer program is stored, and when the computer program is executed by a processor, the following steps are implemented: obtaining the input input to the current convolutional layer of the trained convolutional neural network Feature map, the input feature map is the feature map generated according to the image data, and the first bias matrix of the current convolutional layer is obtained, where the first bias matrix is the matrix generated when the trained convolutional neural network is generated, according to the input feature Figure generates the second bias matrix, obtains the reference adjacency matrix, calculates the sum of the reference adjacency matrix, the first bias matrix and the second bias matrix, obtains the target adjacency matrix, obtains the convolution kernel of the current convolution layer, and calculates the The convolution kernel of the product layer, the target adjacency matrix and the input feature map generate the target output feature map, and the recognition result of the graph data is recognized according to the target output feature map.

在一个实施例中,根据输入特征图生成的第二偏置矩阵,包括:采用已训练的卷积神经网络中的降维函数对输入特征图进行降维,得到降维矩阵,归一化降维矩阵,得到归一化矩阵,归一化矩阵为第二偏置矩阵。In one embodiment, the second bias matrix generated according to the input feature map includes: using the dimensionality reduction function in the trained convolutional neural network to perform dimensionality reduction on the input feature map to obtain a dimensionality reduction matrix, normalized reduction dimension matrix to obtain a normalized matrix, and the normalized matrix is the second bias matrix.

在一个实施例中,降维函数包括两个,输入特征图至少包括三个维度,其中,第一维度为通道数,包括:根据降维函数中的第一降维函数对输入特征图的各个通道的矩阵进行降维,得到各个通道的第一降维矩阵,根据降维函数中的第二降维函数对输入特征图的各个通道的矩阵进行降维,得到各个通道的第二降维矩阵,计算各个通道的第一降维矩阵和第二降维矩阵的乘积,得到各个通道的第一乘积矩阵,归一化各个通道的第一乘积矩阵,得到归一化矩阵对应的通道的矩阵。In one embodiment, the dimension reduction function includes two, and the input feature map includes at least three dimensions, wherein the first dimension is the number of channels, including: according to the first dimension reduction function in the dimension reduction function, each of the input feature maps The matrix of the channel is dimensionally reduced to obtain the first dimensionality reduction matrix of each channel, and the matrix of each channel of the input feature map is reduced according to the second dimensionality reduction function in the dimensionality reduction function to obtain the second dimensionality reduction matrix of each channel , calculate the product of the first dimensionality reduction matrix and the second dimensionality reduction matrix of each channel, obtain the first product matrix of each channel, normalize the first product matrix of each channel, and obtain the matrix of the channel corresponding to the normalization matrix.

在一个实施例中,计算机程序被处理器执行时还实现以下步骤:生成已训练的卷积神经网络的步骤,包括:获取包含多个训练图数据的训练集合,训练图数据携带标签信息,将训练图数据和标签信息输入初始卷积神经网络,通过初始卷积神经网络提取各个训练图数据的特征,根据各个训练图数据的特征,识别出各个训练图数据对应的识别结果,按照预设损失函数计算各个训练图数据的识别结果和标签的损失值,当损失值小于或等于预设损失值时,得到已训练的卷积神经网络。In one embodiment, when the computer program is executed by the processor, the following steps are also implemented: the step of generating a trained convolutional neural network includes: obtaining a training set comprising a plurality of training graph data, the training graph data carrying label information, and The training image data and label information are input into the initial convolutional neural network, and the characteristics of each training image data are extracted through the initial convolutional neural network. According to the characteristics of each training image data, the recognition results corresponding to each training image data are identified. According to the preset loss The function calculates the recognition result of each training image data and the loss value of the label. When the loss value is less than or equal to the preset loss value, a trained convolutional neural network is obtained.

在一个实施例中,计算机程序被处理器执行时还实现以下步骤:当损失值大于预设损失值时,根据损失值通过梯度回传算法更新初始卷积神经网络的网络参数,采用更新了网络参数的初始卷积神经网络作为初始卷积神经网络,进入将训练图数据和标签信息输入初始卷积神经网络,直至按照预设损失函数计算各个训练图数据的识别结果和标签的损失值,小于或等于预设损失值时,得到已训练的卷积神经网络。In one embodiment, when the computer program is executed by the processor, the following steps are also implemented: when the loss value is greater than the preset loss value, update the network parameters of the initial convolutional neural network through the gradient return algorithm according to the loss value, and use the updated network The initial convolutional neural network with parameters is used as the initial convolutional neural network, and the training image data and label information are input into the initial convolutional neural network until the recognition result of each training image data and the loss value of the label are calculated according to the preset loss function, which is less than Or when it is equal to the preset loss value, the trained convolutional neural network is obtained.

在一个实施例中,初始卷积神经网络模型包括至少一个卷积层,卷积层中包括初始偏置矩阵和初始降维函数,根据损失值通过梯度回传算法更新初始卷积神经网络的网络参数,包括:通过梯度回传算法将损失值回传到任意一个卷积层时,得到各个卷积层的回传值,根据各个卷积层的回传值更新初始降维函数的参数和初始偏置矩阵的参数。In one embodiment, the initial convolutional neural network model includes at least one convolutional layer, the convolutional layer includes an initial bias matrix and an initial dimensionality reduction function, and the network of the initial convolutional neural network is updated through a gradient backpropagation algorithm according to the loss value Parameters, including: when the loss value is returned to any convolution layer through the gradient return algorithm, the return value of each convolution layer is obtained, and the parameters of the initial dimensionality reduction function and the initial dimensionality reduction function are updated according to the return value of each convolution layer. Parameters for the bias matrix.

在一个实施例中,根据目标输出特征图,识别出图数据对应的识别结果,包括:当当前卷积层为已训练的卷积神经网络中的最后一个卷积层时,判断多个目标输出特征图中是否存在需要合并的目标特征图,当存在时,合并需要合并的目标输出特征图,得到合并特征图,当合并特征图包含全部目标输出特征图时,对合并特征图进行识别,得到合并特征图对应的识别结果,当合并特征图包含全部目标输出特征图时,对合并特征图进行识别,得到合并特征图对应的识别结果,对未合并的目标输出特征图进行识别,得到未合并的目标输出特征图对应的识别结果,当不存在时,对各个目标输出特征图进行识别,得到各个目标输出特征图对应的识别结果。In one embodiment, according to the target output feature map, identifying the recognition result corresponding to the graph data includes: when the current convolutional layer is the last convolutional layer in the trained convolutional neural network, judging multiple target output Whether there is a target feature map that needs to be merged in the feature map, if it exists, merge the target output feature map that needs to be merged to obtain the merged feature map, when the merged feature map contains all target output feature maps, identify the merged feature map to get The recognition result corresponding to the merged feature map. When the merged feature map contains all target output feature maps, the merged feature map is recognized to obtain the recognition result corresponding to the merged feature map. The unmerged target output feature map is recognized to obtain the unmerged feature map. If there is no recognition result corresponding to the target output feature map, each target output feature map is recognized, and the recognition result corresponding to each target output feature map is obtained.

本领域普通技术人员可以理解实现上述实施例方法中的全部或部分流程,是可以通过计算机程序来指令相关的硬件来完成,所述的程序可存储于一非易失性计算机可读取存储介质中,该程序在执行时,可包括如上述各方法的实施例的流程。其中,本申请所提供的各实施例中所使用的对存储器、存储、数据库或其它介质的任何引用,均可包括非易失性和/或易失性存储器。非易失性存储器可包括只读存储器(ROM)、可编程ROM(PROM)、电可编程ROM(EPROM)、电可擦除可编程ROM(EEPROM)或闪存。易失性存储器可包括随机存取存储器(RAM)或者外部高速缓冲存储器。作为说明而非局限,RAM以多种形式可得,诸如静态RAM(SRAM)、动态RAM(DRAM)、同步DRAM(SDRAM)、双数据率SDRAM(DDRSDRAM)、增强型SDRAM(ESDRAM)、同步链路(Synchlink)DRAM(SLDRAM)、存储器总线(Rambus)直接RAM(RDRAM)、直接存储器总线动态RAM(DRDRAM)、以及存储器总线动态RAM(RDRAM)等。Those of ordinary skill in the art can understand that all or part of the processes in the methods of the above embodiments can be realized through computer programs to instruct related hardware, and the programs can be stored in a non-volatile computer-readable storage medium When the program is executed, it may include the processes of the embodiments of the above-mentioned methods. Wherein, any references to memory, storage, database or other media used in the various embodiments provided in the present application may include non-volatile and/or volatile memory. Nonvolatile memory can include read only memory (ROM), programmable ROM (PROM), electrically programmable ROM (EPROM), electrically erasable programmable ROM (EEPROM), or flash memory. Volatile memory can include random access memory (RAM) or external cache memory. By way of illustration and not limitation, RAM is available in many forms such as Static RAM (SRAM), Dynamic RAM (DRAM), Synchronous DRAM (SDRAM), Double Data Rate SDRAM (DDRSDRAM), Enhanced SDRAM (ESDRAM), Synchronous Chain Synchlink DRAM (SLDRAM), memory bus (Rambus) direct RAM (RDRAM), direct memory bus dynamic RAM (DRDRAM), and memory bus dynamic RAM (RDRAM), etc.

需要说明的是,在本文中,诸如“第一”和“第二”等之类的关系术语仅仅用来将一个实体或者操作与另一个实体或操作区分开来,而不一定要求或者暗示这些实体或操作之间存在任何这种实际的关系或者顺序。而且,术语“包括”、“包含”或者其任何其他变体意在涵盖非排他性的包含,从而使得包括一系列要素的过程、方法、物品或者设备不仅包括那些要素,而且还包括没有明确列出的其他要素,或者是还包括为这种过程、方法、物品或者设备所固有的要素。在没有更多限制的情况下,由语句“包括一个……”限定的要素,并不排除在包括所述要素的过程、方法、物品或者设备中还存在另外的相同要素。It should be noted that in this article, relative terms such as "first" and "second" are only used to distinguish one entity or operation from another entity or operation, and do not necessarily require or imply these No such actual relationship or order exists between entities or operations. Furthermore, the term "comprises", "comprises" or any other variation thereof is intended to cover a non-exclusive inclusion such that a process, method, article, or apparatus comprising a set of elements includes not only those elements, but also includes elements not expressly listed. other elements of or also include elements inherent in such a process, method, article, or device. Without further limitations, an element defined by the phrase "comprising a ..." does not exclude the presence of additional identical elements in the process, method, article or apparatus comprising said element.

以上所述仅是本发明的具体实施方式,使本领域技术人员能够理解或实现本发明。对这些实施例的多种修改对本领域的技术人员来说将是显而易见的,本文中所定义的一般原理可以在不脱离本发明的精神或范围的情况下,在其它实施例中实现。因此,本发明将不会被限制于本文所示的这些实施例,而是要符合与本文所申请的原理和新颖特点相一致的最宽的范围。The above descriptions are only specific embodiments of the present invention, so that those skilled in the art can understand or implement the present invention. Various modifications to these embodiments will be readily apparent to those skilled in the art, and the general principles defined herein may be implemented in other embodiments without departing from the spirit or scope of the invention. Accordingly, the present invention will not be limited to the embodiments shown herein, but is to be accorded the widest scope consistent with the principles and novel features claimed herein.

Claims (10)

1. a kind of diagram data recognition methods, which is characterized in that the described method includes:
The input feature vector figure for the current convolutional layer of convolutional neural networks that input has been trained is obtained, according to the input feature vector figure The characteristic pattern that diagram data generates;
The first bias matrix of the current convolutional layer is obtained, wherein first bias matrix is to generate the volume trained The matrix generated when product neural network;
The second bias matrix is generated according to the input feature vector figure;
It obtains and refers to adjacency matrix, calculate described with reference to adjacency matrix, first bias matrix and second bias matrix Sum, obtain target adjacency matrix;
Obtain the convolution kernel of the current convolutional layer;
It is special that target output is generated according to the convolution kernel of the current convolutional layer, the target adjacency matrix and the input feature vector figure Sign figure;
Characteristic pattern is exported according to the target, identifies the corresponding recognition result of the diagram data.
2. the method stated when institute according to claim 1, which is characterized in that it is described according to the input feature vector figure generate second partially Set matrix, comprising:
Dimensionality reduction is carried out to the input feature vector figure using the dimensionality reduction function in the convolutional neural networks trained, obtains dimensionality reduction Matrix;
The dimensionality reduction matrix is normalized, normalization matrix is obtained, the normalization matrix is second bias matrix.
3. according to the method described in claim 2, it is characterized in that, the dimensionality reduction function include two, the input feature vector figure Including at least three dimensions, wherein the first dimension is port number, which comprises
Dimensionality reduction is carried out according to matrix of the first dimensionality reduction function in the dimensionality reduction function to each channel of the input feature vector figure, Obtain the first dimensionality reduction matrix in each channel;
Dimensionality reduction is carried out according to matrix of the second dimensionality reduction function in the dimensionality reduction function to each channel of the input feature vector figure, Obtain the second dimensionality reduction matrix in each channel;
The first dimensionality reduction matrix in each channel and the product of the second dimensionality reduction matrix are calculated, first product in each channel is obtained Matrix;
The first product matrix for normalizing each channel, obtains the matrix in the corresponding channel of the normalization matrix.
4. the method according to claim 1, wherein the step of generating the convolutional neural networks trained, Include:
The training set comprising multiple trained diagram datas is obtained, the trained diagram data carries label information;
The trained diagram data and the label information are inputted into initial convolutional neural networks, pass through the initial convolution nerve net Network extracts the feature of each trained diagram data;
According to the feature of each trained diagram data, the corresponding recognition result of each trained diagram data is identified;
The recognition result of each trained diagram data and the penalty values of the label are calculated according to default loss function;
When the penalty values are less than or equal to default penalty values, the convolutional neural networks trained are obtained.
5. according to the method described in claim 4, it is characterized in that, the method also includes:
It is described just by gradient passback algorithm update according to the penalty values when the penalty values are greater than the default penalty values The network parameter of beginning convolutional neural networks;
Using having updated the initial convolutional neural networks of network parameter as the initial convolutional neural networks, by the instruction Practice diagram data and the label information inputs initial convolutional neural networks, until described calculate each institute according to default loss function The recognition result of trained diagram data and the penalty values of the label are stated, when being less than or equal to the default penalty values, is obtained described The convolutional neural networks trained.
6. according to the method described in claim 5, it is characterized in that, the initial convolution neural network model includes at least one Convolutional layer, includes initial bias matrix and initial dimensionality reduction function in the convolutional layer, described to pass through gradient according to the penalty values Passback algorithm updates the network parameter of the initial convolutional neural networks, comprising:
When the penalty values being passed back to any one of convolutional layer by gradient passback algorithm, each volume is obtained The return value of lamination;
The network parameter of the convolutional layer is updated according to the return value of each convolutional layer, the network parameter includes initial drop Tie up the parameter of function and the parameter of the initial bias matrix.
7. method according to any one of claim 1 to 6, which is characterized in that described to export feature according to the target Figure, identifies the corresponding recognition result of the diagram data, comprising:
When the current convolutional layer is the last one convolutional layer in the convolutional neural networks trained, multiple institutes are judged State target signature combined with the presence or absence of needs in target output characteristic pattern;
When it is present, merge the target output characteristic pattern for needing to merge, obtain merging characteristic pattern;
When the merging characteristic pattern includes all target output characteristic patterns, the merging characteristic pattern is identified, is obtained To the corresponding recognition result of the merging characteristic pattern;
When the merging characteristic pattern includes all target output characteristic patterns, the merging characteristic pattern is identified, is obtained To the corresponding recognition result of the merging characteristic pattern, the target output characteristic pattern not merged is identified, is obtained described The corresponding recognition result of target output characteristic pattern not merged;
When it be not present, each target output characteristic pattern is identified, obtains each target output characteristic pattern pair The recognition result answered.
8. a kind of diagram data identification device, which is characterized in that described device includes:
Data acquisition module, the input feature vector figure of the current convolutional layer for obtaining the convolutional neural networks that input has been trained, institute Stating input feature vector figure is the characteristic pattern generated according to image data, obtains the first bias matrix of the current convolutional layer, wherein First bias matrix is the matrix generated when generating the convolutional neural networks trained;
Second bias matrix generation module, for generating the second bias matrix according to the input feature vector figure;
Target adjacency matrix life module, refers to adjacency matrix for obtaining, and calculates the reference adjacency matrix, first biasing The sum of matrix and second bias matrix, obtains target adjacency matrix;
Target exports characteristic pattern generation module, for obtaining the convolution kernel of the current convolutional layer, according to the current convolutional layer Convolution kernel, the target adjacency matrix and the input feature vector figure generate target and export characteristic pattern;
Identification module identifies the recognition result of the diagram data for exporting characteristic pattern according to the target.
9. a kind of computer equipment including memory, processor and stores the meter that can be run on a memory and on a processor Calculation machine program, which is characterized in that the processor realizes any one of claims 1 to 7 institute when executing the computer program The step of stating method.
10. a kind of computer readable storage medium, is stored thereon with computer program, which is characterized in that the computer program The step of method described in any one of claims 1 to 7 is realized when being executed by processor.
CN201910503194.XA 2019-06-11 2019-06-11 Diagram data recognition methods, device, computer equipment and storage medium Pending CN110378372A (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN201910503194.XA CN110378372A (en) 2019-06-11 2019-06-11 Diagram data recognition methods, device, computer equipment and storage medium
PCT/CN2019/129708 WO2020248581A1 (en) 2019-06-11 2019-12-30 Graph data identification method and apparatus, computer device, and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910503194.XA CN110378372A (en) 2019-06-11 2019-06-11 Diagram data recognition methods, device, computer equipment and storage medium

Publications (1)

Publication Number Publication Date
CN110378372A true CN110378372A (en) 2019-10-25

Family

ID=68250141

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910503194.XA Pending CN110378372A (en) 2019-06-11 2019-06-11 Diagram data recognition methods, device, computer equipment and storage medium

Country Status (2)

Country Link
CN (1) CN110378372A (en)
WO (1) WO2020248581A1 (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111967479A (en) * 2020-07-27 2020-11-20 广东工业大学 Image target identification method based on convolutional neural network idea
WO2020248581A1 (en) * 2019-06-11 2020-12-17 中国科学院自动化研究所 Graph data identification method and apparatus, computer device, and storage medium
CN114707641A (en) * 2022-03-23 2022-07-05 平安科技(深圳)有限公司 Training method, device, equipment and medium for neural network model of double-view diagram
CN114724249A (en) * 2022-04-20 2022-07-08 北京沃东天骏信息技术有限公司 Action recognition method, device and storage medium
CN115601617A (en) * 2022-11-25 2023-01-13 安徽数智建造研究院有限公司(Cn) Training method and device of banded void recognition model based on semi-supervised learning
CN118504607A (en) * 2024-07-17 2024-08-16 密码与网络空间安全(黄埔)研究院 Reversible model watermark generation method, device, equipment and medium

Families Citing this family (22)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112560712B (en) * 2020-12-18 2023-05-26 西安电子科技大学 Behavior recognition method, device and medium based on time enhancement graph convolutional network
CN112633224B (en) * 2020-12-30 2024-03-26 深圳云天励飞技术股份有限公司 Social relation recognition method and device, electronic equipment and storage medium
CN112965062B (en) * 2021-02-09 2024-02-20 西安电子科技大学 Radar range profile target recognition method based on LSTM-DAM network
CN113269239B (en) * 2021-05-13 2024-04-19 河南大学 Relation network node classification method based on multichannel convolutional neural network
CN113239875B (en) * 2021-06-01 2023-10-17 恒睿(重庆)人工智能技术研究院有限公司 Method, system and device for acquiring face characteristics and computer readable storage medium
CN113468469A (en) * 2021-06-02 2021-10-01 北京迈格威科技有限公司 Convolution processing method and device of feature graph executed by computer and electronic equipment
CN113468980B (en) * 2021-06-11 2024-05-31 浙江大华技术股份有限公司 Human behavior recognition method and related device
CN113741459B (en) * 2021-09-03 2024-06-21 阿波罗智能技术(北京)有限公司 Method for determining training sample and training method and device for automatic driving model
CN113887575B (en) * 2021-09-13 2024-04-05 华南理工大学 Graph dataset enhancement method based on self-adaptive graph rolling network
CN113761771B (en) * 2021-09-16 2024-05-28 中国人民解放军国防科技大学 Porous material sound absorption performance prediction method and device, electronic equipment and storage medium
CN113971319B (en) * 2021-10-12 2023-04-18 浙江腾腾电气有限公司 Voltage stabilizer with precision compensation and compensation method thereof
CN114090651B (en) * 2021-11-10 2023-04-18 哈尔滨工业大学(深圳) Traffic flow abnormal data judgment method based on double-channel graph neural network self-encoder
CN114120453A (en) * 2021-11-29 2022-03-01 北京百度网讯科技有限公司 Living body detection method and device, electronic equipment and storage medium
CN114493432A (en) * 2022-01-20 2022-05-13 温州大学 Route planning method, device and electronic device for express delivery based on UAV
CN114549947A (en) * 2022-01-24 2022-05-27 北京百度网讯科技有限公司 Model training method, device, equipment and storage medium
CN115064285B (en) * 2022-03-25 2024-10-15 中国科学院自动化研究所 Deep learning-based traditional Chinese medicine prescription analysis method and device
CN114861757B (en) * 2022-03-31 2024-12-10 中国人民解放军国防科技大学 A graph deep clustering method based on double correlation reduction
CN115564630B (en) * 2022-09-28 2023-05-30 华能伊敏煤电有限责任公司 Automatic control method and system for excavating flow of bucket wheel excavator
CN116935363B (en) * 2023-07-04 2024-02-23 东莞市微振科技有限公司 Cutter identification method, cutter identification device, electronic equipment and readable storage medium
CN117557244B (en) * 2023-09-27 2024-07-02 国网江苏省电力有限公司信息通信分公司 Electric power operation and maintenance warning system based on knowledge graph
CN117235584B (en) * 2023-11-15 2024-04-02 之江实验室 Graph data classification method, device, electronic device and storage medium
CN117251715B (en) * 2023-11-17 2024-03-19 华芯程(杭州)科技有限公司 Layout measurement area screening method and device, electronic equipment and storage medium

Citations (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101571948A (en) * 2009-06-11 2009-11-04 西安电子科技大学 Method for restoring motion blurred image based on total variational model and neutral network
US20090324107A1 (en) * 2008-06-25 2009-12-31 Gannon Technologies Group, Llc Systems and methods for image recognition using graph-based pattern matching
US20100158352A1 (en) * 2008-12-22 2010-06-24 Electronics And Telecommunications Research Institute Apparatus and method for real-time camera tracking
CN104036474A (en) * 2014-06-12 2014-09-10 厦门美图之家科技有限公司 Automatic adjustment method for image brightness and contrast
CN105095833A (en) * 2014-05-08 2015-11-25 中国科学院声学研究所 Network constructing method for human face identification, identification method and system
US20150350576A1 (en) * 2014-05-30 2015-12-03 Apple Inc. Raw Camera Noise Reduction Using Alignment Mapping
CN106228142A (en) * 2016-07-29 2016-12-14 西安电子科技大学 Face verification method based on convolutional neural networks and Bayesian decision
CN106339753A (en) * 2016-08-17 2017-01-18 中国科学技术大学 Method for effectively enhancing robustness of convolutional neural network
CN107122396A (en) * 2017-03-13 2017-09-01 西北大学 Three-dimensional model searching algorithm based on depth convolutional neural networks
US20180084988A1 (en) * 2016-09-23 2018-03-29 International Business Machines Corporation Early Prediction Of Age Related Macular Degeneration By Image Reconstruction
CN108875827A (en) * 2018-06-15 2018-11-23 广州深域信息科技有限公司 A kind of method and system of fine granularity image classification
CN108985366A (en) * 2018-07-06 2018-12-11 武汉兰丁医学高科技有限公司 B ultrasound image recognition algorithm and system based on convolution depth network
CN109063706A (en) * 2018-06-04 2018-12-21 平安科技(深圳)有限公司 Verbal model training method, character recognition method, device, equipment and medium
CN109086652A (en) * 2018-06-04 2018-12-25 平安科技(深圳)有限公司 Handwritten word model training method, Chinese characters recognition method, device, equipment and medium

Family Cites Families (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2017095948A1 (en) * 2015-11-30 2017-06-08 Pilot Ai Labs, Inc. Improved general object detection using neural networks
CN106372656B (en) * 2016-08-30 2019-05-10 同观科技(深圳)有限公司 Obtain method, image-recognizing method and the device of the disposable learning model of depth
CN108256544B (en) * 2016-12-29 2019-07-23 杭州光启人工智能研究院 Picture classification method and device, robot
CN110363086A (en) * 2019-06-11 2019-10-22 中国科学院自动化研究所南京人工智能芯片创新研究院 Image data recognition method, device, computer equipment and storage medium
CN110378372A (en) * 2019-06-11 2019-10-25 中国科学院自动化研究所南京人工智能芯片创新研究院 Diagram data recognition methods, device, computer equipment and storage medium
CN110390259A (en) * 2019-06-11 2019-10-29 中国科学院自动化研究所南京人工智能芯片创新研究院 Image data recognition method, device, computer equipment and storage medium

Patent Citations (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090324107A1 (en) * 2008-06-25 2009-12-31 Gannon Technologies Group, Llc Systems and methods for image recognition using graph-based pattern matching
US20100158352A1 (en) * 2008-12-22 2010-06-24 Electronics And Telecommunications Research Institute Apparatus and method for real-time camera tracking
CN101571948A (en) * 2009-06-11 2009-11-04 西安电子科技大学 Method for restoring motion blurred image based on total variational model and neutral network
CN105095833A (en) * 2014-05-08 2015-11-25 中国科学院声学研究所 Network constructing method for human face identification, identification method and system
US20150350576A1 (en) * 2014-05-30 2015-12-03 Apple Inc. Raw Camera Noise Reduction Using Alignment Mapping
CN104036474A (en) * 2014-06-12 2014-09-10 厦门美图之家科技有限公司 Automatic adjustment method for image brightness and contrast
CN106228142A (en) * 2016-07-29 2016-12-14 西安电子科技大学 Face verification method based on convolutional neural networks and Bayesian decision
CN106339753A (en) * 2016-08-17 2017-01-18 中国科学技术大学 Method for effectively enhancing robustness of convolutional neural network
US20180084988A1 (en) * 2016-09-23 2018-03-29 International Business Machines Corporation Early Prediction Of Age Related Macular Degeneration By Image Reconstruction
CN107122396A (en) * 2017-03-13 2017-09-01 西北大学 Three-dimensional model searching algorithm based on depth convolutional neural networks
CN109063706A (en) * 2018-06-04 2018-12-21 平安科技(深圳)有限公司 Verbal model training method, character recognition method, device, equipment and medium
CN109086652A (en) * 2018-06-04 2018-12-25 平安科技(深圳)有限公司 Handwritten word model training method, Chinese characters recognition method, device, equipment and medium
CN108875827A (en) * 2018-06-15 2018-11-23 广州深域信息科技有限公司 A kind of method and system of fine granularity image classification
CN108985366A (en) * 2018-07-06 2018-12-11 武汉兰丁医学高科技有限公司 B ultrasound image recognition algorithm and system based on convolution depth network

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
LEI SHI等: "Non-Local Graph Convolutional Networks for Skeleton-Based Action Recognition", 《ARXIV》 *

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2020248581A1 (en) * 2019-06-11 2020-12-17 中国科学院自动化研究所 Graph data identification method and apparatus, computer device, and storage medium
CN111967479A (en) * 2020-07-27 2020-11-20 广东工业大学 Image target identification method based on convolutional neural network idea
CN114707641A (en) * 2022-03-23 2022-07-05 平安科技(深圳)有限公司 Training method, device, equipment and medium for neural network model of double-view diagram
WO2023178793A1 (en) * 2022-03-23 2023-09-28 平安科技(深圳)有限公司 Method and apparatus for training dual-perspective graph neural network model, device, and medium
CN114707641B (en) * 2022-03-23 2024-11-08 平安科技(深圳)有限公司 Training method, device, equipment and medium for dual-view graph neural network model
CN114724249A (en) * 2022-04-20 2022-07-08 北京沃东天骏信息技术有限公司 Action recognition method, device and storage medium
CN115601617A (en) * 2022-11-25 2023-01-13 安徽数智建造研究院有限公司(Cn) Training method and device of banded void recognition model based on semi-supervised learning
CN118504607A (en) * 2024-07-17 2024-08-16 密码与网络空间安全(黄埔)研究院 Reversible model watermark generation method, device, equipment and medium
CN118504607B (en) * 2024-07-17 2024-10-01 密码与网络空间安全(黄埔)研究院 Reversible model watermark generation method, device, equipment and medium

Also Published As

Publication number Publication date
WO2020248581A1 (en) 2020-12-17

Similar Documents

Publication Publication Date Title
CN110378372A (en) Diagram data recognition methods, device, computer equipment and storage medium
CN110363086A (en) Image data recognition method, device, computer equipment and storage medium
CN108304847B (en) Image classification method and device and personalized recommendation method and device
CN110135406B (en) Image recognition method and device, computer equipment and storage medium
CN112651438A (en) Multi-class image classification method and device, terminal equipment and storage medium
WO2019100724A1 (en) Method and device for training multi-label classification model
CN112749723B (en) Sample labeling method, device, computer equipment and storage medium
CN110390259A (en) Image data recognition method, device, computer equipment and storage medium
WO2017088432A1 (en) Image recognition method and device
CN111126339A (en) Gesture recognition method and device, computer equipment and storage medium
CN113255557B (en) Deep learning-based video crowd emotion analysis method and system
WO2021031704A1 (en) Object tracking method and apparatus, computer device, and storage medium
CN113987188B (en) A kind of short text classification method, device and electronic equipment
CN112001399B (en) Image scene classification method and device based on local feature saliency
CN110765882A (en) Video tag determination method, device, server and storage medium
US20230334893A1 (en) Method for optimizing human body posture recognition model, device and computer-readable storage medium
CN111507285A (en) Face attribute recognition method and device, computer equipment and storage medium
CN110163864A (en) Image partition method, device, computer equipment and storage medium
CN110378213A (en) Activity recognition method, apparatus, computer equipment and storage medium
CN110705490A (en) Visual Emotion Recognition Methods
CN111507288A (en) Image detection method, image detection device, computer equipment and storage medium
CN113343981A (en) Visual feature enhanced character recognition method, device and equipment
CN114612681A (en) Multi-label image classification method, model construction method and device based on GCN
CN112232397A (en) Knowledge distillation method and device of image classification model and computer equipment
JP2015036939A (en) Feature extraction program and information processing apparatus

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20191025