CN114492789B - A method and device for constructing a neural network model of data samples - Google Patents
A method and device for constructing a neural network model of data samples Download PDFInfo
- Publication number
- CN114492789B CN114492789B CN202210085874.6A CN202210085874A CN114492789B CN 114492789 B CN114492789 B CN 114492789B CN 202210085874 A CN202210085874 A CN 202210085874A CN 114492789 B CN114492789 B CN 114492789B
- Authority
- CN
- China
- Prior art keywords
- layer
- neural network
- input
- network
- projection
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/06—Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- General Health & Medical Sciences (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- Computational Linguistics (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Artificial Intelligence (AREA)
- Neurology (AREA)
- Image Analysis (AREA)
Abstract
Description
技术领域Technical Field
本发明涉及数据处理技术领域,特别是涉及一种数据样本的神经网络模型构建方法及装置。The present invention relates to the field of data processing technology, and in particular to a method and device for constructing a neural network model of data samples.
背景技术Background technique
人工神经网络是人工智能领域的重要研究方向,它是从信息处理角度对人脑神经网络的一种模拟。人工神经网络可以任意逼近任何非线性函数,进而实现对信号不同模式的区分。人工神经网络在信号处理、模式识别、自动控制、人工智能等众多领域得到了大量应用。人工神经网络目前已有许多模型,比如BP模型、Hopfield模型、超限学习机、深度学习模型等。这些模型通常由输入层、隐藏层和输出层组成。信息在层之间进行处理和传输。在每一层中,神经元通过激活函数对输入信号进行非线性处理,进而实现特征提取。每层由多个神经元节点组成。相邻层神经元节点之间相互连接,但同层和跨层的神经元节点之间通常不连接。神经元之间的连接都有权值,学习过程就是不断修改神经元之间连接权值的过程。BP神经网络通过信息的前向传输和误差的反向传输来调整权值。Artificial neural network is an important research direction in the field of artificial intelligence. It is a simulation of the human brain neural network from the perspective of information processing. Artificial neural network can arbitrarily approximate any nonlinear function, thereby realizing the distinction between different signal patterns. Artificial neural network has been widely used in many fields such as signal processing, pattern recognition, automatic control, artificial intelligence, etc. There are many models of artificial neural network, such as BP model, Hopfield model, extreme learning machine, deep learning model, etc. These models usually consist of input layer, hidden layer and output layer. Information is processed and transmitted between layers. In each layer, neurons perform nonlinear processing on input signals through activation functions to realize feature extraction. Each layer consists of multiple neuron nodes. Neuron nodes in adjacent layers are connected to each other, but neuron nodes in the same layer and across layers are usually not connected. The connections between neurons have weights, and the learning process is the process of constantly modifying the weights of the connections between neurons. BP neural network adjusts weights by forward transmission of information and reverse transmission of errors.
在BP神经网络的基础上,深度学习重新激活了神经网络的研究,它将人工神经网络从浅层推向了深层,开启了深度神经网络(DNN)的新时代。深度神经网络具有更多的隐含层,第一个隐含层从原始数据中提取基本特征,后面的隐含层将基本特征组合成一个更高阶的抽象特征。深度学习能自动提取分类需要的特征,不需要人的参与,在语音识别,图像处理、模式识别等领域取得了巨大成功。Based on the BP neural network, deep learning has reactivated the research of neural networks. It has pushed artificial neural networks from shallow layers to deep layers, opening a new era of deep neural networks (DNN). Deep neural networks have more hidden layers. The first hidden layer extracts basic features from the original data, and the subsequent hidden layers combine the basic features into a higher-order abstract feature. Deep learning can automatically extract the features required for classification without human participation, and has achieved great success in speech recognition, image processing, pattern recognition and other fields.
但是,深度学习和以前的神经网络一样,需要消耗大量的时间来完成网络的训练。虽然是大数据时代,但专家知识样本很稀缺。另一方面,实际应用中计算能力往往受限,有限的计算资源制约着DNN性能的发挥,同时计算资源的节省可节约能源、带来效益。因此,需要研究泛化能力强且学习速度快的人工神经网络。However, deep learning, like previous neural networks, requires a lot of time to complete network training. Although it is the era of big data, expert knowledge samples are scarce. On the other hand, computing power is often limited in practical applications. Limited computing resources restrict the performance of DNN. At the same time, saving computing resources can save energy and bring benefits. Therefore, it is necessary to study artificial neural networks with strong generalization ability and fast learning speed.
发明内容Summary of the invention
针对于上述问题,本发明提供一种数据样本的神经网络模型构建方法及装置,实现了信息在网络内部的快速传递,提升了对样本数据的学习效率。In order to solve the above problems, the present invention provides a method and device for constructing a neural network model of data samples, which realizes the rapid transmission of information within the network and improves the learning efficiency of sample data.
为了实现上述目的,本发明提供了如下技术方案:In order to achieve the above object, the present invention provides the following technical solutions:
一种数据样本的神经网络模型构建方法,包括:A method for constructing a neural network model of a data sample, comprising:
获取预设神经网络模型,所述预设神经网络模型包括输入层、投射网络和学习网络,所述输入层用于获取输入样本信号,并将输入样本信号映射为神经元输出,所述投射网络用于对输入层输出的高维信息进行降维,所述学习网络用于学习投射网络降维后的信息,并输出学习结果;Obtain a preset neural network model, the preset neural network model includes an input layer, a projection network and a learning network, the input layer is used to obtain an input sample signal and map the input sample signal to a neuron output, the projection network is used to reduce the dimension of the high-dimensional information output by the input layer, and the learning network is used to learn the information after the dimension reduction of the projection network and output the learning result;
基于与所述输入样本信号对应的目标任务,确定与所述预设神经网络模型每一层的连接权值;Determining a connection weight with each layer of the preset neural network model based on a target task corresponding to the input sample signal;
根据所述预设神经网络模型每一层的连接权值对所述预设神经网络模型进行训练得到目标神经网络模型。The preset neural network model is trained according to the connection weights of each layer of the preset neural network model to obtain a target neural network model.
可选地,还包括:Optionally, it also includes:
获取目标训练样本;Obtain target training samples;
基于所述目标训练样本的长度,确定所述预设神经网络模型每一层的神经元数量。Based on the length of the target training sample, the number of neurons in each layer of the preset neural network model is determined.
可选地,所述预设神经网络模型的输入层由单层神经网络组成;所述投射网络由若干层神经网络组成,所述投射网络的若干层神经网络具有特定的连接权值,且每一层神经网络都由若干神经元组成,投射网络神经元仅与前一层的有限个神经元相连接;所述学习网络由一个浅层神经网络组成,所述神经网络的第一层神经网络与所述投射网络的第一层神经网络进行连接,所述浅层神经网络是一个全连接网络,且所述浅层神经网络不同层之间的神经元通过权值连接。Optionally, the input layer of the preset neural network model is composed of a single-layer neural network; the projection network is composed of several layers of neural networks, the several layers of neural networks of the projection network have specific connection weights, and each layer of the neural network is composed of several neurons, and the projection network neurons are only connected to a limited number of neurons in the previous layer; the learning network is composed of a shallow neural network, the first layer of the neural network is connected to the first layer of the projection network, the shallow neural network is a fully connected network, and the neurons between different layers of the shallow neural network are connected through weights.
可选地,所述基于与所述输入样本信号对应的目标任务,确定与所述预设神经网络模型每一层的连接权值,包括:Optionally, determining the connection weights with each layer of the preset neural network model based on the target task corresponding to the input sample signal includes:
获取与所述输入样本信号对应的目标训练样本;Acquire a target training sample corresponding to the input sample signal;
将所述目标训练样本输入至所述输入层,并获得所述输入层神经元的输出;Inputting the target training sample into the input layer, and obtaining the output of the neurons in the input layer;
将所述输入层神经元的输出输入至所述投射网络,获得所述投射网络的输出信号;Inputting the output of the input layer neurons into the projection network to obtain an output signal of the projection network;
将所述投射网络的输出信号输入至所述学习网络,并获得所述学习网络的学习结果;Inputting the output signal of the projection network into the learning network, and obtaining the learning result of the learning network;
基于所述目标任务与所述学习结果的比较值,对神经元的连接权值进行调整,得到与所述预设神经网络模型每一层的连接权值。Based on the comparison value between the target task and the learning result, the connection weights of the neurons are adjusted to obtain the connection weights of each layer of the preset neural network model.
可选地,所述方法还包括:Optionally, the method further comprises:
在确定所述预设神经网络模型的每一层的输出信息时,确定所述神经网络模型的每一层的激活函数,以及对每一层输入的信号进行归一化处理;When determining the output information of each layer of the preset neural network model, determining the activation function of each layer of the neural network model, and performing normalization processing on the input signal of each layer;
根据归一化处理后的输入信号以及对应的激活函数,确定每一层的输出信号。The output signal of each layer is determined based on the normalized input signal and the corresponding activation function.
一种数据样本的神经网络模型构建装置,包括:A device for constructing a neural network model of a data sample, comprising:
获取单元,用于获取预设神经网络模型,所述预设神经网络模型包括输入层、投射网络和学习网络,所述输入层用于获取输入样本信号,并将输入样本信号映射为神经元输出,所述投射网络用于对输入层输出的高维信息进行降维,所述学习网络用于学习投射网络降维后的信息,并输出学习结果;An acquisition unit is used to acquire a preset neural network model, wherein the preset neural network model includes an input layer, a projection network and a learning network, wherein the input layer is used to acquire an input sample signal and map the input sample signal to a neuron output, the projection network is used to reduce the dimension of the high-dimensional information output by the input layer, and the learning network is used to learn the information after the dimension reduction of the projection network and output the learning result;
确定单元,用于基于与所述输入样本信号对应的目标任务,确定与所述预设神经网络模型每一层的连接权值;A determination unit, configured to determine a connection weight with each layer of the preset neural network model based on a target task corresponding to the input sample signal;
训练单元,用于根据所述预设神经网络模型每一层的连接权值对所述预设神经网络模型进行训练得到目标神经网络模型。A training unit is used to train the preset neural network model according to the connection weights of each layer of the preset neural network model to obtain a target neural network model.
可选地,还包括:Optionally, it also includes:
神经元数量确定单元,用于获取目标训练样本;A neuron number determination unit, used to obtain target training samples;
基于所述目标训练样本的长度,确定所述预设神经网络模型每一层的神经元数量。Based on the length of the target training sample, the number of neurons in each layer of the preset neural network model is determined.
可选地,所述预设神经网络模型的输入层由单层神经网络组成;所述投射网络由若干层神经网络组成,所述投射网络的若干层神经网络具有特定的连接权值,且每一层神经网络都由若干神经元组成,投射网络神经元仅与前一层的有限个神经元相连接;所述学习网络由一个浅层神经网络组成,所述神经网络的第一层神经网络与所述投射网络的第一层神经网络进行连接,所述浅层神经网络是一个全连接网络,且所述浅层神经网络不同层之间的神经元通过权值连接。Optionally, the input layer of the preset neural network model is composed of a single-layer neural network; the projection network is composed of several layers of neural networks, the several layers of neural networks of the projection network have specific connection weights, and each layer of the neural network is composed of several neurons, and the projection network neurons are only connected to a limited number of neurons in the previous layer; the learning network is composed of a shallow neural network, the first layer of the neural network is connected to the first layer of the projection network, the shallow neural network is a fully connected network, and the neurons between different layers of the shallow neural network are connected through weights.
可选地,所述确定单元具体用于:Optionally, the determining unit is specifically configured to:
获取与所述输入样本信号对应的目标训练样本;Acquire a target training sample corresponding to the input sample signal;
将所述目标训练样本输入至所述输入层,并获得所述输入层神经元的输出;Inputting the target training sample into the input layer, and obtaining the output of the neurons in the input layer;
将所述输入层神经元的输出输入至所述投射网络,获得所述投射网络的输出信号;Inputting the output of the input layer neurons into the projection network to obtain an output signal of the projection network;
将所述投射网络的输出信号输入至所述学习网络,并获得所述学习网络的学习结果;Inputting the output signal of the projection network into the learning network, and obtaining the learning result of the learning network;
基于所述目标任务与所述学习结果的比较值,对神经元的连接权值进行调整,得到与所述预设神经网络模型每一层的连接权值。Based on the comparison value between the target task and the learning result, the connection weights of the neurons are adjusted to obtain the connection weights of each layer of the preset neural network model.
可选地,所述装置还包括:Optionally, the device further comprises:
输出信号确定单元,用于在确定所述预设神经网络模型的每一层的输出信息时,确定所述神经网络模型的每一层的激活函数,以及对每一层输入的信号进行归一化处理;An output signal determination unit, used to determine the activation function of each layer of the neural network model when determining the output information of each layer of the preset neural network model, and to perform normalization processing on the input signal of each layer;
根据归一化处理后的输入信号以及对应的激活函数,确定每一层的输出信号。The output signal of each layer is determined based on the normalized input signal and the corresponding activation function.
相较于现有技术,本发明提供了一种数据样本的神经网络模型构建方法及装置,包括:获取预设神经网络模型,预设神经网络模型包括输入层、投射网络和学习网络,输入层用于获取输入样本信号,并将输入样本信号映射为神经元输出,投射网络用于对输入层输出的高维信息进行降维,学习网络用于学习投射网络降维后的信息,并输出学习结果;基于与输入样本信号对应的目标任务,确定与预设神经网络模型每一层的连接权值;根据预设神经网络模型每一层的连接权值对预设神经网络模型进行训练得到目标神经网络模型。本发明通过构建投射网络可以便于对数据样本的投射学习,实现信息在网络内部的快速传递,进而实现对样本的高效学习。Compared with the prior art, the present invention provides a method and device for constructing a neural network model of a data sample, including: obtaining a preset neural network model, the preset neural network model includes an input layer, a projection network and a learning network, the input layer is used to obtain an input sample signal and map the input sample signal to a neuron output, the projection network is used to reduce the dimension of the high-dimensional information output by the input layer, and the learning network is used to learn the information after the dimension reduction of the projection network, and output the learning result; based on the target task corresponding to the input sample signal, determine the connection weight with each layer of the preset neural network model; according to the connection weight of each layer of the preset neural network model, train the preset neural network model to obtain the target neural network model. The present invention can facilitate the projection learning of data samples by constructing a projection network, realize the rapid transmission of information within the network, and then realize efficient learning of samples.
附图说明BRIEF DESCRIPTION OF THE DRAWINGS
为了更清楚地说明本发明实施例或现有技术中的技术方案,下面将对实施例或现有技术描述中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本发明的实施例,对于本领域普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据提供的附图获得其他的附图。In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings required for use in the embodiments or the description of the prior art will be briefly introduced below. Obviously, the drawings described below are only embodiments of the present invention. For ordinary technicians in this field, other drawings can be obtained based on the provided drawings without paying creative work.
图1为本发明实施例提供的一种数据样本的神经网络模型构建方法的流程示意图;FIG1 is a schematic diagram of a flow chart of a method for constructing a neural network model of a data sample provided by an embodiment of the present invention;
图2为本发明实施例提供的一种投射学习模型的结构示意图;FIG2 is a schematic diagram of the structure of a projection learning model provided by an embodiment of the present invention;
图3为本发明实施例提供的一种数据样本的神经网络模型构建装置的结构示意图。FIG3 is a schematic diagram of the structure of a device for constructing a neural network model of a data sample provided in an embodiment of the present invention.
具体实施方式Detailed ways
下面将结合本发明实施例中的附图,对本发明实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例仅仅是本发明一部分实施例,而不是全部的实施例。基于本发明中的实施例,本领域普通技术人员在没有做出创造性劳动前提下所获得的所有其他实施例,都属于本发明保护的范围。The following will be combined with the drawings in the embodiments of the present invention to clearly and completely describe the technical solutions in the embodiments of the present invention. Obviously, the described embodiments are only part of the embodiments of the present invention, not all of the embodiments. Based on the embodiments of the present invention, all other embodiments obtained by ordinary technicians in this field without creative work are within the scope of protection of the present invention.
本发明的说明书和权利要求书及上述附图中的术语“第一”和“第二”等是用于区别不同的对象,而不是用于描述特定的顺序。此外术语“包括”和“具有”以及他们任何变形,意图在于覆盖不排他的包含。例如包含了一系列步骤或单元的过程、方法、系统、产品或设备没有设定于已列出的步骤或单元,而是可包括没有列出的步骤或单元。The terms "first" and "second" and the like in the specification and claims of the present invention and the above drawings are used to distinguish different objects, rather than to describe a specific order. In addition, the terms "including" and "having" and any variations thereof are intended to cover non-exclusive inclusions. For example, a process, method, system, product or device including a series of steps or units is not limited to the listed steps or units, but may include steps or units that are not listed.
在本发明实施例中提供了一种数据样本的神经网络模型构建方法,参见图1,该方法可以包括以下步骤:In an embodiment of the present invention, a method for constructing a neural network model of a data sample is provided. Referring to FIG1 , the method may include the following steps:
S101、获取预设神经网络模型。S101, obtaining a preset neural network model.
其中,预设神经网络模型包括输入层、投射网络和学习网络,所述输入层用于获取输入样本信号,并将输入样本信号映射为神经元输出,所述投射网络用于对输入层输出的高维信息进行降维,所述学习网络用于学习投射网络降维后的信息,并输出学习结果。Among them, the preset neural network model includes an input layer, a projection network and a learning network. The input layer is used to obtain input sample signals and map the input sample signals to neuron outputs. The projection network is used to reduce the dimension of the high-dimensional information output by the input layer. The learning network is used to learn the information after the dimension reduction of the projection network and output the learning results.
具体的,所述预设神经网络模型的输入层由单层神经网络组成;所述投射网络由若干层神经网络组成,所述投射网络的若干层神经网络具有特定的连接权值,且每一层神经网络都由若干神经元组成,投射网络神经元仅与前一层的有限个神经元相连接;所述学习网络由一个浅层神经网络组成,所述神经网络的第一层神经网络与所述投射网络的第一层神经网络进行连接,所述浅层神经网络是一个全连接网络,且所述浅层神经网络不同层之间的神经元通过权值连接。Specifically, the input layer of the preset neural network model is composed of a single-layer neural network; the projection network is composed of several layers of neural networks, the several layers of neural networks of the projection network have specific connection weights, and each layer of the neural network is composed of several neurons, and the projection network neurons are only connected to a limited number of neurons in the previous layer; the learning network is composed of a shallow neural network, the first layer of the neural network is connected to the first layer of the projection network, the shallow neural network is a fully connected network, and the neurons between different layers of the shallow neural network are connected through weights.
S102、基于与所述输入样本信号对应的目标任务,确定与所述预设神经网络模型每一层的连接权值。S102: Determine a connection weight with each layer of the preset neural network model based on a target task corresponding to the input sample signal.
目标任务可以是基于输入样本信号对应的目标训练样本进行训练的过程中,希望达到的任意目标任务,对此不作限定。The target task may be any target task that is desired to be achieved during the training process based on the target training sample corresponding to the input sample signal, and there is no limitation on this.
可以基于预设神经网络模型中的每一层基于前一层的输出信号进行处理得到最终的输出信号,然后基于输出信号与目标任务预期的输出信号进行比较,基于比较结果调整每一层的连接权值,以获得与目标任务对应的每一层的连接权值。Each layer in the preset neural network model can be processed based on the output signal of the previous layer to obtain the final output signal, and then the output signal is compared with the expected output signal of the target task, and the connection weights of each layer are adjusted based on the comparison result to obtain the connection weights of each layer corresponding to the target task.
在本申请实施例的一种实施方式中,所述基于与所述输入样本信号对应的目标任务,确定与所述预设神经网络模型每一层的连接权值,包括:In one implementation of the embodiment of the present application, determining the connection weights with each layer of the preset neural network model based on the target task corresponding to the input sample signal includes:
获取与所述输入样本信号对应的目标训练样本;Acquire a target training sample corresponding to the input sample signal;
将所述目标训练样本输入至所述输入层,并获得所述输入层神经元的输出;Inputting the target training sample into the input layer, and obtaining the output of the neurons in the input layer;
将所述输入层神经元的输出输入至所述投射网络,获得所述投射网络的输出信号;Inputting the output of the input layer neurons into the projection network to obtain an output signal of the projection network;
将所述投射网络的输出信号输入至所述学习网络,并获得所述学习网络的学习结果;Inputting the output signal of the projection network into the learning network, and obtaining the learning result of the learning network;
基于所述目标任务与所述学习结果的比较值,对神经元的连接权值进行调整,得到与所述预设神经网络模型每一层的连接权值。Based on the comparison value between the target task and the learning result, the connection weights of the neurons are adjusted to obtain the connection weights of each layer of the preset neural network model.
进一步地,所述方法还包括:Furthermore, the method further comprises:
在确定所述预设神经网络模型的每一层的输出信息时,确定所述神经网络模型的每一层的激活函数,以及对每一层输入的信号进行归一化处理;When determining the output information of each layer of the preset neural network model, determining the activation function of each layer of the neural network model, and performing normalization processing on the input signal of each layer;
根据归一化处理后的输入信号以及对应的激活函数,确定每一层的输出信号。The output signal of each layer is determined based on the normalized input signal and the corresponding activation function.
S103、根据所述预设神经网络模型每一层的连接权值对所述预设神经网络模型进行训练得到目标神经网络模型。S103, training the preset neural network model according to the connection weights of each layer of the preset neural network model to obtain a target neural network model.
在获得了每一层的连接权值之后,对预设神经网络模型之间的关联权重值进行调整训练,得到目标神经网络模型。该目标神经网络可以用于后续与目标任务类似的应用场景中对数据的处理。After obtaining the connection weights of each layer, the associated weight values between the preset neural network models are adjusted and trained to obtain the target neural network model. The target neural network can be used to process data in subsequent application scenarios similar to the target task.
本申请实施例提供了一种数据样本的神经网络模型构建方法,包括:获取预设神经网络模型,预设神经网络模型包括输入层、投射网络和学习网络,输入层用于获取输入样本信号,并将输入样本信号映射为神经元输出,投射网络用于对输入层输出的高维信息进行降维,学习网络用于学习投射网络降维后的信息,并输出学习结果;基于与输入样本信号对应的目标任务,确定与预设神经网络模型每一层的连接权值;根据预设神经网络模型每一层的连接权值对预设神经网络模型进行训练得到目标神经网络模型。本发明通过构建投射网络可以便于对数据样本的投射学习,实现信息在网络内部的快速传递,进而实现对样本的高效学习。The embodiment of the present application provides a method for constructing a neural network model of a data sample, including: obtaining a preset neural network model, the preset neural network model includes an input layer, a projection network and a learning network, the input layer is used to obtain an input sample signal and map the input sample signal to a neuron output, the projection network is used to reduce the dimension of the high-dimensional information output by the input layer, and the learning network is used to learn the information after the dimension reduction of the projection network, and output the learning result; based on the target task corresponding to the input sample signal, determine the connection weight with each layer of the preset neural network model; train the preset neural network model according to the connection weight of each layer of the preset neural network model to obtain a target neural network model. The present invention can facilitate the projection learning of data samples by constructing a projection network, realize the rapid transmission of information within the network, and then realize efficient learning of samples.
在本申请实施例中提供的预设神经网络模型为一种投射学习模型,包括输入层、投射网络和学习网络。The preset neural network model provided in the embodiment of the present application is a projection learning model, including an input layer, a projection network and a learning network.
输入层由单层神经网络组成,用于接受输入样本信号,并将输入样本信号影射为神经元的输出;所述的单层神经网络由若干神经元组成,该层神经元之间没有连接。The input layer is composed of a single-layer neural network, which is used to receive input sample signals and map the input sample signals to outputs of neurons; the single-layer neural network is composed of a number of neurons, and there is no connection between the neurons in this layer.
投射网络由若干层神经网络组成,用于将输入层的输出快速投射到学习网络,实现对输入层高维信息的降维;所述的投射网络的若干层神经网络,这些神经网络具有特定的连接权值,这些权值不需要在学习过程进行修改,每一层神经网络都由若干神经元组成;所述投射网络神经元,具有局部视野,也就是说投射网络神经元只跟前一层的有限个神经元相连接;所述的局部视野,该视野可以动态调整。The projection network is composed of several layers of neural networks, which are used to quickly project the output of the input layer to the learning network to achieve dimensionality reduction of the high-dimensional information of the input layer; the several layers of neural networks in the projection network have specific connection weights, which do not need to be modified during the learning process, and each layer of the neural network is composed of several neurons; the projection network neurons have a local field of view, that is, the projection network neurons are only connected to a limited number of neurons in the previous layer; the local field of view can be dynamically adjusted.
学习网络由一个浅层神经网络组成,用于学习记忆投射网络降维后的信息,并输出学习记忆的结果;所述的学习网络,其第一层神经网络与投射网络的最后一层神经网络进行连接;所述的浅层神经网络,是指该神经网络的层数一般不超过五层;所述的浅层神经网络是一个全连接网络;所述的浅层神经网络不同层之间的神经元通过权值连接,学习过程就是修改所述浅层神经网络连接权值的过程;所述的浅层神经网络,最后一层输出学习结果。The learning network consists of a shallow neural network, which is used to learn and memorize the information after dimensionality reduction of the projection network, and output the results of learning and memorization; the first layer of the neural network of the learning network is connected to the last layer of the neural network of the projection network; the shallow neural network refers to a neural network with a number of layers generally not exceeding five; the shallow neural network is a fully connected network; the neurons between different layers of the shallow neural network are connected through weights, and the learning process is the process of modifying the connection weights of the shallow neural network; the last layer of the shallow neural network outputs the learning results.
下面以对一维数据样本进行学习的过程,对本申请实施例进行说明。例如,以CWRU数据为例,CWRU数据为美国凯斯西储大学轴承数据中心测取的数据,需要说明的是,也可以基于其他的数据进行处理。The following describes the embodiment of the present application by taking the process of learning one-dimensional data samples. For example, taking CWRU data as an example, CWRU data is data measured by the Bearing Data Center of Case Western Reserve University in the United States. It should be noted that processing can also be performed based on other data.
首先将时域的CWRU轴承数据变换到频域,建立训练样本库。根据样本的长度,构建投射学习网络模型;这里选用的样本长度为2048,所构建的投射学习模型如图2所示,图中n=2048,p=d=16。这样,所设计的投射学习模型,其输入层由2048个神经元组成,投射网络由两层神经网络组成,第一投射层由2048个神经元,第二投射层有128个神经元,学习网络由三层神经网络组成,第一学习层的神经元个数与第二投射层的神经元个数相同,也就是第一学习层有128个神经元,第二学习层有32个神经元,第三学习层神经元的个数由待学习样本的类别数目来确定。First, transform the CWRU bearing data in the time domain into the frequency domain to establish a training sample library. According to the length of the sample, construct a projection learning network model; the sample length selected here is 2048, and the constructed projection learning model is shown in Figure 2, where n = 2048, p = d = 16. In this way, the designed projection learning model has an input layer composed of 2048 neurons, a projection network composed of two layers of neural networks, the first projection layer has 2048 neurons, and the second projection layer has 128 neurons. The learning network consists of three layers of neural networks, and the number of neurons in the first learning layer is the same as the number of neurons in the second projection layer, that is, the first learning layer has 128 neurons, the second learning layer has 32 neurons, and the number of neurons in the third learning layer is determined by the number of categories of the samples to be learned.
输入的样本数据经输入层影射为神经元的输出,并作为投射网络的输入信号,每个输入层神经元的输出可表示为:The input sample data is mapped to the output of the neuron through the input layer and used as the input signal of the projection network. The output of each input layer neuron can be expressed as:
yi=f(xi) (1) yi =f( xi ) (1)
式中xi是归一化的样本信号,f(·)表示激活函数,输入层采用的是激活函数为优化的Tanh函数,其表达式为第一投射网络的每个神经元接收16个输入层神经元的输出信号,其输出为:Where xi is the normalized sample signal, f(·) represents the activation function, and the input layer uses the optimized Tanh function as the activation function, which is expressed as Each neuron in the first projection network receives the output signal of 16 input layer neurons, and its output is:
式中称为作用系数,这里选用/>激活函数选用sigmoid函数,其表达式为 In the formula It is called the action coefficient, and is selected here/> The activation function uses the sigmoid function, and its expression is
第二投射网络每个神经元接收16个第一投射网络神经元的输出信号,其输出为:Each neuron in the second projection network receives the output signals of 16 neurons in the first projection network, and its output is:
式中也称为作用系数,这里也选用/>激活函数的选用与第一投射网络相同。In the formula Also called the action coefficient, also used here/> The selection of activation function is the same as that of the first projection network.
学习网络由三层神经网络组成,第一层学习网络与第二层投射网络相连接,第一层学习网络神经元接收第二层投射网络神经元的输出信号。三个学习层之间的神经元通过权值连接,学习过程就是修改这三层神经网络连接权值的过程。学习训练的结果由第三层学习网络输出。由此可知,本发明测试所构建的投射学习模型由6层网络组成。The learning network consists of three layers of neural networks. The first layer of learning network is connected to the second layer of projection network. The neurons of the first layer of learning network receive the output signals of the neurons of the second layer of projection network. The neurons between the three learning layers are connected by weights. The learning process is the process of modifying the connection weights of the three layers of neural networks. The results of learning training are output by the third layer of learning network. It can be seen that the projection learning model constructed by the test of the present invention consists of 6 layers of networks.
利用CWRU轴承数据对上述投射学习模型进行测试,并与深度神经网络模型比较,为了便于比较,将所比较的神经网络设计为5层,所比较的深度神经网络模型为2048-414-114-45-N其中N表示输出层神经元的数目。比较结果如表1所示。由该表可知本发明方法的泛化能力比原深度神经网络模型提高很多。本发明的方法在第一个和第三个测试中的识别率都达到了100%,而原方法只有85.31%和90.45%,本发明的方法将识别率至少提高了9.55%。两种方法的学习时间如表2所示,由该表可知本发明方法的训练时间大大减少,不到原神经网络模型训练时间的2%。这表明本发明的方法不仅具有很强的泛化能力,还具有高效的学习速度。The above-mentioned projection learning model was tested using CWRU bearing data and compared with the deep neural network model. For ease of comparison, the compared neural network was designed to be 5 layers, and the compared deep neural network model was 2048-414-114-45-N, where N represents the number of neurons in the output layer. The comparison results are shown in Table 1. It can be seen from the table that the generalization ability of the method of the present invention is much improved than that of the original deep neural network model. The recognition rate of the method of the present invention in the first and third tests reached 100%, while the original method was only 85.31% and 90.45%. The method of the present invention increased the recognition rate by at least 9.55%. The learning time of the two methods is shown in Table 2. It can be seen from the table that the training time of the method of the present invention is greatly reduced, less than 2% of the training time of the original neural network model. This shows that the method of the present invention not only has a strong generalization ability, but also has an efficient learning speed.
表1识别率的测试对比结果Table 1. Comparison results of recognition rate tests
在表1中,F表示故障样本,N表示正常样本。In Table 1, F represents a faulty sample, and N represents a normal sample.
表2学习时间的测试对比结果Table 2 Comparison results of learning time test
本申请实施例提供的数据样本的神经网络模型构建方法,可以实现数据样本的快速学习,节省对传统神经网络模型训练所耗费的大量时间,节约计算资源。与传统的神经网络模型相比,本发明的方法具有更好的泛化能力、更快的学习训练速度,具有广阔的工业应用前景。The method for constructing a neural network model of a data sample provided in the embodiment of the present application can realize rapid learning of data samples, save a lot of time spent on training traditional neural network models, and save computing resources. Compared with traditional neural network models, the method of the present invention has better generalization ability, faster learning and training speed, and has broad industrial application prospects.
在本申请的另一实施例中还提供了一种数据样本的神经网络模型构建装置,参见图3,包括:In another embodiment of the present application, a device for constructing a neural network model of a data sample is also provided, referring to FIG3 , comprising:
获取单元301,用于获取预设神经网络模型,所述预设神经网络模型包括输入层、投射网络和学习网络,所述输入层用于获取输入样本信号,并将输入样本信号映射为神经元输出,所述投射网络用于对输入层输出的高维信息进行降维,所述学习网络用于学习投射网络降维后的信息,并输出学习结果;An acquisition unit 301 is used to acquire a preset neural network model, wherein the preset neural network model includes an input layer, a projection network and a learning network, wherein the input layer is used to acquire an input sample signal and map the input sample signal to a neuron output, the projection network is used to reduce the dimension of the high-dimensional information output by the input layer, and the learning network is used to learn the information after the dimension reduction of the projection network and output the learning result;
确定单元302,用于基于与所述输入样本信号对应的目标任务,确定与所述预设神经网络模型每一层的连接权值;A determination unit 302, configured to determine a connection weight with each layer of the preset neural network model based on a target task corresponding to the input sample signal;
训练单元303,用于根据所述预设神经网络模型每一层的连接权值对所述预设神经网络模型进行训练得到目标神经网络模型。The training unit 303 is used to train the preset neural network model according to the connection weights of each layer of the preset neural network model to obtain a target neural network model.
可选地,还包括:Optionally, it also includes:
神经元数量确定单元,用于获取目标训练样本;A neuron number determination unit, used to obtain target training samples;
基于所述目标训练样本的长度,确定所述预设神经网络模型每一层的神经元数量。Based on the length of the target training sample, the number of neurons in each layer of the preset neural network model is determined.
可选地,所述预设神经网络模型的输入层由单层神经网络组成;所述投射网络由若干层神经网络组成,所述投射网络的若干层神经网络具有特定的连接权值,且每一层神经网络都由若干神经元组成,投射网络神经元仅与前一层的有限个神经元相连接;所述学习网络由一个浅层神经网络组成,所述神经网络的第一层神经网络与所述投射网络的第一层神经网络进行连接,所述浅层神经网络是一个全连接网络,且所述浅层神经网络不同层之间的神经元通过权值连接。Optionally, the input layer of the preset neural network model is composed of a single-layer neural network; the projection network is composed of several layers of neural networks, the several layers of neural networks of the projection network have specific connection weights, and each layer of the neural network is composed of several neurons, and the projection network neurons are only connected to a limited number of neurons in the previous layer; the learning network is composed of a shallow neural network, the first layer of the neural network is connected to the first layer of the projection network, the shallow neural network is a fully connected network, and the neurons between different layers of the shallow neural network are connected through weights.
可选地,所述确定单元具体用于:Optionally, the determining unit is specifically configured to:
获取与所述输入样本信号对应的目标训练样本;Acquire a target training sample corresponding to the input sample signal;
将所述目标训练样本输入至所述输入层,并获得所述输入层神经元的输出;Inputting the target training sample into the input layer, and obtaining the output of the neurons in the input layer;
将所述输入层神经元的输出输入至所述投射网络,获得所述投射网络的输出信号;Inputting the output of the input layer neurons into the projection network to obtain an output signal of the projection network;
将所述投射网络的输出信号输入至所述学习网络,并获得所述学习网络的学习结果;Inputting the output signal of the projection network into the learning network, and obtaining the learning result of the learning network;
基于所述目标任务与所述学习结果的比较值,对神经元的连接权值进行调整,得到与所述预设神经网络模型每一层的连接权值。Based on the comparison value between the target task and the learning result, the connection weights of the neurons are adjusted to obtain the connection weights of each layer of the preset neural network model.
可选地,所述装置还包括:Optionally, the device further comprises:
输出信号确定单元,用于在确定所述预设神经网络模型的每一层的输出信息时,确定所述神经网络模型的每一层的激活函数,以及对每一层输入的信号进行归一化处理;An output signal determination unit, used to determine the activation function of each layer of the neural network model when determining the output information of each layer of the preset neural network model, and to perform normalization processing on the input signal of each layer;
根据归一化处理后的输入信号以及对应的激活函数,确定每一层的输出信号。The output signal of each layer is determined based on the normalized input signal and the corresponding activation function.
本申请实施例提供了一种数据样本的神经网络模型构建装置,包括:获取预设神经网络模型,预设神经网络模型包括输入层、投射网络和学习网络,输入层用于获取输入样本信号,并将输入样本信号映射为神经元输出,投射网络用于对输入层输出的高维信息进行降维,学习网络用于学习投射网络降维后的信息,并输出学习结果;基于与输入样本信号对应的目标任务,确定与预设神经网络模型每一层的连接权值;根据预设神经网络模型每一层的连接权值对预设神经网络模型进行训练得到目标神经网络模型。本发明通过构建投射网络可以便于对数据样本的投射学习,实现信息在网络内部的快速传递,进而实现对样本的高效学习。The embodiment of the present application provides a neural network model construction device for data samples, including: obtaining a preset neural network model, the preset neural network model includes an input layer, a projection network and a learning network, the input layer is used to obtain an input sample signal and map the input sample signal to a neuron output, the projection network is used to reduce the dimension of the high-dimensional information output by the input layer, and the learning network is used to learn the information after the dimension reduction of the projection network, and output the learning result; based on the target task corresponding to the input sample signal, determine the connection weight with each layer of the preset neural network model; according to the connection weight of each layer of the preset neural network model, train the preset neural network model to obtain the target neural network model. The present invention can facilitate the projection learning of data samples by constructing a projection network, realize the rapid transmission of information within the network, and then realize efficient learning of samples.
基于前述实施例,本申请的实施例提供一种计算机可读存储介质,计算机可读存储介质存储有一个或者多个程序,该一个或者多个程序可被一个或者多个处理器执行,以实现如上任一项的数据样本的神经网络模型构建方法的步骤。Based on the aforementioned embodiments, embodiments of the present application provide a computer-readable storage medium, which stores one or more programs, and the one or more programs can be executed by one or more processors to implement the steps of the method for constructing a neural network model of a data sample as described in any of the above items.
本发明实施例还提供了一种电子设备,包括存储器、处理器以及存储在存储器上并可在处理器上运行的计算机程序,所述处理器执行所述程序时实现的数据样本的神经网络模型构建方法的步骤。An embodiment of the present invention also provides an electronic device, including a memory, a processor, and a computer program stored in the memory and executable on the processor, wherein the processor implements the steps of a method for constructing a neural network model of a data sample when executing the program.
需要说明的是,上述处理器或CPU可以为特定用途集成电路(ApplicationSpecific Integrated Circuit,ASIC)、数字信号处理器(Digital Signal Processor,DSP)、数字信号处理装置(Digital Signal Processing Device,DSPD)、可编程逻辑装置(Programmable Logic Device,PLD)、现场可编程门阵列(Field Programmable GateArray,FPGA)、中央处理器(Central Processing Unit,CPU)、控制器、微控制器、微处理器中的至少一种。可以理解地,实现上述处理器功能的电子器件还可以为其它,本申请实施例不作具体限定。It should be noted that the processor or CPU may be at least one of an Application Specific Integrated Circuit (ASIC), a Digital Signal Processor (DSP), a Digital Signal Processing Device (DSPD), a Programmable Logic Device (PLD), a Field Programmable Gate Array (FPGA), a Central Processing Unit (CPU), a controller, a microcontroller, and a microprocessor. It is understandable that the electronic device that implements the above-mentioned processor function may also be other, and the embodiments of the present application are not specifically limited.
需要说明的是,上述计算机存储介质/存储器可以是只读存储器(Read OnlyMemory,ROM)、可编程只读存储器(Programmable Read-Only Memory,PROM)、可擦除可编程只读存储器(Erasable Programmable Read-Only Memory,EPROM)、电可擦除可编程只读存储器(Electrically Erasable Programmable Read-Only Memory,EEPROM)、磁性随机存取存储器(Ferromagnetic Random Access Memory,FRAM)、快闪存储器(Flash Memory)、磁表面存储器、光盘、或只读光盘(Compact Disc Read-Only Memory,CD-ROM)等存储器;也可以是包括上述存储器之一或任意组合的各种终端,如移动电话、计算机、平板设备、个人数字助理等。It should be noted that the above-mentioned computer storage medium/memory can be a read-only memory (ROM), a programmable read-only memory (PROM), an erasable programmable read-only memory (EPROM), an electrically erasable programmable read-only memory (EEPROM), a magnetic random access memory (FRAM), a flash memory (Flash Memory), a magnetic surface memory, an optical disc, or a compact disc read-only memory (CD-ROM) and other memories; it can also be various terminals including one or any combination of the above-mentioned memories, such as mobile phones, computers, tablet devices, personal digital assistants, etc.
在本申请所提供的几个实施例中,应该理解到,所揭露的设备和方法,可以通过其它的方式实现。以上所描述的设备实施例仅仅是示意性的,例如,所述单元的划分,仅仅为一种逻辑功能划分,实际实现时可以有另外的划分方式,如:多个单元或组件可以结合,或可以集成到另一个系统,或一些特征可以忽略,或不执行。另外,所显示或讨论的各组成部分相互之间的耦合、或直接耦合、或通信连接可以是通过一些接口,设备或单元的间接耦合或通信连接,可以是电性的、机械的或其它形式的。In the several embodiments provided in the present application, it should be understood that the disclosed devices and methods can be implemented in other ways. The device embodiments described above are only schematic. For example, the division of the units is only a logical function division. There may be other division methods in actual implementation, such as: multiple units or components can be combined, or can be integrated into another system, or some features can be ignored, or not executed. In addition, the coupling, direct coupling, or communication connection between the components shown or discussed can be through some interfaces, and the indirect coupling or communication connection of the devices or units can be electrical, mechanical or other forms.
上述作为分离部件说明的单元可以是、或也可以不是物理上分开的,作为单元显示的部件可以是、或也可以不是物理单元,即可以位于一个地方,也可以分布到多个网络单元上;可以根据实际的需要选择其中的部分或全部单元来实现本实施例方案的目的。The units described above as separate components may or may not be physically separated, and the components displayed as units may or may not be physical units, that is, they may be located in one place or distributed on multiple network units; some or all of the units may be selected according to actual needs to achieve the purpose of the present embodiment.
另外,在本申请各实施例中的各功能单元可以全部集成在一个处理模块中,也可以是各单元分别单独作为一个单元,也可以两个或两个以上单元集成在一个单元中;上述集成的单元既可以采用硬件的形式实现,也可以采用硬件加软件功能单元的形式实现。本领域普通技术人员可以理解:实现上述方法实施例的全部或部分步骤可以通过程序指令相关的硬件来完成,前述的程序可以存储于一计算机可读取存储介质中,该程序在执行时,执行包括上述方法实施例的步骤;而前述的存储介质包括:移动存储设备、只读存储器(Read-Only Memory,ROM)、随机存取存储器(Random Access Memory,RAM)、磁碟或者光盘等各种可以存储程序代码的介质。In addition, all functional units in the embodiments of the present application can be integrated into one processing module, or each unit can be a separate unit, or two or more units can be integrated into one unit; the above integrated unit can be implemented in the form of hardware or in the form of hardware plus software functional units. A person of ordinary skill in the art can understand that all or part of the steps of implementing the above method embodiments can be completed by hardware related to program instructions, and the aforementioned program can be stored in a computer-readable storage medium, which, when executed, executes the steps of the above method embodiments; and the aforementioned storage medium includes: mobile storage devices, read-only memory (ROM), random access memory (RAM), disks or optical disks, and other media that can store program codes.
本申请所提供的几个方法实施例中所揭露的方法,在不冲突的情况下可以任意组合,得到新的方法实施例。The methods disclosed in several method embodiments provided in this application can be arbitrarily combined without conflict to obtain new method embodiments.
本申请所提供的几个产品实施例中所揭露的特征,在不冲突的情况下可以任意组合,得到新的产品实施例。The features disclosed in several product embodiments provided in this application can be arbitrarily combined without conflict to obtain new product embodiments.
本申请所提供的几个方法或设备实施例中所揭露的特征,在不冲突的情况下可以任意组合,得到新的方法实施例或设备实施例。The features disclosed in several method or device embodiments provided in this application can be arbitrarily combined without conflict to obtain new method embodiments or device embodiments.
以上所述,仅为本申请的具体实施方式,但本申请的保护范围并不局限于此,任何熟悉本技术领域的技术人员在本申请揭露的技术范围内,可轻易想到变化或替换,都应涵盖在本申请的保护范围之内。因此,本申请的保护范围应以所述权利要求的保护范围为准。The above is only a specific implementation of the present application, but the protection scope of the present application is not limited thereto. Any person skilled in the art who is familiar with the present technical field can easily think of changes or substitutions within the technical scope disclosed in the present application, which should be included in the protection scope of the present application. Therefore, the protection scope of the present application should be based on the protection scope of the claims.
本说明书中各个实施例采用递进的方式描述,每个实施例重点说明的都是与其他实施例的不同之处,各个实施例之间相同相似部分互相参见即可。对于实施例公开的装置而言,由于其与实施例公开的方法相对应,所以描述的比较简单,相关之处参见方法部分说明即可。In this specification, each embodiment is described in a progressive manner, and each embodiment focuses on the differences from other embodiments. The same or similar parts between the embodiments can be referred to each other. For the device disclosed in the embodiment, since it corresponds to the method disclosed in the embodiment, the description is relatively simple, and the relevant parts can be referred to the method part.
对所公开的实施例的上述说明,使本领域专业技术人员能够实现或使用本发明。对这些实施例的多种修改对本领域的专业技术人员来说将是显而易见的,本文中所定义的一般原理可以在不脱离本发明的精神或范围的情况下,在其它实施例中实现。因此,本发明将不会被限制于本文所示的这些实施例,而是要符合与本文所公开的原理和新颖特点相一致的最宽的范围。The above description of the disclosed embodiments enables those skilled in the art to implement or use the present invention. Various modifications to these embodiments will be apparent to those skilled in the art, and the general principles defined herein may be implemented in other embodiments without departing from the spirit or scope of the present invention. Therefore, the present invention will not be limited to the embodiments shown herein, but rather to the widest scope consistent with the principles and novel features disclosed herein.
Claims (2)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202210085874.6A CN114492789B (en) | 2022-01-25 | 2022-01-25 | A method and device for constructing a neural network model of data samples |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202210085874.6A CN114492789B (en) | 2022-01-25 | 2022-01-25 | A method and device for constructing a neural network model of data samples |
Publications (2)
Publication Number | Publication Date |
---|---|
CN114492789A CN114492789A (en) | 2022-05-13 |
CN114492789B true CN114492789B (en) | 2024-05-14 |
Family
ID=81475291
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202210085874.6A Active CN114492789B (en) | 2022-01-25 | 2022-01-25 | A method and device for constructing a neural network model of data samples |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN114492789B (en) |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106096727A (en) * | 2016-06-02 | 2016-11-09 | 腾讯科技(深圳)有限公司 | A kind of network model based on machine learning building method and device |
CN106295799A (en) * | 2015-05-12 | 2017-01-04 | 核工业北京地质研究院 | A kind of implementation method of degree of depth study multilayer neural network |
CN107301453A (en) * | 2016-04-15 | 2017-10-27 | 北京中科寒武纪科技有限公司 | The artificial neural network forward operation apparatus and method for supporting discrete data to represent |
CN110197260A (en) * | 2019-06-06 | 2019-09-03 | 百度在线网络技术(北京)有限公司 | A kind of data processing method and device |
CN111626994A (en) * | 2020-05-18 | 2020-09-04 | 江苏远望仪器集团有限公司 | Equipment fault defect diagnosis method based on improved U-Net neural network |
-
2022
- 2022-01-25 CN CN202210085874.6A patent/CN114492789B/en active Active
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106295799A (en) * | 2015-05-12 | 2017-01-04 | 核工业北京地质研究院 | A kind of implementation method of degree of depth study multilayer neural network |
CN107301453A (en) * | 2016-04-15 | 2017-10-27 | 北京中科寒武纪科技有限公司 | The artificial neural network forward operation apparatus and method for supporting discrete data to represent |
CN106096727A (en) * | 2016-06-02 | 2016-11-09 | 腾讯科技(深圳)有限公司 | A kind of network model based on machine learning building method and device |
CN110197260A (en) * | 2019-06-06 | 2019-09-03 | 百度在线网络技术(北京)有限公司 | A kind of data processing method and device |
CN111626994A (en) * | 2020-05-18 | 2020-09-04 | 江苏远望仪器集团有限公司 | Equipment fault defect diagnosis method based on improved U-Net neural network |
Non-Patent Citations (1)
Title |
---|
" 大功率盘型激光焊熔池形态在线识别及焊缝成形预测模型";张艳喜;《中国博士学位论文全文数据库 信息科技辑》;20150315(第03期);第I138-26页 * |
Also Published As
Publication number | Publication date |
---|---|
CN114492789A (en) | 2022-05-13 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110991633B (en) | Residual error neural network model based on memristor network and application method thereof | |
CN111259142B (en) | Specific target emotion classification method based on attention coding and graph convolution network | |
CN111951805B (en) | A text data processing method and device | |
CN113065636B (en) | Pruning processing method, data processing method and equipment for convolutional neural network | |
CN111357051B (en) | Speech emotion recognition method, intelligent device and computer readable storage medium | |
CN114724549B (en) | Intelligent identification method, device, equipment and storage medium for environmental noise | |
CN107705806A (en) | A kind of method for carrying out speech emotion recognition using spectrogram and deep convolutional neural networks | |
CN108846077A (en) | Semantic matching method, device, medium and electronic equipment for question and answer text | |
TWI690862B (en) | Local learning system in artificial intelligence device | |
CN108171328B (en) | A neural network processor and a convolution operation method performed by the same | |
CN113761934B (en) | Word vector representation method based on self-attention mechanism and self-attention model | |
CN114925320B (en) | Data processing method and related device | |
CN112036461B (en) | Handwriting digital image recognition method, device, equipment and computer storage medium | |
CN110188878A (en) | Neural network search method and device | |
Castro-Garcia et al. | Emergent multilingual language acquisition using developmental networks | |
CN118038238A (en) | Visual question-answering method and device, electronic equipment and storage medium | |
CN116047330A (en) | Battery performance prediction method, model training method and related devices | |
CN113326383B (en) | Short text entity linking method, device, computing equipment and storage medium | |
CN114492789B (en) | A method and device for constructing a neural network model of data samples | |
CN112446472A (en) | Method, apparatus and related product for processing data | |
CN112132281B (en) | Model training method, device, server and medium based on artificial intelligence | |
CN118797058A (en) | Data processing method, model training method and device | |
WO2024239927A1 (en) | Model training method and related device | |
CN112183725B (en) | Method of providing neural network, computing device, and computer-readable storage medium | |
CN118052257A (en) | Acoustic emission signal enhancement method and device based on generation countermeasure network |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |