CN108846349A - A kind of face identification method based on dynamic Spiking neural network - Google Patents
A kind of face identification method based on dynamic Spiking neural network Download PDFInfo
- Publication number
- CN108846349A CN108846349A CN201810585279.2A CN201810585279A CN108846349A CN 108846349 A CN108846349 A CN 108846349A CN 201810585279 A CN201810585279 A CN 201810585279A CN 108846349 A CN108846349 A CN 108846349A
- Authority
- CN
- China
- Prior art keywords
- neural network
- spiking neural
- weight
- image
- face
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000012421 spiking Methods 0.000 title claims abstract description 43
- 238000013528 artificial neural network Methods 0.000 title claims abstract description 42
- 238000000034 method Methods 0.000 title claims abstract description 25
- 210000002569 neuron Anatomy 0.000 claims abstract description 33
- 238000000605 extraction Methods 0.000 claims abstract description 6
- 238000012545 processing Methods 0.000 claims abstract description 4
- 230000006870 function Effects 0.000 claims description 5
- 230000003247 decreasing effect Effects 0.000 claims description 3
- 239000000523 sample Substances 0.000 description 7
- 239000013598 vector Substances 0.000 description 6
- 238000004422 calculation algorithm Methods 0.000 description 4
- 238000004364 calculation method Methods 0.000 description 4
- 239000011159 matrix material Substances 0.000 description 4
- 241000282414 Homo sapiens Species 0.000 description 3
- 238000010586 diagram Methods 0.000 description 3
- 230000000946 synaptic effect Effects 0.000 description 3
- 238000012549 training Methods 0.000 description 3
- 238000011160 research Methods 0.000 description 2
- 238000013473 artificial intelligence Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000013529 biological neural network Methods 0.000 description 1
- 230000001419 dependent effect Effects 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 239000000284 extract Substances 0.000 description 1
- 210000002364 input neuron Anatomy 0.000 description 1
- 210000004205 output neuron Anatomy 0.000 description 1
- 238000003909 pattern recognition Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/168—Feature extraction; Face representation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/213—Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods
- G06F18/2135—Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods based on approximation criteria, e.g. principal component analysis
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/049—Temporal neural networks, e.g. delay elements, oscillating neurons or pulsed inputs
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/082—Learning methods modifying the architecture, e.g. adding, deleting or silencing nodes or connections
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/172—Classification, e.g. identification
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Health & Medical Sciences (AREA)
- General Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- General Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- Oral & Maxillofacial Surgery (AREA)
- Evolutionary Computation (AREA)
- General Engineering & Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Molecular Biology (AREA)
- Human Computer Interaction (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- Multimedia (AREA)
- Biomedical Technology (AREA)
- Computing Systems (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Evolutionary Biology (AREA)
- Image Analysis (AREA)
- Image Processing (AREA)
Abstract
本发明公开了一种基于动态Spiking神经网络的人脸识别方法,涉及图像处理技术领域,本发明包括以下步骤:S1、将人脸图像转换为灰度像素,获得灰度图;S2、对灰度图进行特征提取,得到灰度图区域特征关联的低维特征;S3、将低维特征的特征强度转换为脉冲时间序列;S4、多张人脸图像对Spiking神经网络进行训练,得到动态Spiking神经网络;S5、将待识别人脸图像依次经过S1‑S3处理后得到的脉冲时间序列输入至S4得到的动态Spiking神经网络中,根据调整权值与动态Spiking神经网络中已有的神经元的权值进行比较,权值最接近的神经元的类别便是待识别人脸图像的类别,本发明创造性地使用了动态Spiking神经网络,相较于传统的Spiking图像识别方法,显著提高了识别效率。
The invention discloses a face recognition method based on a dynamic Spiking neural network, and relates to the technical field of image processing. The invention comprises the following steps: S1, converting a face image into grayscale pixels to obtain a grayscale image; S2, grayscale feature extraction from the degree map to obtain the low-dimensional features associated with the regional features of the gray-scale image; S3, convert the feature intensity of the low-dimensional features into a pulse time series; S4, train the Spiking neural network with multiple face images, and obtain the dynamic Spiking Neural network; S5, the pulse time series obtained after the face image to be recognized is processed by S1-S3 is input into the dynamic Spiking neural network obtained by S4, according to the adjustment weight and the existing neuron in the dynamic Spiking neural network Weights are compared, and the category of the neuron with the closest weight is the category of the face image to be recognized. The present invention creatively uses the dynamic Spiking neural network, which significantly improves the recognition efficiency compared with the traditional Spiking image recognition method. .
Description
技术领域technical field
本发明涉及图像处理技术领域,更具体的是涉及一种基于动态Spiking神经网络的人脸识别方法。The invention relates to the technical field of image processing, and more specifically relates to a face recognition method based on a dynamic Spiking neural network.
背景技术Background technique
人脸识别是在图形学、计算机科学与技术和模式识别等相关学科领域的基础上出现的一个新的研究方向,通过提取人类特征的方式对人类进行判别时,把人脸作为依据是最简洁的方式。虽然人脸识别技术从出现到现在只有几十年,但已经成为当下比较热门的研究课题之一。尤其是在人工智能时代,随着科学技术的快速发展和人们对于安全又智能生活的追求,对人脸识别技术的有效性、方便性、快捷性等方面的要求越来越高。Face recognition is a new research direction based on graphics, computer science and technology, pattern recognition and other related disciplines. When distinguishing human beings by extracting human characteristics, it is the most concise way to use human faces as the basis. The way. Although face recognition technology has only been around for decades, it has become one of the more popular research topics at present. Especially in the era of artificial intelligence, with the rapid development of science and technology and people's pursuit of a safe and intelligent life, the requirements for the effectiveness, convenience, and speed of face recognition technology are getting higher and higher.
Spiking神经网络作为第三代神经网络,关注脉冲发放的时间,适合在芯片上实现。但是绝大部分监督或非监督用于脉冲神经网络的学习算法均具有固定结构,其中隐藏层和输出层的大小必须事先指定,并且以离线批处理模式训练,因此,这些方法只能应用于类或簇的数量已知的情况下;此外,这些方法不能应用于数据连续改变的问题,因为它们将需要重新训练旧的和新的数据样本。然而,生物神经网络因其连续学习和增量学习的能力而众所周知,这使他们能够持续适应不断变化的非稳定环境,因此,为了允许SNN(SpikingNeuron Networks,脉冲神经网络)与连续变化的环境交互,有必要使其结构和权重动态地适应新数据,此外,当学习新信息时,应避免灾难性干扰或遗忘。As a third-generation neural network, the Spiking neural network focuses on the timing of pulses and is suitable for implementation on a chip. However, most supervised or unsupervised learning algorithms for spiking neural networks have a fixed structure, in which the size of the hidden layer and output layer must be specified in advance, and are trained in offline batch mode. Therefore, these methods can only be applied to class or the number of clusters is known; moreover, these methods cannot be applied to problems with continuously changing data, since they will require retraining on old and new data samples. However, biological neural networks are well known for their ability to learn continuously and incrementally, which enables them to continuously adapt to changing non-stationary environments, therefore, in order to allow SNNs (SpikingNeuron Networks, Spiking Neural Networks) to interact with continuously changing environments , it is necessary to adapt its structure and weights dynamically to new data, moreover, when learning new information, catastrophic interference or forgetting should be avoided.
发明内容Contents of the invention
本发明的目的在于:为了解决现有技术对人脸识别高度依赖于样本的类别的问题,本发明提供一种基于动态Spiking神经网络的人脸识别方法。The purpose of the present invention is: in order to solve the prior art problem that face recognition is highly dependent on the category of samples, the present invention provides a face recognition method based on dynamic Spiking neural network.
本发明为了实现上述目的具体采用以下技术方案:The present invention specifically adopts the following technical solutions in order to achieve the above object:
一种基于动态Spiking神经网络的人脸识别方法,包括以下步骤:A face recognition method based on dynamic Spiking neural network, comprising the following steps:
S1、将人脸图像转换为灰度像素,获得灰度图;S1, converting the face image into grayscale pixels to obtain a grayscale image;
S2、对灰度图进行特征提取,得到灰度图区域特征关联的低维特征;S2. Perform feature extraction on the grayscale image to obtain low-dimensional features associated with regional features of the grayscale image;
S3、将低维特征的特征强度转换为脉冲时间序列;S3, converting the feature intensity of the low-dimensional feature into a pulse time series;
S4、使用多张人脸图像对Spiking神经网络进行训练S4, using multiple face images to train the Spiking neural network
将多张人脸图像依次经过S1-S3处理后,分别得到相对应的脉冲时间序列,根据每个脉冲时间序列中脉冲的精确时间来调整对应的初始权值,得到调整权值,然后进行权值学习,根据每张人脸图像的标签、调整权值以及Spiking神经网络中已有的神经元来共同判断是否增加新的神经元,当所有的人脸图像均输入到Spiking神经网络后,得到稳定的动态Spiking神经网络;After multiple face images are processed through S1-S3 in sequence, the corresponding pulse time series are respectively obtained, and the corresponding initial weights are adjusted according to the precise time of the pulses in each pulse time series, and the adjusted weights are obtained, and then the weights are carried out. Value learning, according to the label of each face image, adjusting the weight and the existing neurons in the Spiking neural network to jointly judge whether to add new neurons, when all the face images are input into the Spiking neural network, get Stable dynamic Spiking neural network;
S5、对待识别人脸图像进行识别S5. Recognize the face image to be recognized
将待识别人脸图像依次经过S1-S3处理后得到的脉冲时间序列输入至S4得到的动态Spiking神经网络中,根据输入的脉冲时间序列中脉冲的精确时间来调整初始权值,得到相应的调整权值,将该调整权值与动态Spiking神经网络中已有的神经元的权值进行比较,找出与该调整权值最接近的权值的神经元,该神经元的类别便是待识别人脸图像的类别。Input the pulse time series obtained after processing the face image to be recognized through S1-S3 into the dynamic Spiking neural network obtained in S4, adjust the initial weight according to the precise time of the pulse in the input pulse time series, and get the corresponding adjustment Weight, compare the adjusted weight with the weight of the existing neurons in the dynamic Spiking neural network, find out the neuron with the closest weight to the adjusted weight, and the category of the neuron is to be identified The category of face images.
进一步的,所述S2中,利用PCA降维的方法对灰度图进行特征提取。Further, in the above S2, feature extraction is performed on the grayscale image using the PCA dimensionality reduction method.
进一步的,所述S3中,利用高斯编码的方法将低维特征的特征强度转换为脉冲时间序列。Further, in S3, the Gaussian encoding method is used to convert the feature intensity of the low-dimensional feature into a pulse time series.
进一步的,所述S4中基于精确时间的权值学习是用一个递减的函数来实现的,即最先到达的脉冲含有最多信息。Further, the precise time-based weight learning in S4 is realized by using a decreasing function, that is, the first arriving pulse contains the most information.
本发明的有益效果如下:The beneficial effects of the present invention are as follows:
1、本发明基于PCA降维的思想对待识别人脸图像的特征进行提取,然后再使用高斯编码的方法对所提取的特征进行编码,在编码后使用动态Spiking神经网络学习算法对编码后的特征序列进行学习,根据脉冲的精确时间调整输入的脉冲时间序列的初始突触权值,根据权值之间的相似性来动态增加或更新神经元,最终得到识别输出结果,创造性地使用了动态Spiking神经网络,相较于传统的Spiking图像识别方法,显著提高了识别效率。1. The present invention extracts the features of the face image to be recognized based on the idea of PCA dimensionality reduction, and then uses the Gaussian encoding method to encode the extracted features, and uses the dynamic Spiking neural network learning algorithm to encode the encoded features after encoding. Sequential learning, adjust the initial synaptic weights of the input pulse time series according to the precise time of the pulse, dynamically increase or update neurons according to the similarity between the weights, and finally obtain the recognition output result, creatively using dynamic Spiking Compared with the traditional Spiking image recognition method, the neural network has significantly improved the recognition efficiency.
2、本发明根据脉冲的精确时间调整输入的脉冲时间序列的初始突触权值,提高了人脸图像识别的准确率。2. The present invention adjusts the initial synaptic weights of the input pulse time series according to the precise time of the pulse, which improves the accuracy of face image recognition.
3、本发明采用相似度来比较权值,进一步提高了人脸图像识别的效率。3. The present invention uses similarity to compare weights, which further improves the efficiency of face image recognition.
4、本发明使用动态Spiking神经网络结构,根据输入的样本动态地增加神经元,对样本的标签不需要提前统计,降低了人脸识别方法对样本的依赖性。4. The present invention uses a dynamic Spiking neural network structure to dynamically increase neurons according to the input samples, and does not need to make statistics on the labels of the samples in advance, which reduces the dependence of the face recognition method on the samples.
附图说明Description of drawings
图1是本发明的识别流程示意图。Fig. 1 is a schematic diagram of the identification process of the present invention.
图2是本发明的整体网络结构示意图。Fig. 2 is a schematic diagram of the overall network structure of the present invention.
图3是本发明动态Spiking神经网络的训练流程图。Fig. 3 is the training flowchart of the dynamic Spiking neural network of the present invention.
图4是ORL数据集样例图。Figure 4 is a sample diagram of the ORL dataset.
具体实施方式Detailed ways
为了本技术领域的人员更好的理解本发明,下面结合附图和以下实施例对本发明作进一步详细描述。In order for those skilled in the art to better understand the present invention, the present invention will be further described in detail below in conjunction with the accompanying drawings and the following embodiments.
实施例1Example 1
如图1至图4所示,本实施例提供一种基于动态Spiking神经网络的人脸识别方法,包括以下步骤:As shown in Figures 1 to 4, the present embodiment provides a face recognition method based on a dynamic Spiking neural network, comprising the following steps:
S1、将人脸图像转换为灰度像素,获得灰度图;S1, converting the face image into grayscale pixels to obtain a grayscale image;
S2、采用PCA降维的方法对灰度图进行特征提取,得到灰度图区域特征关联的低维特征,包括以下步骤:S2. Using the method of PCA dimensionality reduction to perform feature extraction on the grayscale image to obtain low-dimensional features associated with regional features of the grayscale image, including the following steps:
S2.1、把一个N维列向量表示成xi,其中i=1,2,...,L;S2.1. Express an N-dimensional column vector as x i , where i=1,2,...,L;
S2.2、计算L个样本向量的平均值xi;S2.2. Calculate the average value x i of the L sample vectors;
S2.3、计算协方差矩阵;S2.3, calculate the covariance matrix;
S2.4、非对角线元素是每个列向量元素之间的相关性,计算公式为:S2.4. Off-diagonal elements are the correlation between elements of each column vector, and the calculation formula is:
S2.5、对协方差矩阵做特征分解,得到若干特征值;S2.5. Perform eigendecomposition on the covariance matrix to obtain several eigenvalues;
S2.6、对若干特征值从大到小进行排序,取前r个特征值对应的特征向量得到投影矩阵;S2.6. Sorting several eigenvalues from large to small, and taking the eigenvectors corresponding to the first r eigenvalues to obtain a projection matrix;
S2.7、用投影矩阵计算得到新的低维向量,即低维特征;S2.7. Use the projection matrix to calculate a new low-dimensional vector, that is, a low-dimensional feature;
S3、采用高斯编码的方法将第一层神经元的特征强度转换为脉冲时间序列,具体为:S3. Using the method of Gaussian encoding to convert the characteristic intensity of the neurons in the first layer into a pulse time series, specifically:
假设S2.7得到的低维向量具有有m维特征(x1,x2,...,xm),经过群编码后得到m*p个脉冲时间,首先计算第i个特征在第j个接受域的均值和标准差具体公式为:Assuming that the low-dimensional vector obtained in S2.7 has m-dimensional features (x 1 , x 2 ,...,x m ), after group coding, m*p pulse times are obtained, and the i-th feature is first calculated at j mean of receptive fields and standard deviation The specific formula is:
其中,和分别是第i个特征的最小值和最大值;β是一个参数,通过影响标准差来影响高斯接受域的覆盖范围,通过均值和标准差得到高斯函数计算公式为:in, and are the minimum and maximum values of the i-th feature, respectively; β is a parameter that affects the coverage of the Gaussian receptive field by affecting the standard deviation, and by the mean and standard deviation get the Gaussian function The calculation formula is:
利用高斯函数的结果,得到脉冲时间序列即得到每一个输入神经元的脉冲时间,计算公式为:Using the Gaussian function As a result, the pulse time series is obtained That is, the pulse time of each input neuron is obtained, and the calculation formula is:
S4、使用多张人脸图像对Spiking神经网络进行训练,S4, using multiple face images to train the Spiking neural network,
在ORL人脸数据集中选取多张人脸图像依次经过S1-S3处理后,分别得到相对应的脉冲时间序列,根据每个脉冲时间序列中脉冲的精确时间来调整对应的初始权值,使用多种策略进行分类,当所有的人脸图像均输入到Spiking神经网络后,得到稳定的动态Spiking神经网络;Select multiple face images in the ORL face dataset and process them sequentially through S1-S3 to obtain the corresponding pulse time series, adjust the corresponding initial weights according to the precise time of the pulse in each pulse time series, and use multiple Classify with this strategy, when all face images are input to the Spiking neural network, a stable dynamic Spiking neural network is obtained;
S4.1、初始权值微调策略:S4.1. Initial weight fine-tuning strategy:
一个独立的输出层神经元代表了一个输入模式,对于每一个输入样本,输出层将会创建一个新的神经元,并且该神经元与编码层的权值学习是用一个递减的函数来表示的,计算公式为:An independent output layer neuron represents an input pattern. For each input sample, the output layer will create a new neuron, and the weight learning of the neuron and the encoding layer is represented by a decreasing function. , the calculation formula is:
wij=w0+γexp(-ti/τ)w ij =w 0 +γexp(-t i /τ)
其中,wij是输入层神经元i和输出层神经元j之间的突触权重,w0是初始权值,ti是输入层脉冲的精确时间,从而替代脉冲的次序,τ是时间常数;where w ij is the synaptic weight between input layer neuron i and output layer neuron j, w 0 is the initial weight, t i is the precise time of input layer pulses, thus replacing the order of pulses, and τ is the time constant ;
S4.2、神经元调整策略:S4.2. Neuron adjustment strategy:
在训练阶段,训练输入样本将逐个呈现给Spiking神经网络,然后将存储在Spiking神经网络中的信息与输入样本所携带的信息进行比较,这些信息表示输入特征与样本类别标签之间的函数关系,该算法使每个样本选择一种学习策略:In the training phase, the training input samples will be presented to the Spiking neural network one by one, and then the information stored in the Spiking neural network will be compared with the information carried by the input samples, which represent the functional relationship between the input features and the sample category labels, The algorithm makes each sample choose a learning strategy:
S4.2.1、添加神经元策略:当网络中的信息与输入样本所携带的信息之间的差异相对较大时,输出层中的新神经元被添加以记录新信息;S4.2.1. Add neuron strategy: When the difference between the information in the network and the information carried by the input samples is relatively large, new neurons in the output layer are added to record new information;
S4.2.2、合并神经元策略:当输入样本的信息与存储在网络中已有神经元中的信息充分相似时,新神经元将与最相似的神经元合并,训练的输出神经元表示时空尖峰模式的聚类,根据神经元之间的相似性合并神经元并预测类别标签使得实现足够快的学习成为可能,同时为权向量提供精确的时间使得算法足够有效,合并的神经元需要更新权值,将进来的权值添加到原有权值的基础上,得到新的权值;S4.2.2. Merge neuron strategy: When the information of the input sample is sufficiently similar to the information stored in the existing neurons in the network, the new neuron will be merged with the most similar neuron, and the trained output neuron represents a spatiotemporal spike Clustering of patterns, merging neurons according to the similarity between neurons and predicting class labels makes it possible to achieve fast enough learning, and at the same time provide precise time for the weight vector to make the algorithm efficient enough, the merged neurons need to update the weights , add the incoming weight to the original weight to get a new weight;
S5、对待识别人脸图像进行识别S5. Recognize the face image to be recognized
在ORL人脸数据集中选选择一张人脸图像作为待识别人脸图像,将待识别人脸图像依次经过S1-S3处理后得到的脉冲时间序列输入至S4得到的动态Spiking神经网络中,根据输入的脉冲时间序列中脉冲的精确时间来调整初始权值,得到相应的调整权值,将该调整权值与动态Spiking神经网络中已有的神经元的权值进行比较,找出与该调整权值最接近的权值的神经元,该神经元的类别便是待识别人脸图像的类别。Select a face image in the ORL face data set as the face image to be recognized, and input the pulse time series obtained after the face image to be recognized through S1-S3 to the dynamic Spiking neural network obtained in S4, according to The precise time of the pulse in the input pulse time series is used to adjust the initial weight, and the corresponding adjusted weight is obtained, and the adjusted weight is compared with the weight of the existing neuron in the dynamic Spiking neural network, and the adjustment is found. The neuron with the closest weight, the category of the neuron is the category of the face image to be recognized.
以上所述,仅为本发明的较佳实施例,并不用以限制本发明,本发明的专利保护范围以权利要求书为准,凡是运用本发明的说明书及附图内容所作的等同结构变化,同理均应包含在本发明的保护范围内。The above is only a preferred embodiment of the present invention, and is not intended to limit the present invention. The scope of patent protection of the present invention is subject to the claims. Any equivalent structural changes made by using the description and accompanying drawings of the present invention, All should be included in the protection scope of the present invention in the same way.
Claims (4)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810585279.2A CN108846349A (en) | 2018-06-08 | 2018-06-08 | A kind of face identification method based on dynamic Spiking neural network |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810585279.2A CN108846349A (en) | 2018-06-08 | 2018-06-08 | A kind of face identification method based on dynamic Spiking neural network |
Publications (1)
Publication Number | Publication Date |
---|---|
CN108846349A true CN108846349A (en) | 2018-11-20 |
Family
ID=64210289
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201810585279.2A Pending CN108846349A (en) | 2018-06-08 | 2018-06-08 | A kind of face identification method based on dynamic Spiking neural network |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN108846349A (en) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109871940A (en) * | 2019-01-31 | 2019-06-11 | 清华大学 | A Multilayer Training Algorithm for Spiking Neural Networks |
CN110135498A (en) * | 2019-05-17 | 2019-08-16 | 电子科技大学 | Image identification method based on deep evolution neural network |
CN110674928A (en) * | 2019-09-18 | 2020-01-10 | 电子科技大学 | An online learning method integrating artificial neural network and neuromorphic computing |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103279958A (en) * | 2013-05-31 | 2013-09-04 | 电子科技大学 | Image segmentation method based on Spiking neural network |
CN104662526A (en) * | 2012-07-27 | 2015-05-27 | 高通技术公司 | Apparatus and methods for efficient updates in spiking neuron networks |
CN104685516A (en) * | 2012-08-17 | 2015-06-03 | 高通技术公司 | Apparatus and methods for spiking neuron network learning |
CN105404902A (en) * | 2015-10-27 | 2016-03-16 | 清华大学 | Impulsive neural network-based image feature describing and memorizing method |
CN107194426A (en) * | 2017-05-23 | 2017-09-22 | 电子科技大学 | A kind of image-recognizing method based on Spiking neutral nets |
-
2018
- 2018-06-08 CN CN201810585279.2A patent/CN108846349A/en active Pending
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104662526A (en) * | 2012-07-27 | 2015-05-27 | 高通技术公司 | Apparatus and methods for efficient updates in spiking neuron networks |
CN104685516A (en) * | 2012-08-17 | 2015-06-03 | 高通技术公司 | Apparatus and methods for spiking neuron network learning |
CN103279958A (en) * | 2013-05-31 | 2013-09-04 | 电子科技大学 | Image segmentation method based on Spiking neural network |
CN105404902A (en) * | 2015-10-27 | 2016-03-16 | 清华大学 | Impulsive neural network-based image feature describing and memorizing method |
CN107194426A (en) * | 2017-05-23 | 2017-09-22 | 电子科技大学 | A kind of image-recognizing method based on Spiking neutral nets |
Non-Patent Citations (1)
Title |
---|
LIN ZUO ET AL.: ""A Fast Precise-Spike and Weight-Comparison Based Learning Approach for Evolving Spiking Neural Networks", 《ICONIP 2017: NEURAL INFORMATION PROCESSING》 * |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109871940A (en) * | 2019-01-31 | 2019-06-11 | 清华大学 | A Multilayer Training Algorithm for Spiking Neural Networks |
CN109871940B (en) * | 2019-01-31 | 2021-07-27 | 清华大学 | A Multilayer Training Algorithm for Spiking Neural Networks |
CN110135498A (en) * | 2019-05-17 | 2019-08-16 | 电子科技大学 | Image identification method based on deep evolution neural network |
CN110674928A (en) * | 2019-09-18 | 2020-01-10 | 电子科技大学 | An online learning method integrating artificial neural network and neuromorphic computing |
CN110674928B (en) * | 2019-09-18 | 2023-10-27 | 电子科技大学 | Online learning method integrating artificial neural network and nerve morphology calculation |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Zahisham et al. | Food recognition with resnet-50 | |
CN112307958B (en) | Micro-expression Recognition Method Based on Spatiotemporal Appearance-Motion Attention Network | |
Guo et al. | A survey on deep learning based face recognition | |
CN109145979B (en) | Sensitive image identification method and terminal system | |
Islalm et al. | Recognition bangla sign language using convolutional neural network | |
Stuhlsatz et al. | Feature extraction with deep neural networks by a generalized discriminant analysis | |
CN110321967B (en) | Improved method of image classification based on convolutional neural network | |
CN108304826A (en) | Facial expression recognizing method based on convolutional neural networks | |
CN113469236A (en) | Deep clustering image recognition system and method for self-label learning | |
CN108985252A (en) | The image classification method of improved pulse deep neural network | |
Lagani et al. | Comparing the performance of Hebbian against backpropagation learning using convolutional neural networks | |
CN105023006B (en) | Face identification method based on enhanced nonparametric maximal margin criterion | |
CN103065158A (en) | Action identification method of independent subspace analysis (ISA) model based on relative gradient | |
Reddy et al. | Handwritten Hindi character recognition using deep learning techniques | |
CN106980831A (en) | Based on self-encoding encoder from affiliation recognition methods | |
CN108846349A (en) | A kind of face identification method based on dynamic Spiking neural network | |
Bouchain | Character recognition using convolutional neural networks | |
Masood et al. | Scene recognition from image using convolutional neural network | |
CN114780767A (en) | A large-scale image retrieval method and system based on deep convolutional neural network | |
Yang et al. | Multiple features fusion for facial expression recognition based on ELM | |
Le-Tien et al. | Iris-based biometric recognition using modified convolutional neural network | |
CN117079024A (en) | Image class increment learning algorithm integrating uncertainty estimation and increment stage discrimination | |
Yu et al. | Pattern recognition computation in a spiking neural network with temporal encoding and learning | |
Liu et al. | Multi-digit recognition with convolutional neural network and long short-term memory | |
Saabni | Ada-boosting extreme learning machines for handwritten digit and digit strings recognition |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20181120 |
|
RJ01 | Rejection of invention patent application after publication |