[go: up one dir, main page]

CN106934378A - A kind of dazzle light identifying system and method based on video depth study - Google Patents

A kind of dazzle light identifying system and method based on video depth study Download PDF

Info

Publication number
CN106934378A
CN106934378A CN201710156201.4A CN201710156201A CN106934378A CN 106934378 A CN106934378 A CN 106934378A CN 201710156201 A CN201710156201 A CN 201710156201A CN 106934378 A CN106934378 A CN 106934378A
Authority
CN
China
Prior art keywords
deep learning
module
key frame
database
frame
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201710156201.4A
Other languages
Chinese (zh)
Other versions
CN106934378B (en
Inventor
李成栋
丁子祥
许福运
张桂青
郝丽丽
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shandong Jianzhu University
Original Assignee
Shandong Jianzhu University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shandong Jianzhu University filed Critical Shandong Jianzhu University
Priority to CN201710156201.4A priority Critical patent/CN106934378B/en
Publication of CN106934378A publication Critical patent/CN106934378A/en
Application granted granted Critical
Publication of CN106934378B publication Critical patent/CN106934378B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • G06V20/46Extracting features or characteristics from the video content, e.g. video fingerprints, representative shots or key frames
    • G06V20/47Detecting features for summarising video content
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • G06V20/44Event detection

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Biomedical Technology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Multimedia (AREA)
  • Image Analysis (AREA)

Abstract

本发明公开了一种基于视频深度学习的汽车远光灯识别系统及方法,该系统包括以下两部分:前台部分,用于实现对远光灯违章行为的识别与处理,包括依次连接的道路监测设备模块、视频处理与识别模块、识别结果处理模块以及待检违章结果数据库;后台部分,用于视频处理并实现视频的深度学习,包括关键帧提取算法、带标签数据库以及深度学习模块,带标签数据库是对原始视频数据调用关键帧提取算法进行关键帧提取而构建得到,带标签数据库中的数据用于深度学习模块的训练,训练好的深度学习模块与关键帧提取算法一起供视频处理与识别模块调用。本发明对监控视频进行自动分析识别,保证了执法证据的完备性,且与人工判断类似,具有智能性。

The invention discloses a car high beam recognition system and method based on video deep learning. The system includes the following two parts: the front part, which is used to realize the identification and processing of high beam violations, including sequentially connected road monitoring Equipment module, video processing and recognition module, recognition result processing module, and violation result database to be checked; the background part is used for video processing and deep learning of video, including key frame extraction algorithm, tagged database and deep learning module, with tagged The database is constructed by calling the key frame extraction algorithm for the original video data to extract key frames. The data in the labeled database is used for the training of the deep learning module. The trained deep learning module and the key frame extraction algorithm are used together for video processing and recognition. module call. The invention automatically analyzes and recognizes the monitoring video, ensures the completeness of law enforcement evidence, is similar to manual judgment, and has intelligence.

Description

一种基于视频深度学习的汽车远光灯识别系统及方法A car high beam recognition system and method based on video deep learning

技术领域technical field

本发明涉及一种汽车远光灯识别系统,具体涉及一种基于视频深度学习的汽车远光灯识别系统及方法。属于智能交通技术领域。The invention relates to an automobile high beam recognition system, in particular to an automobile high beam recognition system and method based on video deep learning. It belongs to the field of intelligent transportation technology.

背景技术Background technique

自改革开放以来,我国经济得到了持续、稳定、快速的发展,也使得我国人民的生活水平得到了前所未有的提高,越来越多的国人拥有了私家车辆。私家车数量的快速增长在给人们出行带来方便的同时,也使得交通事故的发生频率越来越高。Since the reform and opening up, my country's economy has achieved sustained, stable and rapid development, which has also improved the living standards of our people unprecedentedly, and more and more Chinese people have private vehicles. While the rapid growth of the number of private cars brings convenience to people's travel, it also makes the frequency of traffic accidents higher and higher.

导致交通事故发生的原因有很多,其中很多事故是远光灯使用不当所造成的。目前针对远光灯违规的监管主要依靠交通警察的抓拍,由于警力以及时间的限制,不能保证对所有的远光灯违规现象都进行有效的监管。除此之外,近年来开发的一些远光灯抓拍系统,都是对抓拍图片进行识别,但这些方法具有一定的局限性,表现在:1)抓拍的远光图片数量少且不连贯,这些远光图片很有可能是司机正常使用时产生的,容易被误判为乱用远光,因此以图片作为执法证据不充分;2)为获得这些图片,在同一地点往往要额外架设多台抓拍设备,造价较高;3)对原有布设的视频监控设备不能完全利用,造成资源浪费。There are many reasons for traffic accidents, many of which are caused by improper use of high beams. At present, the supervision of violations of high beam lights mainly relies on the capture of traffic police. Due to the limitation of police force and time, it is impossible to guarantee effective supervision of all violations of high beam lights. In addition, some high-beam capture systems developed in recent years all recognize the captured pictures, but these methods have certain limitations, as shown in: 1) The number of captured high-beam pictures is small and incoherent. The high-beam pictures are likely to be generated by the driver during normal use, and it is easy to be misjudged as using high-beams indiscriminately. Therefore, using pictures as evidence for law enforcement is not sufficient; 2) In order to obtain these pictures, it is often necessary to set up multiple additional capture devices at the same location , the cost is higher; 3) the original video monitoring equipment cannot be fully utilized, resulting in waste of resources.

发明内容Contents of the invention

本发明的目的是为克服上述现有技术的不足,提供一种基于视频深度学习的汽车远光灯识别系统。The purpose of the present invention is to provide a car high beam recognition system based on video deep learning for overcoming the above-mentioned deficiencies in the prior art.

本发明还提供了上述系统对应的一种基于视频深度学习的汽车远光灯识别方法。The present invention also provides a car high beam recognition method based on video deep learning corresponding to the above system.

为实现上述目的,本发明采用下述技术方案:To achieve the above object, the present invention adopts the following technical solutions:

一种基于视频深度学习的汽车远光灯识别系统,包括以下两部分:A car high beam recognition system based on video deep learning, including the following two parts:

前台部分,用于实现对远光灯违章行为的识别与处理,包括依次连接的道路监测设备模块、视频处理与识别模块、识别结果处理模块以及待检违章结果数据库;The foreground part is used to realize the identification and processing of high beam violations, including the road monitoring equipment module, video processing and identification module, recognition result processing module, and pending inspection results database;

后台部分,用于视频处理并实现视频的深度学习,包括关键帧提取算法、带标签数据库以及深度学习模块,带标签数据库是对原始视频数据调用关键帧提取算法进行关键帧提取而构建得到,带标签数据库中的数据用于深度学习模块的训练,训练好的深度学习模块与关键帧提取算法一起供视频处理与识别模块调用。The background part is used for video processing and deep learning of video, including key frame extraction algorithm, tagged database and deep learning module. The data in the label database is used for the training of the deep learning module, and the trained deep learning module and the key frame extraction algorithm are called by the video processing and recognition module.

作为优选技术方案之一,所述关键帧提取算法为基于聚类的关键帧提取算法。As one of the preferred technical solutions, the key frame extraction algorithm is a cluster-based key frame extraction algorithm.

作为优选技术方案之一,所述深度学习模块为基于CNN+LSE(卷积神经网络+最小二乘估计)的深度学习模块。As one of the preferred technical solutions, the deep learning module is a deep learning module based on CNN+LSE (convolutional neural network+least square estimation).

上述系统对应的一种基于视频深度学习的汽车远光灯识别方法,具体步骤如下:A car high beam recognition method based on video deep learning corresponding to the above system, the specific steps are as follows:

(1)道路监测设备模块获得汽车的行驶视频数据,并将其传输给视频处理与识别模块;(1) The road monitoring equipment module obtains the driving video data of the car and transmits it to the video processing and identification module;

(2)视频处理与识别模块调用关键帧提取算法提取视频数据的关键帧,然后进行灰度化操作,将灰度化后的关键帧作为输入,调用根据带标签数据库训练好的基于CNN+LSE的深度学习模块,得到各个关键帧的输出标签,包括近光灯、雾灯或者远光灯,并将标签赋给相应关键帧图像;(2) The video processing and recognition module calls the key frame extraction algorithm to extract the key frames of the video data, and then performs the grayscale operation, takes the grayscaled key frames as input, and calls the CNN+LSE based on the label database training The deep learning module obtains the output label of each key frame, including low beam, fog light or high beam, and assigns the label to the corresponding key frame image;

(3)将视频数据以及步骤(2)所得带标签的关键帧一并作为识别结果处理模块的输入,用于判断车辆是否违章,并且,在识别结果处理模块中嵌入车牌识别系统,当目标车辆存在远光灯违章行为时,对其进行车牌提取,获取车辆信息,将涉嫌违章视频数据导入待检违章结果数据库。(3) The video data and the key frame with the label obtained in step (2) are used as the input of the recognition result processing module to determine whether the vehicle violates regulations, and the license plate recognition system is embedded in the recognition result processing module. When the target vehicle When there is a high beam violation, the license plate is extracted, the vehicle information is obtained, and the suspected violation video data is imported into the violation result database to be checked.

步骤(2)中,关键帧提取算法如下:In step (2), the key frame extraction algorithm is as follows:

(2-1)取原始视频数据库中的第i段Vi,等时间间隔提取n个帧,并用Fi,j命名第i段视频数据的第j个时刻的帧,将相应视频数据的关键帧序列表示为{Fi,1,Fi,2,...,Fi,n},其中Fi,1为首帧,Fi,n为尾帧;定义相邻两帧之间的相似度是相邻两帧直方图的相似度(即直方图特征差别),预定义阈值δ控制聚类的密度;其中,i、j和n均为整数;(2-1) Take the i-th segment V i in the original video database, extract n frames at equal time intervals, and use F i,j to name the frame at the j-th moment of the i-th segment video data, and set the key of the corresponding video data The frame sequence is expressed as {F i,1 ,F i,2 ,...,F i,n }, where F i , 1 is the first frame, and F i,n is the last frame; define the similarity between two adjacent frames The degree is the similarity of the histograms of two adjacent frames (that is, the histogram feature difference), and the predefined threshold δ controls the density of clustering; wherein, i, j and n are all integers;

(2-2)选定首帧Fi,1为初始的聚类中心,并计算帧Fi,j与初始聚类中心间的相似度,如果该值小于δ,则判定该帧与聚类中心帧之间距离过大,因此,Fi,j不能加入该聚类中;如果Fi,j与所有聚类中心相似度均小于δ,则Fi,j形成一个新的聚类,Fi,j为新的聚类中心;否则,将该帧加入到与之相似度最大的聚类中,使该帧与这个聚类的中心之间的距离最小;(2-2) Select the first frame F i,1 as the initial clustering center, and calculate the similarity between the frame F i,j and the initial clustering center, if the value is less than δ, it is determined that the frame and the clustering center The distance between the center frames is too large, therefore, F i, j cannot be added to the cluster; if the similarity between F i, j and all cluster centers is less than δ, then F i, j forms a new cluster, F i, j is the new cluster center; otherwise, add the frame to the cluster with the largest similarity, so that the distance between the frame and the center of this cluster is the smallest;

(2-3)重复(2-2)操作,将原始视频数据Vi中所提取的n个帧,分别归类到不同聚类后,就可以选择关键帧:从每个聚类中抽取离聚类中心最近的帧作为这个聚类的代表帧,所有聚类的代表帧就构成了原始视频数据Vi的关键帧。(2-3) Repeat the operation of (2-2) to classify the n frames extracted from the original video data V i into different clusters, then you can select the key frame: extract the distance from each cluster The nearest frame of the cluster center is used as the representative frame of this cluster, and the representative frames of all clusters constitute the key frame of the original video data V i .

步骤(2)中,带标签数据库的构建方法如下:In step (2), the construction method of the tagged database is as follows:

将大数据背景下的大量车辆行驶视频数据作为原始视频数据,对原始视频数据调用基于聚类的关键帧提取算法进行关键帧提取,人工判定关键帧中车辆的灯光类型,给每个关键帧添加标签使原始关键帧变为带标签数据,其中,标签类别包括:近光灯、雾灯以及远光灯三种,分别用-1,0以及1表示;将带有标签的关键帧数据存入带标签数据库,带标签数据库中的数据为原始视频数据及其带标签关键帧,表示为(Fi,j,k),其中k取-1,0或者1。A large amount of vehicle driving video data under the background of big data is used as the original video data, and the key frame extraction algorithm based on clustering is called on the original video data to extract the key frames, and the lighting type of the vehicle in the key frames is manually determined, and each key frame is added The label turns the original keyframe into labeled data. The label categories include: low beam, fog light, and high beam, which are represented by -1, 0, and 1 respectively; store the keyframe data with labels in Labeled database, the data in the labeled database is original video data and its labeled keyframes, expressed as (F i,j ,k), where k is -1, 0 or 1.

步骤(2)中,基于CNN+LSE的深度学习模块的构建方法是,采用LeNet5卷积神经网络结构,模块共分为八层,前六层为特征提取部分,后两层为分类器部分,其中,特征提取层采用经典的卷积神经网络结构,分类器层采用全连接结构;模块由带标签数据库中的数据作为训练数据,采用CNN+LSE组合算法对深度学习模块进行训练,对于特征提取部分采用CNN方法进行训练,而对于分类器层,则采用LSE方法进行训练,以期实现模块参数的快速学习并增强模块的泛化能力。In step (2), the construction method of the deep learning module based on CNN+LSE is to adopt the LeNet5 convolutional neural network structure. The module is divided into eight layers, the first six layers are the feature extraction part, and the last two layers are the classifier part. Among them, the feature extraction layer adopts the classic convolutional neural network structure, and the classifier layer adopts the fully connected structure; the module uses the data in the labeled database as the training data, and uses the CNN+LSE combination algorithm to train the deep learning module. For feature extraction The CNN method is used for training partly, while for the classifier layer, the LSE method is used for training, in order to realize the fast learning of module parameters and enhance the generalization ability of the module.

具体方法如下:The specific method is as follows:

带标签数据库中的视频关键帧输入基于CNN+LSE的深度学习模块的第一层;在第二层中采用不同的卷积核对上层输出进行卷积操作;第三层对上层输出进行池化(下采样)操作;第四层与第五层重复第二层与第三层的操作;第六层将上层的输出特征依次展开,排成一列;第七层与上层输出特征全互连;最后一层同样与上层之间采用全互连的形式。基于CNN+LSE的深度学习模块的输出为三种情况:近光灯、雾灯以及远光灯,分别用-1,0以及1表示。The key frame of the video in the label database is input to the first layer of the deep learning module based on CNN+LSE; in the second layer, different convolution kernels are used to perform convolution operations on the upper layer output; the third layer performs pooling on the upper layer output ( Downsampling) operation; the fourth layer and the fifth layer repeat the operation of the second layer and the third layer; the sixth layer expands the output features of the upper layer in turn and arranges them in a row; the seventh layer and the upper layer output features are fully interconnected; finally The first layer is also fully interconnected with the upper layer. The output of the deep learning module based on CNN+LSE is three situations: low beam, fog light and high beam, represented by -1, 0 and 1 respectively.

基于CNN+LSE的深度学习模块的训练过程如下:The training process of the deep learning module based on CNN+LSE is as follows:

从带标签数据库中任取一个样本(Fi,j,k),对Fi,j首先进行灰度化操作,使其变为灰度图像,然后将灰度化后的关键帧Fi,j'输入到模块中,即输入数据为(Fi,j',k);对深度学习模块的两部分分别采用CNN与LSE的方法进行训练;其中,特征提取部分的参数训练方法如下:Randomly take a sample (F i,j ,k) from the labeled database, perform grayscale operation on F i,j first to make it a grayscale image, and then convert the grayscaled key frame F i, j ' is input into the module, that is, the input data is (F i,j ',k); the two parts of the deep learning module are trained by CNN and LSE respectively; among them, the parameter training method of the feature extraction part is as follows:

(2-A1)初始化深度学习模块中特征提取部分的所有连接权重参数;(2-A1) Initialize all connection weight parameters of the feature extraction part in the deep learning module;

(2-A2)计算输入关键帧相对应的实际输出标签Ok(2-A2) Calculate the actual output label O k corresponding to the input key frame;

(2-A3)计算实际输出标签Ok与相应理想输出标签k的差值;(2-A3) Calculate the difference between the actual output label O k and the corresponding ideal output label k;

(2-A4)权重学习:通过极小化误差的方法反向传播调节深度学习模块中特征提取部分的连接权重参数矩阵;(2-A4) Weight learning: adjust the connection weight parameter matrix of the feature extraction part in the deep learning module through backpropagation by minimizing the error;

(2-A5)直至遍历所有视频数据的关键帧,参数训练完毕;(2-A5) Until the key frames of all video data are traversed, the parameter training is completed;

分类器部分的参数训练方法如下:The parameter training method of the classifier part is as follows:

(2-B1)光栅化层与全连接层之间的连接权重与偏置随机生成,并将全连接层输出写(2-B1) The connection weight and bias between the rasterization layer and the fully connected layer are randomly generated, and the output of the fully connected layer is written

为矩阵 for the matrix

其中G(·)为激活函数,ai为连接权重,bi为偏置,L为全连接层节点个数,N为所有关键帧的个数,xj为关键帧,i=1,2,…,L,j=1,2,…,N;Where G( ) is the activation function, a i is the connection weight, b i is the bias, L is the number of fully connected layer nodes, N is the number of all key frames, x j is the key frame, i=1,2 ,...,L,j=1,2,...,N;

(2-B2)将相应关键帧的网络输出结果写为输出向量Y=[y1 y2 … yn]T,其中yj为第j个关键帧xj对应的输出标签;(2-B2) Write the network output result of the corresponding key frame as an output vector Y=[y 1 y 2 ... y n ] T , where y j is the output label corresponding to the jth key frame x j ;

(2-B3)计算全连接层与输出层之间的输出权重β=PHY,其中P=(HTH)-1(2-B3) Calculate the output weight β=PHY between the fully connected layer and the output layer, where P=(H T H) −1 .

步骤(3)中,待检违章结果数据库中的数据为经识别结果处理模块判断为违章的视频数据,其中的待检违章结果应当接受人工检查,然后将确认无误的信息导入违章数据库,并对误判的信息进行删除。In step (3), the data in the database of violation results to be checked is the video data judged to be violations by the recognition result processing module, and the results of violations to be checked should be manually checked, and then the confirmed information will be imported into the violation database, and Misjudged information will be deleted.

步骤(3)中,对于是否存在远光灯违章行为的判断方法是:标签为远光灯的关键帧与其下个关键帧之间的时间间隔ΔT=j2-j1,若ΔT≥θ,则该车辆存在远光灯违章使用现象,其中,θ为违章时间阈值。In step (3), the method for judging whether there is a high beam violation is: the key frame labeled as high beam with the next keyframe The time interval between ΔT=j 2 -j 1 , if ΔT≥θ, the vehicle has illegal use of high beams, where θ is the violation time threshold.

本发明的有益效果:Beneficial effects of the present invention:

本发明对监控视频进行自动分析识别,保证了执法证据的完备性,且与人工判断类似,具有智能性,同时所需布设设备简单,并能对原有的监控设备充分利用。具体如下:The invention automatically analyzes and recognizes the monitoring video, ensures the completeness of law enforcement evidence, is similar to manual judgment, has intelligence, and needs simple layout equipment, and can make full use of the original monitoring equipment. details as follows:

(1)通过对视频数据的挖掘,在保证准确率的基础上,执法证据的充分性得到大幅度提升,防止了远光灯违章执法时证据链的缺失;(1) Through the mining of video data, on the basis of ensuring accuracy, the adequacy of law enforcement evidence has been greatly improved, preventing the lack of evidence chains when high beams are illegally enforced;

(2)同一点位对设备数量要求少,且原先布放的大量监控设备可以直接重复使用,降低了成本,提升了设备利用率;(2) The same point requires less equipment, and a large number of monitoring equipment originally deployed can be directly reused, reducing costs and improving equipment utilization;

(3)采用基于视频深度学习的方式进行远光灯违章的智能判定,代替了人工执法,实现了真正的自动化,提高了效率;同时,经过深度学习后,远光灯违章识别效果有望达到或超越人工识别水平,实现了识别系统的真正智能化;(3) The intelligent judgment of high beam violations based on video deep learning is used to replace manual law enforcement, realizing real automation and improving efficiency; at the same time, after deep learning, the recognition effect of high beam violations is expected to reach or Beyond the level of artificial recognition, it realizes the real intelligence of the recognition system;

(4)深度学习模块采用CNN+LSE的方法对系统进行参数学习,不仅使得系统的参数学习速度更加快速,而且还能使得模块泛化能力变得更强,系统鲁棒性得到提升。(4) The deep learning module uses the method of CNN+LSE to learn the parameters of the system, which not only makes the parameter learning speed of the system faster, but also makes the generalization ability of the module stronger and the robustness of the system improved.

附图说明Description of drawings

图1是本发明的系统结构示意图;Fig. 1 is a schematic diagram of the system structure of the present invention;

图2是基于CNN+LSE的深度学习模块结构图。Figure 2 is a structure diagram of a deep learning module based on CNN+LSE.

具体实施方式detailed description

下面结合附图和实施例对本发明进行进一步的阐述,应该说明的是,下述说明仅是为了解释本发明,并不对其内容进行限定。The present invention will be further described below in conjunction with the accompanying drawings and embodiments. It should be noted that the following description is only for explaining the present invention and not limiting its content.

如图1所示,一种基于视频深度学习的汽车远光灯识别系统,包括以下两部分:As shown in Figure 1, a car high beam recognition system based on video deep learning includes the following two parts:

前台部分,用于实现对远光灯违章行为的识别与处理,包括依次连接的道路监测设备模块、视频处理与识别模块、识别结果处理模块以及待检违章结果数据库;The foreground part is used to realize the identification and processing of high beam violations, including the road monitoring equipment module, video processing and identification module, recognition result processing module, and pending inspection results database;

后台部分,用于视频处理并实现视频的深度学习,包括关键帧提取算法、带标签数据库以及深度学习模块,带标签数据库是对原始视频数据调用关键帧提取算法进行关键帧提取而构建得到,带标签数据库中的数据用于深度学习模块的训练,训练好的深度学习模块与关键帧提取算法一起供视频处理与识别模块调用。The background part is used for video processing and deep learning of video, including key frame extraction algorithm, tagged database and deep learning module. The data in the label database is used for the training of the deep learning module, and the trained deep learning module and the key frame extraction algorithm are called by the video processing and recognition module.

其中,关键帧提取算法为基于聚类的关键帧提取算法;深度学习模块为基于CNN+LSE的深度学习模块。Among them, the key frame extraction algorithm is a key frame extraction algorithm based on clustering; the deep learning module is a deep learning module based on CNN+LSE.

上述系统对应的一种基于视频深度学习的汽车远光灯识别方法,具体步骤如下:A car high beam recognition method based on video deep learning corresponding to the above system, the specific steps are as follows:

(1)道路监测设备模块获得汽车的行驶视频数据,并将其传输给视频处理与识别模块。(1) The road monitoring equipment module obtains the driving video data of the car and transmits it to the video processing and recognition module.

(2)视频处理与识别模块调用关键帧提取算法提取原始视频数据的关键帧,然后进行灰度化操作,将灰度化后的关键帧作为输入,调用带标签数据库训练好的基于CNN+LSE的深度学习模块,得到各个关键帧的输出标签,包括近光灯、雾灯或者远光灯,并将标签赋给相应关键帧图像。(2) The video processing and recognition module calls the key frame extraction algorithm to extract the key frames of the original video data, and then performs the grayscale operation. The grayscaled key frames are used as input, and the trained CNN+LSE based on the label database is called The deep learning module obtains the output labels of each key frame, including low beam, fog light or high beam, and assigns the label to the corresponding key frame image.

关键帧提取算法如下:The key frame extraction algorithm is as follows:

(2-1)取原始视频数据库中的第i段Vi,等时间间隔提取n个帧,并用Fi,j命名第i段视频数据的第j个时刻的帧,将相应视频数据的关键帧序列表示为{Fi,1,Fi,2,...,Fi,n},其中Fi,1为首帧,Fi,n为尾帧;定义相邻两帧之间的相似度是相邻两帧直方图的相似度(即直方图特征差别),预定义阈值δ控制聚类的密度;其中,i、j和n均为整数;(2-1) Take the i-th segment V i in the original video database, extract n frames at equal time intervals, and use F i,j to name the frame at the j-th moment of the i-th segment video data, and set the key of the corresponding video data The frame sequence is expressed as {F i,1 ,F i,2 ,...,F i,n }, where F i,1 is the first frame and F i,n is the last frame; define the similarity between two adjacent frames The degree is the similarity of the histograms of two adjacent frames (that is, the histogram feature difference), and the predefined threshold δ controls the density of clustering; wherein, i, j and n are all integers;

(2-2)选定首帧Fi,1为初始的聚类中心,并计算帧Fi,j与初始聚类中心间的相似度,如果该值小于δ,则判定该帧与聚类中心帧之间距离过大,因此,Fi,j不能加入该聚类中;如果Fi,j与所有聚类中心相似度均小于δ,则Fi,j形成一个新的聚类,Fi,j为新的聚类中心;否则,将该帧加入到与之相似度最大的聚类中,使该帧与这个聚类的中心之间的距离最小;(2-2) Select the first frame F i,1 as the initial clustering center, and calculate the similarity between the frame F i,j and the initial clustering center, if the value is less than δ, it is determined that the frame and the clustering center The distance between the center frames is too large, therefore, F i, j cannot be added to the cluster; if the similarity between F i, j and all cluster centers is less than δ, then F i, j forms a new cluster, F i, j is the new cluster center; otherwise, add the frame to the cluster with the largest similarity, so that the distance between the frame and the center of this cluster is the smallest;

(2-3)重复(2-2)操作,将原始视频数据Vi中所提取的n个帧,分别归类到不同聚类后,就可以选择关键帧:从每个聚类中抽取离聚类中心最近的帧作为这个聚类的代表帧,所有聚类的代表帧就构成了原始视频数据Vi的关键帧。(2-3) Repeat the operation of (2-2) to classify the n frames extracted from the original video data V i into different clusters, then you can select the key frame: extract the distance from each cluster The nearest frame of the cluster center is used as the representative frame of this cluster, and the representative frames of all clusters constitute the key frame of the original video data V i .

带标签数据库的构建方法如下:The construction method of the labeled database is as follows:

将大数据背景下的大量车辆行驶视频数据作为原始视频数据,对原始视频数据调用基于聚类的关键帧提取算法进行关键帧提取,人工判定关键帧中车辆的灯光类型,给每个关键帧添加标签使原始关键帧变为带标签数据,其中,标签类别包括:近光灯、雾灯以及远光灯三种,分别用-1,0以及1表示;将带有标签的关键帧数据存入带标签数据库,带标签数据库中的数据为原始视频数据及其带标签关键帧,表示为(Fi,j,k),其中k取-1,0或者1。A large amount of vehicle driving video data under the background of big data is used as the original video data, and the key frame extraction algorithm based on clustering is called on the original video data to extract the key frames, and the lighting type of the vehicle in the key frames is manually determined, and each key frame is added The label changes the original keyframe into labeled data. The label categories include: low beam, fog light, and high beam, which are represented by -1, 0, and 1 respectively; store the keyframe data with labels in Labeled database, the data in the labeled database is original video data and its labeled keyframes, expressed as (F i,j ,k), where k is -1, 0 or 1.

如图2所示,基于CNN+LSE的深度学习模块的构建方法是,采用LeNet5卷积神经网络结构,模块共分为八层,前六层为特征提取部分,后两层为分类器部分,其中,特征提取层采用经典的卷积神经网络结构,分类器层采用全连接结构;模块由带标签数据库中的数据作为训练数据,采用CNN+LSE组合算法对深度学习模块进行训练,对于特征提取部分采用CNN方法进行训练,而对于分类器层,则采用LSE方法进行训练,以期实现模块参数的快速学习并增强模块的泛化能力。具体方法如下:带标签数据库中的视频关键帧输入基于CNN+LSE的深度学习模块的第一层;在第二层中采用不同的卷积核对上层输出进行卷积操作;第三层对上层输出进行池化(下采样)操作;第四层与第五层重复第二层与第三层的操作;第六层将上层的输出特征依次展开,排成一列;第七层与上层输出特征全互连;最后一层同样与上层之间采用全互连的形式。基于CNN+LSE的深度学习模块的输出为三种情况:近光灯、雾灯以及远光灯,分别用-1,0以及1表示。As shown in Figure 2, the construction method of the deep learning module based on CNN+LSE is to adopt the LeNet5 convolutional neural network structure. The module is divided into eight layers. The first six layers are the feature extraction part, and the last two layers are the classifier part. Among them, the feature extraction layer adopts the classic convolutional neural network structure, and the classifier layer adopts the fully connected structure; the module uses the data in the labeled database as the training data, and uses the CNN+LSE combination algorithm to train the deep learning module. For feature extraction The CNN method is used for training partly, while for the classifier layer, the LSE method is used for training, in order to realize the fast learning of module parameters and enhance the generalization ability of the module. The specific method is as follows: the video key frames in the tagged database are input to the first layer of the deep learning module based on CNN+LSE; in the second layer, different convolution kernels are used to perform convolution operations on the upper layer output; Perform pooling (down-sampling) operations; the fourth and fifth layers repeat the operations of the second and third layers; the sixth layer expands the output features of the upper layer in turn and arranges them in a row; the seventh layer and the upper layer output features are all Interconnection; the last layer is also fully interconnected with the upper layer. The output of the deep learning module based on CNN+LSE is three situations: low beam, fog light and high beam, represented by -1, 0 and 1 respectively.

基于CNN+LSE的深度学习模块的训练过程如下:The training process of the deep learning module based on CNN+LSE is as follows:

从带标签数据库中任取一个样本(Fi,j,k),对Fi,j首先进行灰度化操作,使其变为灰度图像,然后将灰度化后的关键帧Fi,j'输入到模块中,即输入数据为(Fi,j',k);对深度学习模块的两部分分别采用CNN与LSE的方法进行训练;其中,特征提取部分的参数训练方法如下:Randomly take a sample (F i,j ,k) from the labeled database, perform grayscale operation on F i,j first to make it a grayscale image, and then convert the grayscaled key frame F i, j ' is input into the module, that is, the input data is (F i,j ',k); the two parts of the deep learning module are trained by CNN and LSE respectively; among them, the parameter training method of the feature extraction part is as follows:

(2-A1)初始化深度学习模块中特征提取部分的所有连接权重参数;(2-A1) Initialize all connection weight parameters of the feature extraction part in the deep learning module;

(2-A2)计算输入关键帧相对应的实际输出标签Ok(2-A2) Calculate the actual output label O k corresponding to the input key frame;

(2-A3)计算实际输出标签Ok与相应理想输出标签k的差值;(2-A3) Calculate the difference between the actual output label O k and the corresponding ideal output label k;

(2-A4)权重学习:通过极小化误差的方法反向传播调节深度学习模块中特征提取部分的连接权重参数矩阵;(2-A4) Weight learning: adjust the connection weight parameter matrix of the feature extraction part in the deep learning module through backpropagation by minimizing the error;

(2-A5)直至遍历所有视频数据的关键帧,参数训练完毕;(2-A5) Until the key frames of all video data are traversed, the parameter training is completed;

分类器部分的参数训练方法如下:The parameter training method of the classifier part is as follows:

(2-B1)光栅化层与全连接层之间的连接权重与偏置随机生成,并将全连接层输出写(2-B1) The connection weight and bias between the rasterization layer and the fully connected layer are randomly generated, and the output of the fully connected layer is written

为矩阵 for the matrix

其中G(·)为激活函数,ai为连接权重,bi为偏置,L为全连接层节点个数,N为所有关键帧的个数,xj为关键帧,i=1,2,…,L,j=1,2,…,N;Where G( ) is the activation function, a i is the connection weight, b i is the bias, L is the number of fully connected layer nodes, N is the number of all key frames, x j is the key frame, i=1,2 ,...,L,j=1,2,...,N;

(2-B2)将相应关键帧的网络输出结果写为输出向量Y=[y1 y2 … yn]T,其中yj为第j个关键帧xj对应的输出标签;(2-B2) Write the network output result of the corresponding key frame as an output vector Y=[y 1 y 2 ... y n ] T , where y j is the output label corresponding to the jth key frame x j ;

(2-B3)计算全连接层与输出层之间的输出权重β=PHY,其中P=(HTH)-1(2-B3) Calculate the output weight β=PHY between the fully connected layer and the output layer, where P=(H T H) −1 .

(3)将原始视频数据以及步骤(2)所得带标签的关键帧一并作为识别结果处理模块的输入,用于判断车辆是否违章,并且,在识别结果处理模块中嵌入车牌识别系统,当目标车辆存在远光灯违章行为时,对其进行车牌提取,获取车辆信息,将涉嫌违章视频数据导入待检违章结果数据库。(3) The original video data and the key frame with the label obtained in step (2) are used as the input of the recognition result processing module to judge whether the vehicle violates regulations, and the license plate recognition system is embedded in the recognition result processing module. When the target When the vehicle has high beam violations, the license plate is extracted, the vehicle information is obtained, and the suspected violation video data is imported into the violation result database to be inspected.

对于是否存在远光灯违章行为的判断方法是:标签为远光灯的关键帧与其下个关键帧之间的时间间隔ΔT=j2-j1,若ΔT≥θ,则该车辆存在远光灯违章使用现象,其中,θ为违章时间阈值。The method for judging whether there is a high beam violation is: the key frame labeled as high beam with the next keyframe The time interval between ΔT=j 2 -j 1 , if ΔT≥θ, the vehicle has illegal use of high beams, where θ is the violation time threshold.

(4)待检违章结果数据库中的数据为经识别结果处理模块判断为违章的视频数据,其中的待检违章结果应当接受人工检查,然后将确认无误的信息导入违章数据库,并对误判的信息进行删除。(4) The data in the database of violation results to be checked is the video data judged to be violations by the identification result processing module. The information is deleted.

上述虽然结合附图对本发明的具体实施方式进行了描述,但并非对本发明保护范围的限制,在本发明的技术方案的基础上,本领域技术人员不需要付出创造性劳动即可做出的各种修改或变形仍在本发明的保护范围以内。Although the specific implementation of the present invention has been described above in conjunction with the accompanying drawings, it does not limit the protection scope of the present invention. On the basis of the technical solution of the present invention, those skilled in the art can make various Modifications or variations are still within the protection scope of the present invention.

Claims (10)

1.一种基于视频深度学习的汽车远光灯识别系统,其特征在于,包括以下两部分:1. A car high beam recognition system based on video deep learning, is characterized in that, comprises the following two parts: 前台部分,用于实现对远光灯违章行为的识别与处理,包括依次连接的道路监测设备模块、视频处理与识别模块、识别结果处理模块以及待检违章结果数据库;The foreground part is used to realize the identification and processing of high beam violations, including the road monitoring equipment module, video processing and identification module, recognition result processing module, and pending inspection results database; 后台部分,用于视频处理并实现视频的深度学习,包括关键帧提取算法、带标签数据库以及深度学习模块,带标签数据库是对原始视频数据调用关键帧提取算法进行关键帧提取而构建得到,带标签数据库中的数据用于深度学习模块的训练,并与训练好的深度学习模块一并供视频处理与识别模块调用。The background part is used for video processing and deep learning of video, including key frame extraction algorithm, tagged database and deep learning module. The data in the label database is used for the training of the deep learning module, and is called by the video processing and recognition module together with the trained deep learning module. 2.根据权利要求1所述的系统,其特征在于,所述关键帧提取算法为基于聚类的关键帧提取算法。2. The system according to claim 1, wherein the key frame extraction algorithm is a cluster-based key frame extraction algorithm. 3.根据权利要求1所述的系统,其特征在于,所述深度学习模块为基于CNN+LSE的深度学习模块。3. The system according to claim 1, wherein the deep learning module is a deep learning module based on CNN+LSE. 4.权利要求1~3中任一项所述系统对应的一种基于视频深度学习的汽车远光灯识别方法,其特征在于,具体步骤如下:4. A kind of automobile high beam recognition method based on video deep learning corresponding to the system according to any one of claims 1 to 3, characterized in that, the specific steps are as follows: (1)道路监测设备模块获得汽车的行驶视频数据,并将其传输给视频处理与识别模块;(1) The road monitoring equipment module obtains the driving video data of the car and transmits it to the video processing and identification module; (2)视频处理与识别模块调用关键帧提取算法提取原始视频数据的关键帧,然后进行灰度化操作,将灰度化后的关键帧作为输入,调用带标签数据库训练好的基于CNN+LSE的深度学习模块,得到各个关键帧的输出标签,包括近光灯、雾灯或者远光灯,并将标签赋给相应关键帧图像;(2) The video processing and recognition module calls the key frame extraction algorithm to extract the key frames of the original video data, and then performs the grayscale operation. The grayscaled key frames are used as input, and the trained CNN+LSE based on the label database is called The deep learning module obtains the output label of each key frame, including low beam, fog light or high beam, and assigns the label to the corresponding key frame image; (3)将原始视频数据以及步骤(2)所得带标签的关键帧一并作为识别结果处理模块的输入,用于判断车辆是否违章,并且,在识别结果处理模块中嵌入车牌识别系统,当目标车辆存在远光灯违章行为时,对其进行车牌提取,获取车辆信息,将涉嫌违章视频数据导入待检违章结果数据库。(3) The original video data and the key frame with the label obtained in step (2) are used as the input of the recognition result processing module to judge whether the vehicle violates regulations, and the license plate recognition system is embedded in the recognition result processing module. When the target When the vehicle has high beam violations, the license plate is extracted, the vehicle information is obtained, and the suspected violation video data is imported into the violation result database to be inspected. 5.根据权利要求4所述的方法,其特征在于,步骤(2)中,关键帧提取算法如下:5. method according to claim 4, is characterized in that, in step (2), key frame extraction algorithm is as follows: (2-1)取原始视频数据库中的第i段Vi,等时间间隔提取n个帧,并用Fi,j命名第i段视频数据的第j个时刻的帧,将相应视频数据的关键帧序列表示为{Fi,1,Fi,2,...,Fi,n},其中Fi,1为首帧,Fi,n为尾帧;定义相邻两帧之间的相似度是相邻两帧直方图的相似度(即直方图特征差别),预定义阈值δ控制聚类的密度;其中,i、j和n均为整数;(2-1) Take the i-th segment V i in the original video database, extract n frames at equal time intervals, and use F i,j to name the frame at the j-th moment of the i-th segment video data, and set the key of the corresponding video data The frame sequence is expressed as {F i,1 ,F i,2 ,...,F i,n }, where F i , 1 is the first frame, and F i,n is the last frame; define the similarity between two adjacent frames The degree is the similarity of the histograms of two adjacent frames (that is, the histogram feature difference), and the predefined threshold δ controls the density of clustering; wherein, i, j and n are all integers; (2-2)选定首帧Fi,1为初始的聚类中心,并计算帧Fi,j与初始聚类中心间的相似度,如果该值小于δ,则判定该帧与聚类中心帧之间距离过大,因此,Fi,j不能加入该聚类中;如果Fi,j与所有聚类中心相似度均小于δ,则Fi,j形成一个新的聚类,Fi,j为新的聚类中心;否则,将该帧加入到与之相似度最大的聚类中,使该帧与这个聚类的中心之间的距离最小;(2-2) Select the first frame F i,1 as the initial clustering center, and calculate the similarity between the frame F i,j and the initial clustering center, if the value is less than δ, it is determined that the frame and the clustering center The distance between the center frames is too large, therefore, F i, j cannot be added to the cluster; if the similarity between F i, j and all cluster centers is less than δ, then F i, j forms a new cluster, F i, j is the new cluster center; otherwise, add the frame to the cluster with the largest similarity, so that the distance between the frame and the center of this cluster is the smallest; (2-3)重复(2-2)操作,将原始视频数据Vi中所提取的n个帧,分别归类到不同聚类后,就可以选择关键帧:从每个聚类中抽取离聚类中心最近的帧作为这个聚类的代表帧,所有聚类的代表帧就构成了原始视频数据Vi的关键帧。(2-3) Repeat the operation of (2-2) to classify the n frames extracted from the original video data V i into different clusters, then you can select the key frame: extract the distance from each cluster The nearest frame of the cluster center is used as the representative frame of this cluster, and the representative frames of all clusters constitute the key frame of the original video data V i . 6.根据权利要求5所述的方法,其特征在于,步骤(2)中,带标签数据库的构建方法如下:6. method according to claim 5, is characterized in that, in step (2), the construction method of band label database is as follows: 将大数据背景下的大量车辆行驶视频数据作为原始视频数据,对原始视频数据调用基于聚类的关键帧提取算法进行关键帧提取,人工判定关键帧中车辆的灯光类型,给每个关键帧添加标签使原始关键帧变为带标签数据,其中,标签类别包括:近光灯、雾灯以及远光灯三种,分别用-1,0以及1表示;将带有标签的关键帧数据存入带标签数据库,带标签数据库中的数据为原始视频数据及其带标签关键帧,表示为(Fi,j,k),其中k取-1,0或者1。A large amount of vehicle driving video data under the background of big data is used as the original video data, and the key frame extraction algorithm based on clustering is called on the original video data to extract the key frames, and the lighting type of the vehicle in the key frames is manually determined, and each key frame is added The label turns the original keyframe into labeled data. The label categories include: low beam, fog light, and high beam, which are represented by -1, 0, and 1 respectively; store the keyframe data with labels in Labeled database, the data in the labeled database is original video data and its labeled keyframes, expressed as (F i,j ,k), where k is -1, 0 or 1. 7.根据权利要求6所述的方法,其特征在于,步骤(2)中,基于CNN+LSE的深度学习模块的构建方法是,采用LeNet5卷积神经网络结构,模块共分为八层,前六层为特征提取部分,后两层为分类器部分,其中,特征提取层采用经典的卷积神经网络结构,分类器层采用全连接结构;由带标签数据库中的数据作为训练数据,采用CNN+LSE组合算法对深度学习模块进行训练,对于特征提取部分采用CNN方法进行训练,而对于分类器层,则采用LSE方法进行训练。7. method according to claim 6, it is characterized in that, in step (2), the construction method based on the deep learning module of CNN+LSE is, adopts LeNet5 convolutional neural network structure, module is divided into eight layers altogether, preceding The six layers are the feature extraction part, and the last two layers are the classifier part. Among them, the feature extraction layer adopts the classic convolutional neural network structure, and the classifier layer adopts the fully connected structure; the data in the labeled database is used as the training data, using CNN The +LSE combination algorithm trains the deep learning module, uses the CNN method for the feature extraction part, and uses the LSE method for the classifier layer. 8.根据权利要求7所述的方法,其特征在于,基于CNN+LSE的深度学习模块的训练过程如下:8. method according to claim 7, is characterized in that, the training process based on the deep learning module of CNN+LSE is as follows: 从带标签数据库中任取一个样本(Fi,j,k),对Fi,j首先进行灰度化操作,使其变为灰度图像,然后将灰度化后的关键帧Fi,j'输入到模块中,即输入数据为(Fi,j',k);对深度学习模块的两部分分别采用CNN与LSE的方法进行训练;其中,特征提取部分的参数训练方法如下:Randomly take a sample (F i,j ,k) from the labeled database, perform grayscale operation on F i,j first to make it a grayscale image, and then convert the grayscaled key frame F i, j ' is input into the module, that is, the input data is (F i,j ',k); the two parts of the deep learning module are trained by CNN and LSE respectively; among them, the parameter training method of the feature extraction part is as follows: (2-A1)初始化深度学习模块中特征提取部分的所有连接权重参数;(2-A1) Initialize all connection weight parameters of the feature extraction part in the deep learning module; (2-A2)计算输入关键帧相对应的实际输出标签Ok(2-A2) Calculate the actual output label O k corresponding to the input key frame; (2-A3)计算实际输出标签Ok与相应理想输出标签k的差值;(2-A3) Calculate the difference between the actual output label O k and the corresponding ideal output label k; (2-A4)权重学习:通过极小化误差的方法反向传播调节深度学习模块中特征提取部分的连接权重参数矩阵;(2-A4) Weight learning: adjust the connection weight parameter matrix of the feature extraction part in the deep learning module through backpropagation by minimizing the error; (2-A5)直至遍历所有视频数据的关键帧,参数训练完毕;(2-A5) Until the key frames of all video data are traversed, the parameter training is completed; 分类器部分的参数训练方法如下:The parameter training method of the classifier part is as follows: (2-B1)光栅化层与全连接层之间的连接权重与偏置随机生成,并将全连接层输出写为矩阵 (2-B1) The connection weight and bias between the rasterization layer and the fully connected layer are randomly generated, and the output of the fully connected layer is written as a matrix 其中G(·)为激活函数,ai为连接权重,bi为偏置,L为全连接层节点个数,N为所有关键帧的个数,xj为关键帧,i=1,2,…,L,j=1,2,…,N;Where G( ) is the activation function, a i is the connection weight, b i is the bias, L is the number of fully connected layer nodes, N is the number of all key frames, x j is the key frame, i=1,2 ,...,L,j=1,2,...,N; (2-B2)将相应关键帧的网络输出结果写为输出向量Y=[y1 y2 … yn]T,其中yj为第j个关键帧xj对应的输出标签;(2-B2) Write the network output result of the corresponding key frame as an output vector Y=[y 1 y 2 ... y n ] T , where y j is the output label corresponding to the jth key frame x j ; (2-B3)计算全连接层与输出层之间的输出权重β=PHY,其中P=(HTH)-1(2-B3) Calculate the output weight β=PHY between the fully connected layer and the output layer, where P=(H T H) −1 . 9.根据权利要求4所述的方法,其特征在于,步骤(3)中,待检违章结果数据库中的数据为经识别结果处理模块判断为违章的视频数据,其中的待检违章结果应当接受人工检查,然后将确认无误的信息导入违章数据库,并对误判的信息进行删除。9. method according to claim 4, it is characterized in that, in step (3), the data in the results database of violations to be checked is the video data that is judged as violations by the identification result processing module, and the results of violations to be checked wherein should accept Manual inspection, and then import the confirmed correct information into the violation database, and delete the misjudged information. 10.根据权利要求8所述的方法,其特征在于,步骤(3)中,对于是否存在远光灯违章行为的判断方法是:标签为远光灯的关键帧与其下个关键帧之间的时间间隔ΔT=j2-j1,若ΔT≥θ,则该车辆存在远光灯违章使用现象,其中,θ为违章时间阈值。10. The method according to claim 8, characterized in that, in step (3), the judging method for whether there is a high beam violation is: the label is the key frame of the high beam with the next keyframe The time interval between ΔT=j 2 -j 1 , if ΔT≥θ, the vehicle has illegal use of high beams, where θ is the violation time threshold.
CN201710156201.4A 2017-03-16 2017-03-16 Automobile high beam identification system and method based on video deep learning Active CN106934378B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201710156201.4A CN106934378B (en) 2017-03-16 2017-03-16 Automobile high beam identification system and method based on video deep learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710156201.4A CN106934378B (en) 2017-03-16 2017-03-16 Automobile high beam identification system and method based on video deep learning

Publications (2)

Publication Number Publication Date
CN106934378A true CN106934378A (en) 2017-07-07
CN106934378B CN106934378B (en) 2020-04-24

Family

ID=59432614

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710156201.4A Active CN106934378B (en) 2017-03-16 2017-03-16 Automobile high beam identification system and method based on video deep learning

Country Status (1)

Country Link
CN (1) CN106934378B (en)

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108229447A (en) * 2018-02-11 2018-06-29 陕西联森电子科技有限公司 A kind of high beam detection method based on video flowing
CN108921060A (en) * 2018-06-20 2018-11-30 安徽金赛弗信息技术有限公司 Motor vehicle based on deep learning does not use according to regulations clearance lamps intelligent identification Method
CN108932853A (en) * 2018-06-22 2018-12-04 安徽科力信息产业有限责任公司 A kind of method and device recording more motor vehicle illegal parking behaviors
CN108986476A (en) * 2018-08-07 2018-12-11 安徽金赛弗信息技术有限公司 Motor vehicle does not use according to regulations high beam recognition methods, system and storage medium
CN109191419A (en) * 2018-06-25 2019-01-11 国网山东省电力公司检修公司 The detection of real-time pressing plate and state recognition system and method based on machine learning
JP2019028616A (en) * 2017-07-27 2019-02-21 トヨタ自動車株式会社 Identification apparatus
CN109934106A (en) * 2019-01-30 2019-06-25 长视科技股份有限公司 A kind of user behavior analysis method based on video image deep learning
CN110046547A (en) * 2019-03-06 2019-07-23 深圳市麦谷科技有限公司 Report method, system, computer equipment and storage medium violating the regulations
CN111680638A (en) * 2020-06-11 2020-09-18 深圳北斗应用技术研究院有限公司 Passenger path identification method and passenger flow clearing method based on same

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103942751A (en) * 2014-04-28 2014-07-23 中央民族大学 Method for extracting video key frame
US20160019458A1 (en) * 2014-07-16 2016-01-21 Deep Learning Analytics, LLC Systems and methods for recognizing objects in radar imagery
CN105590102A (en) * 2015-12-30 2016-05-18 中通服公众信息产业股份有限公司 Front car face identification method based on deep learning
CN106407931A (en) * 2016-09-19 2017-02-15 杭州电子科技大学 Novel deep convolution neural network moving vehicle detection method

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103942751A (en) * 2014-04-28 2014-07-23 中央民族大学 Method for extracting video key frame
US20160019458A1 (en) * 2014-07-16 2016-01-21 Deep Learning Analytics, LLC Systems and methods for recognizing objects in radar imagery
CN105590102A (en) * 2015-12-30 2016-05-18 中通服公众信息产业股份有限公司 Front car face identification method based on deep learning
CN106407931A (en) * 2016-09-19 2017-02-15 杭州电子科技大学 Novel deep convolution neural network moving vehicle detection method

Cited By (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2019028616A (en) * 2017-07-27 2019-02-21 トヨタ自動車株式会社 Identification apparatus
CN108229447A (en) * 2018-02-11 2018-06-29 陕西联森电子科技有限公司 A kind of high beam detection method based on video flowing
CN108229447B (en) * 2018-02-11 2021-06-11 陕西联森电子科技有限公司 High beam light detection method based on video stream
CN108921060A (en) * 2018-06-20 2018-11-30 安徽金赛弗信息技术有限公司 Motor vehicle based on deep learning does not use according to regulations clearance lamps intelligent identification Method
CN108932853A (en) * 2018-06-22 2018-12-04 安徽科力信息产业有限责任公司 A kind of method and device recording more motor vehicle illegal parking behaviors
CN109191419A (en) * 2018-06-25 2019-01-11 国网山东省电力公司检修公司 The detection of real-time pressing plate and state recognition system and method based on machine learning
CN108986476A (en) * 2018-08-07 2018-12-11 安徽金赛弗信息技术有限公司 Motor vehicle does not use according to regulations high beam recognition methods, system and storage medium
CN109934106A (en) * 2019-01-30 2019-06-25 长视科技股份有限公司 A kind of user behavior analysis method based on video image deep learning
CN110046547A (en) * 2019-03-06 2019-07-23 深圳市麦谷科技有限公司 Report method, system, computer equipment and storage medium violating the regulations
CN111680638A (en) * 2020-06-11 2020-09-18 深圳北斗应用技术研究院有限公司 Passenger path identification method and passenger flow clearing method based on same
CN111680638B (en) * 2020-06-11 2020-12-29 深圳北斗应用技术研究院有限公司 Passenger path identification method and passenger flow clearing method based on same

Also Published As

Publication number Publication date
CN106934378B (en) 2020-04-24

Similar Documents

Publication Publication Date Title
CN106934378A (en) A kind of dazzle light identifying system and method based on video depth study
CN106935035B (en) Parking offense vehicle real-time detection method based on SSD neural network
CN111310583B (en) Vehicle abnormal behavior identification method based on improved long-term and short-term memory network
WO2024051296A1 (en) Method and apparatus for obstacle detection in complex weather
CN109657552B (en) Vehicle type recognition device and method for realizing cross-scene cold start based on transfer learning
CN103279756B (en) Vehicle detection based on integrated classifier analyzes system and determination method thereof
CN103810505A (en) Vehicle identification method and system based on multilayer descriptors
CN105868700A (en) Vehicle type recognition and tracking method and system based on monitoring video
CN107563372A (en) A kind of license plate locating method based on deep learning SSD frameworks
CN105956626A (en) Deep learning based vehicle license plate position insensitive vehicle license plate recognition method
CN109993138A (en) A kind of car plate detection and recognition methods and device
CN107316010A (en) A kind of method for recognizing preceding vehicle tail lights and judging its state
CN105354568A (en) Convolutional neural network based vehicle logo identification method
CN111209905B (en) Defect shielding license plate recognition method based on combination of deep learning and OCR technology
CN103279738B (en) Automatic identification method and system for vehicle logo
CN106446834A (en) An image-based vehicle identification method and device
CN111724408A (en) Validation experiment method of abnormal driving behavior algorithm model based on 5G communication
CN115019279A (en) A contextual feature fusion method based on MobileNet lightweight network
CN114463715A (en) Lane line detection method
CN106886757A (en) A kind of multiclass traffic lights detection method and system based on prior probability image
CN108960074A (en) Small size pedestrian target detection method based on deep learning
CN110555125A (en) Vehicle retrieval method based on local features
CN117333852A (en) Deep learning-based driver safety belt detection method
CN106919925A (en) A kind of Ford Motor's detection method based on Wavelet Entropy Yu artificial neural network
CN115116035A (en) A road traffic light recognition system and method based on neural network

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant