CN110674674A - A rotating target detection method based on YOLO V3 - Google Patents
A rotating target detection method based on YOLO V3 Download PDFInfo
- Publication number
- CN110674674A CN110674674A CN201910707178.2A CN201910707178A CN110674674A CN 110674674 A CN110674674 A CN 110674674A CN 201910707178 A CN201910707178 A CN 201910707178A CN 110674674 A CN110674674 A CN 110674674A
- Authority
- CN
- China
- Prior art keywords
- box
- detection target
- target
- loss function
- detection
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/10—Terrestrial scenes
- G06V20/13—Satellite images
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/26—Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
- G06V10/267—Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion by performing operations on regions, e.g. growing, shrinking or watersheds
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V2201/00—Indexing scheme relating to image or video recognition or understanding
- G06V2201/07—Target detection
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- General Engineering & Computer Science (AREA)
- Computing Systems (AREA)
- Multimedia (AREA)
- Molecular Biology (AREA)
- Computational Linguistics (AREA)
- Biophysics (AREA)
- Biomedical Technology (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Evolutionary Biology (AREA)
- Astronomy & Astrophysics (AREA)
- Remote Sensing (AREA)
- Bioinformatics & Computational Biology (AREA)
- Image Analysis (AREA)
Abstract
本发明公开了一种基于YOLO V3的旋转目标检测方法,本发明先通过对具有任意方向的遥感舰船训练集图像进行信息提取,使用改进的YOLO V3算法对遥感舰船训练集进行训练,最终可以在测试集中完成对具有任意角度的遥感舰船的检测。本发明对原YOLO V3算法的anchor box的产生、IOU的计算、损失函数的计算方法进行重新设计,增加训练集图像中检测目标的角度信息,使得可以在测试集中完成对具有任意角度的目标的检测。
The invention discloses a rotating target detection method based on YOLO V3. The invention first extracts information from images of remote sensing ship training set images with arbitrary directions, uses the improved YOLO V3 algorithm to train the remote sensing ship training set, and finally trains the remote sensing ship training set. The detection of remote sensing ships with arbitrary angles can be done in the test set. The invention redesigns the generation of the anchor box, the calculation of the IOU, and the calculation method of the loss function of the original YOLO V3 algorithm, and increases the angle information of the detection target in the training set image, so that the target with any angle can be completed in the test set. detection.
Description
技术领域technical field
本发明属于深度学习领域,涉及一种基于YOLO V3的旋转目标检测方法。The invention belongs to the field of deep learning, and relates to a rotating target detection method based on YOLO V3.
背景技术Background technique
目前,目标检测已被广泛应用于军事和民用等领域中。深度卷积神经网络可以利用目标数据集对要检测的目标进行自主学习,并完善自己的模型。YOLO V3是一种单步目标检测算法,此算法不需要使用区域候选网络(RPN)来提取候选目标信息,而是直接通过网络来产生目标的位置和类别信息,是一种端到端的目标检测算法。因此,单步目标检测算法具有更快的检测速度。At present, target detection has been widely used in military and civilian fields. A deep convolutional neural network can use the target dataset to autonomously learn the target to be detected and improve its own model. YOLO V3 is a single-step target detection algorithm. This algorithm does not need to use a regional candidate network (RPN) to extract candidate target information, but directly generates target location and category information through the network. It is an end-to-end target detection. algorithm. Therefore, the single-step object detection algorithm has a faster detection speed.
YOLO V3模型是采用分网格直接回归目标坐标和分类目标的方法进行目标检测,主要利用水平矩形边界框定义目标位置,通过边界框参数的回归对目标进行定位。这种方法对于目标对象是人,车等目标是足够准确的,而对于目标对象是文本,舰船等任意方向的目标是不适合的。The YOLO V3 model uses the method of directly regressing the target coordinates and classification targets by dividing the grid for target detection. It mainly uses the horizontal rectangular bounding box to define the target position, and locates the target through the regression of the bounding box parameters. This method is accurate enough for targets such as people and vehicles, but it is not suitable for targets with arbitrary directions such as text and ships.
发明内容SUMMARY OF THE INVENTION
本发明针对现有技术的不足,提供了一种基于YOLO V3的旋转目标检测方法,该方法通过对具有任意方向的遥感舰船训练集图像进行信息提取,使用改进的YOLO V3算法对遥感舰船训练集进行训练,最终可以在测试集中完成对具有任意角度的遥感舰船的检测。Aiming at the deficiencies of the prior art, the present invention provides a rotating target detection method based on YOLO V3. The method extracts information from images of remote sensing ship training sets with arbitrary directions, and uses the improved YOLO V3 algorithm to detect remote sensing ships. The training set is used for training, and finally the detection of remote sensing ships with arbitrary angles can be completed in the test set.
步骤(1)、对训练集图像中的检测目标进行位置信息,角度信息的提取;Step (1), extracting the position information and angle information of the detection target in the training set image;
训练集图像进行处理,将检测目标表示为一个5维向量{x,y,w,h,θ};其中:x表示检测目标的x轴方向的坐标,y表示检测目标的y轴方向的坐标,w表示检测目标的长度,h表示检测目标的宽度,θ表示检测目标的倾斜角度;The training set image is processed, and the detection target is represented as a 5-dimensional vector {x, y, w, h, θ}; where: x represents the coordinates of the detection target in the x-axis direction, and y represents the coordinates of the y-axis direction of the detection target. , w represents the length of the detection target, h represents the width of the detection target, and θ represents the inclination angle of the detection target;
步骤(2)、改进YOLO V3算法框架Step (2), improve the YOLO V3 algorithm framework
2.1锚点框的设置2.1 Setting the anchor box
使用带有旋转角度的三维度变量锚点,设置N个不同的角度来控制检测目标疑似区域的提取;设置多组锚点框长度和宽度,根据已知的训练集图像检测目标设置三组锚点框大小;根据卷积神经网络中卷积核的步幅和输入图像的大小会产生不同的尺度;在每个尺度上使用3个锚点来产生3个边界框,得到3*3*N个锚点框来与检测目标的位置区域进行拟合;Use a three-dimensional variable anchor point with a rotation angle to set N different angles to control the extraction of the suspected area of the detection target; set the length and width of multiple sets of anchor boxes, and set three sets of anchors according to the known training set image detection target Point box size; different scales are generated according to the stride of the convolution kernel in the convolutional neural network and the size of the input image; use 3 anchors at each scale to generate 3 bounding boxes, resulting in 3*3*N An anchor box is used to fit the location area of the detection target;
2.2损失函数的设置;2.2 Setting of loss function;
正负样本区域提取的划定准则为:锚点框与GT框的IOU大于0.7,同时与GT框的角度夹角小于π/12为正样本;锚点框与GT框的IOU小于0.3或是与GT框的IOU大于0.7,但是与GT框的角度夹角大于π/12为负样本;没有被归为正样本和负样本的在训练过程中不会被使用;GT框即检测目标的位置区域;IOU表示锚点框与GT框拟合的重叠面积大小;The delineation criteria for the extraction of positive and negative sample regions are: the IOU between the anchor box and the GT box is greater than 0.7, and the angle between the anchor box and the GT box is less than π/12, which is a positive sample; the IOU between the anchor box and the GT box is less than 0.3 or The IOU with the GT box is greater than 0.7, but the angle with the GT box is greater than π/12 is a negative sample; the samples that are not classified as positive samples and negative samples will not be used in the training process; the GT box is the location of the detection target area; IOU represents the overlapping area of the anchor box and the GT box fitted;
定义损失函数:Define the loss function:
Loss=Lcls(p,l)+Lreg(v*,v)+Lconf Loss=L cls (p,l)+L reg (v * ,v)+L conf
其中:Lcls(p,l)表示进行分类的损失函数,l表示类别标签,l=1表示选择的是目标物,l=0表示选择的是背景,如果选择背景则不存在回归的损失函数,当l=1时,采用交叉熵的方式计算真实类别与预测类别的损失,p为线性分类的概率;Lreg(v*,v)表示进行回归的损失函数;v*为步骤(1)中检测目标的信息(x,y,w,h,θ),v为步骤(2)中锚点框对应的预测检测目标的信息(x',y',w',h',θ'),采用均方差的方式计算预测值v与真实值v*的损失;Lconf表示置信度的损失函数,当IOU低于0.3时视为图片中不存在目标,当高于0.3时需要让预测的目标尽可能靠近真实值,没有目标的部分需要尽可能靠近背景真实值,采用交叉熵的方式来计算置信度的损失;Among them: L cls (p, l) represents the loss function for classification, l represents the category label, l=1 means the target is selected, l=0 means the background is selected, if the background is selected, there is no regression loss function , when l=1, the loss of the real category and the predicted category is calculated by means of cross entropy, p is the probability of linear classification; L reg (v * ,v) represents the loss function for regression; v * is step (1) The information of the detection target in (x, y, w, h, θ), v is the information of the predicted detection target corresponding to the anchor box in step (2) (x', y', w', h', θ') , the loss between the predicted value v and the real value v * is calculated by means of the mean square error; L conf represents the loss function of the confidence degree. When the IOU is lower than 0.3, it is considered that there is no target in the picture. When it is higher than 0.3, the predicted value needs to be The target is as close to the real value as possible, and the part without the target needs to be as close to the real value of the background as possible, and the loss of confidence is calculated by means of cross entropy;
步骤(3)、使用改进后的YOLO V3算法对训练集图像进行训练,迭代至损失函数不再下降为止,获取权重文件;Step (3), use the improved YOLO V3 algorithm to train the training set images, iterate until the loss function no longer decreases, and obtain the weight file;
步骤(4)、利用步骤(3)中得到的权重文件,对测试集图像进行测试。Step (4), use the weight file obtained in step (3) to test the test set images.
本发明的有益效果如下:The beneficial effects of the present invention are as follows:
本发明对原YOLO V3算法的anchor box的产生、IOU的计算、损失函数的计算方法进行重新设计,增加训练集图像中检测目标的角度信息,使得可以在测试集中完成对具有任意角度的目标的检测。The invention redesigns the generation of the anchor box, the calculation of the IOU, and the calculation method of the loss function of the original YOLO V3 algorithm, and increases the angle information of the detection target in the training set image, so that the target with any angle can be completed in the test set. detection.
附图说明Description of drawings
图1为本发明的流程图。FIG. 1 is a flow chart of the present invention.
具体实施方式Detailed ways
下面结合图1对本发明做进一步的分析。The present invention is further analyzed below in conjunction with FIG. 1 .
本实验将一组采集的舰船目标图像划分为训练集和测试集。在基于YOLO V3的旋转目标检测任务中的具体步骤如下:In this experiment, a group of collected ship target images are divided into training set and test set. The specific steps in the rotating target detection task based on YOLO V3 are as follows:
步骤(1)、对训练集图像中的检测目标进行位置信息,角度信息的提取。Step (1), extracting the position information and angle information of the detection target in the training set image.
旋转目标检测方法相比于原YOLO V3算法需要增加一个角度信息。对训练集图像进行处理,将检测目标表示为一个5维向量{x,y,w,h,θ}。其中x表示检测目标的x轴方向的坐标,y表示检测目标的y轴方向的坐标,w表示检测目标的长度,w表示检测目标的宽度,θ表示检测目标的倾斜角度。Compared with the original YOLO V3 algorithm, the rotating target detection method needs to add an angle information. The training set images are processed to represent the detection target as a 5-dimensional vector {x, y, w, h, θ}. Where x represents the x-axis coordinate of the detection target, y represents the y-axis coordinate of the detection target, w represents the length of the detection target, w represents the width of the detection target, and θ represents the inclination angle of the detection target.
步骤(2)、改进YOLO V3算法框架。Step (2), improve the YOLO V3 algorithm framework.
2.1anchor box(锚点框)的设置。2.1 Anchor box (anchor box) settings.
使用带有旋转角度的三维度变量锚点,设置多个不同的角度来控制检测目标疑似区域的提取,可以使用6个不同的角度:-π/6,0,π/6,π/3,π/2,2π/3。设置多组锚点框长度和宽度,可以根据已知的训练集图像检测目标设置三组anchor box大小:[(10,13),(16,30),(33,23)],[(30,61),(62,45),(59,119)],[(116,90),(156,198),(373,326)]。根据卷积神经网络中卷积核的步幅和输入图像的大小会产生不同的尺度,原YOLO V3特征图步幅分别使32、16、8,输入图像的大小416*416,产生的尺度为13*13、26*26、52*52。在每个尺度上使用3个anchor来产生3个边界框,这样就可以的到3*3*6=54个anchor box来与检测目标的位置区域进行拟合Use a three-dimensional variable anchor point with a rotation angle to set multiple different angles to control the extraction of the suspected area of the detection target. You can use 6 different angles: -π/6, 0, π/6, π/3, π/2, 2π/3. To set the length and width of multiple sets of anchor boxes, three sets of anchor box sizes can be set according to the known training set image detection targets: [(10, 13), (16, 30), (33, 23)], [(30 , 61), (62, 45), (59, 119)], [(116, 90), (156, 198), (373, 326)]. According to the stride of the convolution kernel in the convolutional neural network and the size of the input image, different scales will be generated. The stride of the original YOLO V3 feature map is 32, 16, and 8 respectively, and the size of the input image is 416*416. The resulting scale is 13*13, 26*26, 52*52. Use 3 anchors at each scale to generate 3 bounding boxes, so that 3*3*6=54 anchor boxes can be fitted to the location area of the detection target
2.2损失函数的设置。2.2 Setting of loss function.
正负样本区域提取的划定准则为:anchor box与GT框(检测目标的位置区域)的IOU(anchor box与GT框拟合的重叠面积大小)大于0.7,同时与GT框的角度夹角小于π/12为正样本;anchor box与GT框的IOU小于0.3或是与GT框的IOU大于0.7,但是与GT框的角度夹角大于π/12为负样本;没有被归为正样本和负样本的在训练过程中不会被使用。The delineation criteria for the extraction of positive and negative sample regions are: the IOU (the size of the overlap area fitted by the anchor box and the GT box) of the anchor box and the GT box (the location area of the detection target) is greater than 0.7, and the angle with the GT box is less than 0.7. π/12 is a positive sample; the IOU of the anchor box and the GT box is less than 0.3 or the IOU of the GT box is greater than 0.7, but the angle with the GT box is greater than π/12 is a negative sample; not classified as positive samples and negative samples Samples are not used during training.
定义损失函数:Define the loss function:
Loss=Lcls(p,l)+Lreg(v*,v)+Lconf Loss=L cls (p,l)+L reg (v * ,v)+L conf
其中:Lcls(p,l)表示进行分类的损失函数,l表示类别标签(l=1表示选择的是目标物,l=0表示选择的是背景,如果选择背景则不存在回归的损失函数),当l=1时,采用交叉熵的方式计算真实类别与预测类别的损失(p为线性分类的概率);Lreg(v*,v)表示进行回归的损失函数。v*为步骤(1)中检测目标的信息(x,y,w,h,θ),v为步骤(2)中anchorbox对应的预测检测目标的信息(x',y',w',h',θ'),采用均方差的方式计算预测值v与真实值v*的损失;Lconf表示置信度的损失函数。当IOU低于0.3时视为图片中不存在目标,当高于0.3时需要让预测的目标尽可能靠近真实值,没有目标的部分需要尽可能靠近背景真实值,采用交叉熵的方式来计算置信度的损失。Among them: L cls (p, l) represents the loss function for classification, l represents the category label (l=1 means the target is selected, l=0 means the background is selected, if the background is selected, there is no regression loss function ), when l=1, the loss between the real category and the predicted category is calculated by means of cross entropy (p is the probability of linear classification); L reg (v * ,v) represents the loss function for regression. v * is the information of the detection target in step (1) (x, y, w, h, θ), v is the information of the predicted detection target corresponding to the anchorbox in step (2) (x', y', w', h) ', θ'), the loss between the predicted value v and the true value v * is calculated by means of the mean square error; L conf represents the loss function of the confidence. When the IOU is lower than 0.3, it is considered that there is no target in the picture. When it is higher than 0.3, the predicted target needs to be as close to the real value as possible, and the part without the target needs to be as close to the real value of the background as possible. Cross entropy is used to calculate the confidence. degree loss.
步骤(3)、使用改进后的YOLO V3算法对训练集图像进行训练,迭代至损失函数不再下降为止,获取权重文件。Step (3): Use the improved YOLO V3 algorithm to train the training set images, iterate until the loss function no longer decreases, and obtain the weight file.
步骤(4)、利用步骤(3)中得到的权重文件,对测试集图像进行测试。Step (4), use the weight file obtained in step (3) to test the test set images.
上述实施例并非是对于本发明的限制,本发明并非仅限于上述实施例,只要符合本发明要求,均属于本发明的保护范围。The above-mentioned embodiments are not intended to limit the present invention, and the present invention is not limited to the above-mentioned embodiments. As long as the requirements of the present invention are met, they all belong to the protection scope of the present invention.
Claims (1)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910707178.2A CN110674674A (en) | 2019-08-01 | 2019-08-01 | A rotating target detection method based on YOLO V3 |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910707178.2A CN110674674A (en) | 2019-08-01 | 2019-08-01 | A rotating target detection method based on YOLO V3 |
Publications (1)
Publication Number | Publication Date |
---|---|
CN110674674A true CN110674674A (en) | 2020-01-10 |
Family
ID=69068473
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201910707178.2A Pending CN110674674A (en) | 2019-08-01 | 2019-08-01 | A rotating target detection method based on YOLO V3 |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN110674674A (en) |
Cited By (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111462235A (en) * | 2020-03-31 | 2020-07-28 | 武汉卓目科技有限公司 | Inclined target detection method and device based on yolo v3 algorithm and storage medium |
CN112364805A (en) * | 2020-11-21 | 2021-02-12 | 西安交通大学 | Rotary palm image detection method |
CN112488113A (en) * | 2020-11-06 | 2021-03-12 | 杭州电子科技大学 | Remote sensing image rotating ship target detection method based on local straight line matching |
CN112580435A (en) * | 2020-11-25 | 2021-03-30 | 厦门美图之家科技有限公司 | Face positioning method, face model training and detecting method and device |
CN112966587A (en) * | 2021-03-02 | 2021-06-15 | 北京百度网讯科技有限公司 | Training method of target detection model, target detection method and related equipment |
CN113033303A (en) * | 2021-02-09 | 2021-06-25 | 北京工业大学 | Method for realizing SAR image rotating ship detection based on RCIoU loss |
CN113723217A (en) * | 2021-08-09 | 2021-11-30 | 南京邮电大学 | Object intelligent detection method and system based on yolo improvement |
CN113920436A (en) * | 2021-11-22 | 2022-01-11 | 江苏科技大学 | A remote sensing image marine ship identification system and method based on improved YOLOv4 algorithm |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109492561A (en) * | 2018-10-29 | 2019-03-19 | 北京遥感设备研究所 | A kind of remote sensing image Ship Detection based on improved YOLO V2 model |
CN109978036A (en) * | 2019-03-11 | 2019-07-05 | 华瑞新智科技(北京)有限公司 | Target detection deep learning model training method and object detection method |
-
2019
- 2019-08-01 CN CN201910707178.2A patent/CN110674674A/en active Pending
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109492561A (en) * | 2018-10-29 | 2019-03-19 | 北京遥感设备研究所 | A kind of remote sensing image Ship Detection based on improved YOLO V2 model |
CN109978036A (en) * | 2019-03-11 | 2019-07-05 | 华瑞新智科技(北京)有限公司 | Target detection deep learning model training method and object detection method |
Non-Patent Citations (2)
Title |
---|
JIANQI MA等: "Arbitrary-Oriented Scene Text Detection via Rotation Proposals", 《IEEE TRANSACTIONS ON MULTIMEDIA》 * |
JOSEPH REDMON等: "YOLOv3: An Incremental Improvement", 《ARXIV》 * |
Cited By (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111462235A (en) * | 2020-03-31 | 2020-07-28 | 武汉卓目科技有限公司 | Inclined target detection method and device based on yolo v3 algorithm and storage medium |
CN112488113A (en) * | 2020-11-06 | 2021-03-12 | 杭州电子科技大学 | Remote sensing image rotating ship target detection method based on local straight line matching |
CN112364805A (en) * | 2020-11-21 | 2021-02-12 | 西安交通大学 | Rotary palm image detection method |
CN112580435A (en) * | 2020-11-25 | 2021-03-30 | 厦门美图之家科技有限公司 | Face positioning method, face model training and detecting method and device |
CN112580435B (en) * | 2020-11-25 | 2024-05-31 | 厦门美图之家科技有限公司 | Face positioning method, face model training and detecting method and device |
CN113033303A (en) * | 2021-02-09 | 2021-06-25 | 北京工业大学 | Method for realizing SAR image rotating ship detection based on RCIoU loss |
CN113033303B (en) * | 2021-02-09 | 2024-03-01 | 北京工业大学 | SAR image rotation ship detection implementation method based on RCIoU loss |
CN112966587A (en) * | 2021-03-02 | 2021-06-15 | 北京百度网讯科技有限公司 | Training method of target detection model, target detection method and related equipment |
CN112966587B (en) * | 2021-03-02 | 2022-12-20 | 北京百度网讯科技有限公司 | Training method of target detection model, target detection method and related equipment |
CN113723217A (en) * | 2021-08-09 | 2021-11-30 | 南京邮电大学 | Object intelligent detection method and system based on yolo improvement |
CN113920436A (en) * | 2021-11-22 | 2022-01-11 | 江苏科技大学 | A remote sensing image marine ship identification system and method based on improved YOLOv4 algorithm |
CN113920436B (en) * | 2021-11-22 | 2024-12-03 | 江苏科技大学 | A remote sensing image marine ship recognition system and method based on improved YOLOv4 algorithm |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110674674A (en) | A rotating target detection method based on YOLO V3 | |
CN101887586B (en) | Self-adaptive angular-point detection method based on image contour sharpness | |
CN103390164B (en) | Method for checking object based on depth image and its realize device | |
CN103714541B (en) | A Method for Identifying and Locating Buildings Using Mountain Contour Area Constraints | |
CN107657279A (en) | A kind of remote sensing target detection method based on a small amount of sample | |
CN106228125A (en) | Method for detecting lane lines based on integrated study cascade classifier | |
CN106384079A (en) | RGB-D information based real-time pedestrian tracking method | |
CN105426870A (en) | Face key point positioning method and device | |
CN103697883B (en) | A kind of aircraft horizontal attitude defining method based on skyline imaging | |
CN104899590A (en) | Visual target tracking method and system for unmanned aerial vehicle | |
CN107463890A (en) | A kind of Foregut fermenters and tracking based on monocular forward sight camera | |
CN107239740A (en) | A kind of SAR image automatic target recognition method of multi-source Fusion Features | |
CN104537689B (en) | Method for tracking target based on local contrast conspicuousness union feature | |
CN107507170A (en) | A kind of airfield runway crack detection method based on multi-scale image information fusion | |
CN103886325A (en) | Cyclic matrix video tracking method with partition | |
CN115115672B (en) | Dynamic Visual SLAM Method Based on Object Detection and Velocity Constraints of Feature Points | |
CN114926747A (en) | Remote sensing image directional target detection method based on multi-feature aggregation and interaction | |
CN102945374A (en) | Method for automatically detecting civil aircraft in high-resolution remote sensing image | |
CN110826485B (en) | Target detection method and system for remote sensing image | |
Mseddi et al. | YOLOv5 based visual localization for autonomous vehicles | |
CN106503663A (en) | A kind of signal lighties duration detection method based on deep learning | |
CN104200199A (en) | TOF (Time of Flight) camera based bad driving behavior detection method | |
CN104268574A (en) | A SAR Image Change Detection Method Based on Genetic Kernel Fuzzy Clustering | |
CN105335688B (en) | A kind of aircraft model recognition methods of view-based access control model image | |
CN112084860A (en) | Target object detection method and device and thermal power plant detection method and device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
WD01 | Invention patent application deemed withdrawn after publication | ||
WD01 | Invention patent application deemed withdrawn after publication |
Application publication date: 20200110 |