CN106886754B - Object identification method and system under a kind of three-dimensional scenic based on tri patch - Google Patents
Object identification method and system under a kind of three-dimensional scenic based on tri patch Download PDFInfo
- Publication number
- CN106886754B CN106886754B CN201710034186.6A CN201710034186A CN106886754B CN 106886754 B CN106886754 B CN 106886754B CN 201710034186 A CN201710034186 A CN 201710034186A CN 106886754 B CN106886754 B CN 106886754B
- Authority
- CN
- China
- Prior art keywords
- triangular
- dimensional
- eigenvalues
- feature vector
- eigenvectors
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/20—Scenes; Scene-specific elements in augmented reality scenes
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
- G06F18/2411—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on the proximity to a decision surface, e.g. support vector machines
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T17/00—Three dimensional [3D] modelling, e.g. data description of 3D objects
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- Life Sciences & Earth Sciences (AREA)
- Software Systems (AREA)
- Geometry (AREA)
- Computer Graphics (AREA)
- Artificial Intelligence (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Evolutionary Biology (AREA)
- Evolutionary Computation (AREA)
- General Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Image Analysis (AREA)
Abstract
本发明公开了一种基于三角面片的三维场景下的物体识别方法。该方法在传统的二维图像识别方法的基础上,考虑到三维场景重建信息的三角网格存储方式,将三维图像进行三角面片划分,经过颜色邻域直方图特征统计,将空间金字塔多尺度信息和邻域空间信息进行融合,用支撑向量机进行训练,获得分类模型对三维场景中的草地面片进行识别。本发明还提供了一种基于三角面片的三维场景下的物体识别系统,本发明技术方案解决了三维重建场景中识别问题中识别区域难确定的问题及图像中局部信息不能完全表达图像特征的问题,对于表面重建和场景理解等意义重大。
The invention discloses an object recognition method in a three-dimensional scene based on triangular facets. Based on the traditional two-dimensional image recognition method, this method takes into account the triangular grid storage method of the three-dimensional scene reconstruction information, divides the three-dimensional image into triangular patches, and uses the color neighborhood histogram feature statistics to divide the spatial pyramid into multi-scale. The information and the neighborhood space information are fused, and the support vector machine is used for training to obtain a classification model to identify the grass patches in the three-dimensional scene. The present invention also provides an object recognition system in a three-dimensional scene based on triangular facets. The technical solution of the present invention solves the problem that the recognition area is difficult to determine in the recognition problem in the three-dimensional reconstruction scene, and the local information in the image cannot fully express the image features. The problem is of great significance for surface reconstruction and scene understanding.
Description
技术领域technical field
本发明属于计算机视觉技术领域,更具体地,涉及一种基于三角面片的三维场景下的物体识别方法及系统。The invention belongs to the technical field of computer vision, and more particularly, relates to an object recognition method and system in a three-dimensional scene based on triangular facets.
背景技术Background technique
随着人类社会的发展,图像识别已经迅速发展成为一项极为重要的科技手段,其研究目标是赋予计算机人类的视觉能力,使其通过图像认知周边环境信息,包括识别环境中三维目标的几何形状、位置和姿态等。图像识别需要综合运用计算机科学、模式识别、机器视觉及图像理解等学科,并随着这些学科的发展而前进。With the development of human society, image recognition has rapidly developed into an extremely important scientific and technological means. Its research goal is to give computers the visual ability of human beings, so that they can recognize the surrounding environment information through images, including recognizing the geometry of three-dimensional objects in the environment. shape, position, and attitude, etc. Image recognition requires a combination of disciplines such as computer science, pattern recognition, machine vision, and image understanding, and advances with the development of these disciplines.
一般来说,图像识别系统由以下几个部分组成。第一部分是图像信息的获取,对图像识别来说就是把图片、底片和文字图形等用光电扫描设备转换为电信号以备后续处理。第二部分是图像的预处理,主要目的是消除无关特征并加强感兴趣特征。第三步是特征提取,常用的特征包括颜色,纹理,尺度不变特征变换(Scale-invariant featuretransform,SIFT),梯度直方图特征(Histogram of Oriented Gradient,HOG)等。第四部分是判决或分类。其中训练方法可以采用支持向量机方法(Support Vector Machine,SVM)、改进的迭代训练方法(Adaptive Boostint,AdaBoost)等。Generally speaking, an image recognition system consists of the following parts. The first part is the acquisition of image information. For image recognition, it is to convert pictures, negatives and text graphics into electrical signals for subsequent processing. The second part is image preprocessing, the main purpose is to eliminate irrelevant features and strengthen interesting features. The third step is feature extraction. Commonly used features include color, texture, Scale-invariant feature transform (SIFT), Histogram of Oriented Gradient (HOG), etc. The fourth part is judgment or classification. The training method may adopt a support vector machine method (Support Vector Machine, SVM), an improved iterative training method (Adaptive Boostint, AdaBoost) and the like.
目前,最主流的三种三维物体识别研究思路是:1)基于模型或几何的方法;2)基于外观和视图的方法;3)基于局部特征匹配的方法。At present, the three most mainstream 3D object recognition research ideas are: 1) method based on model or geometry; 2) method based on appearance and view; 3) method based on local feature matching.
基于模型或几何的方法的原理就是利用传感器获得真实目标的三维信息并对信息进行分析处理,得到一种表面、边界及连接关系的描述。这种方法优点是比较直观和易于理解,但是一般使用的算法的运算量都较大,并且需要人工借助计算机辅助设计(ComputerAided Design,CAD)等软件产生模型。The principle of model-based or geometry-based methods is to use sensors to obtain three-dimensional information of real targets and analyze and process the information to obtain a description of surfaces, boundaries and connections. The advantage of this method is that it is relatively intuitive and easy to understand, but the computational load of the commonly used algorithm is relatively large, and the model needs to be manually generated by software such as Computer Aided Design (CAD).
基于视图的方法通过视觉相似性来识别物体,识别系统设计相对简单,无需显式地计算物体三维模型。该方法一般分为两个步骤:首先,通过不同光照条件和三维物体在二维图像中呈现出的不同姿态,来自动地学习物体的表示或训练系统;然后,在一幅未知的二维图像中判断是否存在目标物。该方法的限制条件是感兴趣的对象必须与背景能够较好的区分开来,因此对物体间的重叠较为敏感,且需要较好的图像分割。但当物体的几何建模很困难或根本不可能得到时,则可利用基于视图的方法来识别三维物体。View-based methods identify objects by visual similarity, and the design of the recognition system is relatively simple, without the need to explicitly calculate the 3D model of the object. The method is generally divided into two steps: first, automatically learn the representation of the object or train the system through different lighting conditions and different poses of the 3D object in the 2D image; then, in an unknown 2D image to determine whether there is a target. The limitation of this method is that the object of interest must be well differentiated from the background, so it is sensitive to overlap between objects and requires better image segmentation. But when geometric modeling of objects is difficult or impossible, view-based methods can be used to identify three-dimensional objects.
基于局部特征匹配的方法主要包括点、线和面的局部特征,它的优点是,因为视角改变引起的物体外观的形变,全局看来尽管非常复杂,但在局部的尺度上可通过简单的变化来估计;同时因为无需所有的局部特征得到匹配,这种方法在物体有重叠和复杂背景情况下都有较好的稳健性。因为建立了区域间的相似性,物体识别也做到了局部化;二维图像识别中在图像的预处理阶段往往通过超像素分割成不同的区域,然后对区域进行特征提取并进行分类,而三维场景由于其结构复杂性,不能通过划分超像素区域进行识别。The method based on local feature matching mainly includes the local features of points, lines and surfaces. At the same time, because all local features do not need to be matched, this method has better robustness in the case of overlapping objects and complex backgrounds. Because the similarity between regions is established, object recognition is also localized; in two-dimensional image recognition, in the preprocessing stage of the image, it is often divided into different regions by superpixels, and then the regions are extracted and classified. Scenes cannot be identified by dividing superpixel regions due to their structural complexity.
发明内容SUMMARY OF THE INVENTION
针对现有技术的以上缺陷或改进需求,本发明提供了一种基于三角面片的三维场景下的物体识别方法及系统,其目的在于将待识别三维场景划分为不同大小的三角面片,利用三角面片中的目标的特征进行识别,由此将二维图像识别方法进行扩展,提出一种基于三角面片的草地识别方法。In view of the above defects or improvement requirements of the prior art, the present invention provides a method and system for object recognition in a three-dimensional scene based on triangular patches, the purpose of which is to divide the three-dimensional scene to be identified into triangular patches of different sizes, using The feature of the target in the triangular patch is identified, thus the two-dimensional image recognition method is extended, and a grass recognition method based on the triangular patch is proposed.
为实现上述目的,按照本发明的一个方面,提供了一种基于三角面片的三维场景下的物体识别方法,该方法包括以下步骤:In order to achieve the above object, according to an aspect of the present invention, a method for object recognition in a three-dimensional scene based on a triangular patch is provided, and the method includes the following steps:
(1)将待识别三维场景进行三维重建,按三维场景表面的三角网格划分三角面片;(1) 3D reconstruction is performed on the 3D scene to be identified, and triangular patches are divided according to the triangular mesh of the surface of the 3D scene;
(2)提取所有三角面片单元的HSI特征;(2) Extract the HSI features of all triangular patch units;
(3)根据选定三角面片和其共享边的相邻三角面片的H特征值和S特征值进行直方图统计,将统计数据放入特征向量中,将三角面片的特征向量和其相邻三角面片的特征向量进行叠加(说明书中详细说明),叠加结果作为选定三角面片的特征向量;(3) Perform histogram statistics according to the H eigenvalues and S eigenvalues of the selected triangular facet and its adjacent triangular facets with shared edges, put the statistical data into the eigenvectors, and combine the eigenvectors of the triangular facets with its eigenvalues. The eigenvectors of adjacent triangular patches are superimposed (detailed in the specification), and the superposition result is used as the eigenvectors of the selected triangular patch;
(4)将选定三角面片的特征向量输入到SVM识别模型中,获取识别结果。(4) Input the feature vector of the selected triangular facet into the SVM recognition model to obtain the recognition result.
进一步地,所述步骤(3)中进行直方图具体设置为:Further, in the step (3), the histogram is specifically set as:
设置H特征值直方图和S特征值直方图的组数为8,组距分别为H特征值取值范围的1/8和S特征值取值范围的1/8。Set the number of groups of H eigenvalue histogram and S eigenvalue histogram to 8, and the group spacing is 1/8 of the value range of H eigenvalue and 1/8 of the value range of S eigenvalue.
进一步地,所述步骤(3)中将统计数据放入特征向量中具体为:Further, putting the statistical data into the feature vector in the step (3) is specifically:
将三角面片的H特征值进行直方图统计得到一组8维向量,将三角面片的S特征值进行直方图统计也得到一组8维向量,将H特征值的8维向量和S特征值的8维向量首尾相连组成16维的特征向量。Perform histogram statistics on the H eigenvalues of the triangular facets to obtain a set of 8-dimensional vectors, and perform histogram statistics on the S eigenvalues of the triangular facets to obtain a set of 8-dimensional vectors. The 8-dimensional vectors of the H feature values and the S feature The 8-dimensional vector of values is connected end to end to form a 16-dimensional feature vector.
进一步地,所述步骤(3)中特征向量进行叠加具体为:Further, in the step (3), the feature vector is superimposed as follows:
其中,{H11,H12,H13,...,H18,S11,S12,...,S18}和{H21,H22,H23,...,H28,S21,S22,...,S28}表示16维的特征向量。Among them, {H 11 , H 12 , H 13 ,...,H 18 ,S 11 ,S 12 ,...,S 18 } and {H 21 ,H 22 ,H 23 ,...,H 28 , S 21 , S 22 ,...,S 28 } represent 16-dimensional feature vectors.
进一步地,所述步骤(4)中SVM识别模型的训练过程具体为:Further, the training process of the SVM recognition model in the step (4) is specifically:
(41)收集设定张正样本和设定张负样本作为样本集,正样本选取全为待识别物体的二维图片,负样本选取不含待识别物体的二维图片;(41) collecting and setting positive samples and setting negative samples as sample sets, the positive samples are selected as two-dimensional pictures of objects to be identified, and the negative samples are selected as two-dimensional pictures without objects to be identified;
(42)将所有样本的尺寸调整为一致,选择多个预设的尺度对样本进行三角面片的划分,提取所有三角面片的HSI特征;(42) adjusting the sizes of all samples to be consistent, selecting a plurality of preset scales to divide the samples into triangular patches, and extracting the HSI features of all triangular patches;
(43)根据选定三角面片和其共享边的相邻三角面片的H特征值和S特征值进行直方图统计,将统计数据放入特征向量中,将三角面片的特征向量和其相邻三角面片的特征向量进行叠加,叠加结果作为选定三角面片的特征向量;(43) Perform histogram statistics according to the H eigenvalues and S eigenvalues of the selected triangular facet and its adjacent triangular facets with shared edges, put the statistical data into the eigenvectors, and combine the eigenvectors of the triangular facets with its eigenvalues. The eigenvectors of adjacent triangular patches are superimposed, and the superposition result is used as the eigenvectors of the selected triangular patch;
(44)将所有样本的特征向量输入到SVM进行训练;SVM的决策函数为:(44) Input the feature vectors of all samples into SVM for training; the decision function of SVM is:
其中,L表示分类总数,取值2;ci表示函数系数,ci∈R;γ表示交叉校验的参数,γ∈R+;表示被选用为训练的支撑向量;h0表示训练的非支撑向量;采用直方图的卡方距离作为核函数,定义为:Among them, L represents the total number of classifications, and the value is 2; ci represents the function coefficient, ci ∈ R; γ represents the parameter of cross-checking, γ ∈ R + ; represents the support vector selected for training; h 0 represents the non-support vector for training; the chi-square distance of the histogram is used As a kernel function, it is defined as:
其中,N表示样本集中特征向量总数量;表示第k个特征向量对应的支撑向量;h0(k)表示第k个特征向量;Among them, N represents the total number of feature vectors in the sample set; represents the support vector corresponding to the kth eigenvector; h 0 (k) represents the kth eigenvector;
(45)训练后得到SVM识别模型参数,构建SVM识别模型。(45) After training, the parameters of the SVM recognition model are obtained, and the SVM recognition model is constructed.
按照本发明的另一方面,提供了一种基于三角面片的三维场景下的物体识别系统,改系统包括以下模块:According to another aspect of the present invention, an object recognition system based on a triangular patch in a three-dimensional scene is provided, and the system includes the following modules:
三角面片划分模块,用于将待识别三维场景进行三维重建,按三维场景表面的三角网格划分三角面片;The triangular patch division module is used to perform 3D reconstruction of the 3D scene to be identified, and divide triangular patches according to the triangular mesh of the surface of the 3D scene;
HSI特征提取模块,用于提取所有三角面片单元的HSI特征;The HSI feature extraction module is used to extract the HSI features of all triangular patch units;
特征向量计算模块,用于根据选定三角面片和其共享边的相邻三角面片的H特征值和S特征值进行直方图统计,将统计数据放入特征向量中,将三角面片的特征向量和其相邻三角面片的特征向量进行叠加,叠加结果作为选定三角面片的特征向量;The eigenvector calculation module is used to perform histogram statistics according to the H eigenvalues and S eigenvalues of the selected triangular facet and its adjacent triangular facets with shared edges, put the statistical data into the eigenvectors, and calculate the The eigenvectors and the eigenvectors of its adjacent triangular patches are superimposed, and the superimposed result is used as the eigenvectors of the selected triangular patch;
识别模块,用于将选定三角面片的特征向量输入到SVM识别模型中,获取识别结果。The recognition module is used to input the feature vector of the selected triangular facet into the SVM recognition model to obtain the recognition result.
进一步地,所述特征向量计算模块中进行直方图具体设置为:Further, the specific setting of the histogram in the feature vector calculation module is:
设置H特征值直方图和S特征值直方图的组数为8,组距分别为H特征值取值范围的1/8和S特征值取值范围的1/8。Set the number of groups of H eigenvalue histogram and S eigenvalue histogram to 8, and the group spacing is 1/8 of the value range of H eigenvalue and 1/8 of the value range of S eigenvalue.
进一步地,所述特征向量计算模块中将统计数据放入特征向量中具体为:Further, putting the statistical data into the feature vector in the feature vector calculation module is specifically:
将三角面片的H特征值进行直方图统计得到一组8维向量,将三角面片的S特征值进行直方图统计也得到一组8维向量,将H特征值的8维向量和S特征值的8维向量首尾相连组成16维的特征向量。Perform histogram statistics on the H eigenvalues of the triangular facets to obtain a set of 8-dimensional vectors, and perform histogram statistics on the S eigenvalues of the triangular facets to obtain a set of 8-dimensional vectors. The 8-dimensional vectors of the H feature values and the S feature The 8-dimensional vector of values is connected end to end to form a 16-dimensional feature vector.
进一步地,所述步骤(3)中特征向量进行叠加具体为:Further, in the step (3), the feature vector is superimposed as follows:
其中,{H11,H12,H13,...,H18,S11,S12,...,S18}和{H21,H22,H23,...,H28,S21,S22,...,S28}表示16维的特征向量。Among them, {H 11 , H 12 , H 13 ,...,H 18 ,S 11 ,S 12 ,...,S 18 } and {H 21 ,H 22 ,H 23 ,...,H 28 , S 21 , S 22 ,...,S 28 } represent 16-dimensional feature vectors.
进一步地,所述识别模块中SVM识别模型的训练系统包括以下单元:Further, the training system of the SVM recognition model in the recognition module includes the following units:
样本采集单元,用于收集多张正样本和多张负样本作为样本集,正样本选取全为待识别物体的二维图片,负样本选取不含待识别物体的二维图片;The sample collection unit is used to collect multiple positive samples and multiple negative samples as a sample set, the positive samples are selected as two-dimensional pictures of the objects to be recognized, and the negative samples are selected as two-dimensional pictures without the objects to be identified;
三角面片划分单元,用于将所有样本的尺寸调整为一致,选择多个预设的尺度对样本进行三角面片的划分,提取所有三角面片的HSI特征;The triangular patch division unit is used to adjust the size of all samples to be consistent, select multiple preset scales to divide the samples into triangular patches, and extract the HSI features of all triangular patches;
特征向量计算单元,用于根据选定三角面片和其共享边的相邻三角面片的H特征值和S特征值进行直方图统计,将统计数据放入特征向量中,将三角面片的特征向量和其相邻三角面片的特征向量进行叠加,叠加结果作为选定三角面片的特征向量;The eigenvector calculation unit is used to perform histogram statistics according to the H eigenvalues and S eigenvalues of the selected triangular facet and its adjacent triangular facets with shared edges, put the statistical data into the eigenvectors, and calculate the The eigenvectors and the eigenvectors of its adjacent triangular patches are superimposed, and the superimposed result is used as the eigenvectors of the selected triangular patch;
训练单元,用于将所有样本的特征向量输入到SVM进行训练;SVM的决策函数为:The training unit is used to input the feature vectors of all samples into the SVM for training; the decision function of the SVM is:
其中,L表示分类总数,取值2;ci表示函数系数,ci∈R;γ表示交叉校验的参数,γ∈R+;表示被选用为训练的支撑向量;h0表示训练的非支撑向量;采用直方图的卡方距离作为核函数,定义为:Among them, L represents the total number of classifications, and the value is 2; ci represents the function coefficient, ci ∈ R; γ represents the parameter of cross-checking, γ ∈ R + ; represents the support vector selected for training; h 0 represents the non-support vector for training; the chi-square distance of the histogram is used As a kernel function, it is defined as:
其中,N表示样本集中特征向量总数量;表示第k个特征向量对应的支撑向量;h0(k)表示第k个特征向量;Among them, N represents the total number of feature vectors in the sample set; represents the support vector corresponding to the kth eigenvector; h 0 (k) represents the kth eigenvector;
模型构建单元,用于训练后得到SVM识别模型参数,构建SVM识别模型。The model building unit is used to obtain the parameters of the SVM recognition model after training, and construct the SVM recognition model.
总体而言,通过本发明所构思的以上技术方案与现有技术相比,具有以下技术特征及有益效果:In general, compared with the prior art, the above technical solutions conceived by the present invention have the following technical features and beneficial effects:
(1)本发明技术方案利用了重建出的三维场景中的三角网格,将三维场景划分为多个二维三角面片,避免在三维场景中进行超像素提取过于复杂的问题;(1) The technical scheme of the present invention utilizes the triangular mesh in the reconstructed three-dimensional scene, and divides the three-dimensional scene into a plurality of two-dimensional triangular patches, so as to avoid the problem that superpixel extraction is too complicated in the three-dimensional scene;
(2)本发明技术方案采用多尺度特征信息和邻域信息融合,简单有效,识别率高,还能够进一步用于三维场景表面重建和场景理解分析等问题。(2) The technical solution of the present invention adopts the fusion of multi-scale feature information and neighborhood information, which is simple and effective, has a high recognition rate, and can be further used for three-dimensional scene surface reconstruction and scene understanding and analysis.
附图说明Description of drawings
图1是本发明实施例草地识别方法流程;1 is a flowchart of a grass identification method according to an embodiment of the present invention;
图2是本发明实施例草地识别的SVM训练的流程图;Fig. 2 is the flow chart of the SVM training of grass field recognition according to the embodiment of the present invention;
图3a是本发明实施例草地样本H数据统计直方图;Fig. 3a is a statistical histogram of grass sample H data according to an embodiment of the present invention;
图3b是本发明实施例天空样本H数据统计直方图。FIG. 3b is a statistical histogram of sky sample H data according to an embodiment of the present invention.
具体实施方式Detailed ways
为了使本发明的目的、技术方案及优点更加清楚明白,以下结合附图及实施例,对本发明进行进一步详细说明。应当理解,此处所描述的具体实施例仅仅用以解释本发明,并不用于限定本发明。此外,下面所描述的本发明各个实施方式中所涉及到的技术特征只要彼此之间未构成冲突就可以相互组合。In order to make the objectives, technical solutions and advantages of the present invention clearer, the present invention will be further described in detail below with reference to the accompanying drawings and embodiments. It should be understood that the specific embodiments described herein are only used to explain the present invention, but not to limit the present invention. In addition, the technical features involved in the various embodiments of the present invention described below can be combined with each other as long as they do not conflict with each other.
如图1所示为采用本发明技术方案的三维场景的草地识别方法流程,包括以下步骤:As shown in FIG. 1, the process flow of the grass recognition method of the three-dimensional scene using the technical solution of the present invention includes the following steps:
(1)将待识别三维场景进行三维重建,按三维场景表面的三角网格划分三角面片。(1) 3D reconstruction is performed on the 3D scene to be recognized, and triangular patches are divided according to the triangular mesh of the surface of the 3D scene.
(2)提取所有三角面片单元的HSI特征。(2) Extract the HSI features of all triangular patch units.
(3)根据选定三角面片和其共享边的相邻三角面片的H特征值和S特征值进行直方图统计,将统计数据放入特征向量中,将三角面片的特征向量和其相邻三角面片的特征向量进行叠加,叠加结果作为选定三角面片的特征向量。(3) Perform histogram statistics according to the H eigenvalues and S eigenvalues of the selected triangular facet and its adjacent triangular facets with shared edges, put the statistical data into the eigenvectors, and combine the eigenvectors of the triangular facets with its eigenvalues. The eigenvectors of adjacent triangular patches are superimposed, and the superposition result is used as the eigenvectors of the selected triangular patch.
(4)将选定三角面片的特征向量输入到SVM识别模型中,获取识别结果。(4) Input the feature vector of the selected triangular facet into the SVM recognition model to obtain the recognition result.
如图2所示为采用本发明技术方案的构建三维场景的草地SVM识别模型的流程:As shown in Figure 2, the process of adopting the technical solution of the present invention to construct the grass SVM recognition model of the three-dimensional scene:
(1)搜集50张正样本和200张负样本,选择全为草地的图像作为正样本以便于进行训练集采样和特征提取,由于负样本要涵盖的种类复杂,这里选择帕斯卡数据集(PASCALVOC)中的各类非草地目标作为负样本。(1) Collect 50 positive samples and 200 negative samples, and select all grass images as positive samples to facilitate training set sampling and feature extraction. Since the types of negative samples to be covered are complex, the Pascal dataset (PASCALVOC) is selected here. Various types of non-grass targets in are used as negative samples.
(2)将正负训练样本的尺寸调整为162*162,将训练样本作规则直角三角面片划分处理,三角面片的尺寸为3种,分别是直角边长3个像素点、6个像素点和9个像素点;由于样本图像是RGB模式,但由于RGB模式在不同条件下,测得RGB的颜色值分布分散,在RGB空间两点的欧式距离和颜色距离不成线性关系,不利于进行颜色特征的分类。为提高识别效果,将颜色表示从RGB颜色空间转化为更接近人眼观察方式的HSI颜色空间,从而提高不同光照下环境下识别系统的鲁棒性。在HSI颜色空间中:H代表色度;S代表饱和度;I代表亮度。3个分量中,受光照影响最大的是I分量,而H和S通常是构成物体的原材料的光线吸收和反射特征决定。为了减少光照对识别精度的影响,在识别目标时,以H和S作为提取的特征。从RGB到HSI的转换公式如下:(2) Adjust the size of the positive and negative training samples to 162*162, and divide the training samples into regular right-angled triangular patches. There are three sizes of triangular patches, which are 3 pixels and 6 pixels on a right-angled side. point and 9 pixels; because the sample image is in RGB mode, but because the RGB mode is under different conditions, the measured RGB color value distribution is scattered, and the Euclidean distance and color distance between two points in the RGB space are not linear, which is not conducive to the Classification of color features. In order to improve the recognition effect, the color representation is converted from the RGB color space to the HSI color space that is closer to the way the human eye observes, thereby improving the robustness of the recognition system under different lighting environments. In the HSI color space: H stands for chroma; S stands for saturation; I stands for lightness. Among the three components, the I component is most affected by light, while H and S are usually determined by the light absorption and reflection characteristics of the raw materials that constitute the object. In order to reduce the influence of illumination on the recognition accuracy, when recognizing the target, H and S are used as the extracted features. The conversion formula from RGB to HSI is as follows:
I=(R+G+B)/3I=(R+G+B)/3
S=1-3min(R,G,B)/IS=1-3min(R,G,B)/I
若B>G,H=360-HIf B>G, H=360-H
(3)根据选定三角面片和其共享边的相邻三角面片的H特征值和S特征值进行直方图统计,将统计数据放入特征向量中,将三角面片的特征向量和其相邻三角面片的特征向量进行叠加,叠加结果作为选定三角面片的特征向量。设置H特征值直方图和S特征值直方图的组数为8,组距分别为H特征值取值范围的1/8和S特征值取值范围的1/8。将三角面片的H特征值进行直方图统计得到一组8维向量,将三角面片的S特征值进行直方图统计也得到一组8维向量,将H特征值的8维向量和S特征值的8维向量首尾相连组成16维的特征向量。将三角面片的特征向量保存成mat文件供SVM使用。如图3a所示为天空样本中某个三角面片的H直方图统计结果,如图3b所示为草地样本中某个三角面片的H直方图统计结果。(3) Perform histogram statistics according to the H eigenvalues and S eigenvalues of the selected triangular facet and its adjacent triangular facets with shared edges, put the statistical data into the eigenvectors, and combine the eigenvectors of the triangular facets with its eigenvalues. The eigenvectors of adjacent triangular patches are superimposed, and the superposition result is used as the eigenvectors of the selected triangular patch. Set the number of groups of H eigenvalue histogram and S eigenvalue histogram to 8, and the group spacing is 1/8 of the value range of H eigenvalue and 1/8 of the value range of S eigenvalue. Perform histogram statistics on the H eigenvalues of the triangular facets to obtain a set of 8-dimensional vectors, and perform histogram statistics on the S eigenvalues of the triangular facets to obtain a set of 8-dimensional vectors. The 8-dimensional vectors of the H feature values and the S feature The 8-dimensional vector of values is connected end to end to form a 16-dimensional feature vector. Save the feature vector of the triangular patch as a mat file for use by SVM. Figure 3a shows the statistical result of the H histogram of a triangular patch in the sky sample, and Figure 3b shows the statistical result of the H histogram of a triangular patch in the grass sample.
(4)将所有样本的特征向量输入到SVM进行训练;SVM的决策函数为:(4) Input the feature vectors of all samples into SVM for training; the decision function of SVM is:
其中,L表示分类总数,取值2;ci表示函数系数,ci∈R;γ表示交叉校验的参数,γ∈R+;表示被选用为训练的支撑向量;h0表示训练的非支撑向量;采用直方图的卡方距离作为核函数,定义为:Among them, L represents the total number of classifications, and the value is 2; ci represents the function coefficient, ci ∈ R; γ represents the parameter of cross-checking, γ ∈ R + ; represents the support vector selected for training; h 0 represents the non-support vector for training; the chi-square distance of the histogram is used As a kernel function, it is defined as:
其中,N表示样本集中特征向量总数量;表示第k个特征向量对应的支撑向量;h0(k)表示第k个特征向量。Among them, N represents the total number of feature vectors in the sample set; represents the support vector corresponding to the kth eigenvector; h 0 (k) represents the kth eigenvector.
根据直方图的卡方距离计算协方差矩阵空间。对这个协方差矩阵空间进行镜像变换得到对应的超平面矩阵,也就是说将一个向量变换为由一个超平面反射的镜像,是一种线性变换,SVM就是在特征空间中构造最优超平面。得到协方差矩阵以及其对应的超平面矩阵,根据这两个矩阵分别计算每个特征的特征系数,并将特征系数对协方差矩阵进行缩放。对缩放后的协方差矩阵求逆后,计算模型参数。将训练好的SVM模型的参数存入矩阵文件中以供后续三维场景草地识别调用。Computes the covariance matrix space from the chi-square distance of the histogram. Mirror transformation is performed on this covariance matrix space to obtain the corresponding hyperplane matrix, that is to say, transforming a vector into a mirror reflected by a hyperplane is a linear transformation, and SVM constructs the optimal hyperplane in the feature space. The covariance matrix and its corresponding hyperplane matrix are obtained, the characteristic coefficients of each feature are calculated according to these two matrices, and the characteristic coefficients are scaled to the covariance matrix. After inverting the scaled covariance matrix, the model parameters are calculated. The parameters of the trained SVM model are stored in the matrix file for subsequent 3D scene grass recognition calls.
以上所述仅为本发明的较佳实施例而已,并不用以限制本发明,凡在本发明的精神和原则之内所作的任何修改、等同替换和改进等,均应包含在本发明的保护范围之内。The above descriptions are only preferred embodiments of the present invention and are not intended to limit the present invention. Any modifications, equivalent replacements and improvements made within the spirit and principles of the present invention shall be included in the protection of the present invention. within the range.
Claims (6)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710034186.6A CN106886754B (en) | 2017-01-17 | 2017-01-17 | Object identification method and system under a kind of three-dimensional scenic based on tri patch |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710034186.6A CN106886754B (en) | 2017-01-17 | 2017-01-17 | Object identification method and system under a kind of three-dimensional scenic based on tri patch |
Publications (2)
Publication Number | Publication Date |
---|---|
CN106886754A CN106886754A (en) | 2017-06-23 |
CN106886754B true CN106886754B (en) | 2019-07-09 |
Family
ID=59176518
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201710034186.6A Active CN106886754B (en) | 2017-01-17 | 2017-01-17 | Object identification method and system under a kind of three-dimensional scenic based on tri patch |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN106886754B (en) |
Families Citing this family (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108921952B (en) * | 2018-06-15 | 2022-09-06 | 深圳大学 | Object functionality prediction method, device, computer equipment and storage medium |
CN109344798B (en) * | 2018-10-24 | 2022-08-12 | 艾凯克斯(嘉兴)信息科技有限公司 | Three-dimensional object identification and detection method based on three-dimensional grid and neural network |
CN110059213A (en) * | 2019-03-20 | 2019-07-26 | 杭州电子科技大学 | A kind of threedimensional model classification retrieving method based on Density Estimator |
CN114882272B (en) * | 2022-04-22 | 2024-10-18 | 成都飞机工业(集团)有限责任公司 | All-angle projection image patch attribute fusion analysis method for aviation manufacturing part |
Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104598889A (en) * | 2015-01-30 | 2015-05-06 | 北京信息科技大学 | Human action recognition method and device |
CN105006021A (en) * | 2015-06-30 | 2015-10-28 | 南京大学 | Color mapping method and device suitable for rapid point cloud three-dimensional reconstruction |
CN105608421A (en) * | 2015-12-18 | 2016-05-25 | 中国科学院深圳先进技术研究院 | Human movement recognition method and device |
CN105678278A (en) * | 2016-02-01 | 2016-06-15 | 国家电网公司 | Scene recognition method based on single-hidden-layer neural network |
CN105912999A (en) * | 2016-04-05 | 2016-08-31 | 中国民航大学 | Human behavior identification method based on depth information |
CN106022228A (en) * | 2016-05-11 | 2016-10-12 | 东南大学 | Three-dimensional face recognition method based on vertical and horizontal local binary pattern on the mesh |
CN106204749A (en) * | 2016-07-07 | 2016-12-07 | 北京航空航天大学 | The threedimensional model of sparse low-rank feature representation is divided into segmentation method |
-
2017
- 2017-01-17 CN CN201710034186.6A patent/CN106886754B/en active Active
Patent Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104598889A (en) * | 2015-01-30 | 2015-05-06 | 北京信息科技大学 | Human action recognition method and device |
CN105006021A (en) * | 2015-06-30 | 2015-10-28 | 南京大学 | Color mapping method and device suitable for rapid point cloud three-dimensional reconstruction |
CN105608421A (en) * | 2015-12-18 | 2016-05-25 | 中国科学院深圳先进技术研究院 | Human movement recognition method and device |
CN105678278A (en) * | 2016-02-01 | 2016-06-15 | 国家电网公司 | Scene recognition method based on single-hidden-layer neural network |
CN105912999A (en) * | 2016-04-05 | 2016-08-31 | 中国民航大学 | Human behavior identification method based on depth information |
CN106022228A (en) * | 2016-05-11 | 2016-10-12 | 东南大学 | Three-dimensional face recognition method based on vertical and horizontal local binary pattern on the mesh |
CN106204749A (en) * | 2016-07-07 | 2016-12-07 | 北京航空航天大学 | The threedimensional model of sparse low-rank feature representation is divided into segmentation method |
Non-Patent Citations (1)
Title |
---|
多特征和SVM相融合的三维物体识别方法;湛宁;《计算机仿真》;20130331;第30卷(第3期);第380-383、407页 |
Also Published As
Publication number | Publication date |
---|---|
CN106886754A (en) | 2017-06-23 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN107578418B (en) | Indoor scene contour detection method fusing color and depth information | |
CN108416307B (en) | An aerial image pavement crack detection method, device and equipment | |
CN105574534B (en) | Conspicuousness object detection method based on sparse subspace clustering and low-rank representation | |
CN104850850B (en) | A kind of binocular stereo vision image characteristic extracting method of combination shape and color | |
CN103632132B (en) | Face detection and recognition method based on skin color segmentation and template matching | |
Zhang et al. | Lighting and pose robust face sketch synthesis | |
CN104182765B (en) | Internet image driven automatic selection method of optimal view of three-dimensional model | |
CN110837768B (en) | An online detection and identification method for rare animal protection | |
CN106682598A (en) | Multi-pose facial feature point detection method based on cascade regression | |
CN111178208A (en) | Pedestrian detection method, device and medium based on deep learning | |
CN107194937A (en) | Tongue image partition method under a kind of open environment | |
CN105678278A (en) | Scene recognition method based on single-hidden-layer neural network | |
Wang et al. | Recognition and localization of occluded apples using K-means clustering algorithm and convex hull theory: a comparison | |
CN103473759B (en) | The twilight image method for extracting remarkable configuration that a kind of WKPCA homogeneity degree correction nCRF suppresses | |
CN106886754B (en) | Object identification method and system under a kind of three-dimensional scenic based on tri patch | |
CN111091129B (en) | Image salient region extraction method based on manifold ordering of multiple color features | |
CN108875655A (en) | A kind of real-time target video tracing method and system based on multiple features | |
CN103440035A (en) | Gesture recognition system in three-dimensional space and recognition method thereof | |
CN103971122B (en) | Three-dimensional face based on depth image describes method | |
CN112836651A (en) | Gesture image feature extraction method based on dynamic fusion mechanism | |
CN108280421A (en) | Human bodys' response method based on multiple features Depth Motion figure | |
CN114612802A (en) | System and method for classifying fine granularity of ship target based on MBCNN | |
Dawkins et al. | Automatic scallop detection in benthic environments | |
CN103150575A (en) | Real-time three-dimensional unmarked human body gesture recognition method and system | |
CN105139013B (en) | An object recognition method that fuses shape features and interest points |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |