[go: up one dir, main page]

CN113012298A - Curved MARK three-dimensional registration augmented reality method based on region detection - Google Patents

Curved MARK three-dimensional registration augmented reality method based on region detection Download PDF

Info

Publication number
CN113012298A
CN113012298A CN202011563089.4A CN202011563089A CN113012298A CN 113012298 A CN113012298 A CN 113012298A CN 202011563089 A CN202011563089 A CN 202011563089A CN 113012298 A CN113012298 A CN 113012298A
Authority
CN
China
Prior art keywords
mark
point
camera
pose
natural texture
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202011563089.4A
Other languages
Chinese (zh)
Other versions
CN113012298B (en
Inventor
张明敏
陈忠庆
潘志庚
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhejiang University ZJU
Original Assignee
Zhejiang University ZJU
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhejiang University ZJU filed Critical Zhejiang University ZJU
Priority to CN202011563089.4A priority Critical patent/CN113012298B/en
Publication of CN113012298A publication Critical patent/CN113012298A/en
Application granted granted Critical
Publication of CN113012298B publication Critical patent/CN113012298B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T19/00Manipulating 3D models or images for computer graphics
    • G06T19/006Mixed reality
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T15/003D [Three Dimensional] image rendering
    • G06T15/005General purpose rendering architectures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • G06T7/73Determining position or orientation of objects or cameras using feature-based methods
    • G06T7/75Determining position or orientation of objects or cameras using feature-based methods involving models
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/80Analysis of captured images to determine intrinsic or extrinsic camera parameters, i.e. camera calibration
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10024Color image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30204Marker
    • G06T2207/30208Marker matrix

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Graphics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Computer Hardware Design (AREA)
  • General Engineering & Computer Science (AREA)
  • Software Systems (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a curved MARK three-dimensional registration augmented reality method based on region detection. Partial occlusion can be done for the curved MARK without affecting the final effect. The method provided by the invention overcomes the problems that the traditional plane MARK can not be bent, so that the consistency of a cylindrical object on an augmented reality scene is damaged, the robustness of the natural texture MARK is low, the real-time performance is low and the like.

Description

Curved MARK three-dimensional registration augmented reality method based on region detection
Technical Field
The invention belongs to the field of intersection of computer vision technology and graphics, and particularly relates to a curved MARK three-dimensional registration augmented reality method based on region detection.
Background
With the increasing development and maturity of the internet and the iterative update of multimedia technology, Augmented Reality (AR) is more and more common in our daily life and learning. Augmented reality is a very practical technology combining computer graphics and computer vision, and can overlay information such as virtual objects, videos and characters to a real scene so that a user can acquire more information in the scene, and the user can understand the scene more deeply and clearly.
The augmented reality has wide applications in daily life, such as teaching demonstrations, tour navigation, virtual shopping, workshop guidance, and the like. In the teaching field, the augmented reality can bring safer and more interesting teaching experiments for students, and improve the interest and the practical ability of the students. Because the dangerousness or the imperceptibility of many experiments, these experiments are often neglected in the teaching process, this has promoted the application of augmented reality in the teaching process, and the student can use the interaction of virtual object and real object that superpose in the augmented reality to accomplish the dangerousness experiment and observe more detailed experimental effect, all has very big promotion to student's hands-on ability and theoretical cognition. For the tourism industry, the augmented reality can also enable the user to obtain more direct and vivid explanation on the mobile terminal, and the interestingness and the interactivity of navigation are enhanced.
The augmented reality system mainly relates to technologies such as three-dimensional registration, user interaction, virtual-real fusion and the like, wherein the three-dimensional registration plays a decisive role in development and popularization of the augmented reality system, and the method mainly has the function of estimating the relative pose of a camera in a scene and then superimposing a virtual object on the real scene. Three-dimensional registration is not satisfactory in the experience degree of users at present due to the problems of instantaneity, robustness, stability, attractiveness and the like, so that the deep exploration of the three-dimensional registration technology has a profound significance in the development process of augmented reality for the research and development of the three-dimensional registration technology to be a hot topic in the field of augmented reality nowadays.
In the augmented reality system, in order to generate the visual effect of virtual-real fusion, registration alignment of virtual-real environment is firstly ensured. The most commonly used method in the augmented reality system is that the virtual and real environments share the same spatial coordinate system, so that virtual objects can be rendered in a scene to achieve the effect of virtual-real interaction. In augmented reality systems, cameras are generally used as main sensors, and the positions of virtual objects to be rendered are acquired by estimating the relative poses of the cameras in real time through a three-dimensional registration technology.
The three-dimensional registration technique most commonly used today is the vision-based registration technique, with planar MARKs being the most commonly used. However, for a planar MARK with a curved surface, such as a cylinder, the aesthetic property of the MARK is damaged, so that the immersion of a user is greatly reduced, and therefore, the three-dimensional registration based on the curved MARK is not very significant for the development of augmented reality. The category of the MARK mainly includes an artificial MARK and a MARK based on a natural texture, wherein the artificial MARK such as a hamming code, a two-dimensional code, and the like has the disadvantages of being not shelterable and not being able to obtain a correct pose after being bent, so the MARK based on the natural texture becomes a unique choice for realizing the bent MARK.
Disclosure of Invention
The invention aims to apply a curved MARK three-dimensional registration technology to the field of augmented reality, and provides a curved MARK three-dimensional registration augmented reality method based on region detection. The region where the MARK is located is obtained through the neural network model, three-dimensional registration is carried out on the bent MARK, meanwhile, the MARK can be partially shielded, and the attractiveness and the real-time performance of augmented reality cannot be damaged.
The method is based on a region detection technology, the region where the bent MARK is located in the scene is obtained, and a three-dimensional model formed by attaching the MARK to the cylinder after bending is built according to the radius of the cylinder object and the coordinates of the MARK plane feature points. Coordinates of the curved MARK feature points are acquired in a scene, and the relative pose between the camera and the MARK is restored through a PnP algorithm, so that the virtual object is rendered in the scene.
The technical scheme adopted by the invention for solving the technical problems is as follows:
a curved MARK three-dimensional registration augmented reality method based on region detection comprises the following steps:
step (1), calibrating a camera:
acquiring internal parameters and distortion parameters of the RGB monocular camera by using a Zhang Zhen friend camera calibration method;
step (2), constructing a data set:
the natural texture MARK to be identified is shot by more than 300 pictures under different angles, different distances, different illumination, partial shielding and non-shielding conditions respectively, wherein 80% of the pictures are used as a training set, and the rest 20% of the pictures are used as a verification set. Calibrating a border (bounding box) and a class (classes) of a natural texture MARK in a picture by using labelImg software to generate a corresponding xml format file, framing out an area where the natural texture MARK is located by using a rectangular frame in the calibration process, and then calibrating the class corresponding to the natural texture MARK;
step (3), a neural network model of Yolov5 is used as a natural texture MARK target detection model, the training set constructed in the step (2) is adopted for training, the accuracy of the model is verified through a verification set, the trained model is extracted, and the frame of MARK in a scene can be identified and a specific type can be identified through the trained natural texture MARK target detection model;
step (4), printing the MARK, pasting the MARK on a cylindrical object, measuring the radius r of the cylinder, the width and height of a natural texture MARK picture, marker _ w and marker _ h, extracting the MARK picture characteristic points by using a Fast algorithm, calculating the three-dimensional coordinate of each characteristic point relative to the central point of the MARK, and calculating the included angle between the characteristic point with the coordinate (x, y) of the MARK picture and the connecting line of the characteristic point and the center of the cylinder
Figure RE-GDA0003039726560000031
The pixel to millimeter conversion scale is pixel2mm, solving for the equation (1):
Figure RE-GDA0003039726560000032
the corresponding three-dimensional coordinates are (the coordinate units here are all millimeters):
Figure RE-GDA0003039726560000033
Figure RE-GDA0003039726560000034
Figure RE-GDA0003039726560000035
storing the three-dimensional coordinates of all the feature points in a dictionary, wherein the keys of the dictionary are the coordinates (x, y) of a MARK picture, and the values are the coordinates obtained by the formulas (2), (3) and (4);
and (5) for a scene picture obtained by a camera, extracting a natural texture MARK in the scene picture by using a natural texture MARK target detection model to generate a Region Of Interest (ROI Region), extracting all feature points in the ROI Region by adopting a Fast algorithm, calculating a descriptor by using an ORB algorithm, matching the extracted feature points Of the ROI Region with original MARK feature points by using a RANSAC and K nearest neighbor classification algorithm (KNN) on the basis Of calculating the hamming distance Of the descriptor, obtaining 30 most matched feature point matching pairs, obtaining three-dimensional coordinates Of the feature points Of the MARK picture from a dictionary in the step (4), and estimating the relative pose Of the camera and the curved MARK by using a PNP algorithm, wherein the specific implementation is as follows:
point X of world coordinate systemw=(xw,yw,zw1), with its projection coordinates X on the image planei=(xi,yiThe relation of 1) is expressed by the following formula, wherein fx, fy, cx and cy are camera internal parameters calibrated by Zhangyingyou, and rijRepresenting a rotational variable, tiRepresents the translation variables:
Figure RE-GDA0003039726560000041
the method is simplified as follows:
Figure RE-GDA0003039726560000042
wherein, λ represents a scale factor, the matrix K is a camera internal parameter matrix, and the matrix M is a model viewpoint matrix. Randomly selecting 4 characteristic point matching point pairs from 30 characteristic point matching pairs, calculating 4 groups of different solutions by using 3 point pairs, substituting the rest 1 point pairs into a formula, solving the solution with the minimum reprojection error into a final solution, and optimizing the final solution by using a random sample consensus (RANSAC) algorithm in the process;
step (6), in the MARK moving process, tracking the moving state of the feature points by using an optical flow method, judging the moving quantity of the feature points of the same two frames, when the moving quantity is less than or equal to ten percent of the total feature point quantity, considering that the pose of the marker does not change relative to the previous frame, and when the moving quantity is greater than ten percent of the total feature point quantity, considering that the pose of the marker changes relative to the previous frame, and then following the step (5) to obtain the pose of the current natural texture MARK for three-dimensional registration, specifically realizing the following steps:
and setting I and J as the gray level images of the previous frame and the current frame, then:
Figure RE-GDA0003039726560000043
wherein, the point A is any point in the image, and the coordinate vector is (x, y)TFor a point u on the previous frame I ═ ux,uy]TThe purpose of feature point tracking is to find its position v + u + d in the current frame imagex+dx,uy+dy]TD is ═ dx,dy]TIs the image velocity at point a, i.e. the optical flow at point a. Defining the concept of similarity in the two-dimensional neighborhood sense due to the influence of the aperture, setting ωxAnd ωyFor two integer values, the minimized residual function for the velocity vector d is defined as follows:
Figure RE-GDA0003039726560000051
the similarity definition can be obtained by the above formulaThe definition of similarity is based on the size of the image neighborhood as (2 omega)x+1)×(2ωy+1), solving d to obtain the corresponding position of the point u in the image J; omegaxAnd ωyIs 2, 3, 4, 5, 6 or 7.
Comparing the position of the feature point calculated by the current frame with the position of the feature point corresponding to the previous frame, judging whether the feature points of two adjacent frames of the camera move or not, counting the number of the moved feature points, if the number of the feature points of the current frame is less than or equal to ten percent of the number of the total feature points, considering that the object does not move relative to the object of the previous frame, and directly acquiring the pose of the previous frame, if the number of the feature points of the current frame is greater than ten percent of the number of the total feature points, considering that the object moves relative to the object of the previous frame, and recalculating the relative pose of the camera relative to the object;
and (7) in the process of estimating the relative pose of the camera in the step (5), predicting and correcting the 6D pose of the MARK by using Kalman filtering.
First defining the displacement of the camera with respect to the natural texture MARK (t)x,ty,tz) And the rotation angles (psi, theta, phi), the first derivative of the coordinates being (t)x',ty',tz') and the second derivative of the coordinates is (t)x”,t'y', tz") where the first derivative represents the speed at which the natural texture MARK moves and the second derivative represents the acceleration at which the natural texture MARK moves, the first derivative of the rotation angle is (ψ ', θ ', φ '), and the second derivative is (ψ", θ ", φ"), where the first derivative represents the speed at which the MARK rotates and the second derivative represents the acceleration at which the natural texture MARK rotates. The kalman filtering may be used for estimation and correction, and the specific formula is as follows:
Kalman=(tx,ty,tz,tx′,ty′,tz′,tx″,ty″,tz″,ψ,θ,φ,ψ′,θ′,φ′,ψ″,θ″,φ″) (9)
and (8) eliminating the frame with the wrong pose estimation by using a sliding window.
And judging whether the current estimated camera pose is correct or not by the camera pose coordinates of the last two frames and the first two frames relative to the current frame, so as to eliminate the frame with wrong estimation of the camera pose caused by blurring in the motion process of the natural texture MARK. The displacement in the 6D pose estimation of the current frame camera is (x)t,yt,zt) (parameters with different meanings need to be represented by different symbols), calculating the average displacement (x ', y ', z ') of the cameras of the first two frames and the average displacement (x ", y", z ") of the cameras of the second two frames, wherein the current displacement satisfies:
x"-dt<xt<x'+dt orx'-dt<xt<x"+dt
y"-dt<yt<y'+dt ory'-dt<yt<y"+dt (10)
z"-dt<zt<z'+dt or z'-dt<zt<z"+dt
considering the current pose as an effective pose, otherwise considering the current frame as a fuzzy frame, and continuously using the last effective pose, wherein dtThreshold adjusted for translation, dt=3。
And (9) after the relative pose between the camera and the natural texture MARK is obtained through the steps (5), (6), (7) and (8), the virtual object needing three-dimensional registration is subjected to translation and rotation transformation, and the virtual object is rendered into a scene through OpenGL and OpenCV to achieve the effect of augmented reality.
The invention has the beneficial effects that:
the method comprises the steps of attaching a two-dimensional natural texture MARK to a cylinder to form a curved MARK, processing the curved MARK through a neural network model to obtain the region where the curved MARK is located in the current scene, calculating the relative pose between a camera and an object through feature point matching, and rendering a virtual object into an augmented reality scene. Partial occlusion can be done for the curved MARK without affecting the final effect. The method solves the problems that the traditional plane MARK can not be bent, so that the consistency of the cylindrical object to the augmented reality scene is damaged, the robustness of the natural texture MARK is low, the real-time performance is low and the like.
Drawings
FIG. 1 is a picture of a natural texture MARK according to an embodiment of the present invention;
FIG. 2 is a diagram illustrating feature points detected in a MARK picture according to an embodiment of the present invention;
FIG. 3 is a schematic diagram illustrating the calculation of three-dimensional coordinates of feature points on a MARK according to an embodiment of the present invention;
FIG. 4 is a comparison diagram of feature points between two adjacent frames according to an embodiment of the present invention;
FIG. 5 is an effect diagram of a virtual object rendered to an assigned pose in a scene according to an embodiment of the present invention;
FIG. 6 is a flowchart of a method according to an embodiment of the present invention.
Detailed Description
The method of the present invention is further described below with reference to the accompanying drawings.
The experimental environment is a monocular RGB video camera (640 x 480), a cylindrical object, a natural texture MARK picture is printed and attached to the cylindrical object, and the part of the MARK on the cylindrical object is always aligned to the monocular camera in the experimental process.
As shown in fig. 1, a picture with sharp corners and irregular natural texture features with multiple features is used as a MARK, and a symmetrical picture is not selected, so that a large number of pictures are selected, and the pictures have obvious differences.
As shown in fig. 2, all feature points in the MARK picture are calculated by using Fast algorithm, and the specific steps are as follows:
step (a), selecting a pixel Q from the MARK picture, and setting the brightness value of the pixel Q as I to judge whether the pixel is a characteristic point or notq
Step (b), a Bresenham circle is obtained by taking the pixel point Q as the center and the radius of the Bresenham circle as 3, and the circle has 16 pixels;
step (c), on the circle with the size of 16 pixels, if the pixel values of 9 continuous pixel points are all larger than Iq+ t or both less than IqAnd t, the pixel point Q is considered as a characteristic point, and the t is a set threshold value.
Step (d), in order to improve the judgment efficiency of the angular points to eliminate pixels of non-angular points in the image, checking corresponding pixels according to four positions of 1, 9, 5 and 13, and when a pixel point Q is an angular point, at least 3 pixel values of the pixel points of the four positions are all larger than Iq+ t is greater than or less than IqAnd (c) if the pixel values of the pixel points at the four positions do not meet the condition, judging and screening all the pixel points which are not the angular points if the pixel values of the pixel points at the four positions are not the angular points, and judging and screening the rest pixel points to obtain the final angular points by performing the operation judgment in the step (c).
As in fig. 3, for each feature point, its three-dimensional coordinates relative to the MARK center point are calculated. After attaching the MARK to the cylindrical object, a three-dimensional model can be obtained. After the radius r of the cylinder is obtained, for the feature point with the coordinate (x, y) of the MARK picture, the included angle between the feature point and the central connecting line of the cylinder is
Figure RE-GDA0003039726560000071
The pixel to millimeter conversion scale is pixel2mm, and the calculation formula is as follows:
Figure RE-GDA0003039726560000072
the corresponding three-dimensional coordinates are (the coordinate units here are all pixels):
Figure RE-GDA0003039726560000081
Figure RE-GDA0003039726560000082
Figure RE-GDA0003039726560000083
storing the three-dimensional coordinates of all the feature points in a dictionary, wherein the keys of the dictionary are the coordinates (x, y) of the MARK picture, and the values are obtainedThree dimensional coordinates (3 d)x,3dy,3dz)。
Shooting under different angles, different distances, different illumination, partial shielding and non-shielding conditions to obtain a training set picture; 300 pictures are shot in total, wherein 240 pictures are used as a training set, the rest pictures are used as a verification set, a border (bounding box) and a category (classes) of the pictures are calibrated by using labelImg to generate an xml file, and then the corresponding pictures and the marked xml are put into corresponding paths of a Yolov5 model code (step (3)).
Detecting a natural texture MARK in a scene through a Yolov5 target detection model, wherein the model can extract a region where the MARK is located in the scene and the confidence coefficient of the region in real time, if the confidence coefficient is less than 20, the region is not considered to be the region where the MARK is located, if the confidence coefficient is greater than or equal to 20, a bounding box (bounding box) of the MARK in the scene is obtained, and masking operation is performed on an image of the region by using OPENCV so that the RGB value of pixels except the other part of the MARK region is (0,0, 0).
As shown in fig. 4, tracking all feature points in a scene by an optical flow method (optical flow) is specifically implemented as follows:
and setting I and J as the gray level images of the previous frame and the current frame, then:
Figure RE-GDA0003039726560000084
wherein, the point A is any point in the image, and the coordinate vector is (x, y)TFor a point u on the previous frame I ═ ux,uy]TThe purpose of feature point tracking is to find its position v + u + d in the current frame imagex+dx,uy+dy]TD is ═ dx,dy]TIs the image velocity at point a, i.e. the optical flow at point a. Defining the concept of similarity in the two-dimensional neighborhood sense due to the influence of the aperture, setting ωxAnd ωyFor two integer values, the minimized residual function for the velocity vector d is defined as follows:
Figure RE-GDA0003039726560000091
the similarity definition can be obtained through the formula, and the similarity definition is based on the image neighborhood size of (2 omega)x+1)×(2ωy+1), solving for d, resulting in the corresponding position of point u in image J, for ωxAnd ωyTypical values are 2, 3, 4, 5, 6, 7.
Comparing the position of the feature point calculated by the current frame with the position of the feature point corresponding to the previous frame, judging whether the feature points of two adjacent frames of the camera move or not, counting the number of the moved feature points, if the number of the feature points of the current frame is less than or equal to ten percent of the number of the total feature points, considering that the object does not move relative to the object of the previous frame, and directly acquiring the pose of the previous frame, if the number of the feature points of the current frame is greater than ten percent of the number of the total feature points, considering that the object moves relative to the object of the previous frame, and recalculating the relative pose of the camera relative to the object;
and calculating the obtained descriptor of the feature point by using an ORB algorithm, and specifically comprising the following steps of:
setting the center of a key point O, and using O as the center of a circlerThe size of the pixel is that the radius is made into a circle;
taking N point pairs in the circle, wherein N is 512;
step (g), defining operation M (wherein I)AExpressing the gray scale of A, IBGrayscale for B):
Figure RE-GDA0003039726560000092
and (h) carrying out the operation of the step (g) on the selected key points to obtain a descriptor combination consisting of 0 and 1.
The ORB implementation in OPENCV adopts an image pyramid to solve the problem that descriptors are not sensitive to illumination and have no scale consistency. For rotation consistency, the principal direction of each feature point is calculated by adopting a gray scale centroid method, the gray scale centroid coordinate is calculated in the circular area range with the radius r of the feature point, and the direction vector from the center position to the centroid position is determined as the principal direction.
And carrying out similarity matching on the feature point descriptor of the natural texture MARK in the scene and the original MARK feature point descriptor. Hamming distance is used to calculate the similarity between two descriptors, with dkHamming distance, D, between rBRIEF descriptors representing feature points A and BADescriptor representing characteristic point A, DBA descriptor representing a feature point B, i representing the bit value of the i-th position of the descriptor:
Figure RE-GDA0003039726560000101
and after obtaining a feature point pair matched with the natural texture MARK and the reference MARK in the scene, carrying out outlier rejection through a ratio test, regarding the feature point p of the natural texture MARK in the scene, the distance between the two feature points closest to the reference image is d1 and d2, and when d1/d2> ratio (ratio is preferably 0.8), considering the feature point p as an outlier to carry out rejection. The use of random sample consensus (RANSAC) algorithm on the ratio-tested valid feature points (inliers) further eliminates possible outliers. In the matching process, cross validation (namely that the feature point p and the feature point q are mutually the most matched feature points) and a nearest neighbor algorithm are used for further screening out the point pairs with wrong matching, and finally the camera pose is calculated through a PnP algorithm. As shown in fig. 5, for the effect of augmented reality by rendering the virtual object into the scene, the virtual object completely covers the cup attached with the natural texture MARK, so that the cup in the scene is replaced.
As shown in fig. 6, which is a flow chart of the practical application of the method of the present invention, the steps are as follows:
step (1) acquiring a natural texture MARK Region (ROI) in a scene by using Yolov 5;
step (2) comparing the feature points of the previous frame with the optical flow method, if the object is judged not to move, directly acquiring the pose of the camera of the previous frame, and executing step (4), otherwise, executing step (3);
step (3) extracting feature points of an ROI (region of interest) by using a Fast algorithm to match with feature points of a MARK picture, finding out 30 feature points which are most matched by a KNN (nearest neighbor algorithm), recovering three-dimensional coordinates of the 30 feature points by a dictionary, calculating to obtain a MARK pose by using a PnP (neighbor nearest neighbor algorithm) and a RANSAC (random sample consensus) algorithm, comparing the MARK pose with an average pose of a sliding window to judge whether the current pose is effective, updating the sliding window if the MARK pose is effective, and executing the step (4), otherwise, using the pose of a previous frame of camera;
and (4) rendering the virtual object to the acquired pose through OPENGL and OPENCV to perform augmented reality.

Claims (6)

1.一种基于区域检测的弯曲MARK三维注册增强现实方法,其特征在于,步骤如下:1. a curved MARK three-dimensional registration augmented reality method based on area detection, is characterized in that, step is as follows: 一种基于区域检测的弯曲MARK三维注册增强现实方法,步骤如下:An augmented reality method for curved MARK 3D registration based on region detection, the steps are as follows: 步骤(1)、标定相机:Step (1), calibrate the camera: 使用张正友相机标定法获取RGB单目相机的内参和畸变参数;Use Zhang Zhengyou's camera calibration method to obtain the internal parameters and distortion parameters of the RGB monocular camera; 步骤(2)、构造数据集:Step (2), construct the data set: 将所要识别的自然纹理MARK各自以不同角度、不同距离、不同光照、部分遮挡和非遮挡条件下总共拍摄300张以上的图片,其中80%的图片作为训练集,剩下的20%的图片用来做验证集;使用labelImg软件标定图片中的自然纹理MARK的边框(bounding box)和类别(classes)生成对应的xml格式文件,在标定过程中,先将自然纹理MARK所在的区域用矩形框框出,然后标定该自然纹理MARK对应的类别;Take a total of more than 300 pictures of the natural texture MARK to be identified under different angles, different distances, different lighting, partial occlusion and non-occlusion conditions, of which 80% of the pictures are used as the training set, and the remaining 20% of the pictures are used for To make a validation set; use labelImg software to calibrate the bounding box and classes of the natural texture MARK in the picture to generate the corresponding xml format file. During the calibration process, first frame the area where the natural texture MARK is located with a rectangular frame , and then calibrate the category corresponding to the natural texture MARK; 步骤(3)、使用Yolov5的神经网络模型做为自然纹理MARK目标检测模型,采用步骤(2)构造的训练集进行训练,并通过验证集验证模型的精确度,将训练好的模型提取出来,通过训练好的自然纹理MARK目标检测模型能够识别场景中的MARK的边框并识别出具体类别;Step (3), use the neural network model of Yolov5 as the natural texture MARK target detection model, use the training set constructed in step (2) for training, and verify the accuracy of the model through the verification set, and extract the trained model, Through the trained natural texture MARK target detection model, the frame of the MARK in the scene can be recognized and the specific category can be identified; 步骤(4)、打印MARK,并贴到圆柱体物体上,同时测量出圆柱体半径r,自然纹理MARK图片的宽和高marker_w,marker_h,使用Fast算法提取MARK图片特征点,对于每个特征点计算其相对于MARK中心点的三维坐标,对于MARK图片坐标为(x,y)的特征点,其与圆柱体中心连线夹角为
Figure RE-FDA0003039726550000011
像素到毫米的转换尺度为pixel2mm,求解公式为(1):
Step (4), print the MARK, and paste it on the cylinder object, and measure the cylinder radius r, the width and height of the natural texture MARK image marker_w, marker_h, and use the Fast algorithm to extract the MARK image feature points. For each feature point Calculate its three-dimensional coordinates relative to the MARK center point. For the feature point whose MARK image coordinates are (x, y), the angle between the line connecting it and the center of the cylinder is
Figure RE-FDA0003039726550000011
The conversion scale from pixel to millimeter is pixel2mm, and the solution formula is (1):
Figure RE-FDA0003039726550000012
Figure RE-FDA0003039726550000012
对应的的三维坐标为(这里的坐标单位都是毫米):The corresponding three-dimensional coordinates are (the coordinate units here are all millimeters):
Figure RE-FDA0003039726550000013
Figure RE-FDA0003039726550000013
Figure RE-FDA0003039726550000021
Figure RE-FDA0003039726550000021
Figure RE-FDA0003039726550000022
Figure RE-FDA0003039726550000022
将所有特征点的三维坐标存在一个字典中,字典的键为MARK图片的坐标(x,y),值为公式(2),(3),(4)所求得的坐标;Store the three-dimensional coordinates of all feature points in a dictionary, the keys of the dictionary are the coordinates (x, y) of the MARK picture, and the values are the coordinates obtained by formulas (2), (3), (4); 步骤(5)、对于相机获取的场景图片,使用自然纹理MARK目标检测模型将其中的自然纹理标志提取出来产生感兴趣区域,即ROI区域(Region Of Interest),采用Fast算法提取ROI区域中所有特征点,并用ORB算法计算描述子,并在计算描述子汉明距离的基础上使用RANSAC和K最近邻分类算法(KNN)对提取出的ROI区域的特征点与原始MARK特征点之间进行匹配,获取的最匹配的30个特征点匹配对,从步骤(4)的字典中,获取特征点MARK图片特征点的三维坐标,通过PNP算法估计相机与弯曲的MARK的相对位姿,具体实现为:Step (5), for the scene picture obtained by the camera, use the natural texture MARK target detection model to extract the natural texture mark to generate a region of interest, that is, the ROI region (Region Of Interest), and use the Fast algorithm to extract all the features in the ROI region point, and use the ORB algorithm to calculate the descriptor, and use RANSAC and K nearest neighbor classification algorithm (KNN) on the basis of calculating the descriptor Hamming distance to match the extracted feature points of the ROI area with the original MARK feature points, Obtain the most matching 30 feature point matching pairs, from the dictionary in step (4), obtain the three-dimensional coordinates of the feature point MARK picture feature point, and use the PNP algorithm to estimate the relative pose of the camera and the curved MARK. The specific implementation is as follows: 世界坐标系的点Xw=(xw,yw,zw,1),与其在像平面上的投影坐标Xi=(xi,yi,1)的关系采用如下公式表示,其中fx,fy,cx,cy为张正友标定的相机内参,rij表示旋转变量,ti表示平移变量:The point X w =(x w , y w , z w , 1) of the world coordinate system, and the relationship between its projected coordinate X i =(x i , y i , 1) on the image plane is expressed by the following formula, where fx ,fy,cx,cy are the camera internal parameters calibrated by Zhang Zhengyou, r ij represents the rotation variable, t i represents the translation variable:
Figure RE-FDA0003039726550000023
Figure RE-FDA0003039726550000023
简化为:Simplifies to:
Figure RE-FDA0003039726550000024
Figure RE-FDA0003039726550000024
其中,λ表示比例因子,矩阵K为相机内部参数矩阵,矩阵M为模型视点矩阵;从30个特征点匹配对随机选择4个特征点匹配点对,使用其中3个点对来计算出4组不同的解,然后将剩下的1个点对代入公式,求取重投影误差最小的解为最终的解,在这个过程中使用随机采样一致性算法(RANSAC)优化最终解;Among them, λ represents the scale factor, the matrix K is the camera internal parameter matrix, and the matrix M is the model viewpoint matrix; 4 feature point matching point pairs are randomly selected from 30 feature point matching pairs, and 3 point pairs are used to calculate 4 groups. Different solutions, and then substitute the remaining 1 point pair into the formula to obtain the solution with the smallest reprojection error as the final solution, and use the random sampling consistency algorithm (RANSAC) to optimize the final solution in this process; 步骤(6)、在MARK移动过程中,使用光流法跟踪特征点移动状态,判断相同两帧特征点移动数量,当移动数量小于等于总特征点数量的百分之十时,认为标志物位姿相对于上一帧没有发生变化,当移动数量大于总特征点数量的百分之十时,认为标志物的位姿对于上一帧发生变化,则遵循步骤(5)获取到当前自然纹理MARK的位姿进行三维注册;Step (6), in the process of MARK moving, use the optical flow method to track the moving state of the feature points, and judge the moving quantity of the feature points in the same two frames. When the moving quantity is less than or equal to 10% of the total feature points, it is considered that the marker level The pose does not change relative to the previous frame. When the number of moves is greater than ten percent of the total feature points, it is considered that the pose of the marker has changed for the previous frame, and the current natural texture MARK is obtained by following step (5). The pose of the 3D registration is performed; 步骤(7)、在步骤(5)估计相机相对位姿的过程中,使用卡尔曼滤波对自然纹理MARK的6D位姿进行预测和矫正;Step (7), in the process of estimating the relative pose of the camera in step (5), use Kalman filtering to predict and correct the 6D pose of the natural texture MARK; 首先定义相机相对于自然纹理MARK位移的(tx,ty,tz)和旋转角度(ψ,θ,φ),坐标的一阶导数为(tx',ty',tz'),坐标的二阶导数为(tx”,t″y,tz”),其中一阶导数表示自然纹理MARK移动的速度,而二阶导数表示自然纹理MARK移动的加速度,旋转角度的一阶导数为(ψ’,θ’,φ’),二阶导数为(ψ”,θ”,φ”),其中一阶导数表示MARK旋转的速度,二阶导数表示自然纹理MARK旋转的加速度;可以利用卡尔曼滤波进行估计和校正,具体公式如下:First define the displacement (t x , t y , t z ) and rotation angle (ψ, θ, φ) of the camera relative to the natural texture MARK, and the first derivative of the coordinates is (t x ', t y ', t z ') , the second derivative of the coordinates is (t x ”, t” y , t z ”), where the first derivative represents the moving speed of the natural texture MARK, and the second derivative represents the acceleration of the natural texture MARK moving, the first order of the rotation angle The derivative is (ψ', θ', φ'), and the second derivative is (ψ", θ", φ"), where the first derivative represents the speed of MARK's rotation, and the second-order derivative represents the acceleration of natural texture MARK rotation; The Kalman filter is used for estimation and correction, and the specific formula is as follows: Kalman=(tx,ty,tz,tx′,ty′,tz′,tx″,ty″,tz″,ψ,θ,φ,ψ′,θ′,φ′,ψ″,θ″,φ″) (7)Kalman=(t x ,t y ,t z ,t x ′,t y ′,t z ′,t x ″,t y ″,t z ″,ψ,θ,φ,ψ′,θ′,φ′ ,ψ″,θ″,φ″) (7) 步骤(8)、使用滑动窗口对位姿估计错误的帧进行剔除;Step (8), use the sliding window to eliminate the frame with wrong pose estimation; 通过相对于当前帧的后两帧和前两帧的相机位姿坐标来判断当前所估计的相机位姿是否正确,从而剔除自然纹理MARK运动过程中由于模糊而导致相机位姿估计错误的帧;当前帧相机的6D位姿估计中位移为(xt,yt,zt)(不同含义的参数需要用不同符号表示),计算前两帧的相机的平均位移(x',y',z')和后两帧相机的平均位移(x",y",z"),当前位移满足:Judging whether the currently estimated camera pose is correct by relative to the camera pose coordinates of the last two frames and the first two frames of the current frame, so as to eliminate the frames that cause the wrong camera pose estimation due to blurring during the movement of the natural texture MARK; The displacement in the 6D pose estimation of the camera in the current frame is (x t , y t , z t ) (parameters with different meanings need to be represented by different symbols), and the average displacement of the camera in the first two frames is calculated (x', y', z ') and the average displacement of the camera in the next two frames (x", y", z"), the current displacement satisfies:
Figure RE-FDA0003039726550000031
Figure RE-FDA0003039726550000031
则认为当前位姿为有效位姿,否则认为当前帧为模糊帧,继续使用上一个有效位姿,其中dt为平移所调节的阈值,dt=3;Then the current pose is considered to be a valid pose, otherwise the current frame is considered to be a fuzzy frame, and the last valid pose is continued, where d t is the threshold adjusted by translation, d t =3; 步骤(9)、通过步骤(5)、(6)、(7)、(8)获取到相机与自然纹理MARK的相对位姿后,对需要进行三维注册的虚拟物体进行平移和旋转变换,并通过OpenGL和OpenCV将虚拟物体渲染到场景中完成增强现实的效果。Step (9), after obtaining the relative pose of the camera and the natural texture MARK through steps (5), (6), (7), and (8), translate and rotate the virtual object that needs to be registered in three dimensions, and Render virtual objects into the scene through OpenGL and OpenCV to complete the augmented reality effect.
2.根据权利要求1所述的一种基于区域检测的弯曲MARK三维注册增强现实方法,其特征在于,使用棱角分明、非规则、具有多特征的自然纹理特征的图片作为MARK,不选择对称的图片,选择的图片中有大量的图形、各个图形之间有明显的差异。2. a kind of curved MARK three-dimensional registration augmented reality method based on area detection according to claim 1, it is characterized in that, use sharp-edged, irregular, the picture with the natural texture feature of many features as MARK, do not choose symmetrical MARK. Pictures, there are a large number of graphics in the selected pictures, and there are obvious differences between the graphics. 3.根据权利要求1所述的一种基于区域检测的弯曲MARK三维注册增强现实方法,其特征在于,使用Fast算法计算出MARK图片中所有的特征点,具体步骤如下:3. a kind of curved MARK three-dimensional registration augmented reality method based on regional detection according to claim 1, is characterized in that, uses Fast algorithm to calculate all feature points in MARK picture, concrete steps are as follows: 步骤(a)、从MARK图片中选取一个像素Q,为了判断该像素点是否为特征点,首先将它的亮度值设为IqStep (a), select a pixel Q from the MARK picture, in order to judge whether this pixel is a feature point, at first its brightness value is set as Iq ; 步骤(b)、以像素点Q为中心,半径为3获取一个Bresenham圆,这个圆上有16个像素;Step (b), take the pixel point Q as the center, and obtain a Bresenham circle with a radius of 3, and there are 16 pixels on this circle; 步骤(c)、在这个大小为16个像素的圆上,如果有9个连续的像素点的像素值都大于Iq+t或都小于Iq+t,则认为像素点Q是一个特征点,所述的t为设定的阈值;Step (c), on this circle with a size of 16 pixels, if the pixel values of 9 consecutive pixel points are all greater than I q + t or less than I q + t, the pixel point Q is considered to be a feature point. , the t is the set threshold; 步骤(d)、为了提高角点的判断效率来排除图像中非角点的像素,按照1、9、5、13四个位置检查对应的像素,当像素点Q为角点时,则这四个位置的像素点的像素值至少有3个都大于Iq+t大或小于Iq+t,如果四个位置的像素点的像素值不满足这个条件,则像素点Q就不是角点,对所有的像素点都进行判断筛选,排除不是角点的像素点,将剩下的像素点在进行步骤(c)的操作判断获得最终的角点。Step (d), in order to improve the judgment efficiency of corner points to exclude non-corner pixels in the image, check the corresponding pixels according to the four positions of 1, 9, 5, and 13. When the pixel point Q is a corner point, then these four At least 3 of the pixel values of the pixel points in each position are larger than I q +t or smaller than I q + t. If the pixel values of the pixel points in the four positions do not meet this condition, then the pixel point Q is not a corner point, All pixel points are judged and screened, pixels that are not corner points are excluded, and the remaining pixel points are judged by the operation of step (c) to obtain the final corner point. 4.根据权利要求1或3所述的一种基于区域检测的弯曲MARK三维注册增强现实方法,其特征在于,使用ORB算法计算得到的特征点的描述子,具体步骤如下:4. a kind of curved MARK three-dimensional registration augmented reality method based on regional detection according to claim 1 and 3, is characterized in that, uses the descriptor of the feature point that ORB algorithm calculates, concrete steps are as follows: 步骤(e)、设关键点O圆心,以Or像素大小为半径做圆;Step (e), set the center of the key point O, and make a circle with the size of O r pixel as the radius; 步骤(f)、在圆内取N个点对,N=512;Step (f), take N point pairs in the circle, N=512; 步骤(g)、定义操作M,其中IA表示A的灰度,IB表示B的灰度:Step (g), define operation M, wherein IA represents the grayscale of A , and IB represents the grayscale of B:
Figure RE-FDA0003039726550000041
Figure RE-FDA0003039726550000041
步骤(h)、对已经选取的关键点进行步骤(g)操作得到由0,1构成的描述子组合。In step (h), step (g) is performed on the selected key points to obtain a descriptor combination composed of 0 and 1.
5.根据权利要求4所述的一种基于区域检测的弯曲MARK三维注册增强现实方法,其特征在于,步骤(6)具体实现如下:5. a kind of curved MARK three-dimensional registration augmented reality method based on area detection according to claim 4, is characterized in that, step (6) is specifically realized as follows: 设I和J为上一帧和当前帧的灰度图像,则:Let I and J be the grayscale images of the previous frame and the current frame, then:
Figure RE-FDA0003039726550000051
Figure RE-FDA0003039726550000051
其中点A为图像中任意一点,其坐标向量为(x,y)T,对于上一帧图像I上的一点u=[ux,uy]T,特征点跟踪的目的是找到其在当前帧图像中的位置v=u+d=[ux+dx,uy+dy]T,向量d=[dx,dy]T为点A的图像速度,即点A处的光流;由于光圈的影响,在二维邻域意义上定义相似性的概念,设置ωx和ωy为两个整数值,则对于速度向量d的最小化残差函数定义如下:The point A is any point in the image, and its coordinate vector is (x, y) T . For a point u=[u x , u y ] T on the previous frame of image I, the purpose of feature point tracking is to find its current The position v=u+d=[u x +d x , u y +d y ] T in the frame image, the vector d=[d x , dy ] T is the image speed of point A, that is, the light at point A flow; due to the influence of the aperture, the concept of similarity is defined in the sense of two-dimensional neighborhood, and ω x and ω y are set as two integer values, then the minimized residual function for the velocity vector d is defined as follows:
Figure RE-FDA0003039726550000052
Figure RE-FDA0003039726550000052
通过上述公式能够得到相似度定义,相似度定义是根据图像邻域大小为(2ωx+1)×(2ωy+1),对d的求解,得到点u在图像J中对应的位置;The similarity definition can be obtained through the above formula. The similarity definition is based on the image neighborhood size of (2ω x +1)×(2ω y +1), the solution of d, and the corresponding position of point u in the image J is obtained; 将当前帧计算得到的特征点位置与上一帧对应的特征点的位置进行比较,判断相机相邻两帧的特征点之间是否发生移动并对移动的特征点数量进行统计,如果当前帧特征点移动数量小于或等于总特征点数量的百分之十,则认为该物体相对于上一帧的物体并没有发生移动,直接获取上一帧的位姿即可,如果当前帧特征点移动的数量大于总特征点数量的百分之十,则认为该物体相对于上一帧的物体发生了移动,需要重新计算相机相对于物体的相对位姿。Compare the position of the feature point calculated in the current frame with the position of the feature point corresponding to the previous frame, determine whether there is movement between the feature points of the two adjacent frames of the camera, and count the number of moving feature points. The number of points moving is less than or equal to 10% of the total number of feature points, it is considered that the object has not moved relative to the object in the previous frame, and the pose of the previous frame can be directly obtained. If the feature point of the current frame moves If the number is greater than ten percent of the total number of feature points, it is considered that the object has moved relative to the object in the previous frame, and the relative pose of the camera relative to the object needs to be recalculated.
6.根据权利要求5所述的一种基于区域检测的弯曲MARK三维注册增强现实方法,其特征在于,ωx和ωy的取值为2,3,4,5,6或7。6 . The augmented reality method for curved MARK three-dimensional registration based on region detection according to claim 5 , wherein the values of ω x and ω y are 2, 3, 4, 5, 6 or 7. 7 .
CN202011563089.4A 2020-12-25 2020-12-25 Curved MARK three-dimensional registration augmented reality method based on region detection Active CN113012298B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011563089.4A CN113012298B (en) 2020-12-25 2020-12-25 Curved MARK three-dimensional registration augmented reality method based on region detection

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011563089.4A CN113012298B (en) 2020-12-25 2020-12-25 Curved MARK three-dimensional registration augmented reality method based on region detection

Publications (2)

Publication Number Publication Date
CN113012298A true CN113012298A (en) 2021-06-22
CN113012298B CN113012298B (en) 2022-04-08

Family

ID=76383738

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011563089.4A Active CN113012298B (en) 2020-12-25 2020-12-25 Curved MARK three-dimensional registration augmented reality method based on region detection

Country Status (1)

Country Link
CN (1) CN113012298B (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113643367A (en) * 2021-07-22 2021-11-12 上海智能网联汽车技术中心有限公司 Indoor positioning method and system based on augmented reality
CN118365765A (en) * 2024-06-14 2024-07-19 中南大学 A method for loading large-scale building scene models based on mixed reality

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110098056A1 (en) * 2009-10-28 2011-04-28 Rhoads Geoffrey B Intuitive computing methods and systems
US20110098029A1 (en) * 2009-10-28 2011-04-28 Rhoads Geoffrey B Sensor-based mobile search, related methods and systems
CN102142055A (en) * 2011-04-07 2011-08-03 上海大学 True three-dimensional design method based on augmented reality interactive technology
US20170206691A1 (en) * 2014-03-14 2017-07-20 Magic Leap, Inc. Augmented reality systems and methods utilizing reflections
CN110288657A (en) * 2019-05-23 2019-09-27 华中师范大学 A 3D Registration Method for Augmented Reality Based on Kinect

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110098056A1 (en) * 2009-10-28 2011-04-28 Rhoads Geoffrey B Intuitive computing methods and systems
US20110098029A1 (en) * 2009-10-28 2011-04-28 Rhoads Geoffrey B Sensor-based mobile search, related methods and systems
CN102142055A (en) * 2011-04-07 2011-08-03 上海大学 True three-dimensional design method based on augmented reality interactive technology
US20170206691A1 (en) * 2014-03-14 2017-07-20 Magic Leap, Inc. Augmented reality systems and methods utilizing reflections
CN110288657A (en) * 2019-05-23 2019-09-27 华中师范大学 A 3D Registration Method for Augmented Reality Based on Kinect

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
MINGMIN ZHANG等: "Dream-Experiment: A MR User Interface with Natural Multi-channel Interaction for Virtual Experiments", 《IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS》 *

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113643367A (en) * 2021-07-22 2021-11-12 上海智能网联汽车技术中心有限公司 Indoor positioning method and system based on augmented reality
CN118365765A (en) * 2024-06-14 2024-07-19 中南大学 A method for loading large-scale building scene models based on mixed reality
CN118365765B (en) * 2024-06-14 2024-08-16 中南大学 Large-scale building scene model loading method based on mixed reality

Also Published As

Publication number Publication date
CN113012298B (en) 2022-04-08

Similar Documents

Publication Publication Date Title
CN107292965B (en) Virtual and real shielding processing method based on depth image data stream
CN110458161B (en) Mobile robot doorplate positioning method combined with deep learning
CN110288657B (en) A Kinect-based Augmented Reality 3D Registration Method
CN111401266B (en) Method, equipment, computer equipment and readable storage medium for positioning picture corner points
CN103106688B (en) Based on the indoor method for reconstructing three-dimensional scene of double-deck method for registering
CN108122256B (en) A method of it approaches under state and rotates object pose measurement
CN110246168A (en) A kind of feature matching method of mobile crusing robot binocular image splicing
CN103839277A (en) Mobile augmented reality registration method of outdoor wide-range natural scene
CN112053447A (en) Augmented reality three-dimensional registration method and device
CN109919971B (en) Image processing method, apparatus, electronic device, and computer-readable storage medium
CN112819892B (en) Image processing method and device
CN107038758B (en) An Augmented Reality 3D Registration Method Based on ORB Operator
CN109003307B (en) Size design method of fishing nets based on underwater binocular vision measurement
CN117593650B (en) Moving point filtering vision SLAM method based on 4D millimeter wave radar and SAM image segmentation
CN111998862A (en) Dense binocular SLAM method based on BNN
CN113614735A (en) Dense 6-DoF gesture object detector
CN113012298B (en) Curved MARK three-dimensional registration augmented reality method based on region detection
CN113240656B (en) Visual positioning method and related device and equipment
Yang et al. Research on Tracking and Registration Algorithm Based on Natural Feature Point.
CN113723432B (en) Intelligent identification and positioning tracking method and system based on deep learning
CN115830135A (en) Image processing method and device and electronic equipment
CN113570535A (en) Visual positioning method and related device and equipment
Brunken et al. Incorporating Plane-Sweep in Convolutional Neural Network Stereo Imaging for Road Surface Reconstruction.
CN118089753B (en) Monocular semantic SLAM positioning method and system based on three-dimensional target
CN113920160B (en) Method, device, computer-readable medium and program product for estimating initial pose

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant