CN106568394A - Hand-held three-dimensional real-time scanning method - Google Patents
Hand-held three-dimensional real-time scanning method Download PDFInfo
- Publication number
- CN106568394A CN106568394A CN201510662328.4A CN201510662328A CN106568394A CN 106568394 A CN106568394 A CN 106568394A CN 201510662328 A CN201510662328 A CN 201510662328A CN 106568394 A CN106568394 A CN 106568394A
- Authority
- CN
- China
- Prior art keywords
- measurement
- point cloud
- point
- angle
- dimensional
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 80
- 238000005259 measurement Methods 0.000 claims abstract description 112
- 238000005286 illumination Methods 0.000 claims abstract description 7
- 230000010363 phase shift Effects 0.000 claims abstract description 7
- 238000001314 profilometry Methods 0.000 claims abstract description 7
- 238000002310 reflectometry Methods 0.000 claims abstract description 6
- 230000009466 transformation Effects 0.000 claims description 19
- 239000011159 matrix material Substances 0.000 claims description 18
- 238000005457 optimization Methods 0.000 claims description 6
- 238000000354 decomposition reaction Methods 0.000 claims description 4
- 206010069747 Burkholderia mallei infection Diseases 0.000 claims description 3
- 238000004458 analytical method Methods 0.000 claims description 3
- 238000000605 extraction Methods 0.000 claims description 3
- 239000004576 sand Substances 0.000 claims description 3
- FGUUSXIOTUKUDN-IBGZPJMESA-N C1(=CC=CC=C1)N1C2=C(NC([C@H](C1)NC=1OC(=NN=1)C1=CC=CC=C1)=O)C=CC=C2 Chemical compound C1(=CC=CC=C1)N1C2=C(NC([C@H](C1)NC=1OC(=NN=1)C1=CC=CC=C1)=O)C=CC=C2 FGUUSXIOTUKUDN-IBGZPJMESA-N 0.000 claims description 2
- 238000005070 sampling Methods 0.000 claims description 2
- 238000006243 chemical reaction Methods 0.000 claims 1
- HOWHQWFXSLOJEF-MGZLOUMQSA-N systemin Chemical compound NCCCC[C@H](N)C(=O)N[C@@H](CCSC)C(=O)N[C@@H](CCC(N)=O)C(=O)N[C@@H]([C@@H](C)O)C(=O)N[C@@H](CC(O)=O)C(=O)OC(=O)[C@@H]1CCCN1C(=O)[C@H]1N(C(=O)[C@H](CC(O)=O)NC(=O)[C@H](CCCN=C(N)N)NC(=O)[C@H](CCCCN)NC(=O)[C@H](CO)NC(=O)[C@H]2N(CCC2)C(=O)[C@H]2N(CCC2)C(=O)[C@H](CCCCN)NC(=O)[C@H](CO)NC(=O)[C@H](CCC(N)=O)NC(=O)[C@@H](NC(=O)[C@H](C)N)C(C)C)CCC1 HOWHQWFXSLOJEF-MGZLOUMQSA-N 0.000 claims 1
- 108010050014 systemin Proteins 0.000 claims 1
- 238000000691 measurement method Methods 0.000 description 4
- 238000001514 detection method Methods 0.000 description 2
- 238000010586 diagram Methods 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 241000282320 Panthera leo Species 0.000 description 1
- 238000006073 displacement reaction Methods 0.000 description 1
- 238000010191 image analysis Methods 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 238000011084 recovery Methods 0.000 description 1
- 230000002123 temporal effect Effects 0.000 description 1
- 238000011426 transformation method Methods 0.000 description 1
- 238000012795 verification Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01B—MEASURING LENGTH, THICKNESS OR SIMILAR LINEAR DIMENSIONS; MEASURING ANGLES; MEASURING AREAS; MEASURING IRREGULARITIES OF SURFACES OR CONTOURS
- G01B11/00—Measuring arrangements characterised by the use of optical techniques
- G01B11/24—Measuring arrangements characterised by the use of optical techniques for measuring contours or curvatures
- G01B11/25—Measuring arrangements characterised by the use of optical techniques for measuring contours or curvatures by projecting a pattern, e.g. one or more lines, moiré fringes on the object
- G01B11/254—Projection of a pattern, viewing through a pattern, e.g. moiré
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01B—MEASURING LENGTH, THICKNESS OR SIMILAR LINEAR DIMENSIONS; MEASURING ANGLES; MEASURING AREAS; MEASURING IRREGULARITIES OF SURFACES OR CONTOURS
- G01B11/00—Measuring arrangements characterised by the use of optical techniques
- G01B11/002—Measuring arrangements characterised by the use of optical techniques for measuring two or more coordinates
Landscapes
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Engineering & Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Length Measuring Devices By Optical Means (AREA)
Abstract
The invention provides a hand-held three-dimensional real-time scanning method. The method comprises steps that (1), a measurement angle of a measured object is selected, and an adjacent angle superposition rate is made to be greater than 60%; (2), a measurement mode is selected, and inching measurement or continuous measurement is carried out; (3), the exposure time of a camera is adjusted according to the information of different-angle illumination and reflectivity; (4), scanning measurement is carried out by the camera according to the selected measurement mode and the selected measurement angle, namely shooting; (5), the three frequency colorful phase shift profilometry and a colorful stripe decoupling method are employed to acquire three-dimensional point cloud and a texture graph of each angle; and (6), the acquired different-angle three-dimensional point cloud acquired in the step (5) are spliced to be an integral model. The method is advantaged in that the method is applicable to most measurement objects, especially dynamic objects, the method needs no artificial mark points, and a multi-angle measurement splicing problem can be excellently solved.
Description
Technical Field
The invention relates to the field of optical detection, in particular to a full-field rapid three-dimensional measurement method based on a novel multi-angle splicing method.
Background
The portable scanners in the market have smaller volumes and are convenient to carry, but the three-dimensional point cloud can be reconstructed after a multi-frame continuous exposure acquisition deformation fringe pattern is needed in each measurement, and the operation mode is generally as follows: selecting a position, fixing an instrument, measuring, moving the position, fixing and measuring. The working mode has low measuring efficiency, cannot realize handheld continuous scanning, and is particularly not suitable for scanning dynamic objects. Although some portable scanners can adopt a handheld scanning mode by means of a stereoscopic vision technology, mark points need to be marked on a measured object for positioning before measurement, so that the measurement efficiency is low, the workload is large, and points reconstructed by a machine vision method are related to the number and the positions of the marked mark points, so that dense point clouds cannot be reconstructed. The laser three-dimensional scanner can also be made into a portable scanner, which also requires the scanner to be kept still during the measurement process, so that the handheld scanning cannot be realized, and in addition, the laser scanner is point scanning or line scanning, so that the scanning efficiency is low.
Multi-angle stitching of handheld three-dimensional scanners is also a concern with respect to the several scanning techniques described above. On one hand, splicing can be achieved by calibrating the position of the multiple measurement scanner or the position of the measured object, but calibration requires auxiliary equipment, such as a displacement table, which increases cost and limits application occasions. On the other hand, the splicing is assisted by attaching some mark points on the surface of the measured object, which can cause the measurement period to be lengthened and the measurement to be complicated.
In view of the foregoing, there is a need for a handheld three-dimensional real-time scanner that can meet both handheld measurements and dynamic measurements, and that is better able to match up the angular measurements.
Disclosure of Invention
The invention aims to realize a portable and operational handheld scanning method which is suitable for most measuring objects, especially dynamic objects, and can well realize the splicing problem of multi-angle measurement without manual marking points.
The invention provides a hand-held three-dimensional real-time scanning method, which comprises the following steps:
(1) selecting a measurement angle of a measured object, and enabling the coincidence rate of adjacent angles to be larger than 60%;
(2) selection of measuring mode, jog or continuous measurement
Inching measurement, namely, performing measurement once when triggering is performed once under the condition of meeting the measurement angle in the step (1); continuous measurement, namely moving the scanner at a constant speed around a target object, and simultaneously continuously exposing the camera to realize continuous measurement;
(3) the camera automatically adjusts the exposure time according to the information of the illumination and the reflectivity at different angles;
(4) the camera performs scanning measurement according to the selected measurement mode and the selected measurement angle, namely photographing;
(5) obtaining three-dimensional point clouds and texture maps of all angles by adopting a three-frequency color phase-shift profilometry and a color fringe decoupling method;
(6) and (5) automatically splicing the three-dimensional point clouds with different angles obtained in the step (5) into a complete model.
In one embodiment of the invention, when a person carries out handheld measurement, the trigger speed of continuous measurement is set to be 5-10 frames/s.
In one embodiment of the present invention, the exposure automatic adjustment method of step (3) is as follows:
obtaining a high-frequency component of the three-frequency color stripes collected by the camera by adopting empirical mode decomposition and two-dimensional Hilbert transform;
extracting two-dimensional amplitude of a high-frequency component of a picture acquired by a camera by adopting two-dimensional Hilbert, and calculating the maximum value and the minimum value of the two-dimensional amplitude;
and calculating the difference between the maximum value and the minimum value of the fringe pattern acquired by the camera and the maximum value and the minimum value of the projection fringe pattern respectively, and adjusting the exposure time to ensure that the sum of the difference is minimum.
In one embodiment of the present invention, the tri-band color phase shift profilometry and texture retrieval method of step (5) is as follows:
projecting a three-frequency color sine fringe pattern on the surface of a measured object by adopting a color grating projection method;
setting RGB three-channel frequency ratio, and shooting a deformed fringe pattern projected on the surface of an object by using a color camera;
using Fourier transform to separate three carrier frequency fringe patterns of high, middle and low of the deformed fringe pattern, namely three frequencies;
respectively unfolding the phases of the three-frequency carrier frequency fringe pattern, and unwrapping the phases to recover the high-frequency carrier frequency phase;
obtaining three-dimensional point cloud through parameters calibrated by a system;
and obtaining a three-frequency sine fringe component by adopting a frequency domain minimization method for the deformed fringe pattern, and subtracting the three-frequency sine fringe component from the color deformed fringe pattern so as to recover the texture pattern of the surface of the measured object.
In one embodiment of the present invention, the RGB three channel frequency ratio is 1: 4: 12.
In one embodiment of the invention, the method of obtaining a three-dimensional point cloud is as follows:
(a) generating a three-frequency color fringe pattern by a computer, and projecting three-frequency color fringes onto the surface of a measured object by a projection device;
(b) shooting a measured object by a color camera to generate a monocular color deformation fringe pattern;
(c) separating high, medium and low three carrier frequencies by Fourier change to obtain three deformation fringe patterns of high, medium and low of a single unit;
(d) performing stripe self-adaptive analysis and phase extraction by using Hilbert-Huang transform HHT to obtain high, medium and low frequency wrapping phases;
(e) obtaining a high-precision absolute phase by a time domain variable precision unwrapping method;
(f) obtaining three-dimensional point cloud of the measured object by system calibration parameters according to the absolute phase obtained in the step (e);
(g) obtaining a texture map: respectively subtracting the high, middle and low three-frequency components obtained in the step (c) from the obtained color deformation fringe pattern to obtain background components; separating and recovering the obtained background components through texture illumination to obtain original texture information of the measured object;
(h) obtaining three-dimensional point cloud containing texture information: and (d) assigning the texture map obtained in the step (g) to the point cloud obtained in the step (f), so as to obtain the high-precision texture-containing three-dimensional point cloud of the measured object.
In one embodiment of the present invention, the automatic splicing method of step (6) is as follows:
assuming that a world coordinate system is coincident with a camera coordinate system under the 1 st angle, point clouds obtained by n angles measured each time are converted into the world coordinate system, and multi-angle splicing can be realized; the world coordinate system is expressed as (X)w,Yw,Zw,Ow) The camera coordinate system at each angle is represented as (X)i,Yi,Zi,Oi) Simultaneously (X)w,Yw,Zw,Ow)=(X1,Y1,Z1,O1) I.e. the world coordinates coincide with the camera coordinate system of the first location; the representation of the point cloud isWherein i represents the point cloud obtained by the ith measurement, j represents j points in the point cloud, and l represents coordinates under the l angular coordinate system;
1) splicing results of two adjacent angle scans
For the splicing of exposure measurement results of two adjacent angles, a method for splicing point clouds obtained by two adjacent scans is explained by taking the splicing of three-dimensional point clouds measured at the ith angle and the (i + 1) th angle as an example; wherein the point cloud obtained by the ith measurement is represented ask represents the number of points of the point cloud i; the point cloud obtained from the (i + 1) th measurement is represented ass represents the number of points of the point cloud i + 1; the specific splicing method comprises the following steps:
a) finding sparse corresponding feature points in the texture maps i and i +1 by using a scale invariant feature transformation method, wherein the sparse corresponding feature points are marked as { a1, a 2., am } and { b1, b 2., bm }, and m is the number of corresponding points found in the texture maps;
b) points with larger epipolar line errors are removed by a random sampling consistency method, the reserved points are recorded as { a1, a 2., ar } and { b1, b 2., br }, and r is the reserved corresponding points found in the texture map;
c) finding the spatial point coordinates { A1, A2, A3,. Ar } and { B1, B2, B3,. Br } corresponding to the matching feature points, wherein,representing the feature points in the ith point cloud,representing the corresponding point in the (i + 1) th point cloud, wherein r is less than or equal to s, k;
d) the iterative method solves the formula 1 to obtain p1, wherein p1 represents the (i + 1) th coordinate system (X)i+1,Yi+1,Zi+1,Oi1) In the i-th coordinate system (X)i,Yi,Zi,Oi) Pose potential matrix of (X)0,Y0,Z0) And represents the position of the position, gamma, theta,representing the attitude, i.e. the (i + 1) th coordinate system (X)i+1,Yi+1,Zi+1,Oi+1) With the i-th coordinate system (X)i,Yi,Zi,Oi) Angle between coordinate axes
Wherein p1 ═ R T],R=RotX·RotY·RotZ,T=[X0,Y0,,Z0,1]i,
e) All points in the point cloud i +1The matrix p1 is multiplied to the left, and the transformed matrix is converted into the coordinate system of the point cloud i, namely
f) Solving a space coordinate matrix p2 according to the consistency of the curvature of the corresponding point and the normal direction, wherein the meaning represented by p2 is the same as that of p 1;
g) iterative optimization is carried out on the splicing result by adopting an iterative closest point ICP iterative method, an optimization parameter p3 in a formula (2) is solved, the meaning of p3 is the same as that of p1, and coordinate transformation is carried out on the point cloud i +1 obtained in e) again;
f(p3)=∑||Mi,j-p3·Mi+1,j||2=min (2)
to this end, a transformation matrix p (i) ═ p3 · p2 · p1 is obtained by stitching the i +1 th point cloud and the i-th point cloud;
2) stitching of multiple exposure measurements
Obtaining P (1), P (2), P (3), and P (n-1) according to the method described in 1), splicing the ith point cloud with the 1 st point cloud obtained by the first exposure measurement, and performing coordinate transformationWherein,representing transformation of the i +1 th angle measurement to the world coordinate system (X)w,Yw,Zw,Ow) I.e., the coordinates of the 1 st measurement angle coordinate system,represents the coordinates of the (i + 1) th angle measurement result in the original coordinate system, i.e. the (i + 1) th measurement angle coordinate system, and is recorded as Hi=P(1)·P(2)·P(3).....P(i-1);
And (5) converting the 2 nd to n th angle measurement results into a world coordinate system according to the steps, completing the splicing of multi-angle measurement, and reconstructing a complete three-dimensional model.
In one embodiment of the present invention, the step of solving for p2 in step f) is:
the method comprises the following steps: initializing p2 to 0;
step two: respectively carrying out triangular patch division on the point cloud i and the point cloud i +1, namely constructing a triangular patch model with a connected topological relation; the point cloud i and the divided patch are marked as a model i and expressed asSimilarly, point cloud i +1 and its divided patch are denoted as model i +1, and are represented as
Step three: for model i +1, i.e.Performing a coordinate transformation of the transformation matrix P2, i.e.And calculating the curvature and normal direction of each point after coordinate transformation, and expressing as ri,1,ri,2,ri,3,...,ri,k},{ri+1,1,ri+1,2,ri+1,3,...,ri+1,sAnd { N }i,1,Ni,2,Ni,3,...,Ni,k},{Ni+1,1,Ni+1,2,Ni+1,3,...,Ni+1,s};
Step four: each point of the surface model i +1 is along its normal direction { Ni+1,1,Ni+1,2,Ni+1,3,...,Ni+1,sFinding the corresponding point on the surface model i;
define Q as a point on the surface model i +1If a point Q' on the surface model is found to satisfy the following two conditions, it is the corresponding point on the surface model i.
i.Q' connecting the Q line direction and the normal direction N of Qn+1,jAt very small included angles, i.e.
angle[|Q′-Q|,Nn+1,j]≤ω;
The distance between Q 'and Q is less than a certain range, namely distance | Q' -Q | is less than or equal to d;
of the points satisfying the i and ii conditions, the point having the closest curvature is selected as the matching point, i.e., the point having the closest curvature
f(Q′)=min|rQ′-ri+1,j|;
If the point meeting the conditions i and ii is not found, the point Q is considered to not find a matching point; wherein, omega and d are obtained by experimental calibration; corresponding points on the model i found by the model i +1 are respectively represented as a point set { Q }1,Q2,...,QvAnd { Q }1′,Q2′,...,Qv' }, v denotes the number of found corresponding point pairs, the curvatures of which are expressed asAnd
step five, calculating the curvature variance of the corresponding points
Step six, if g is less than or equal to sigma, outputting p 2; if g > σ, change p2 and repeat steps two through six.
The invention has the advantages that:
1) three-dimensional point cloud measurement can be completed by realizing single-frame exposure by using a method of projecting a three-frequency color sine fringe pattern, and the measurement result point cloud is dense and high in precision;
2) automatically adjusting the exposure time of the camera according to the information such as color, texture, reflectivity and the like of the measured object;
3) the two measurement modes of inching measurement and continuous measurement can be realized, and in the continuous measurement mode, the scanner can continuously expose at a certain frame rate to realize continuous measurement;
4) the single-angle three-dimensional measurement of the motion curved surface can be realized;
5) obtaining a three-frequency sine stripe component by a frequency domain minimization method, and subtracting the three-frequency sine stripe component from the color deformation stripe graph to recover the texture graph of the surface of the measured object;
6) and realizing the splicing of multi-angle and multi-frame exposure measurement according to the texture map recovered from the deformed stripe map, the curvature of the curved surface, the normal direction and other information.
Drawings
FIG. 1 illustrates a general flow diagram of the handheld three-dimensional scanning method of the present invention;
FIG. 2 shows a projected color sine-fringe pattern (R: G: B: 1: 4: 12);
FIG. 3 shows a flow chart of a three-frequency color sinusoidal fringe profiling and color texture restoration method;
FIG. 4 shows a flow chart of a method for stitching two adjacent angle measurements;
FIG. 5 is a diagram illustrating a multi-angle scanning result stitching method;
FIG. 6 illustrates a face 6 angular scan model;
FIG. 7 shows a multi-angle presentation of the stitched model.
Detailed Description
The present invention will be described in detail below with reference to the accompanying drawings.
1. Overall measurement method
The invention relates to a hand-held three-dimensional scanning method which is applied to a mobile, hand-held and portable three-dimensional scanner. The handheld three-dimensional scanning method has two measurement modes, namely inching measurement and continuous measurement. In both measurement modes, the measurement process and the measurement method are the same, and the overall measurement method is shown in fig. 1.
(1) And selecting the measurement angle of the measured object to ensure that the coincidence rate of adjacent angles is more than 60 percent.
For example, the measurement angles are sequentially selected to be about 0, 30, 60, 90,. and 330 degrees.
(2) Selecting a measuring mode, a jog measurement or a continuous measurement.
Inching measurement, namely, performing measurement once when triggering is performed once under the condition of meeting the measurement angle in the step (1); during continuous measurement, the scanner is moved at a constant speed around the target object, and the camera is continuously exposed, so that continuous measurement is realized, for example, for a field size of 0.5m × 0.5m, if the moving speed of an operator is 1m/s, and the trigger speed of the signal generator is 4 frames/s, the coincidence rate between two adjacent measurements can be ensured to be more than 60%. When people hold the measuring tool, in order to avoid overlarge data amount caused by too much overlapped parts, the triggering speed of continuous measurement can be set to be 5-10 frames/s through experimental verification.
(3) The camera automatically adjusts the exposure time according to the information of illumination, reflectivity and the like at different angles. The camera auto-adjustment method is described in detail in the following 2.
(4) And the camera performs scanning measurement, namely photographing according to the selected measurement mode and the selected measurement angle.
(5) And obtaining three-dimensional point clouds and texture maps of all angles by adopting a three-frequency color phase-shift profilometry and a color fringe decoupling method.
The detailed steps are as follows and shown in fig. 3.
(6) And (4) automatically splicing the three-dimensional point clouds with different angles obtained in the step (5) into a complete model, wherein the automatic splicing method is as shown in the following and in fig. 4 and 5.
2. Automatic exposure adjustment
The high-frequency components (Zou H-H, Zhou X, ZHao H, et al. color front-project detection for measuring two dimensional object based on two dimensional spatial temporal composition [ J ]. Applied optics, 2012, 51 (16): 3622) 3630.) of the tri-band color stripes collected by the camera are obtained by empirical mode decomposition and two-dimensional Hilbert transform, the two-dimensional amplitude of the high-frequency components of the photos collected by the camera is extracted by two-dimensional Hilbert, the maximum value and the minimum value of the two-dimensional amplitude are calculated, the difference between the maximum value and the minimum value of the stripe pattern collected by the camera and the maximum value and the minimum value of the projection stripe pattern are calculated, and the exposure time is adjusted to minimize the sum of the difference, so as to achieve the purpose of adjusting the exposure.
3. Tri-frequency color phase shift profilometry and texture recovery
The hand-held three-dimensional scanning method of the invention adopts a color grating projection method, projects a three-frequency color sine stripe graph shown in figure 2 to the surface of an object to be measured, the ratio of RGB three-channel frequencies is 1: 4: 12, shoots a deformed stripe graph projected on the surface of the object by a color camera, separates three carrier frequencies (Takeda M, Mutoh K. fourier transform profile for the automatic measurement of 3-D objects shape [ J ]. Applied optics, 1983, 22 (24): 397 < 7 > 3982.) (hereinafter referred to as "three-frequency"), then spreads the three-frequency phases respectively, performs phase unwrapping (Zhao H, Chen W, Tan Y. unwrapped for the three-dimensional measurement of three-dimensional shapes [ J ]. application, 449, 33, 4500), and finally obtains a three-dimensional carrier frequency point cloud (4500 ) through a point cloud calibration X system, xi J, Jin Y, et al, accurate calibration for a camera-projector system based on structured lighting project [ J ]. Optics and Lasers in Engineering, 2009, 47 (3): 310-319.). The deformed fringe pattern adopts a frequency domain minimization method (Zou H-H, Zhou X, ZHao H, et al color fringe-projected technical for measuring dynamic object based on two dimensional empirical mode decomposition [ J ]. Applied optics, 2012, 51 (16): 3622 and 3630.) to obtain a three-frequency sine fringe component, and the texture pattern of the measured object surface is recovered by subtracting the three-frequency sine fringe component from the color deformed fringe pattern.
The specific steps for obtaining a three-dimensional point cloud are shown in fig. 3 and explained as follows.
(a) Generating a three-frequency color fringe pattern by a computer, and projecting the three-frequency color fringe pattern onto the surface of a measured object by a projection device, as shown in FIG. 2;
(b) shooting a measured object by a color camera to generate a monocular color deformation fringe pattern;
(c) separating high, medium and low three carrier frequencies by Fourier change to obtain three deformation fringe patterns of high, medium and low of a single unit;
(d) performing stripe self-adaptive analysis and phase extraction by using Hilbert-Huang transform (HHT) to obtain high, medium and low three-frequency wrapping phases;
(e) obtaining a high-precision absolute phase by a time domain variable precision unwrapping method;
(f) and (e) obtaining the three-dimensional point cloud of the measured object by the system calibration parameters according to the absolute phase obtained in the step (e).
(g) Obtaining a texture map: respectively subtracting the high, middle and low three-frequency components obtained in the step (c) from the obtained color deformation fringe pattern to obtain background components; and the obtained background component is separated and recovered through texture illumination to obtain the original texture information of the measured object.
(h) Obtaining three-dimensional point cloud containing texture information: and (d) assigning the texture map obtained in the step (g) to the point cloud obtained in the step (f) to obtain the high-precision texture-containing three-dimensional point cloud of the measured object.
4. Automatic splicing method
The hand-held three-dimensional scanning method can realize automatic seamless splicing without mark points, and the splicing mainly depends on texture information recovered from a deformed fringe pattern, takes the curvature and the normal direction of measured data as constraints, searches the position relation between two adjacent frames of measurement results, and finally realizes global splicing.
The coordinate system of the point cloud scanned at each angle is superposed with the camera coordinate system of the scanning position at that time, so if the point cloud coordinates scanned at different angles are not in the same coordinate system, the point clouds scanned at different angles are converted into the point cloud under the same coordinate system, and the process is the splicing of the point clouds. And if the world coordinate system is overlapped with the camera coordinate system under the angle 1, converting point clouds obtained by n angles measured each time into the world coordinate system, and realizing multi-angle splicing. The world coordinate system is expressed as (X)w,Yw,Zw,Ow) The camera coordinate system at each angle is represented as (X)i,Yi,Zi,Oi) Simultaneously (X)w,Yw,Zw,Ow)=(X1,Y1,Z1,O1) I.e. the world coordinates coincide with the camera coordinate system of the first location; the representation of the point cloud isWherein i represents the point cloud obtained by the ith measurement, j represents j points in the point cloud, and l represents coordinates under the ith angular coordinate system.
1) Splicing results of two adjacent angle scans
For the splicing of exposure measurement results of two adjacent angles, the splicing method of point clouds obtained by two adjacent scans is described in detail below by taking the splicing of three-dimensional point clouds measured at the ith angle and the (i + 1) th angle as an example. Wherein the point cloud obtained by the ith measurement is represented ask represents the number of points of the point cloud i; point cloud obtained by i +1 th measurementIs shown ass represents the number of points of the point cloud i + 1; the splicing method is shown in fig. 4, and the specific steps are as follows.
a) A method of extracting Feature points using Scale-invariant Feature Transform (SIFT) (Lowe dg. discrete image features from innovative keys [ J ]. International journal of computer vision, 2004, 60 (2): 91-110.) find corresponding feature points sparse in texture maps i and i +1, which are marked as { a1, a2,. said., am } and { b1, b2,. said., bm }, where m is the corresponding point number found in the texture maps;
b) random Sample Consensus (RANSAC) method (Fischler MA, balls RC. A partner for Model Fitting with Applications to image analysis and Automated graphics (reprinted in reading in computer vision, ed. MA Fischler, "[ J ]. Comm ACM, 1981, 24 (6): 381) 395) was used. Removing points with larger epipolar line errors, and recording the reserved points as { a1, a2,. multidot., ar } and { b1, b2,. multidot.,. br }, wherein r is the reserved corresponding points found in the texture map;
c) finding the spatial point coordinates { A1, A2, A3,. Ar } and { B1, B2, B3,. Br } corresponding to the matching feature points, wherein,representing the feature points in the ith point cloud,and (3) representing the corresponding point in the (i + 1) th point cloud, wherein r is less than or equal to s, k.
d) The iterative method solves the formula 1 to obtain p1, wherein p1 represents the (i + 1) th coordinate system (X)i+1,Yi+1,Zi+1,Oi1) In the i-th coordinate system (X)i,Yi,Zi,Oi) Pose potential matrix of (X)0,Y0,Z0) And represents the position of the position, gamma, theta,representing the attitude, i.e. the (i + 1) th coordinate system (X)i+1,Yi+1,Zi+1,Oi+1) With the i-th coordinate system (X)i,Yi,Zi,Oi) The included angle of each coordinate axis.
Wherein p1 ═ R T],R=RotX·RotY·RotZ,T=[X0,Y0,,Z0,1]i,
e) All points in the point cloud i +1The matrix p1 is multiplied to the left, and the transformed matrix is converted into the coordinate system of the point cloud i, namely
f) According to the consistency of the curvature of the corresponding point and the normal direction, solving a space coordinate matrix p2, wherein the meaning represented by p2 is the same as that of p1, and the step of solving p2 is as follows:
the method comprises the following steps: initializing p2 to 0;
step two: respectively carrying out triangular patch division on the point cloud i and the point cloud i +1, namely constructing a triangular patch model with a connected topological relation; the point cloud i and the divided patch are marked as a model i and expressed asSimilarly, point cloud i +1 and its divided patch are denoted as model i +1, and are represented as
Step three: for model i +1, i.e.Performing a coordinate transformation of the transformation matrix to p2, i.e.And calculating the curvature and normal direction of each point after coordinate transformation, and expressing as ri,1,ri,2,ri,3,...,ri,k},{ri+1,1,ri+1,2,ri+1,3,...,ri+1,sAnd { N }i,1,Ni,2,Ni,3,...,Ni,k},{Ni+1,1,Ni+1,2,Ni+1,3,...,Ni+1,s};
Step four: each point of the surface model i +1 is along its normal direction { Ni+1,1,Ni+1,2,Ni+1,3,...,Ni+1,sFind the corresponding point on the surface model i. Define Q as a point on the surface model i +1If a point Q' on the surface model is found to satisfy the following two conditions, it is the corresponding point on the surface model i.
i.Q' connecting the Q line direction and the normal direction N of Qn+1,jAt very small included angles, i.e.
angle[|Q′-Q|,Nn+1,j]≤ω;
The distance between Q 'and Q is less than a certain range, namely distance | Q' -Q | is less than or equal to d;
among the points satisfying the conditions of i and ii, the point having the closest curvature is selected as a matching point, that is, the point having the closest curvature is selected as the matching point
f(Q′)=min|rQ′-ri+1,j|;
If the point meeting the conditions i and ii is not found, the point Q is considered to not find a matching point; wherein, omega and d are obtained by experimental calibration. Corresponding points on the model i found by the model i +1 are respectively represented as a point set { Q }1,Q2,...,Qv- } and { Q1′,Q2′,...,Qv' }, v denotes the number of found corresponding point pairs, the curvatures of which are expressed asAnd
step five, calculating the curvature variance of the corresponding points
Step six, if g is less than or equal to sigma, outputting p 2; if g > σ, change p2 and repeat steps two through six.
g) Iterative optimization is carried out on the splicing result by adopting an iterative Closest point ICP (iterative Closest point) iteration method, an optimization parameter p3 in a formula (2) is solved, the meaning represented by p3 is the same as that of p1, and coordinate transformation is carried out on the point cloud i +1 obtained in e) again;
f(p3)=∑||Mi,j-p3·Mi+1,j||2=min (2)
to this end, a transformation matrix p (i) ═ p3 · p2 · p1 is obtained by stitching the i +1 th point cloud and the i-th point cloud;
2) stitching of multiple exposure measurements
The stitching process for the multiple exposure measurements is shown in fig. 5.
According to the method described in 1), P (2), P (3), and P (n-1) can be obtained, and for the ith point cloud, the ith point cloud is spliced with the 1 st point cloud obtained by the first exposure measurement, and the required coordinate transformation is carried outWherein,representing transformation of the i +1 th angle measurement to the world coordinate system (X)w,Yw,Zw,Ow) I.e., the coordinates of the 1 st measurement angle coordinate system,represents the coordinates of the (i + 1) th angle measurement result in the original coordinate system, i.e. the (i + 1) th measurement angle coordinate system, and is recorded as Hi=P(1)·P(2)·P(3).....P(i-1)。
And (5) converting the 2 nd to n th angle measurement results into a world coordinate system according to the steps, completing the splicing of multi-angle measurement, and reconstructing a complete three-dimensional model.
Experimental cases:
the hand-held three-dimensional scanning equipment described by the invention is used for scanning the human face, and the human face has uneven reflectivity due to complex color, so that the problem is well solved by using the automatic exposure adjustment technology of the invention. Fig. 6 is a three-dimensional scan of a person with six angles, and fig. 7 is a multi-angle presentation of the results of the automatic stitching after scanning.
Claims (8)
1. A hand-held three-dimensional real-time scanning method comprises the following steps:
(1) selecting a measurement angle of a measured object, and enabling the coincidence rate of adjacent angles to be larger than 60%;
(2) selection of measuring mode, jog or continuous measurement
Inching measurement, namely, performing measurement once when triggering is performed once under the condition of meeting the measurement angle in the step (1); continuous measurement, namely moving the scanner at a constant speed around a target object, and simultaneously continuously exposing the camera to realize continuous measurement;
(3) the camera automatically adjusts the exposure time according to the information of the illumination and the reflectivity at different angles;
(4) the camera performs scanning measurement according to the selected measurement mode and the selected measurement angle, namely photographing;
(5) obtaining three-dimensional point clouds and texture maps of all angles by adopting a three-frequency color phase-shift profilometry and a color fringe decoupling method;
(6) and (5) automatically splicing the three-dimensional point clouds with different angles obtained in the step (5) into a complete model.
2. The handheld three-dimensional real-time scanning method as claimed in claim 1, wherein the trigger speed of continuous measurement is set to 5-10 frames/s during handheld measurement of a person.
3. The handheld three-dimensional real-time scanning method as claimed in claim 1, wherein the exposure automatic adjustment method of step (3) is as follows:
obtaining a high-frequency component of the three-frequency color stripes collected by the camera by adopting empirical mode decomposition and two-dimensional Hilbert transform;
extracting two-dimensional amplitude of a high-frequency component of a picture acquired by a camera by adopting two-dimensional Hilbert, and calculating the maximum value and the minimum value of the two-dimensional amplitude;
and calculating the difference between the maximum value and the minimum value of the fringe pattern acquired by the camera and the maximum value and the minimum value of the projection fringe pattern respectively, and adjusting the exposure time to ensure that the sum of the difference is minimum.
4. The handheld three-dimensional real-time scanning method as claimed in claim 1, wherein the tri-band color phase shift profilometry and texture retrieval method of step (5) is as follows:
projecting a three-frequency color sine fringe pattern on the surface of a measured object by adopting a color grating projection method;
setting RGB three-channel frequency ratio, and shooting a deformed fringe pattern projected on the surface of an object by using a color camera;
using Fourier transform to separate three carrier frequency fringe patterns of high, middle and low of the deformed fringe pattern, namely three frequencies;
respectively unfolding the phases of the three-frequency carrier frequency fringe pattern, and unwrapping the phases to recover the high-frequency carrier frequency phase;
obtaining three-dimensional point cloud through parameters calibrated by a system;
and obtaining a three-frequency sine fringe component by adopting a frequency domain minimization method for the deformed fringe pattern, and subtracting the three-frequency sine fringe component from the color deformed fringe pattern so as to recover the texture pattern of the surface of the measured object.
5. The handheld three-dimensional real-time scanning method of claim 4, wherein the RGB three-channel frequency ratio is 1: 4: 12.
6. The handheld three-dimensional real-time scanning method of claim 4, wherein the three-dimensional point cloud is obtained by:
(a) generating a three-frequency color fringe pattern by a computer, and projecting three-frequency color fringes onto the surface of a measured object by a projection device;
(b) shooting a measured object by a color camera to generate a monocular color deformation fringe pattern;
(c) separating high, medium and low three carrier frequencies by Fourier change to obtain three deformation fringe patterns of high, medium and low of a single unit;
(d) performing stripe self-adaptive analysis and phase extraction by using Hilbert-Huang transform HHT to obtain high, medium and low frequency wrapping phases;
(e) obtaining a high-precision absolute phase by a time domain variable precision unwrapping method;
(f) obtaining three-dimensional point cloud of the measured object by system calibration parameters according to the absolute phase obtained in the step (e);
(g) obtaining a texture map: respectively subtracting the high, middle and low three-frequency components obtained in the step (c) from the obtained color deformation fringe pattern to obtain background components; separating and recovering the obtained background components through texture illumination to obtain original texture information of the measured object;
(h) obtaining three-dimensional point cloud containing texture information: and (d) assigning the texture map obtained in the step (g) to the point cloud obtained in the step (f), so as to obtain the high-precision texture-containing three-dimensional point cloud of the measured object.
7. The handheld three-dimensional real-time scanning method as claimed in claim 1, wherein the automatic stitching method of step (6) is as follows:
assuming that a world coordinate system is coincident with a camera coordinate system under the 1 st angle, point clouds obtained by n angles measured each time are converted into the world coordinate system, and multi-angle splicing can be realized; the world coordinate system is expressed as (X)w,Yw,Zw,Ow) The camera coordinate system at each angle is represented as (X)i,Yi,Zi,Oi) At the same time (X)w,Yw,Zw,Ow)=(X1,Y1,Z1,O1) I.e. the world coordinates coincide with the camera coordinate system of the first location; the representation of the point cloud isWherein i represents the point cloud obtained by the ith measurement, j represents j points in the point cloud, and l represents coordinates under the l angular coordinate system;
1) splicing results of two adjacent angle scans
For the splicing of exposure measurement results of two adjacent angles, a method for splicing point clouds obtained by two adjacent scans is explained by taking the splicing of three-dimensional point clouds measured at the ith angle and the (i + 1) th angle as an example; wherein the point cloud obtained by the ith measurement is represented ask represents the number of points of the point cloud i; the point cloud obtained from the (i + 1) th measurement is represented ass represents the number of points of the point cloud i + 1; the specific splicing method comprises the following steps:
a) finding sparse corresponding feature points in the texture map i and i +1 by using a scale invariant feature conversion method, marking as { a1, a2, …, am } and { b1, b2, …, bm }, wherein m is the number of corresponding points found in the texture map;
b) points with larger epipolar line errors are removed by a random sampling consistency method, the reserved points are marked as { a1, a2, …, ar } and { b1, b2, …, br }, and r is the reserved corresponding point number found in the texture map;
c) finding the spatial point coordinates { A1, A2, A3, … Ar } and { B1, B2, B3, … Br } corresponding to the matching feature points, wherein,representing the feature points in the ith point cloud,representing the corresponding point in the (i + 1) th point cloud, wherein r is less than or equal to s, k;
d) solving the formula 1 by an iterative method to obtain p1, wherein p1 represents the (i + 1) th coordinate systemIn the i-th coordinate system (X)i,Yi,Zi,Oi) Pose potential matrix of (X)0,Y0,Z0) And represents the position of the position, gamma, theta,representing the attitude, i.e. the (i + 1) th coordinate system (X)i+1,Yi+1,Zi+1,Oi+1) With the i-th coordinate system (X)i,Yi,Zi,Oi) Angle between coordinate axes
Wherein p1 ═ R T],R=RotX·RotY·RotZ,T=[X0.Y0.,Z0.1]l,
e) All points in the point cloud i +1The matrix p1 is multiplied to the left, and the transformed matrix is converted into the coordinate system of the point cloud i, namely
f) Solving a space coordinate matrix p2 according to the consistency of the curvature of the corresponding point and the normal direction, wherein the meaning represented by p2 is the same as that of p 1;
g) iterative optimization is carried out on the splicing result by adopting an iterative closest point ICP iterative method, an optimization parameter p3 in a formula (2) is solved, the meaning of p3 is the same as that of p1, and coordinate transformation is carried out on the point cloud i +1 obtained in e) again;
f(p3)=Σ||Mi,j-p3·Mi+1,j||2=min (2)
to this end, a transformation matrix p (i) ═ p3 · p2 · p1 is obtained by stitching the i +1 th point cloud and the i-th point cloud;
2) stitching of multiple exposure measurements
Obtaining P (1), P (2), P (3), and P (n-1) according to the method described in 1), splicing the ith point cloud with the 1 st point cloud obtained by the first exposure measurement, and performing coordinate transformationWherein,representing transformation of the i +1 th angle measurement to the world coordinate system (X)w,Yw,Zw,Ow) I.e., the coordinates of the 1 st measurement angle coordinate system,represents the coordinates of the (i + 1) th angle measurement result in the original coordinate system, i.e. the (i + 1) th measurement angle coordinate system, and is recorded as Hi=P(1)·P(2)·P(3)……P(i-1);
And (5) converting the 2 nd to n th angle measurement results into a world coordinate system according to the steps, completing the splicing of multi-angle measurement, and reconstructing a complete three-dimensional model.
8. The handheld three-dimensional real-time scanning method as recited in claim 7, wherein the step of solving for p2 in step f) is:
the method comprises the following steps: initializing p2 to 0;
step two: respectively carrying out triangular patch division on the point cloud i and the point cloud i +1, namely constructing a triangular patch model with a connected topological relation; the point cloud i and the divided patch are marked as a model i and expressed asSimilarly, point cloud i +1 and its divided patch are denoted as model i +1, and are represented as
Step three: for model i +1, i.e.Performing a coordinate transformation of the transformation matrix to p2, i.e.And calculating the curvature and normal direction of each point after coordinate transformation, and expressing as ri,1,ri,2,ri,3,…,ri,k},{ri+1,1,ri+1,2,ri+1,3,…,ri+1,sAnd { N }i,1,Ni,2,Ni,3,…,Ni,k},{Ni+1,1,Ni+1,2,Ni+1,3,…,Ni+1,s};
Step four: each point of the surface model i +1 is along its normal direction { Ni+1,1,Ni+1,2,Ni+1,3,…,Ni-1,sFinding the corresponding point on the surface model i;
define Q as a point on the surface model i +1If a point Q' on the surface model is found to satisfy the following two conditions, it is the corresponding point on the surface model i.
The direction of the line connecting Q' and Q and the normal direction N of Qn+1,jAt very small included angles, i.e.
angle[|Q′-Q|,Nn+1,j]≤ω;
The distance between Q 'and Q is less than a certain range, namely distance | Q' -Q | is less than or equal to d;
of the points satisfying the conditions i and ii, the point having the closest curvature is selected as the matching point, i.e., f (Q') ═ min | rQ′-ri+1,j|;
If the point meeting the conditions i and ii is not found, the point Q is considered to not find a matching point; wherein, omega and d are obtained by experimental calibration; corresponding points on the model i found by the model i +1 are respectively represented as a point set { Q }1,Q2,...,QvAnd { Q }1′,Q2′,...,Qv' }, v denotes the number of found corresponding point pairs, the curvatures of which are expressed asAnd
step five, calculating the curvature variance of the corresponding points
Step six, if g is less than or equal to sigma, outputting p 2; if g > σ, change p2 and repeat steps two through six.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201510662328.4A CN106568394A (en) | 2015-10-09 | 2015-10-09 | Hand-held three-dimensional real-time scanning method |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201510662328.4A CN106568394A (en) | 2015-10-09 | 2015-10-09 | Hand-held three-dimensional real-time scanning method |
Publications (1)
Publication Number | Publication Date |
---|---|
CN106568394A true CN106568394A (en) | 2017-04-19 |
Family
ID=58508832
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201510662328.4A Pending CN106568394A (en) | 2015-10-09 | 2015-10-09 | Hand-held three-dimensional real-time scanning method |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN106568394A (en) |
Cited By (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106705885A (en) * | 2017-01-17 | 2017-05-24 | 天津微深通用科技有限公司 | Color three-dimensional optical scanning method capable of processing transparent and highlight materials |
WO2019015154A1 (en) * | 2017-07-17 | 2019-01-24 | 先临三维科技股份有限公司 | Monocular three-dimensional scanning system based three-dimensional reconstruction method and apparatus |
CN109472741A (en) * | 2018-09-30 | 2019-03-15 | 先临三维科技股份有限公司 | Three-dimensional splicing method and device |
WO2019100700A1 (en) * | 2017-11-24 | 2019-05-31 | Guangdong Kang Yun Technologies Limited | Portable 3d scanning systems and scanning methods |
CN110243309A (en) * | 2019-06-18 | 2019-09-17 | 西安交通大学 | A high-efficiency and high-precision structured light three-dimensional measurement system recalibration method |
CN111023970A (en) * | 2019-12-17 | 2020-04-17 | 杭州思看科技有限公司 | Multi-mode three-dimensional scanning method and system |
CN111649690A (en) * | 2019-12-12 | 2020-09-11 | 天目爱视(北京)科技有限公司 | Handheld 3D information acquisition equipment and method |
CN118258309A (en) * | 2024-05-29 | 2024-06-28 | 华东交通大学 | Three-dimensional displacement measurement method and system based on polarization fringe projection |
CN118882517A (en) * | 2024-07-05 | 2024-11-01 | 西安知象光电科技有限公司 | A three-dimensional scanning system and method based on head mounted display device |
CN118882517B (en) * | 2024-07-05 | 2025-04-15 | 西安知象光电科技有限公司 | A three-dimensional scanning system and method based on head mounted display device |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN1374620A (en) * | 2002-03-27 | 2002-10-16 | 深圳市特得维技术有限公司 | Computerized 3D visual color scan system and its scanning mode |
CN101739717A (en) * | 2009-11-12 | 2010-06-16 | 天津汇信软件有限公司 | Non-contact scanning method for three-dimensional colour point clouds |
CN102628676A (en) * | 2012-01-19 | 2012-08-08 | 东南大学 | Adaptive window Fourier phase extraction method in optical three-dimensional measurement |
CN104299211A (en) * | 2014-09-25 | 2015-01-21 | 周翔 | Free-moving type three-dimensional scanning method |
-
2015
- 2015-10-09 CN CN201510662328.4A patent/CN106568394A/en active Pending
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN1374620A (en) * | 2002-03-27 | 2002-10-16 | 深圳市特得维技术有限公司 | Computerized 3D visual color scan system and its scanning mode |
CN101739717A (en) * | 2009-11-12 | 2010-06-16 | 天津汇信软件有限公司 | Non-contact scanning method for three-dimensional colour point clouds |
CN102628676A (en) * | 2012-01-19 | 2012-08-08 | 东南大学 | Adaptive window Fourier phase extraction method in optical three-dimensional measurement |
CN104299211A (en) * | 2014-09-25 | 2015-01-21 | 周翔 | Free-moving type three-dimensional scanning method |
Cited By (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106705885A (en) * | 2017-01-17 | 2017-05-24 | 天津微深通用科技有限公司 | Color three-dimensional optical scanning method capable of processing transparent and highlight materials |
WO2019015154A1 (en) * | 2017-07-17 | 2019-01-24 | 先临三维科技股份有限公司 | Monocular three-dimensional scanning system based three-dimensional reconstruction method and apparatus |
US10783651B2 (en) | 2017-07-17 | 2020-09-22 | Shining 3D Tech Co., Ltd. | Three-dimensional reconstruction method and device based on monocular three-dimensional scanning system |
WO2019100700A1 (en) * | 2017-11-24 | 2019-05-31 | Guangdong Kang Yun Technologies Limited | Portable 3d scanning systems and scanning methods |
CN109472741A (en) * | 2018-09-30 | 2019-03-15 | 先临三维科技股份有限公司 | Three-dimensional splicing method and device |
CN110243309A (en) * | 2019-06-18 | 2019-09-17 | 西安交通大学 | A high-efficiency and high-precision structured light three-dimensional measurement system recalibration method |
CN111649690A (en) * | 2019-12-12 | 2020-09-11 | 天目爱视(北京)科技有限公司 | Handheld 3D information acquisition equipment and method |
CN111023970A (en) * | 2019-12-17 | 2020-04-17 | 杭州思看科技有限公司 | Multi-mode three-dimensional scanning method and system |
WO2021121320A1 (en) * | 2019-12-17 | 2021-06-24 | 杭州思看科技有限公司 | Multi-mode three-dimensional scanning method and system |
US12260595B2 (en) | 2019-12-17 | 2025-03-25 | Scantech (Hangzhou) Co., Ltd. | Multi-mode three-dimensional scanning method and system |
CN118258309A (en) * | 2024-05-29 | 2024-06-28 | 华东交通大学 | Three-dimensional displacement measurement method and system based on polarization fringe projection |
CN118882517A (en) * | 2024-07-05 | 2024-11-01 | 西安知象光电科技有限公司 | A three-dimensional scanning system and method based on head mounted display device |
CN118882517B (en) * | 2024-07-05 | 2025-04-15 | 西安知象光电科技有限公司 | A three-dimensional scanning system and method based on head mounted display device |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN104299211B (en) | Free-moving type three-dimensional scanning method | |
CN106568394A (en) | Hand-held three-dimensional real-time scanning method | |
CN110288642B (en) | Three-dimensional object rapid reconstruction method based on camera array | |
CN110880185B (en) | High-precision dynamic real-time 360-degree omnidirectional point cloud acquisition method based on fringe projection | |
CN106931910B (en) | A kind of efficient acquiring three-dimensional images method based on multi-modal composite coding and epipolar-line constraint | |
JP5337243B2 (en) | Adaptive 3D scanning system for surface features | |
DK3144881T3 (en) | PROCEDURE FOR 3D PANORAMA MOSAIC CREATION OF A SCENE | |
CN102506757B (en) | Self-positioning method in multi-angle measurement of binocular stereo measurement system | |
Ben-Hamadou et al. | Flexible calibration of structured-light systems projecting point patterns | |
CN101608908A (en) | The three-dimension digital imaging method that digital speckle projection and phase measuring profilometer combine | |
CN102184566A (en) | Micro projector mobile phone platform-based portable three-dimensional scanning system and method | |
CN111189416A (en) | Structural light 360-degree three-dimensional surface shape measuring method based on characteristic phase constraint | |
CN114170284A (en) | Multi-view point cloud registration method based on active landmark projection assistance | |
CN117073579A (en) | Structured light binocular three-dimensional measurement system and method based on stripe projection | |
CN117450955B (en) | Three-dimensional measurement method of thin objects based on spatial annular features | |
US7046839B1 (en) | Techniques for photogrammetric systems | |
CN117635843A (en) | Three-dimensional human posture reconstruction method based on multi-camera self-calibration technology | |
D Apuzzo | Measurement and modeling of human faces from multi images | |
Skabek et al. | Comparison of photgrammetric techniques for surface reconstruction from images to reconstruction from laser scanning | |
Barazzetti | Planar metric rectification via parallelograms | |
CN112325799A (en) | High-precision three-dimensional face measurement method based on near-infrared light projection | |
Croci et al. | Dimensional accuracy assessment of 3D models based on photogrammetry and 3D scanner: a case study from the Museo Egizio of Turin | |
D'Apuzzo | Human face modeling from multi images | |
CN113375593B (en) | Visual measurement method and related device for mirror plate shape of reversible hydroelectric generating set | |
WO2019051728A1 (en) | Three-dimensional digital imaging method and device for wrapped phase based on phase mapping |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
WD01 | Invention patent application deemed withdrawn after publication |
Application publication date: 20170419 |
|
WD01 | Invention patent application deemed withdrawn after publication |