CN109360267B - Rapid three-dimensional reconstruction method for thin object - Google Patents
Rapid three-dimensional reconstruction method for thin object Download PDFInfo
- Publication number
- CN109360267B CN109360267B CN201811147802.XA CN201811147802A CN109360267B CN 109360267 B CN109360267 B CN 109360267B CN 201811147802 A CN201811147802 A CN 201811147802A CN 109360267 B CN109360267 B CN 109360267B
- Authority
- CN
- China
- Prior art keywords
- base
- point
- sample
- transformation matrix
- points
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000000034 method Methods 0.000 title claims abstract description 24
- 230000009466 transformation Effects 0.000 claims abstract description 39
- 239000011159 matrix material Substances 0.000 claims abstract description 33
- 230000000007 visual effect Effects 0.000 claims abstract description 18
- 239000013598 vector Substances 0.000 claims description 10
- 238000005516 engineering process Methods 0.000 description 3
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000006243 chemical reaction Methods 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000012804 iterative process Methods 0.000 description 1
- 239000000463 material Substances 0.000 description 1
- 238000005457 optimization Methods 0.000 description 1
- 238000012800 visualization Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T17/00—Three dimensional [3D] modelling, e.g. data description of 3D objects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformations in the plane of the image
- G06T3/60—Rotation of whole images or parts thereof
- G06T3/604—Rotation of whole images or parts thereof using coordinate rotation digital computer [CORDIC] devices
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/50—Depth or shape recovery
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02P—CLIMATE CHANGE MITIGATION TECHNOLOGIES IN THE PRODUCTION OR PROCESSING OF GOODS
- Y02P10/00—Technologies related to metal processing
- Y02P10/25—Process efficiency
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Computer Graphics (AREA)
- Geometry (AREA)
- Software Systems (AREA)
- Image Processing (AREA)
Abstract
The invention discloses a method for quickly reconstructing three dimensions of a thin object, which comprises the following steps: and fixing the thin object to be reconstructed on a base with marking points, acquiring a depth image and a color image by a depth camera, identifying the marking points, obtaining a transformation matrix, carrying out coordinate transformation on the point cloud of the object to be reconstructed according to the transformation matrix, and finally splicing the point clouds of two visual angles of the transformed object, thereby completing the three-dimensional reconstruction of the sample. The three-dimensional reconstruction of the thin object is realized by splicing the point clouds of the two visual angles based on the mark points, so that the reconstruction speed is high, the reconstruction precision is high and the operation is simple for the thin object with the thickness of about 2 mm-30 mm. Meanwhile, the invention only needs one depth camera and one base with simple mark points, and has lower cost.
Description
Technical Field
The invention relates to the technical field of computer vision, in particular to a rapid three-dimensional reconstruction method for a thin object.
Background
With the development of computer vision, the three-dimensional reconstruction technology is increasingly widely applied in the fields of computer-aided medical treatment, reverse engineering, industrial automatic detection and the like. The three-dimensional reconstruction technology of the real object comprises reconstruction based on professional software, reconstruction by a computer vision method and the like. The modeling technology by special software is very mature, has wide application and good modeling effect, but the special modeling software for application needs to be trained by professionals. A lot of manpower and material resources are consumed. In the field of computer vision, an Iterative Closest Point (ICP) algorithm is a common method in point cloud registration, and accurate registration of two point sets can be realized through continuous iterative optimization matrix, but the algorithm has strong dependence on the given initial position and the corresponding relation in the iterative process, and has the problems of large calculated amount, long modeling time, local optimum sinking of iteration and the like.
Disclosure of Invention
Aiming at the defects, the invention provides a rapid three-dimensional reconstruction method for a thin object, which has the advantages of high speed, high precision, low cost and simple operation, by using a mark point and a depth camera.
In order to achieve the above purpose, the technical scheme adopted by the invention is as follows: a method for quickly reconstructing three dimensions of a thin object comprises the following steps:
and fixing the thin object to be reconstructed on a base with marking points, acquiring a depth image and a color image by a depth camera, identifying the marking points, obtaining a transformation matrix, carrying out coordinate transformation on the point cloud with the reconstructed object according to the transformation matrix, and finally splicing the point clouds of two visual angles of the transformed object, thereby completing the three-dimensional reconstruction of the sample.
Further, the method comprises the following steps:
(1) Fixing a sample to be three-dimensionally reconstructed on a base with marking points, wherein the shape of the front surface and the back surface of the base are completely the same;
(2) Collecting a color image and a depth image of a first visual angle (front) comprising a base and a sample by using a depth camera, and converting the color image and the depth image into point clouds by combining internal and external parameters of the camera;
(3) Detecting mark points in the base according to the color image;
(4) Repeating the step (2) -the step (3) by rotating the base to obtain a point cloud and a marked point of a second view angle (back);
(5) Calculating transformation matrices T of two views by using the mark points detected in the step (3) and the step (4), respectively 1 ,T 2 The method comprises the steps of carrying out a first treatment on the surface of the Rotating both the front and back of the base to be parallel to the xoy plane through the transformation matrix, and translating the center of the base to the origin of the camera coordinate system;
(6) Using a transformation matrix T 1 ,T 2 Respectively carrying out coordinate transformation on point clouds of two visual angles of the sample, and enabling the point clouds of a second visual angle of the sample to pass through a transformation matrix T 3 And rotating the sample by 180 degrees around the Y axis, translating the sample along the Z axis by a value which is the same as the thickness of the base, and finally splicing the transformed point cloud to finish the three-dimensional reconstruction of the sample.
Further, the marked point on the base in the step (3) is marked as (p 0 ,p 1 ,p 2 ,p 3 )。
Further, the transform matrix of the two views in the step (5) is obtained as follows:
by calculating the center point coordinates (x c ,y c ,z c ) Thereby obtaining the front S of the base 1 The relative distance between the center point and the origin point can obtain a translation vector t;
front S of base 1 The coordinates of the center point of (c) are:
(x c ,y c ,z c )=(p 5 +p 6 )/2 (2)
wherein: p is p 5 For marking point p 0 ,p 1 At the center of two points, p is 5 =(p 0 +p 1 )/2;
p 6 For marking point p 2 ,p 3 At the center of two points, p is 6 =(p 2 +p 3 )/2;
Front S of the base 1 Rotated to be parallel to the xoy plane, the rotation matrix R is shown as formula (3):
wherein r is 3 =[a,b,c]The method comprises the steps of carrying out a first treatment on the surface of the Vectors [ a, b, c ]]Is S 1 Normal vector of the plane;
r 1 =(p 5 -p 6 ).normalized();
r 2 =r 3 .cross(r 1 );
the transformation matrix for the first viewing angle of the base is shown in the following formula:
similarly, a transformation matrix T of the second view angle of the base can be obtained 2 。
Further, the step (6) specifically includes the following steps:
the three-dimensional reconstruction model of the sample is specifically:
C=T 1 *C 1 +T 3 *T 2 *C 2 (5)
T 3 as shown in formula (6):
wherein l is the thickness of the base; c is the reconstructed point cloud, C 1 ,C 2 The point clouds of the front and back sides of the sample before reconstruction, respectively.
The invention has the beneficial effects that: and fixing the thin object to be reconstructed on a base with marking points, acquiring a depth image and a color image by a depth camera, identifying the marking points, obtaining a transformation matrix, carrying out coordinate transformation on the point cloud with the reconstructed object according to the transformation matrix, and finally splicing the point clouds of two visual angles of the transformed object, thereby completing the three-dimensional reconstruction of the sample. The three-dimensional reconstruction of the object can be realized by only collecting two visual angles of the object to be reconstructed, so the reconstruction speed is high. Meanwhile, only one depth camera and a base with marking points are needed, and the cost is low. Compared with the traditional ICP algorithm, the method can complete reconstruction more quickly, and has the advantages of high reconstruction accuracy, simple cost, strong expandability, visualization and the like. The three-dimensional reconstruction of the thin object is realized by splicing the point clouds of the two visual angles based on the mark points, so that the reconstruction speed is high, the reconstruction precision is high and the operation is simple for the thin object with the thickness of about 2 mm-30 mm. Meanwhile, the invention only needs one depth camera and one base with simple mark points, and has lower cost.
Drawings
FIG. 1 is a flow chart of the present invention;
FIG. 2 is a schematic view of a base marked with a mark point;
FIG. 3 is a schematic view of the base before and after coordinate transformation;
fig. 4 is a schematic diagram of the three-dimensional reconstruction result of a thin object.
Detailed Description
Embodiments of the technical scheme of the present invention are described in detail below with reference to the accompanying drawings. The following examples are given only for the purpose of more clearly illustrating the technical solutions of the present invention, and are therefore only exemplary and not intended to limit the scope of the present invention.
As shown in fig. 1, the invention provides a rapid three-dimensional reconstruction method for a thin object, which is characterized in that the thin object to be reconstructed is fixed on a base with marking points, a depth image and a color image are acquired through a depth camera, the marking points are identified, a transformation matrix can be obtained, coordinate transformation is carried out on point clouds with the reconstructed object according to the transformation matrix, and finally, the point clouds with two visual angles of the transformed object are spliced, so that the three-dimensional reconstruction of a sample is completed.
The method specifically comprises the following steps:
step 1: fixing a sample to be three-dimensionally reconstructed on a base with marking points, wherein the shape of the front surface and the back surface of the base are completely the same, and the specific structure size is known;
step 2: collecting a color image and a depth image of a first visual angle (front) comprising a base and a sample by using a depth camera, and converting the color image and the depth image into point clouds by combining internal and external parameters of the camera;
step 3: from the color image, the mark points in the base are detected, and the detected mark points provided in this embodiment are sequentially denoted as (p 0 ,p 1 ,p 2 ,p 3 ) Of course, 3 or more are possible;
step 4: repeating the step 2-3 by rotating the base to obtain a point cloud and a mark point of a second view angle (back);
step 5: because the width of the base is known, and the shape of the front surface is the same as that of the back surface, the reconstruction of the base can be realized by enabling two point clouds on the front surface and the back surface to coincide through coordinate conversion and then moving the back surface along the z-axis direction by the unit with the same thickness as the base. Since the sample is fixed on the base, a three-dimensional reconstruction model of the sample can be realized by only making the same rotation and translation as the base. For this purpose use is made ofThe mark points detected in the step 3 and the step 4 respectively calculate transformation matrixes T of two visual angles 1 ,T 2 The method comprises the steps of carrying out a first treatment on the surface of the Rotating both the front and back sides of the mount to be parallel to the xoy plane by a transformation matrix, as shown in fig. 3, and translating the center of the mount to the origin of the camera coordinate system; specifically, taking the front side as an example:
by calculating the center point coordinates (x c ,y c ,z c ) Thereby obtaining the front S of the base 1 The relative distance between the center point and the origin point can obtain a translation vector t.
To accurately calculate S 1 Is to identify the centers of four mark points (p 0 ,p 1 ,p 2 ,p 3 (ii) as shown in figure 2. Deducing S 1 The coordinates of the center point of (c) are:
(x c ,y c ,z c )=(p 5 +p 6 )/2 (2)
wherein: p is p 5 For marking point p 0 ,p 1 At the center of two points, p is 5 =(p 0 +p 1 )/2;
p 6 For marking point p 2 ,p 3 At the center of two points, p is 6 =(p 2 +p 3 )/2;
Due to the front S of the base 1 Rotated to be parallel to the xoy plane, the rotation matrix R is shown in equation 3:
wherein r is 3 =[a,b,c]The method comprises the steps of carrying out a first treatment on the surface of the Vectors [ a, b, c ]]Is S 1 Normal vector of the plane.
r 1 =(p 5 -p 6 ).normalized();
r 2 =r 3 .cross(r 1 );
The transformation matrix for the front of the base is shown as:
wherein R represents a rotation matrix, and t represents a translation vector;
similarly, a transformation matrix T on the back of the base can be obtained 2 ;
Step 6: using a transformation matrix T 1 ,T 2 Respectively carrying out coordinate transformation on point clouds of two visual angles of the sample, and enabling the point clouds of a second visual angle of the sample to pass through a transformation matrix T 3 And rotating the sample by 180 degrees around the Y axis, translating the sample along the Z axis by a value which is the same as the thickness of the base, and finally splicing the transformed point cloud to finish the three-dimensional reconstruction of the sample. (as shown in fig. 4), specifically:
C=T 1 *C 1 +T 3 *T 2 *C 2 (5)
T 3 as shown in equation 6:
wherein l is the thickness of the base; t (T) 1 ,T 2 Representing the transformation matrix for the front and back of the chassis, respectively. C is the reconstructed point cloud, C 1 ,C 2 The point clouds of the front and back sides of the sample before reconstruction, respectively.
The foregoing description is only of the preferred embodiments of the present invention and is not intended to limit the scope of the invention, and all equivalent structures or equivalent processes using the descriptions and drawings of the present invention or directly or indirectly applied to other related technical fields are included in the scope of the invention.
Claims (2)
1. A method for quickly reconstructing three dimensions of a thin object is characterized in that the method comprises the following steps:
fixing a thin object to be reconstructed on a base with marking points, collecting a depth image and a color image through a depth camera, identifying the marking points to obtain a transformation matrix, carrying out coordinate transformation on point clouds of the object to be reconstructed according to the transformation matrix, and finally splicing the point clouds of two visual angles of the transformed object to finish three-dimensional reconstruction of a sample;
wherein the method comprises the following steps:
(1) Fixing a sample to be three-dimensionally reconstructed on a base with marking points, wherein the shape of the front surface and the back surface of the base are completely the same;
(2) Collecting a color image and a depth image of a first visual angle comprising a base and a sample by using a depth camera, and converting the color image and the depth image into point cloud by combining internal and external parameters of the camera;
(3) Detecting a mark point in the base from the color image, the mark point on the base being denoted (p 0 ,p 1 ,p 2 ,p 3 );
(4) Rotating the base to repeat the step (2) -the step (3) to obtain a point cloud and a marked point of the second view angle;
(5) Calculating transformation matrices T of two views by using the mark points detected in the step (3) and the step (4), respectively 1 ,T 2 The method comprises the steps of carrying out a first treatment on the surface of the Rotating both the front and back of the base to be parallel to the xoy plane through the transformation matrix, and translating the center of the base to the origin of the camera coordinate system; the process of solving the transformation matrix of the two view angles is as follows:
by calculating the center point coordinates (x c ,y c ,z c ) Thereby obtaining the front S of the base 1 The relative distance between the center point and the origin is obtained, namely a translation vector t is obtained;
front S of base 1 The coordinates of the center point of (c) are:
(x c ,y c ,z c )=(p 5 +p 6 )/2 (2)
wherein: p is p 5 For marking point p 0 ,p 1 At the center of two points, p is 5 =(p 0 +p 1 )/2;
p 6 For marking point p 2 ,p 3 At the center of two points, p is 6 =(p 2 +p 3 )/2;
Front S of the base 1 Rotated to be parallel to the xoy plane, the rotation matrix R is shown as formula (3):
wherein r is 3 =[a,b,c]The method comprises the steps of carrying out a first treatment on the surface of the Vectors [ a, b, c ]]Is S 1 Normal vector of the plane;
r 1 =(p 5 -p 6 ).normalized();
r 2 =r 3 .cross(r 1 );
the transformation matrix for the first viewing angle of the base is shown in the following formula:
similarly, a transformation matrix T of a second view angle of the base is obtained 2 ;
(6) Using a transformation matrix T 1 ,T 2 Respectively carrying out coordinate transformation on point clouds of two visual angles of the sample, and enabling the point clouds of a second visual angle of the sample to pass through a transformation matrix T 3 And rotating the sample by 180 degrees around the Y axis, translating the sample along the Z axis by a value which is the same as the thickness of the base, and finally splicing the transformed point cloud to finish the three-dimensional reconstruction of the sample.
2. The method for rapid three-dimensional reconstruction of a thin object according to claim 1, wherein the step (6) is specifically as follows:
the three-dimensional reconstruction model of the sample is specifically:
C=T 1 *C 1 +T 3 *T 2 *C 2 (5)
T 3 as shown in formula (6):
wherein l is the thickness of the base; c is the reconstructed point cloud, C 1 ,C 2 The point clouds of the front and back sides of the sample before reconstruction, respectively.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201811147802.XA CN109360267B (en) | 2018-09-29 | 2018-09-29 | Rapid three-dimensional reconstruction method for thin object |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201811147802.XA CN109360267B (en) | 2018-09-29 | 2018-09-29 | Rapid three-dimensional reconstruction method for thin object |
Publications (2)
Publication Number | Publication Date |
---|---|
CN109360267A CN109360267A (en) | 2019-02-19 |
CN109360267B true CN109360267B (en) | 2023-06-06 |
Family
ID=65348194
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201811147802.XA Active CN109360267B (en) | 2018-09-29 | 2018-09-29 | Rapid three-dimensional reconstruction method for thin object |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN109360267B (en) |
Families Citing this family (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111768490B (en) * | 2020-05-14 | 2023-06-27 | 华南农业大学 | Plant three-dimensional modeling method and system based on iteration closest point and manual intervention |
CN112069923B (en) * | 2020-08-18 | 2024-07-12 | 广东正扬传感科技股份有限公司 | 3D face point cloud reconstruction method and system |
CN112132971B (en) * | 2020-09-08 | 2024-04-23 | 合肥的卢深视科技有限公司 | Three-dimensional human modeling method, three-dimensional human modeling device, electronic equipment and storage medium |
Family Cites Families (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103198523B (en) * | 2013-04-26 | 2016-09-21 | 清华大学 | A kind of three-dimensional non-rigid body reconstruction method based on many depth maps and system |
CN104330074B (en) * | 2014-11-03 | 2017-01-18 | 广州欧科信息技术股份有限公司 | Intelligent surveying and mapping platform and realizing method thereof |
CN105627948B (en) * | 2016-01-31 | 2018-02-06 | 山东科技大学 | A kind of method that large complicated carved measuring system carries out complex-curved sampling |
CN105989604A (en) * | 2016-02-18 | 2016-10-05 | 合肥工业大学 | Target object three-dimensional color point cloud generation method based on KINECT |
CN106556356A (en) * | 2016-12-07 | 2017-04-05 | 西安知象光电科技有限公司 | A kind of multi-angle measuring three-dimensional profile system and measuring method |
CN107240129A (en) * | 2017-05-10 | 2017-10-10 | 同济大学 | Object and indoor small scene based on RGB D camera datas recover and modeling method |
CN107631700B (en) * | 2017-09-07 | 2019-06-21 | 西安电子科技大学 | The three-dimensional vision information method that spatial digitizer is combined with total station |
-
2018
- 2018-09-29 CN CN201811147802.XA patent/CN109360267B/en active Active
Also Published As
Publication number | Publication date |
---|---|
CN109360267A (en) | 2019-02-19 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN108759665B (en) | Spatial target three-dimensional reconstruction precision analysis method based on coordinate transformation | |
CN103530880B (en) | Based on the camera marking method of projection Gaussian network pattern | |
CN109919984A (en) | A kind of point cloud autoegistration method based on local feature description's | |
CN111121655A (en) | Visual detection method for pose and aperture of coplanar workpiece with equal large hole patterns | |
CN109360267B (en) | Rapid three-dimensional reconstruction method for thin object | |
CN113744351B (en) | Underwater structure light measurement calibration method and system based on multi-medium refraction imaging | |
CN105716542A (en) | Method for three-dimensional data registration based on flexible feature points | |
CN111784778A (en) | Binocular camera external parameter calibration method and system based on linear solving and nonlinear optimization | |
CN110866969A (en) | Engine blade reconstruction method based on neural network and point cloud registration | |
CN103558850A (en) | Laser vision guided welding robot full-automatic movement self-calibration method | |
CN109035327B (en) | Panoramic camera attitude estimation method based on deep learning | |
WO2018201677A1 (en) | Bundle adjustment-based calibration method and device for telecentric lens-containing three-dimensional imaging system | |
CN109583377B (en) | Control method and device for pipeline model reconstruction and upper computer | |
CN110009667A (en) | Multi-viewpoint cloud global registration method based on Douglas Rodríguez transformation | |
Li et al. | Research on the calibration technology of an underwater camera based on equivalent focal length | |
CN114332348B (en) | Track three-dimensional reconstruction method integrating laser radar and image data | |
CN104697463B (en) | The blanking feature constraint scaling method and device of a kind of binocular vision sensor | |
CN106097433A (en) | Object industry and the stacking method of Image model and system | |
CN108180834A (en) | A kind of industrial robot is the same as three-dimensional imaging instrument position orientation relation scene real-time calibration method | |
CN113870366B (en) | Calibration method and calibration system of three-dimensional scanning system based on pose sensor | |
CN103900504A (en) | Nano-scale real-time three-dimensional visual information feedback method | |
CN109342008B (en) | Wind tunnel test model attack angle single-camera video measuring method based on homography matrix | |
Langming et al. | A flexible method for multi-view point clouds alignment of small-size object | |
CN113628170A (en) | Laser line extraction method and system based on deep learning | |
CN107773246A (en) | A kind of defect of auricle measuring method based on optical three-dimensional measurement |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant | ||
EE01 | Entry into force of recordation of patent licensing contract |
Application publication date: 20190219 Assignee: Zhejiang Meier Micro Vision Technology Co.,Ltd. Assignor: HANGZHOU LANXIN TECHNOLOGY CO.,LTD. Contract record no.: X2024980010898 Denomination of invention: A fast 3D reconstruction method for thin objects Granted publication date: 20230606 License type: Common License Record date: 20240730 |
|
EE01 | Entry into force of recordation of patent licensing contract |