CN113096187A - Method for automatically acquiring relative position of vehicle and obstacle - Google Patents
Method for automatically acquiring relative position of vehicle and obstacle Download PDFInfo
- Publication number
- CN113096187A CN113096187A CN202110487005.1A CN202110487005A CN113096187A CN 113096187 A CN113096187 A CN 113096187A CN 202110487005 A CN202110487005 A CN 202110487005A CN 113096187 A CN113096187 A CN 113096187A
- Authority
- CN
- China
- Prior art keywords
- vehicle
- coordinate system
- calibration
- sphere
- coordinates
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 43
- 238000013519 translation Methods 0.000 claims abstract description 5
- 238000012545 processing Methods 0.000 claims description 20
- 238000001914 filtration Methods 0.000 claims description 19
- 239000011159 matrix material Substances 0.000 claims description 18
- 238000009499 grossing Methods 0.000 claims description 17
- 238000004364 calculation method Methods 0.000 claims description 6
- 238000006243 chemical reaction Methods 0.000 claims description 6
- 230000003068 static effect Effects 0.000 claims description 5
- 238000003384 imaging method Methods 0.000 claims description 3
- 239000000203 mixture Substances 0.000 claims description 3
- 238000011410 subtraction method Methods 0.000 claims 1
- 230000004888 barrier function Effects 0.000 abstract 1
- 238000010586 diagram Methods 0.000 description 3
- 230000008859 change Effects 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 230000008569 process Effects 0.000 description 2
- 238000010923 batch production Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000010276 construction Methods 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 238000009826 distribution Methods 0.000 description 1
- 238000005286 illumination Methods 0.000 description 1
- 239000003550 marker Substances 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000006855 networking Effects 0.000 description 1
- 230000009467 reduction Effects 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/70—Determining position or orientation of objects or cameras
- G06T7/73—Determining position or orientation of objects or cameras using feature-based methods
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S7/00—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
- G01S7/48—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S17/00
- G01S7/497—Means for monitoring or calibrating
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/13—Edge detection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/174—Segmentation; Edge detection involving the use of two or more images
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/194—Segmentation; Edge detection involving foreground-background segmentation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/80—Analysis of captured images to determine intrinsic or extrinsic camera parameters, i.e. camera calibration
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20112—Image segmentation details
- G06T2207/20164—Salient point detection; Corner detection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20212—Image combination
- G06T2207/20224—Image subtraction
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30248—Vehicle exterior or interior
- G06T2207/30252—Vehicle exterior; Vicinity of vehicle
- G06T2207/30261—Obstacle
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Theoretical Computer Science (AREA)
- Computer Networks & Wireless Communication (AREA)
- Radar, Positioning & Navigation (AREA)
- Remote Sensing (AREA)
- Traffic Control Systems (AREA)
Abstract
The invention provides a method for automatically acquiring the relative position of a vehicle and an obstacle, which is characterized in that a camera is arranged above the middle position of a vehicle-mounted laser radar calibration field, spherical characteristic calibration objects are respectively arranged at two opposite corners of the calibration field, the world coordinate of the center of a sphere of the spherical characteristic calibration objects in the calibration field is detected, the outline of a calibrated vehicle to be acquired is detected, the external rectangle of the outline of the vehicle is calculated, the world coordinate of the center of the external rectangle is acquired, the center of the external rectangle of the outline of the vehicle is taken as the origin of a vehicle coordinate system, the X axis and the Y axis of the vehicle coordinate system are respectively parallel to two sides of the external rectangle of the outline of the vehicle, and the coordinates of the center of the spherical characteristic calibration objects in the vehicle coordinate system are finally calculated through the rotational translation rotating system of the vehicle coordinate system and the world coordinate system. The invention can automatically acquire the vehicle body pose and the relative position relation between the vehicle body pose and the barrier, reduces the requirement on the vehicle placing pose when calibrating the laser radar, and improves the laser radar calibration efficiency and precision.
Description
Technical Field
The invention relates to the technical field of automatic driving and intelligent networking automobiles, in particular to the field of sensor calibration, and specifically relates to a method for automatically acquiring the relative position of a vehicle and an obstacle.
Background
The calibration of the vehicle-mounted laser radar of the intelligent networked automobile or the automatic driving automobile is usually completed in a specific scene, and the calibration of the vehicle-mounted sensor is a preorder step of development and test of the intelligent networked automobile or the automatic driving automobile. Because the calibration scene and the characteristic marker are fixed in position, the parking according to the specified pose is required to be ensured when the laser radar is calibrated, and the effectiveness of the relative position relation between the prestored specific calibration object and the vehicle body can be ensured; the prior method is that when a vehicle is placed, manual operation is usually carried out, a standardized and automatic mode for acquiring the relative position relation between a vehicle body and a characteristic calibration object is lacked, the vehicle pose is adjusted manually, the efficiency is low, the difficulty is high, and the precision of a calibration result has certain uncertainty, even the calibration fails.
Disclosure of Invention
The invention aims to provide an automatic acquisition method for relative positions of a vehicle and an obstacle, which aims to solve the technical problems that in the process of calibrating a vehicle-mounted laser radar of an intelligent networked vehicle or an automatic driving vehicle in a specific scene, a standardized and automatic acquisition mode for relative position relation between a vehicle body and a characteristic calibration object is lacked, the vehicle body placement difficulty is high, the calibration efficiency and the calibration precision of the vehicle-mounted laser radar are low, the calibration failure is easily caused, and the like.
In order to solve the technical problem, the invention provides an automatic acquisition method for the relative position of a vehicle and an obstacle, which is characterized in that a camera for acquiring the state in a calibration field is arranged above the middle position of the calibration field of a vehicle-mounted laser radar, respectively arranging sphere feature calibration objects at two opposite corners of a calibration field, then detecting world coordinates of the sphere center of the sphere feature calibration objects in the calibration field, detecting the outline of a calibration vehicle to be obtained, then, the external rectangle of the vehicle outline is calculated, the world coordinate of the center of the external rectangle is obtained, the center of the external rectangle of the vehicle outline is taken as the origin of the vehicle coordinate system, the X axis and the Y axis of the vehicle coordinate system are set to be parallel to two sides of the external rectangle of the vehicle outline, and finally the coordinate of the sphere center of the sphere characteristic calibration object in the vehicle coordinate system is calculated through the rotation translation relation between the vehicle coordinate system and the world coordinate system.
Preferably, the method for automatically acquiring the relative position of the vehicle and the obstacle specifically comprises the following steps:
(1) respectively arranging a group of sphere feature calibration objects at two opposite angles of a vehicle-mounted laser radar calibration field, wherein each group of sphere feature calibration objects are formed by three spheres with high and matte surfaces which are distributed in a regular triangle;
(2) placing a camera above the middle of a vehicle-mounted laser radar calibration field, wherein the height of the camera needs to ensure that the camera has a complete and clear view field of the calibration field, keeping the imaging plane of the camera horizontal to the plane of the calibration field, and then parking a vehicle in the middle area of the calibration field and ensuring that the heading angle of a vehicle head is between 0 and 180 degrees;
(3) calibrating an internal reference matrix and an external reference matrix of a camera positioned above the middle of a vehicle-mounted laser radar calibration site, and storing the internal and external reference matrices of the camera in a calibration operation computer;
(4) loading and storing an image when a calibrated vehicle does not enter a calibration field, carrying out gray processing on the image, carrying out Gaussian filtering and smoothing processing on the obtained gray image, then detecting a circle in the image after the Gaussian filtering and smoothing processing, extracting a circle center coordinate, namely a coordinate of a sphere center under a two-dimensional pixel coordinate system, converting the coordinate of the sphere center under the two-dimensional pixel coordinate system into a world coordinate system, obtaining a coordinate (two-dimensional coordinate) of the sphere center under the world coordinate system, and storing the coordinate (two-dimensional coordinate) in a calibration computer;
(5) loading an image when a vehicle to be calibrated does not enter a calibration field to obtain a world coordinate (two-dimensional coordinate) of the sphere center of the sphere, performing background modeling on the image subjected to Gaussian filtering and smoothing processing in the step (4) through a Gaussian mixture model, and storing a modeling result, namely a background image;
(6) on the premise of ensuring that the heading angle of the vehicle head is within 0-180 degrees, the vehicle is parked in a calibration field at any pose, and a current foreground frame image after the vehicle to be calibrated enters the calibration field is obtained and loaded;
(7) acquiring and loading a current scene frame image P of a vehicle parked in a calibration fieldvehicleThen, the current foreground frame image P is used as the foreground frame imagevehicleAfter Gaussian filtering processing, the background frame image P is stored in a calibration computer in advanceemptyAs an input, processing the image by using a background difference method, calculating a pixel area deviating from the background by more than a certain threshold value as a region where the vehicle to be calibrated is located in the image, and acquiring a foreground pixel image only containing the vehicle to be calibrated;
(8) performing binarization, Gaussian filtering and smoothing on the foreground pixel map obtained in the step (7), then detecting the outer contour information of the calibrated vehicle to be obtained, extracting two diagonal points at the outermost periphery in the outer contour of the vehicle and obtaining the coordinates of the two diagonal points in a pixel coordinate system, and constructing an external rectangle of the outer contour of the calibrated vehicle by using the two diagonal points;
(9) setting the origin of the vehicle body coordinate system on a plane which is higher than the ground and can ensure that the camera has a clear view covering the whole calibration field by taking the geometric central point of the circumscribed rectangle constructed in the step (8) as the origin of the vehicle body coordinate system, respectively drawing perpendicular lines from the origin of the vehicle body coordinate system to two sides of the circumscribed rectangle, and acquiring coordinates of intersection points of the perpendicular lines and the sides of the circumscribed rectangle in a pixel coordinate system;
(10) respectively calculating coordinates of the origin of the vehicle body coordinate system, the intersection point of the perpendicular line and the side of the circumscribed rectangle in the step (9) in the world coordinate system through coordinate conversion formulas under the pixel coordinate system and the world coordinate system, and then calculating the current heading angle of the vehicle body (namely the rotation angle of the vehicle body coordinate system relative to the world coordinate system);
(11) and finally, calculating the coordinates of the sphere center of the sphere feature calibration object under the vehicle body coordinate system.
Preferably, in the step (1), the radius of each sphere is 0.6m, the distance between the centers of adjacent spheres in each set of sphere feature calibration objects is 2.2m, and the center of each set of sphere feature calibration objects is placed at the distance of 8-10m from the origin of the vehicle coordinate system.
Preferably, the step (3) is to calibrate an internal reference matrix and an external reference matrix of the camera which are arranged above the middle position of the calibration field by using a Zhangnyou chessboard pattern calibration method; and the coordinate conversion formulas under the pixel coordinate system and the world coordinate system are as follows:
u, v are the coordinates of a point in the pixel coordinate system, Xw,Yw,ZwThe coordinates of the point in the world coordinate system are K, an internal reference matrix of the camera and R, T, an external reference matrix of the camera.
Preferably, the step (4) is to load and save an image of the calibration vehicle when the calibration vehicle does not enter the calibration field by using a cvLoadImage () function in OpenCV;
the step (4) is to carry out graying processing on the image through a cvtColor () function;
detecting a circle in the image subjected to Gaussian filtering and smoothing by using a Hough gradient method;
the step (4) is to output the coordinates of the centers of all circles through an ellipse fitting function fitEllipse () in OpenCV, and the coordinates are expressed as a matrix Ocam:
(u1,v1)..(u6,v6) Respectively representing the coordinates of the centers of six spheres in the two groups of sphere feature calibration objects under a pixel coordinate system, and the coordinate of the center of the sphere under a world coordinate system is represented as Oword:
(xw1,yw1)..(xw6,yw6) The horizontal and vertical coordinates of the sphere centers of six spheres in the sphere feature calibration object under a two-dimensional world coordinate system are shown, and r is the radius of the sphere.
Preferably, in the step (5), a gaussian background model building function background modeling () in OpenCV is used to perform background modeling on the image after the gaussian filtering and smoothing processing in the step (4).
Preferably, the step (8) is to detect the outer contour information of the calibration vehicle to be acquired by using a Canny algorithm.
Preferably, the calculation formula for calculating the current heading angle θ of the vehicle body in the step (10) is as follows:
at this time, the coordinate calculation formula of the sphere center of each sphere under the vehicle body coordinate system is as follows:
wherein, (x, y) is the coordinate of the sphere center in the two-dimensional world coordinate system, then the coordinate of the sphere center in the vehicle body coordinate system is (x ', y', r-k), r is the sphere radius, k is the distance from the XY plane of the vehicle body coordinate system to the ground, and r-k is the coordinate of the sphere center in the Z axis of the vehicle body coordinate system.
Preferably, in the step (11), coordinates of the center of the sphere of the spherical feature calibration object in the vehicle body coordinate system are calculated through the formula (2) and the formula (3), the ordinate of the centers of all the spherical feature calibration objects in the vehicle body coordinate system is k-r, and a coordinate set of the centers of the spherical feature calibration objects in the vehicle body coordinate system is represented as Ovehicle:
(xV1,yV1,zV1)..(xV6,yV6,zV6) And calibrating coordinates of the centers of six spheres in the two groups of sphere characteristic calibration objects under the vehicle body coordinate system.
Preferably, the calibration scene of the vehicle-mounted laser radar is a static scene.
By adopting the technical scheme, the invention has the following beneficial effects:
the method for automatically acquiring the relative position of the vehicle and the obstacle has reasonable conception and simple operation flow, and is a method for quickly determining the coordinate of a vehicle body coordinate system and the coordinate of a characteristic calibration object in the vehicle body coordinate system when an intelligent networked vehicle or an automatic driving vehicle-mounted laser radar is calibrated; the invention utilizes the image processing technology to acquire the image information in the calibration field in real time to determine the relative position relationship between the vehicle and the characteristic calibration object, can randomly place the vehicle in the specified moving range, greatly reduces the requirement on the placement position during the calibration of the vehicle-mounted laser radar, and lays a technical foundation for the standardization, the flow and the automation of the laser radar calibration step.
The invention determines the relative position between the vehicle and the characteristic calibration object based on the camera, allows the vehicle to be parked randomly in a specified range, does not need to strictly ensure the unique vehicle body pose, greatly improves the efficiency and the accuracy of vehicle-mounted laser radar calibration, and can provide a feasible technical scheme and engineering landing conditions for the fast and full-automatic calibration of the vehicle-mounted laser radar of the automatic driving vehicle in batch production.
The invention can effectively reduce the requirement of vehicle position and pose placement when the vehicle-mounted laser radar is calibrated, improve the calibration efficiency and precision of the laser radar, and solve the technical problems that the traditional laser radar calibration method based on the characteristic markers needs to be placed according to the specified vehicle body position and pose, has higher difficulty in accurate placement, higher error and the like.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings used in the embodiments or the description in the prior art will be briefly described below, and it is obvious that the drawings in the following description are some embodiments of the present invention, and other drawings can be obtained by those skilled in the art without creative efforts.
FIG. 1 is a diagram showing a state of a camera in the method for automatically acquiring the relative position between a vehicle and an obstacle according to the present invention;
FIG. 2 is a top view of a calibration site for the method for automatically acquiring the relative position of a vehicle and an obstacle according to the present invention;
FIG. 3 is a schematic view of a process for determining world coordinates of the center of sphere of a sphere feature calibration object in the method for automatically acquiring relative positions of a vehicle and an obstacle according to the present invention;
FIG. 4 is a schematic diagram of a background difference processing procedure in the method for automatically acquiring relative positions of a vehicle and an obstacle according to the present invention;
FIG. 5 is a schematic diagram of a vehicle outer contour detection output flow in the method for automatically acquiring the relative position between a vehicle and an obstacle according to the present invention.
Detailed Description
The technical solutions of the present invention will be described clearly and completely with reference to the accompanying drawings, and it should be understood that the described embodiments are some, but not all embodiments of the present invention. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
The present invention will be further explained with reference to specific embodiments.
The invention relates to a method for automatically acquiring the relative position of a vehicle and an obstacle, which is implemented by arranging a camera for acquiring the state in a calibration field above the middle of the calibration field of a vehicle-mounted laser radar, respectively arranging spherical characteristic calibration objects at two opposite corners of the calibration field, then detecting the world coordinate of the center of the sphere of the spherical characteristic calibration object in the calibration field, detecting the outline of the calibrated vehicle to be acquired, then calculating the external rectangle of the outline of the vehicle and acquiring the world coordinate of the center of the external rectangle, then setting the X axis and the Y axis of a vehicle body coordinate system to be respectively parallel to two sides of the external rectangle of the outline of the vehicle by taking the center of the external rectangle of the outline of the vehicle as the origin of the vehicle body coordinate system, finally calculating the coordinate of the center of the spherical characteristic calibration object in the vehicle body coordinate system through the rotational translation relationship between the vehicle body coordinate system and the world coordinate system, and obtaining the relative position relation between the spherical characteristic calibration object and the vehicle body.
The invention relates to a method for automatically acquiring the relative position of a vehicle and an obstacle, which specifically comprises the following steps:
(1) a group of sphere feature calibration objects are respectively arranged at two opposite corners of a vehicle-mounted laser radar calibration field (as shown in fig. 1, in the embodiment, the upper left corner and the lower right corner of a vehicle-mounted laser radar calibration scene of an automatic driving vehicle), each group of sphere feature calibration objects is formed by three spheres with high and matte surfaces and distributed in a regular triangle, and the specific distribution form is shown in fig. 1; wherein the radius of each sphere is 0.6m, the distance between the centers of adjacent spheres in each group of sphere feature calibration objects is 2.2m, and the center of each group of sphere feature calibration objects is arranged at the distance of 8-10m from the origin of the vehicle coordinate system.
(2) A high-definition camera is arranged above the middle position of a vehicle-mounted laser radar calibration field, the imaging plane of the camera and the calibration field plane are kept horizontal, the height of the camera is h meters (h is a numerical value in all feasible ranges), namely the height of the camera needs to ensure that the camera has a complete and clear view field of the calibration field, as shown in fig. 1; and then the vehicle is parked in the middle area of the calibration site shown in figure 2, and the heading angle of the vehicle head is ensured to be between 0 and 180 degrees.
(3) Calibrating an internal reference matrix K and external reference matrices R and T (namely rotation and translation matrices from a camera coordinate system to a world coordinate system) of the camera positioned above the middle position of a calibration field by using a Zhangyingyou chessboard lattice calibration method, and storing the internal and external reference matrices of the camera in a calibration operation computer; the coordinate conversion formulas under the pixel coordinate system and the world coordinate system are as follows:
wherein u and v are the coordinates of a certain point in the pixel coordinate system, Xw,Yw,ZwThe coordinates of the point in the world coordinate system are K, an internal reference matrix of the camera and R, T, an external reference matrix of the camera.
(4) Loading and storing an image of a calibration vehicle when the calibration vehicle does not enter a calibration field by using a function cvLoadimage () in OpenCV, carrying out gray processing on the image by using a cvtColor () function, carrying out Gaussian filtering and smoothing processing (by using a GaussionBlur () function) on the obtained gray image, detecting circles in the image after the Gaussian filtering and smoothing processing by using a Hough gradient method (cvHougcircle () function), and outputting the coordinates of the centers of all circles by using an ellipse fitting function fitEllipse () in OpenCV, wherein the coordinates are represented as a matrix Ocam:
(u1,v1)..(u6,v6) Respectively representing two sets of sphere feature calibrationCoordinates of the centers of six spheres in the object under a two-dimensional pixel coordinate system, and coordinates (u) of the centers of the spheres under the pixel coordinate system1,v1)..(u6,v6) Converting into world coordinate system by coordinate conversion formula shown in formula (1) and storing in calibration computer, and extracting only the coordinates of the sphere center in two-dimensional world coordinate system (not considering Z axis temporarily), and its coordinates in world coordinate system is represented as Oword:
The flow is shown in FIG. 3; wherein (x)w1,yw1)..(xw6,yw6) The method is characterized in that the horizontal and vertical coordinates of a calibration object sphere in a two-dimensional world coordinate system are characterized, r is the radius of the sphere, the operation is not needed to be carried out in the subsequent operation after the debugging stage is completed, and the data stored in a calibration computer can be directly used in the subsequent operation.
(5) Loading an image when a vehicle to be calibrated does not enter a calibration field to obtain the world coordinates of the sphere center of the sphere, taking the image subjected to Gaussian filtering and smoothing in the step (4) as an input for constructing a background model, performing background modeling on the image subjected to Gaussian filtering and smoothing in the step (4) by using a mixed Gaussian background model construction function background modeling MOG () in OpenCV, and storing a modeling result, namely a background frame image Pempty(ii) a Because the calibration scene is a fixed scene, the background frame image P constructed based on the Gaussian mixture modelemptyNo real-time updating is required.
(6) On the premise of ensuring that the heading angle of the vehicle head is within 0-180 degrees, the vehicle is parked in the calibration field at any pose (position + posture), and a current foreground frame image P after the vehicle to be calibrated enters the calibration field is acquired and loadedvehicleAnd load the foreground frame image PvehicleThen, before the calibration is completed, the pose of the vehicle can not change any more; in which the vehicle exits the parking area from above as it enters the parking area from below, as shown in fig. 2.
(7) Acquiring and loading a current scene frame image P of a vehicle parked in a calibration fieldvehicleThen, the current foreground frame image P is used as the foreground frame imagevehicleBackground frame image P after noise reduction and pre-stored in calibration computeremptyAs input, the image is processed by using a background difference method, namely, the current foreground frame image PvehicleAnd the background frame image PemptyAnd (3) performing subtraction operation, calculating a pixel area which deviates from the background by more than a certain threshold value T (T is a numerical value within all reasonable ranges) as an area where the vehicle to be calibrated is located in the image, and directly outputting information such as the position, size and shape of the vehicle to be calibrated in the image, namely obtaining a foreground pixel map only containing the vehicle to be calibrated, wherein the flow is shown in fig. 4. The calibration field scene is a static scene, so that the calibration scene with static characteristics and extremely small change degree of illumination conditions can be provided for the background difference method, the camera can be ensured not to shake in a static state, the technical defects of the background difference method can be avoided, and the output precision of a target area is improved.
(8) And (3) performing binarization, gaussian filtering and smoothing on the foreground pixel map obtained in the step (7), detecting the outer contour information of the vehicle to be calibrated by using a Canny algorithm, extracting two diagonal points which are positioned at the outermost periphery in the outer contour of the vehicle, obtaining coordinates of the two diagonal points in a pixel coordinate system, and constructing a circumscribed rectangle of the outer contour of the vehicle to be calibrated by using the two diagonal points, wherein the flow is shown in fig. 5.
(9) Taking the geometric central point C of the circumscribed rectangle constructed in the step (8) as the original point of the vehicle body coordinate system, setting the original point of the vehicle body coordinate system on a plane with the height being k meters away from the ground (k meters is a value in a reasonable range, namely, the camera is ensured to have a clear view covering the whole calibration field), and setting the coordinate of the point C in the image coordinate system as (u mcen,vcen) Drawing perpendicular lines from the point C to the two sides of the circumscribed rectangle, respectively, and obtaining coordinates D (u) in the image coordinate system of the intersection point of the perpendicular line and the sides of the circumscribed rectangletrs,vtrs)、F(ulon,vlonB) of the group A and B); wherein, the point D is the intersection point of a perpendicular line made towards the direction of the vehicle head and the external rectangleThe point F is an intersection of a perpendicular line made to the right vehicle body direction and the circumscribed rectangle, the direction from the point C to the point F is the direction of the X axis of the vehicle body coordinate system, and the direction from the point C to the point D is the direction of the Y axis of the vehicle body coordinate system, as shown in fig. 2.
(10) Respectively calculating the coordinates (x) of the point C, the point D and the point F in the world coordinate system in the step (9) through a formula (1)C,yC)、(xD,yD)、(xF,yF) (ii) a Then calculating the course angle theta of the vehicle body at the moment, wherein the calculation formula is as follows:
at this time, the coordinate calculation formula of the sphere center of each sphere under the vehicle body coordinate system is as follows:
wherein, (x, y) is the coordinate of the sphere center in the two-dimensional world coordinate system, then the coordinate of the sphere center in the vehicle body coordinate system is (x ', y', r-k), r is the sphere radius, k is the distance from the XY plane of the vehicle body coordinate system to the ground, and r-k is the coordinate of the sphere center in the Z axis of the vehicle body coordinate system.
(11) Calculating the coordinates of the center of the sphere characteristic calibration object under the vehicle body coordinate system according to the formula (2) and the formula (3); the coordinate set of the sphere center of the sphere feature calibration object under the vehicle body coordinate system is represented as Ovehicle:
(xV1,yV1,zV1)...(xV6,yV6,zV6) Coordinates of the centers of six spheres in the two groups of sphere feature calibration objects under the vehicle body coordinate system are calibrated, the origin C of the vehicle body coordinate system is on a plane with the height of k meters away from the ground, and the values of the centers of all the spheres on the Z coordinate axis of the vehicle body coordinate system are k-r。
The method has reasonable concept and simple operation flow, utilizes the image processing technology to acquire the image information in the calibration field in real time to determine the relative position relationship between the vehicle body and the characteristic calibration object, can reduce the difficulty of placing the vehicle, improves the calibration efficiency and precision of the laser radar, and can place the vehicle at will in the specified moving range so as to achieve the calibration purpose.
Finally, it should be noted that: the above embodiments are only used to illustrate the technical solution of the present invention, and not to limit the same; while the invention has been described in detail and with reference to the foregoing embodiments, it will be understood by those skilled in the art that: the technical solutions described in the foregoing embodiments may still be modified, or some or all of the technical features may be equivalently replaced; and the modifications or the substitutions do not make the essence of the corresponding technical solutions depart from the scope of the technical solutions of the embodiments of the present invention.
Claims (10)
1. An automatic acquisition method for the relative position of a vehicle and an obstacle is characterized in that a camera for acquiring the state in a calibration field is arranged above the middle position of the calibration field of a vehicle-mounted laser radar, respectively arranging sphere feature calibration objects at two opposite corners of a calibration field, then detecting world coordinates of the sphere center of the sphere feature calibration objects in the calibration field, detecting the outline of a calibration vehicle to be obtained, then, the external rectangle of the vehicle outline is calculated, the world coordinate of the center of the external rectangle is obtained, the center of the external rectangle of the vehicle outline is taken as the origin of the vehicle coordinate system, the X axis and the Y axis of the vehicle coordinate system are set to be parallel to two sides of the external rectangle of the vehicle outline, and finally the coordinate of the sphere center of the sphere characteristic calibration object in the vehicle coordinate system is calculated through the rotation translation rotating system of the vehicle coordinate system and the world coordinate system.
2. The method for automatically acquiring the relative position of the vehicle and the obstacle according to claim 1, specifically comprising the steps of:
(1) respectively arranging a group of sphere feature calibration objects at two opposite angles of a vehicle-mounted laser radar calibration field, wherein each group of sphere feature calibration objects are formed by three spheres with high and matte surfaces which are distributed in a regular triangle;
(2) placing a camera above the middle of a vehicle-mounted laser radar calibration field, wherein the height of the camera needs to ensure that the camera has a complete and clear view field of the calibration field, keeping the imaging plane of the camera horizontal to the plane of the calibration field, and then parking a vehicle in the middle area of the calibration field and ensuring that the heading angle of a vehicle head is between 0 and 180 degrees;
(3) calibrating an internal reference matrix and an external reference matrix of a camera positioned above the middle of a vehicle-mounted laser radar calibration site, and storing the internal and external reference matrices of the camera in a calibration operation computer;
(4) loading and storing an image when a calibrated vehicle does not enter a calibration field, carrying out gray processing on the image, carrying out Gaussian filtering and smoothing processing on the obtained gray image, then detecting a circle in the image after the Gaussian filtering and smoothing processing, extracting a circle center coordinate, namely a coordinate of a sphere center under a two-dimensional pixel coordinate system, converting the coordinate of the sphere center under the two-dimensional pixel coordinate system into a world coordinate system, obtaining the coordinate of the sphere center under the world coordinate system, and storing the coordinate in a calibration computer;
(5) loading an image when a vehicle to be calibrated does not enter a calibration field to obtain the world coordinates of the sphere center of the sphere, performing background modeling on the image subjected to Gaussian filtering and smoothing in the step (4) through a Gaussian mixture model, and storing a modeling result, namely a background image;
(6) on the premise of ensuring that the heading angle of the vehicle head is within 0-180 degrees, the vehicle is parked in a calibration field at any pose, and a current foreground frame image after the vehicle to be calibrated enters the calibration field is obtained and loaded;
(7) acquiring and loading a current scene frame image P of a vehicle parked in a calibration fieldvehicleThen, the current foreground frame image P is used as the foreground frame imagevehicleAfter Gaussian filtering processing, the background frame image P is stored in a calibration computer in advanceemptyAs input, the image is processed by using a background subtraction method, and the deviation from the background is calculatedTaking the pixel area exceeding a certain threshold value as the area where the vehicle to be calibrated is located in the image, and acquiring a foreground pixel map only containing the vehicle to be calibrated;
(8) performing binarization, Gaussian filtering and smoothing on the foreground pixel map obtained in the step (7), then detecting the outer contour information of the calibrated vehicle to be obtained, extracting two diagonal points at the outermost periphery in the outer contour of the vehicle and obtaining the coordinates of the two diagonal points in a pixel coordinate system, and constructing an external rectangle of the outer contour of the calibrated vehicle by using the two diagonal points;
(9) setting the origin of the vehicle body coordinate system on a plane which is higher than the ground and can ensure that the camera has a clear view covering the whole calibration field by taking the geometric central point of the circumscribed rectangle constructed in the step (8) as the origin of the vehicle body coordinate system, respectively drawing perpendicular lines from the origin of the vehicle body coordinate system to two sides of the circumscribed rectangle, and acquiring coordinates of the intersection points of the perpendicular lines and the sides of the circumscribed rectangle in an image coordinate system;
(10) respectively calculating coordinates of the origin of the vehicle body coordinate system, the intersection point of the perpendicular line and the side of the circumscribed rectangle in the step (9) in the world coordinate system through coordinate conversion formulas under the pixel coordinate system and the world coordinate system, and then calculating the course angle of the vehicle body at the moment;
(11) and finally, calculating the coordinates of the sphere center of the sphere feature calibration object under the vehicle body coordinate system.
3. The method for automatically acquiring the relative position of the vehicle and the obstacle according to claim 2, characterized in that: the radius of each sphere in the step (1) is 0.6m, the distance between the centers of adjacent spheres in each group of sphere feature calibration objects is 2.2m, and the center of each group of sphere feature calibration objects is placed at the distance of 8-10m from the origin of the vehicle coordinate system.
4. The method for automatically acquiring the relative position of the vehicle and the obstacle according to claim 2, characterized in that: the step (3) is to calibrate an internal reference matrix and an external reference matrix of the camera which are arranged above the middle position of the calibration field by using a Zhang-Zhengyou calibration method; and the coordinate conversion formulas under the pixel coordinate system and the world coordinate system are as follows:
in the above formula (1), u and v are coordinates of a certain point in a pixel coordinate system, and Xw,Yw,ZwThe coordinates of the point in the world coordinate system are K, an internal reference matrix of the camera and R, T, an external reference matrix of the camera.
5. The method for automatically acquiring the relative position of the vehicle and the obstacle according to claim 2, characterized in that: the step (4) is to load and store an image when a calibration vehicle does not enter a calibration field by using a function cvLoadimage () in OpenCV;
the step (4) is to carry out graying processing on the image through a cvtColor () function;
detecting a circle in the image subjected to Gaussian filtering and smoothing by using a Hough gradient method;
the step (4) is to output the coordinates of the centers of all circles through an ellipse fitting function fitEllipse () in OpenCV, and the coordinates are expressed as a matrix Ocam:
(u1,v1)..(u6,v6) Respectively representing the coordinates of the centers of six spheres in the two groups of sphere feature calibration objects under a pixel coordinate system, and the coordinate of the center of the sphere under a world coordinate system is represented as Oword:
(xw1,yw1)..(xw6,yw6) For calibrating the horizontal and vertical coordinates of the centers of six spheres in the sphere feature under a two-dimensional world coordinate systemAnd r is the radius of the sphere.
6. The method for automatically acquiring the relative position of the vehicle and the obstacle according to claim 2, characterized in that: and (5) performing background modeling on the image subjected to Gaussian filtering and smoothing processing in the step (4) by using a mixed Gaussian background model building function background modeling MOG () in OpenCV.
7. The method for automatically acquiring the relative position of the vehicle and the obstacle according to claim 2, characterized in that: and (8) detecting the outer contour information of the calibration vehicle to be acquired by using a Canny algorithm.
8. The method for automatically acquiring the relative position of the vehicle and the obstacle according to claim 2, characterized in that: the calculation formula for calculating the current heading angle theta of the vehicle body in the step (10) is as follows:
at this time, the coordinate calculation formula of the sphere center of each sphere under the vehicle body coordinate system is as follows:
wherein, (x, y) is the coordinate of the sphere center in the two-dimensional world coordinate system, then the coordinate of the sphere center in the vehicle body coordinate system is (x ', y', r-k), r is the sphere radius, k is the distance from the XY plane of the vehicle body coordinate system to the ground, and r-k is the coordinate of the sphere center in the Z axis of the vehicle body coordinate system.
9. The method for automatically acquiring the relative position of the vehicle and the obstacle according to claim 8, characterized in that: the step (11) is to calculate the coordinates of the sphere centers of the sphere feature calibration objects under the vehicle body coordinate system through the formula (2) and the formula (3), and the sphere centers of all the sphere feature calibration objects are located at the vehicle body seatThe vertical coordinates under the standard system are all k-r, and the coordinate set of the sphere center of the sphere feature calibration object under the vehicle body coordinate system is represented as Ovehicle:
(xV1,yV1,zV1)..(xV6,yV6,zV6) And calibrating coordinates of the centers of six spheres in the two groups of sphere characteristic calibration objects under the vehicle body coordinate system.
10. The method for automatically acquiring the relative position of the vehicle and the obstacle according to claim 1, characterized in that: and the calibration scene of the vehicle-mounted laser radar is a static scene.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110487005.1A CN113096187B (en) | 2021-05-03 | 2021-05-03 | Method for automatically acquiring relative position of vehicle and obstacle |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110487005.1A CN113096187B (en) | 2021-05-03 | 2021-05-03 | Method for automatically acquiring relative position of vehicle and obstacle |
Publications (2)
Publication Number | Publication Date |
---|---|
CN113096187A true CN113096187A (en) | 2021-07-09 |
CN113096187B CN113096187B (en) | 2022-05-17 |
Family
ID=76681258
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202110487005.1A Expired - Fee Related CN113096187B (en) | 2021-05-03 | 2021-05-03 | Method for automatically acquiring relative position of vehicle and obstacle |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN113096187B (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115083208A (en) * | 2022-07-20 | 2022-09-20 | 深圳市城市交通规划设计研究中心股份有限公司 | Human-vehicle conflict early warning method, early warning analysis method, electronic device and storage medium |
CN118397109A (en) * | 2024-06-27 | 2024-07-26 | 杭州海康威视数字技术股份有限公司 | Vehicle-mounted camera calibration method and device, electronic equipment and machine-readable storage medium |
Citations (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20110102592A1 (en) * | 2009-10-30 | 2011-05-05 | Valeo Vision | System of gauging a camera suitable for equipping a vehicle |
CN103871071A (en) * | 2014-04-08 | 2014-06-18 | 北京经纬恒润科技有限公司 | Method for camera external reference calibration for panoramic parking system |
CN106405555A (en) * | 2016-09-23 | 2017-02-15 | 百度在线网络技术(北京)有限公司 | Obstacle detecting method and device used for vehicle-mounted radar system |
CN107133988A (en) * | 2017-06-06 | 2017-09-05 | 科大讯飞股份有限公司 | The scaling method and calibration system of camera in vehicle-mounted panoramic viewing system |
CN107966700A (en) * | 2017-11-20 | 2018-04-27 | 天津大学 | A kind of front obstacle detecting system and method for pilotless automobile |
CN108490830A (en) * | 2018-03-16 | 2018-09-04 | 合肥工业大学 | A kind of control device and its control method of the automatic car washing based on machine vision |
CN109557525A (en) * | 2019-01-31 | 2019-04-02 | 浙江工业大学 | A kind of automatic calibration method of laser radar formula vehicle overall dimension measuring instrument |
CN109767473A (en) * | 2018-12-30 | 2019-05-17 | 惠州华阳通用电子有限公司 | A kind of panorama parking apparatus scaling method and device |
CN109849782A (en) * | 2017-11-30 | 2019-06-07 | 比亚迪股份有限公司 | Virtual panoramic auxiliary driving device and its display methods, vehicle |
CN110488234A (en) * | 2019-08-30 | 2019-11-22 | 北京百度网讯科技有限公司 | Outer ginseng scaling method, device, equipment and the medium of vehicle-mounted millimeter wave radar |
CN111145260A (en) * | 2019-08-30 | 2020-05-12 | 广东星舆科技有限公司 | Vehicle-mounted binocular calibration method |
CN111862672A (en) * | 2020-06-24 | 2020-10-30 | 北京易航远智科技有限公司 | Parking lot vehicle self-positioning and map construction method based on top view |
CN112224132A (en) * | 2020-10-28 | 2021-01-15 | 武汉极目智能技术有限公司 | Vehicle panoramic all-around obstacle early warning method |
CN112233188A (en) * | 2020-10-26 | 2021-01-15 | 南昌智能新能源汽车研究院 | Laser radar-based roof panoramic camera and calibration method thereof |
-
2021
- 2021-05-03 CN CN202110487005.1A patent/CN113096187B/en not_active Expired - Fee Related
Patent Citations (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20110102592A1 (en) * | 2009-10-30 | 2011-05-05 | Valeo Vision | System of gauging a camera suitable for equipping a vehicle |
CN103871071A (en) * | 2014-04-08 | 2014-06-18 | 北京经纬恒润科技有限公司 | Method for camera external reference calibration for panoramic parking system |
CN106405555A (en) * | 2016-09-23 | 2017-02-15 | 百度在线网络技术(北京)有限公司 | Obstacle detecting method and device used for vehicle-mounted radar system |
CN107133988A (en) * | 2017-06-06 | 2017-09-05 | 科大讯飞股份有限公司 | The scaling method and calibration system of camera in vehicle-mounted panoramic viewing system |
CN107966700A (en) * | 2017-11-20 | 2018-04-27 | 天津大学 | A kind of front obstacle detecting system and method for pilotless automobile |
CN109849782A (en) * | 2017-11-30 | 2019-06-07 | 比亚迪股份有限公司 | Virtual panoramic auxiliary driving device and its display methods, vehicle |
CN108490830A (en) * | 2018-03-16 | 2018-09-04 | 合肥工业大学 | A kind of control device and its control method of the automatic car washing based on machine vision |
CN109767473A (en) * | 2018-12-30 | 2019-05-17 | 惠州华阳通用电子有限公司 | A kind of panorama parking apparatus scaling method and device |
CN109557525A (en) * | 2019-01-31 | 2019-04-02 | 浙江工业大学 | A kind of automatic calibration method of laser radar formula vehicle overall dimension measuring instrument |
CN110488234A (en) * | 2019-08-30 | 2019-11-22 | 北京百度网讯科技有限公司 | Outer ginseng scaling method, device, equipment and the medium of vehicle-mounted millimeter wave radar |
CN111145260A (en) * | 2019-08-30 | 2020-05-12 | 广东星舆科技有限公司 | Vehicle-mounted binocular calibration method |
CN111862672A (en) * | 2020-06-24 | 2020-10-30 | 北京易航远智科技有限公司 | Parking lot vehicle self-positioning and map construction method based on top view |
CN112233188A (en) * | 2020-10-26 | 2021-01-15 | 南昌智能新能源汽车研究院 | Laser radar-based roof panoramic camera and calibration method thereof |
CN112224132A (en) * | 2020-10-28 | 2021-01-15 | 武汉极目智能技术有限公司 | Vehicle panoramic all-around obstacle early warning method |
Non-Patent Citations (2)
Title |
---|
朱政泽 等: "基于延迟补偿的网联式自主驾驶车辆协同控制", 《系统仿真学报》 * |
汪佩 等: "基于单线激光雷达与视觉融合的负障碍检测算法", 《计算机工程》 * |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115083208A (en) * | 2022-07-20 | 2022-09-20 | 深圳市城市交通规划设计研究中心股份有限公司 | Human-vehicle conflict early warning method, early warning analysis method, electronic device and storage medium |
CN115083208B (en) * | 2022-07-20 | 2023-02-03 | 深圳市城市交通规划设计研究中心股份有限公司 | Human-vehicle conflict early warning method, early warning analysis method, electronic device and storage medium |
CN118397109A (en) * | 2024-06-27 | 2024-07-26 | 杭州海康威视数字技术股份有限公司 | Vehicle-mounted camera calibration method and device, electronic equipment and machine-readable storage medium |
Also Published As
Publication number | Publication date |
---|---|
CN113096187B (en) | 2022-05-17 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN108873943B (en) | Image processing method for centimeter-level accurate landing of unmanned aerial vehicle | |
CN113096187B (en) | Method for automatically acquiring relative position of vehicle and obstacle | |
US20220148213A1 (en) | Method for fully automatically detecting chessboard corner points | |
CN112223285B (en) | Robot hand-eye calibration method based on combined measurement | |
CN106326892B (en) | Visual landing pose estimation method of rotary wing type unmanned aerial vehicle | |
CN106340044B (en) | Join automatic calibration method and caliberating device outside video camera | |
CN112308916B (en) | Target pose recognition method based on image target | |
CN110823252B (en) | Automatic calibration method for multi-line laser radar and monocular vision | |
CN108305264A (en) | A kind of unmanned plane precision landing method based on image procossing | |
CN111508027B (en) | Method and device for calibrating external parameters of camera | |
CN107194399A (en) | A kind of vision determines calibration method, system and unmanned plane | |
CN104835173A (en) | Positioning method based on machine vision | |
CN113156411A (en) | Vehicle-mounted laser radar calibration method | |
WO2021017352A1 (en) | Laser radar-camera joint calibration target and joint calibration method | |
CN113793270A (en) | Aerial image geometric correction method based on unmanned aerial vehicle attitude information | |
CN113379712A (en) | Steel bridge bolt disease detection method and system based on computer vision | |
CN110969661B (en) | Image processing device and method, and position calibration system and method | |
CN115014371A (en) | Positioning and mapping method and device for grain transfer vehicle of intelligent grain depot and storage medium | |
CN117315046A (en) | Method and device for calibrating looking-around camera, electronic equipment and storage medium | |
CN114972531B (en) | Corner detection method, equipment and readable storage medium | |
CN115792865A (en) | Camera and mechanical laser radar-based external parameter calibration method, system, medium and vehicle | |
CN110543612B (en) | Card collection positioning method based on monocular vision measurement | |
CN117723563B (en) | Chip pin defect detection method based on point cloud registration | |
CN112907677B (en) | Camera calibration method and device for single-frame image and storage medium | |
CN112102396B (en) | Method, device, equipment and storage medium for positioning vehicle under bridge crane |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant | ||
CF01 | Termination of patent right due to non-payment of annual fee |
Granted publication date: 20220517 |
|
CF01 | Termination of patent right due to non-payment of annual fee |