CN108527319B - Robot teaching method and system based on vision system - Google Patents
Robot teaching method and system based on vision system Download PDFInfo
- Publication number
- CN108527319B CN108527319B CN201810264284.3A CN201810264284A CN108527319B CN 108527319 B CN108527319 B CN 108527319B CN 201810264284 A CN201810264284 A CN 201810264284A CN 108527319 B CN108527319 B CN 108527319B
- Authority
- CN
- China
- Prior art keywords
- robot
- track
- teaching
- data platform
- vision system
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000000034 method Methods 0.000 title claims abstract description 9
- 238000012545 processing Methods 0.000 claims abstract description 21
- 238000012937 correction Methods 0.000 claims abstract description 16
- 238000009434 installation Methods 0.000 claims abstract description 7
- 238000005457 optimization Methods 0.000 claims abstract description 6
- 230000000007 visual effect Effects 0.000 claims description 11
- 238000004519 manufacturing process Methods 0.000 claims description 7
- 238000005094 computer simulation Methods 0.000 claims description 3
- 238000013461 design Methods 0.000 claims description 3
- 238000003754 machining Methods 0.000 description 2
- 238000013473 artificial intelligence Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
Classifications
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B25—HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
- B25J—MANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
- B25J9/00—Programme-controlled manipulators
- B25J9/0081—Programme-controlled manipulators with leader teach-in means
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B25—HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
- B25J—MANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
- B25J9/00—Programme-controlled manipulators
- B25J9/16—Programme controls
- B25J9/1656—Programme controls characterised by programming, planning systems for manipulators
- B25J9/1671—Programme controls characterised by programming, planning systems for manipulators characterised by simulation, either to verify existing program or to create and verify new program, CAD/CAM oriented, graphic oriented programming systems
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B25—HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
- B25J—MANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
- B25J9/00—Programme-controlled manipulators
- B25J9/16—Programme controls
- B25J9/1694—Programme controls characterised by use of sensors other than normal servo-feedback from position, speed or acceleration sensors, perception control, multi-sensor controlled systems, sensor fusion
- B25J9/1697—Vision controlled systems
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02P—CLIMATE CHANGE MITIGATION TECHNOLOGIES IN THE PRODUCTION OR PROCESSING OF GOODS
- Y02P90/00—Enabling technologies with a potential contribution to greenhouse gas [GHG] emissions mitigation
- Y02P90/02—Total factory control, e.g. smart factories, flexible manufacturing systems [FMS] or integrated manufacturing systems [IMS]
Landscapes
- Engineering & Computer Science (AREA)
- Robotics (AREA)
- Mechanical Engineering (AREA)
- Numerical Control (AREA)
Abstract
The invention discloses a robot teaching method and a system based on a vision system, which are used for eliminating equipment installation errors in the robot offline program teaching by means of acquisition analysis of the vision system and matching optimization of a data platform, and firstly controlling the robot to run according to a track provided by an offline program; the vision system moves along the rough track of the robot, and images are acquired; analyzing the track offset relation between the robot and the workpiece according to the image; the data platform analyzes and determines the optimal track and the track correction parameters by matching the processing data stored in the database. The teaching system comprises a robot processing system, a vision system and a data platform. The invention applies the vision system to the teaching of the robot processing system, solves the problem that the error is eliminated by manually optimizing the track by the aid of the demonstrator after the offline program is imported in the past, and the robot acquires the track by the vision system, analyzes and judges the track by the aid of big data, and autonomously selects and optimizes the teaching track, so that the teaching efficiency is high, and the track optimization is accurate.
Description
Technical Field
The invention relates to the field of automobile manufacturing, in particular to a robot teaching method and system based on a vision system.
Background
The industrial robot is used for realizing automatic production, replaces workers to engage in complex and repeated labor, and is widely applied to the field of automobile manufacturing. Industrial robots require teaching prior to use.
The robot teaching is commonly realized by combining an offline program with a demonstrator, and the motion trail of the robot reaches the processing requirement under manual intervention. However, the use of the demonstrator to optimize the track has high requirements on the skill level of the operator, the time and effort spent by manual teaching are high, and the manual teaching is only carried out by naked eye observation and experience, so that the teaching precision is low.
Disclosure of Invention
The invention aims to provide a robot teaching method and a robot teaching system based on a vision system.
In order to achieve the above purpose, the technical scheme of the invention is as follows: a robot teaching method based on a vision system eliminates equipment installation errors in the robot offline program teaching by means of acquisition and analysis of the vision system and matching optimization of a data platform, and specifically comprises the following steps:
step one, offline programming, namely generating a robot offline program based on 3D digital-analog dynamic simulation of a robot processing system;
installing equipment, namely installing the robot and the clamp installation design data in place; the offline program is imported into the robot controller;
step three, the coarse track is operated, and the robot is controlled to operate according to the track provided by the off-line program;
step four, image acquisition, wherein a vision system moves along a rough track of a robot, and depth images of workpieces mounted on a clamp are acquired;
fifthly, calculating the track offset, wherein the visual system analyzes the spatial position relation between the robot and the workpiece according to the depth image, and calculates the track offset of the robot through the visual system;
step six, matching the optimal track, wherein the visual system feeds back the track offset of the robot and the workpiece to a data platform, and the data platform analyzes and determines the optimal track by matching the processing data stored in the database;
step seven, track correction parameters are calculated, the data platform compares the space deviation of the optimal track and the coarse track, and track correction parameters are determined;
and step eight, track correction, wherein the data platform feeds back the track correction parameters to the robot controller, and the robot is controlled by the data platform to automatically optimize the track correction, so that teaching is completed.
Further, the data platform is connected with a factory data center and is used for storing or calling existing parameters of the mass production vehicle type, such as workpiece parameters, machining parameters, material forming parameters and the like, so that enough reference can be provided for robot teaching.
The robot teaching system based on the vision system comprises a robot processing system, a vision system and a data platform, wherein the robot processing system comprises a robot controller, a robot, a clamp and a teaching workpiece arranged on the clamp; the visual system comprises an image acquisition device and an image processing industrial personal computer; the image acquisition device is arranged at the tail end of the robot and moves along with the robot; a spatial position analysis module is arranged in the image processing industrial personal computer; the image processing industrial personal computer is connected with the data platform, the data platform is connected with the robot controller, and the data platform stores the existing parameters of the vehicle model.
The beneficial effects of the invention are as follows:
the invention applies the vision system to the teaching of the robot processing system, solves the problem that the error is eliminated by manually optimizing the track by the aid of the demonstrator after the offline program is imported in the past, and the robot acquires the track by the vision system, analyzes and judges the track by the aid of big data, and autonomously selects and optimizes the teaching track, so that the teaching efficiency is high, and the track optimization is accurate.
Drawings
FIG. 1 is a schematic diagram of the structure of the present invention.
Detailed Description
The technical solutions of the embodiments of the present invention will be clearly and completely described below with reference to the accompanying drawings.
A robot teaching method based on a vision system eliminates equipment installation errors in the robot offline program teaching by means of acquisition and analysis of the vision system and matching optimization of a data platform, and specifically comprises the following steps:
step one, offline programming, namely generating a robot offline program based on 3D digital-analog dynamic simulation of a robot processing system;
installing equipment, namely installing the robot and the clamp installation design data in place; the offline program is imported into the robot controller;
step three, the coarse track is operated, and the robot is controlled to operate according to the track provided by the off-line program;
step four, image acquisition, wherein a vision system moves along a rough track of a robot, and depth images of workpieces mounted on a clamp are acquired;
fifthly, calculating the track offset, wherein the visual system analyzes the spatial position relation between the robot and the workpiece according to the depth image, and calculates the track offset of the robot through the visual system;
step six, matching the optimal track, wherein the visual system feeds back the track offset of the robot and the workpiece to a data platform, and the data platform analyzes and determines the optimal track by matching the processing data stored in the database;
step seven, track correction parameters are calculated, the data platform compares the space deviation of the optimal track and the coarse track, and track correction parameters are determined;
and step eight, track correction, wherein the data platform feeds back the track correction parameters to the robot controller, and the robot is controlled by the data platform to automatically optimize the track correction, so that teaching is completed.
The data platform is connected with the factory data center and is used for storing or calling existing parameters of the mass production vehicle type, such as workpiece parameters, machining parameters, material forming parameters and the like, so that enough references for robot teaching can be provided.
As shown in fig. 1, a robot teaching system based on a vision system comprises a robot processing system, a vision system and a data platform, wherein the robot processing system comprises a robot controller, a robot, a clamp and a teaching workpiece arranged on the clamp; the visual system comprises an image acquisition device and an image processing industrial personal computer; the image acquisition device is arranged at the tail end of the robot and moves along with the robot; a spatial position analysis module is arranged in the image processing industrial personal computer; the image processing industrial personal computer is connected with the data platform, the data platform is connected with the robot controller, and the data platform stores the existing parameters of the vehicle model.
The data platform is built based on big data ideas, needs to have the functions of accessing factory-level and industry-level data platforms, has the autonomous learning capability of updating data periodically or irregularly, and has the capability of acquiring network data by connecting the Internet in real time, the capability of finishing and summarizing production practice data, and is suitable for the development of the Internet of things in the artificial intelligence direction.
The described embodiments are only some, but not all, embodiments of the invention. All other embodiments, which can be made by those skilled in the art based on the embodiments of the invention without making any inventive effort, are intended to be within the scope of the invention.
Claims (2)
1. The robot teaching method based on the vision system is characterized by eliminating equipment installation errors in the robot offline program teaching by means of acquisition and analysis of the vision system and matching optimization of a data platform, and comprises the following specific steps:
step one, offline programming, namely generating a robot offline program based on 3D digital-analog dynamic simulation of a robot processing system;
installing equipment, namely installing the robot and the clamp installation design data in place; the offline program is imported into the robot controller;
step three, the coarse track is operated, and the robot is controlled to operate according to the track provided by the off-line program;
step four, image acquisition, wherein a vision system moves along a rough track of a robot, and depth images of workpieces mounted on a clamp are acquired;
fifthly, calculating the track offset, wherein the visual system analyzes the spatial position relation between the robot and the workpiece according to the depth image, and calculates the track offset of the robot through the visual system;
step six, matching the optimal track, wherein the visual system feeds back the track offset of the robot and the workpiece to a data platform, and the data platform analyzes and determines the optimal track by matching the processing data stored in the database;
step seven, track correction parameters are calculated, the data platform compares the space deviation of the optimal track and the coarse track, and track correction parameters are determined;
and step eight, track correction, wherein the data platform feeds back the track correction parameters to the robot controller, and the robot is controlled by the data platform to automatically optimize the track correction, so that teaching is completed.
2. The vision system-based robot teaching method according to claim 1, wherein the data platform is connected to a factory data center for storing or calling existing parameters of a model of a production vehicle, and providing a sufficient reference for robot teaching.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810264284.3A CN108527319B (en) | 2018-03-28 | 2018-03-28 | Robot teaching method and system based on vision system |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810264284.3A CN108527319B (en) | 2018-03-28 | 2018-03-28 | Robot teaching method and system based on vision system |
Publications (2)
Publication Number | Publication Date |
---|---|
CN108527319A CN108527319A (en) | 2018-09-14 |
CN108527319B true CN108527319B (en) | 2024-02-13 |
Family
ID=63485363
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201810264284.3A Active CN108527319B (en) | 2018-03-28 | 2018-03-28 | Robot teaching method and system based on vision system |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN108527319B (en) |
Families Citing this family (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109483532A (en) * | 2018-11-01 | 2019-03-19 | 东莞市中天自动化科技有限公司 | Lathe and its corase grinding robot control method |
CN109530984A (en) * | 2018-11-19 | 2019-03-29 | 施努卡(苏州)智能装备有限公司 | Vision positioning welding and assembling method |
CN109676583B (en) * | 2018-12-03 | 2021-08-24 | 深圳市越疆科技有限公司 | Deep learning visual acquisition method based on target posture, learning system and storage medium |
CN111002317A (en) * | 2019-11-20 | 2020-04-14 | 希美埃(芜湖)机器人技术有限公司 | Novel spraying method and novel spraying device based on robot vision in door and window spraying industry |
CN110977260B (en) * | 2019-12-17 | 2021-08-17 | 易思维(杭州)科技有限公司 | Intelligent repair welding system and follow-up repair welding method for body-in-white |
CN111283685A (en) * | 2020-03-05 | 2020-06-16 | 广州市斯睿特智能科技有限公司 | Vision teaching method of robot based on vision system |
CN113050424A (en) * | 2021-03-19 | 2021-06-29 | 安徽理工大学环境友好材料与职业健康研究院(芜湖) | Intelligent self-adaptive control method and system based on image recognition |
CN114170314B (en) * | 2021-12-07 | 2023-05-26 | 群滨智造科技(苏州)有限公司 | Intelligent 3D vision processing-based 3D glasses process track execution method |
CN114683283B (en) * | 2022-03-25 | 2024-04-16 | 中铁科工集团有限公司 | Teaching-free welding method and device for welding robot |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP1462224A2 (en) * | 2003-02-27 | 2004-09-29 | Fanuc Ltd | Taught position modification device |
CN102785246A (en) * | 2012-08-24 | 2012-11-21 | 电子科技大学 | Robot calibration method capable of achieving automatic trajectory correction |
CN103286782A (en) * | 2013-06-07 | 2013-09-11 | 上海发那科机器人有限公司 | Flexible tracking and positioning system and flexible tracking and positioning method of robot |
CN203345084U (en) * | 2013-06-07 | 2013-12-18 | 上海发那科机器人有限公司 | Automatic assembly system for car windshield glass |
CN104741739A (en) * | 2013-12-30 | 2015-07-01 | 唐山长城电焊机总厂有限公司 | Position correcting system of welding robot |
CN105364927A (en) * | 2015-12-15 | 2016-03-02 | 天津立德尔智能装备科技有限公司 | Robot carrying visual system and robot carrying quick positioning method |
CN106607907A (en) * | 2016-12-23 | 2017-05-03 | 西安交通大学 | Mobile vision robot and measurement and control method thereof |
CN208117838U (en) * | 2018-03-28 | 2018-11-20 | 广州瑞松北斗汽车装备有限公司 | The robot teaching system of view-based access control model system |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP6188440B2 (en) * | 2013-06-17 | 2017-08-30 | キヤノン株式会社 | Robot apparatus and robot control method |
-
2018
- 2018-03-28 CN CN201810264284.3A patent/CN108527319B/en active Active
Patent Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP1462224A2 (en) * | 2003-02-27 | 2004-09-29 | Fanuc Ltd | Taught position modification device |
CN102785246A (en) * | 2012-08-24 | 2012-11-21 | 电子科技大学 | Robot calibration method capable of achieving automatic trajectory correction |
CN103286782A (en) * | 2013-06-07 | 2013-09-11 | 上海发那科机器人有限公司 | Flexible tracking and positioning system and flexible tracking and positioning method of robot |
CN203345084U (en) * | 2013-06-07 | 2013-12-18 | 上海发那科机器人有限公司 | Automatic assembly system for car windshield glass |
CN104741739A (en) * | 2013-12-30 | 2015-07-01 | 唐山长城电焊机总厂有限公司 | Position correcting system of welding robot |
CN105364927A (en) * | 2015-12-15 | 2016-03-02 | 天津立德尔智能装备科技有限公司 | Robot carrying visual system and robot carrying quick positioning method |
CN106607907A (en) * | 2016-12-23 | 2017-05-03 | 西安交通大学 | Mobile vision robot and measurement and control method thereof |
CN208117838U (en) * | 2018-03-28 | 2018-11-20 | 广州瑞松北斗汽车装备有限公司 | The robot teaching system of view-based access control model system |
Also Published As
Publication number | Publication date |
---|---|
CN108527319A (en) | 2018-09-14 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN108527319B (en) | Robot teaching method and system based on vision system | |
CN110694828A (en) | Robot spraying track planning method based on large complex curved surface model | |
US11119470B2 (en) | Industrial robot process cloud system and working method thereof | |
CN111152229B (en) | Manipulator guiding method and device for 3D mechanical vision | |
CN113110328A (en) | Production process full-period intelligent workshop system based on digital twin technology and solution method | |
CN111596614B (en) | Motion control error compensation system and method based on cloud edge cooperation | |
EP3733355A1 (en) | Robot motion optimization system and method | |
CN114161048B (en) | 3D vision-based parameterized welding method and device for tower legs of iron tower | |
CN103406905A (en) | Robot system with visual servo and detection functions | |
CN103713579A (en) | Industrial robot operation method | |
CN111310318A (en) | Digital twinning-based process margin processing method and system and mechanical manufacturing assembly | |
KR20170006397A (en) | System and method for compensating robot teaching | |
CN116079732A (en) | Cabin assembly method based on laser tracker and binocular vision mixed guidance | |
CN108472808B (en) | Synchronization of multiple robots | |
CN103144109A (en) | Substation type precision compensation for robot system with additional external shaft | |
CN110153582B (en) | Welding scheme generation method and device and welding system | |
CN110967042A (en) | Industrial robot positioning precision calibration method, device and system | |
CN208117838U (en) | The robot teaching system of view-based access control model system | |
CN112257252B (en) | Method for simulating and analyzing influence of machine tool space error on workpiece machining precision | |
CN114800523B (en) | Mechanical arm track correction method, system, computer and readable storage medium | |
CN110850879A (en) | Automatic cruising operation method of wall-climbing robot | |
CN116834009A (en) | Online compensation system for pose errors of distributed virtual-real twin robot | |
CN113492071B (en) | Path planning method applied to window frame spraying | |
CN109048925A (en) | A kind of flexible assembly robot and its Magnetic template assembly method | |
CN103115565A (en) | Laser measuring system and laser measuring method thereof |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |