CN110992410B - Robot vision guiding method and device based on RGB-D data fusion - Google Patents

Robot vision guiding method and device based on RGB-D data fusion Download PDF

Info

Publication number
CN110992410B
CN110992410B CN201911196215.4A CN201911196215A CN110992410B CN 110992410 B CN110992410 B CN 110992410B CN 201911196215 A CN201911196215 A CN 201911196215A CN 110992410 B CN110992410 B CN 110992410B
Authority
CN
China
Prior art keywords
rgb
data
point
point cloud
target
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201911196215.4A
Other languages
Chinese (zh)
Other versions
CN110992410A (en
Inventor
刁世普
郑振兴
秦磊
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
FOSHAN DILE VISION TECHNOLOGY Co.,Ltd.
Original Assignee
Guangdong Polytechnic Normal University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guangdong Polytechnic Normal University filed Critical Guangdong Polytechnic Normal University
Priority to CN201911196215.4A priority Critical patent/CN110992410B/en
Publication of CN110992410A publication Critical patent/CN110992410A/en
Priority to PCT/CN2020/101335 priority patent/WO2021103558A1/en
Application granted granted Critical
Publication of CN110992410B publication Critical patent/CN110992410B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/30Determination of transform parameters for the alignment of images, i.e. image registration
    • G06T7/33Determination of transform parameters for the alignment of images, i.e. image registration using feature-based methods
    • GPHYSICS
    • G05CONTROLLING; REGULATING
    • G05BCONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
    • G05B19/00Programme-control systems
    • G05B19/02Programme-control systems electric
    • G05B19/18Numerical control [NC], i.e. automatically operating machines, in particular machine tools, e.g. in a manufacturing environment, so as to execute positioning, movement or co-ordinated operations by means of programme data in numerical form
    • G05B19/19Numerical control [NC], i.e. automatically operating machines, in particular machine tools, e.g. in a manufacturing environment, so as to execute positioning, movement or co-ordinated operations by means of programme data in numerical form characterised by positioning or contouring control systems, e.g. to control position from one programmed point to another or to control movement along a programmed continuous path
    • G05B19/21Numerical control [NC], i.e. automatically operating machines, in particular machine tools, e.g. in a manufacturing environment, so as to execute positioning, movement or co-ordinated operations by means of programme data in numerical form characterised by positioning or contouring control systems, e.g. to control position from one programmed point to another or to control movement along a programmed continuous path using an incremental digital measuring device
    • G05B19/25Numerical control [NC], i.e. automatically operating machines, in particular machine tools, e.g. in a manufacturing environment, so as to execute positioning, movement or co-ordinated operations by means of programme data in numerical form characterised by positioning or contouring control systems, e.g. to control position from one programmed point to another or to control movement along a programmed continuous path using an incremental digital measuring device for continuous-path control
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/50Depth or shape recovery
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • G06T7/73Determining position or orientation of objects or cameras using feature-based methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10028Range image; Depth image; 3D point clouds
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30108Industrial image inspection

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Theoretical Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Manufacturing & Machinery (AREA)
  • Automation & Control Theory (AREA)
  • Image Analysis (AREA)
  • Manipulator (AREA)
  • Image Processing (AREA)

Abstract

The invention discloses a robot vision guiding method and device based on RGB-D data fusion, which are based on an RGB-D composite sensor, and a machining guide point sequence is finally obtained by sequentially carrying out the steps of machining target identification, machining target segmentation, machining path point acquisition and machining guide point conversion from the beginning of machining target data acquisition, so that the calculation time is reduced, the requirement on real-time processing is met, the requirement on the performance of software and hardware is reduced, the cost can be saved, the development difficulty is reduced, and the requirement on a high-speed large-scale production mode is met.

Description

Robot vision guiding method and device based on RGB-D data fusion
Technical Field
The invention relates to the field of robot vision, in particular to a robot vision guiding method and device based on RGB-D data fusion.
Background
Automation equipment (robot system) for manufacturing strong country instruments is required to be accelerated and intelligentized. An important means of intellectualization of automation equipment is to equip the machine with an "eye" and a "brain" able to cooperate with this eye. The "eye" may be a monocular camera, a binocular camera, a multi-view camera, a three-dimensional scanner, or an RGB-D (RGB + Depth) sensor. The core work content of the automation equipment intellectualization comprises the following steps: the image data acquired by the "eye" is analyzed (e.g., image recognition), and the robotic system is guided to perform a particular machining or assembly operation based on the analysis. Therefore, the image data analysis based on two-dimensional images, which is widely adopted at present, is a key basic core technology. However, the method based on two-dimensional image data analysis is easily interfered by external environmental factors such as light and the like, so that the recognition accuracy is low, the robustness is poor, and the precision does not meet the requirement. In order to meet the current requirements for high speed and high precision of image data analysis, the RGB-D sensor, which can acquire the depth information of the corresponding two-dimensional image data that is not easily interfered by external environmental factors such as light while acquiring the conventional two-dimensional image data, is popularized and applied. The 3D data (point cloud data) corresponding to the target object may be obtained by processing the RGB image data and the depth data acquired by the RGB-D sensor. The existing point cloud analysis method has the following disadvantages: the method for segmenting the processing target by the point cloud is time-consuming and cannot meet the requirement of current high-speed production; the 3D processing target positioning method based on pure 3D point cloud and deep learning has poor accuracy and low resolution precision, and does not meet the requirement of current high-precision processing. In summary, the existing point cloud analysis method is not highly universal and is only suitable for a specific automatic processing system. In addition, since no key information based on RGB-D data for a specific machining or assembly operation task is currently available to guide the research application of the robot system. Therefore, the invention discloses a method and a device for determining a target direction of a robot based on RGB-D data, thereby providing a visual guidance method and a visual guidance device which can meet the detection precision requirement of a processing target required by current processing and are suitable for a processing target workpiece.
Disclosure of Invention
The invention mainly aims to provide a robot vision guiding method and device based on RGB-D data fusion, aiming at solving the technical problems that the existing point cloud segmentation processing target method is time-consuming and cannot meet the requirement of current high-speed production, the 3D processing target positioning method based purely on 3D point cloud and deep learning is poor in accuracy, low in resolution precision and not in line with the requirement of current high-precision processing, and the existing point cloud analysis method is not high in universality and only suitable for a specific automatic processing system. In view of these disadvantages, the object of the present invention is to provide a method and an apparatus for recognizing a target orientation, which can satisfy the accuracy of detection of a processing target required for current processing, and which are suitable for a workpiece to be processed, which can satisfy the requirements of detection and processing of a current processing target, and which can satisfy the requirements of accuracy for medium accuracy, and which are suitable for a workpiece to be processed.
In order to solve the above problems, the present invention provides a robot vision guiding method based on RGB-D data fusion, comprising:
step 1, acquiring an RGB two-dimensional image and depth data containing a target to be processed through an RGB-D composite sensor, and registering the RGB two-dimensional image and the depth data according to preset registration parameters of the RGB-D composite sensor to generate a registered two-dimensional image I corresponding to the RGB two-dimensional imageRGBAnd registered depth data I corresponding to said depth dataD
Step 2, using the image segmentation model trained in advance to obtain the registered two-dimensional image I as inputRGBDividing the region S corresponding to the target to be processedRGBAccording to said region SRGBFrom the registered depth data IDExtracting 3D point cloud data S of the target to be processed3D
Step 3, from the 3D point cloud data S3DExtracting significant feature point sequence { KXi}i=1->mUsing the significant feature point sequence { KXi}i=1->m(ii) assigning the 3D point cloud data S for a cut-off point3DDivided into local point cloud sets { S3D-j}j=1->nSaid KXiIs a significant feature point, i is the corresponding significant feature point KXiThe value range of i is [1, m ]]M is the significant feature point KXiTotal number of said S3D-jIs a local point cloud, j is the local point cloud S3D-jThe value range of j is [1, n ]]N is the local point cloud S3D-jThe total number of (c);
step 4, sequentially setting the j to be from 1 to n, and searching and extracting the point cloud S vertical to the local point by using a local search algorithm3D-jAnd the plane of the bounding box of (A) and the corresponding local point cloud S3D-jMachining path point SX at the boundary of (1)jThereby obtaining the machining path point sequence { SXj}j=1->n
Step 5, setting the j to be from 1 to n in sequence, and using a conversion matrix BET to convert the processing path point SXjConversion to a machining guide point BXjSo as to process the path point sequence { SXj}j=1->nConversion into a sequence of machining guide points { BXj}j=1->n
Preferably, in the step 4, the j is sequentially set to be from 1 to n, and the local point cloud S is obtained through a surface fitting algorithm3D-jFitting to curved surfaces SS3D-jThen searching and extracting the point cloud S vertical to the local point by using the local search algorithm3D-jThe plane of the bounding box and the corresponding curved surface SS3D-jThe machining path point SX at the junction ofjThereby obtaining the machining path point sequence { SXj}j=1->n
Preferably, said step 2, registering said two-dimensional images IRGBAnd corresponding said registration depth data IDAre merged to generate fused data IRGB-DUsing said fused data I as input by a pre-trained image segmentation modelRGB-DDividing the region S corresponding to the target to be processedRGB-DAccording to said region SRGB-DFrom the registered depth data IDExtracting 3D point cloud data S of the target to be processed3D
Preferably, the training samples required for the pre-training of the image segmentation model of the step 2 are obtained by acquiring the registered two-dimensional image I containing the object to be processed of the same type using the RGB-D composite sensorRGBAnd corresponding said registration depth data IDAnd (4) generating.
Preferably, the conversion matrix BET of step 5 is generated by a previously calibrated Denavit-Hartenberg (D-H) parameter.
The invention further provides a robot vision guiding device based on RGB-D data fusion, which comprises:
the processing target data acquisition module is used for acquiring an RGB two-dimensional image and depth data containing a target to be processed through an RGB-D composite sensor, and registering the RGB two-dimensional image and the depth data according to preset registration parameters of the RGB-D composite sensor so as to generate a registered two-dimensional image I corresponding to the RGB two-dimensional imageRGBAnd registered depth data I corresponding to said depth dataD
A processing target recognition module for using a pre-trained image segmentation model to extract the registered two-dimensional image I from the image as inputRGBDividing the region S corresponding to the target to be processedRGBAccording to said region SRGBFrom the registered depth data IDExtracting 3D point cloud data S of the target to be processed3D
A processing target segmentation module for segmenting the 3D point cloud data S3DExtracting significant feature point sequence { KXi}i=1->mUsing the significant feature point sequence { KXi}i=1->m(ii) assigning the 3D point cloud data S for a cut-off point3DDivided into local point cloud sets { S3D-j}j=1->nSaid KXiIs a significant feature point, i is the corresponding significant feature point KXiThe value range of i is [1, m ]]M is the significant feature point KXiTotal number of said S3D-jIs a local point cloud, j is the local point cloud S3D-jThe value range of j is [1, n ]]N is the local point cloud S3D-jThe total number of (c);
a processing path point acquisition module for sequentially setting the j from 1 to n, searching and extracting the point cloud S perpendicular to the local point by using a local search algorithm3D-jAnd the plane of the bounding box of (A) and the corresponding local point cloud S3D-jMachining path point SX at the boundary of (1)jThereby obtaining the machining path point sequence { SXj}j=1->n
A processing guide point conversion module for setting the j from 1 to n in sequence, wherein the processing guide point conversion module comprises a BE (least squares element) passing through a conversion matrixT will the machining path point SXjConversion to a machining guide point BXjSo as to process the path point sequence { SXj}j=1->nConversion into a sequence of machining guide points { BXj}j=1->n
Preferably, the processing path point obtaining module is configured to set j to be from 1 to n in sequence, and use a surface fitting algorithm to obtain the local point cloud S3D-jFitting to curved surfaces SS3D-jThen searching and extracting the point cloud S vertical to the local point by using the local search algorithm3D-jThe plane of the bounding box and the corresponding curved surface SS3D-jThe machining path point SX at the junction ofjThereby obtaining the machining path point sequence { SXj}j=1->n
Preferably, the processing target identification module is used for registering the two-dimensional image IRGBAnd corresponding said registration depth data IDAre merged to generate fused data IRGB-DUsing said fused data I as input by a pre-trained image segmentation modelRGB-DDividing the region S corresponding to the target to be processedRGB-DAccording to said region SRGB-DFrom the registered depth data IDExtracting 3D point cloud data S of the target to be processed3D
Preferably, training samples required for the pre-training of the image segmentation model among the processing target recognition modules are acquired by using the RGB-D composite sensor to acquire a registered two-dimensional image I containing the target to be processed of the same kindRGBAnd corresponding said registration depth data IDAnd (4) generating.
Preferably, the conversion matrix BET among the machining guide point conversion modules is generated by a previously calibrated Denavit-Hartenberg (D-H) parameter.
According to the technical scheme, the method can meet the requirement of the detection precision of the machining target required by the current machining, greatly reduces the calculation amount, reduces the calculation complexity, accelerates the processing speed, reduces the calculation time, meets the requirement of real-time processing, reduces the requirement on the performance of software and hardware, can save the cost, reduces the development difficulty, and meets the requirement on a high-speed large-scale production mode.
Drawings
FIG. 1 is a schematic flow chart of a robot vision guiding method based on RGB-D data fusion according to a first embodiment of the present invention;
FIG. 2 is a functional block diagram of a first embodiment of the robot vision guiding device based on RGB-D data fusion according to the present invention;
fig. 3 is a schematic diagram of an RGB-D composite sensor embodying the present invention.
The implementation, functional features and advantages of the objects of the present invention will be further explained with reference to the accompanying drawings.
Detailed Description
It should be understood that the specific embodiments described herein are merely illustrative of the invention and are not intended to limit the invention.
Various embodiments for implementing the present invention will now be described with reference to the accompanying drawings. In the following description, suffixes such as "module", "component", or "unit" used to denote elements are used only for facilitating the explanation of the present invention, and have no specific meaning in themselves. Thus, "module" and "component" may be used in a mixture.
Referring to fig. 1, fig. 1 is a flowchart illustrating a robot vision guiding method based on RGB-D data fusion according to a first embodiment of the present invention. In the embodiment shown in fig. 1, the method for guiding the robot vision based on RGB-D data fusion includes the following steps:
and S10, acquiring processing target data.
The method comprises the steps of acquiring an RGB two-dimensional image and depth data containing a target to be processed through an RGB-D composite sensor, and carrying out registration processing on the RGB two-dimensional image and the depth data according to preset registration parameters of the RGB-D composite sensor, so as to generate a registration two-dimensional image I corresponding to the RGB two-dimensional imageRGBAnd registered depth data I corresponding to said depth dataD
Referring to fig. 3, the RGB-D composite sensor is placed on top of the robot arm D40 with the RGB camera D20 in the middle of the RGB-D composite vision sensor, and the color image data is compressed before being transferred to the computer to ensure speed in analyzing the RGB data. The sensors D10 and D30 on the left and right sides of the RGB-D composite vision sensor are responsible for emitting and receiving infrared rays, respectively: firstly, infrared rays are emitted to an object O10 to be processed through an infrared emitter D10 on the left side, and the infrared rays have high randomness, so that light spots formed by the reflection of the infrared rays at any two different positions in space are different, and a three-dimensional 'light code' is formed for the environment; then, an infrared image in the visual field is collected through the infrared receiver D30 on the right side; finally, a series of complex calculations are carried out on the infrared image by using the parameters of the RGB-D composite vision sensor, and the depth data in the visual field can be obtained.
And S20, identifying the processing target.
I.e. from the registered two-dimensional image I as input using a pre-trained image segmentation modelRGBDividing the region S corresponding to the target to be processedRGBAccording to said region SRGBFrom the registered depth data IDExtracting 3D point cloud data S of the target to be processed3D
The image segmentation model based on the deep learning framework realizes a neural network model for semantic segmentation of the target to be processed, and has the characteristics of high precision, high processing speed and real-time processing.
And S30, dividing the processing target.
I.e. from the 3D point cloud data S3DExtracting significant feature point sequence { KXi}i=1->mUsing the significant feature point sequence { KXi}i=1->m(ii) assigning the 3D point cloud data S for a cut-off point3DDivided into local point cloud sets { S3D-j}j=1->nSaid KXiIs a significant feature point, i is the corresponding significant feature point KXiThe value range of i is [1, m ]]M is the significant feature point KXiOf saidS3D-jIs a local point cloud, j is the local point cloud S3D-jThe value range of j is [1, n ]]N is the local point cloud S3D-jThe total number of (a).
The KXiIs a coordinate vector of the corresponding salient feature point, said S3D-jA set of all points comprising the local point cloud.
And S40, acquiring machining path points.
Namely, sequentially setting the j from 1 to n, searching and extracting the point cloud S vertical to the local point by using a local search algorithm3D-jAnd the plane of the bounding box of (A) and the corresponding local point cloud S3D-jMachining path point SX at the boundary of (1)jThereby obtaining the machining path point sequence { SXj}j=1->n
The step S30 is to obtain the 3D point cloud data S of the target to be processed3DWith a sequence of salient feature points { KXi}i=1->mDividing the boundary point into local point cloud sets (S)3D-j}j=1->nIn step S40, the corresponding machining path point sequence { SX can be obtained simply and effectivelyj}j=1->n. Especially for very complex said 3D point cloud data S3DMore beneficial to processing path point sequence { SXj}j=1->nAnd stability and robustness are greatly enhanced.
And S50, converting the machining guide point.
Setting j to from 1 to n in sequence, using a process comprising converting the machining path points SX by means of a transformation matrix BETjConversion to a machining guide point BXjSo as to process the path point sequence { SXj}j=1->nConversion into a sequence of machining guide points { BXj}j=1->n
Machining path point SXjThe position coordinate information in the coordinate system corresponding to the RGB-D composite vision sensor needs to be converted into the position coordinate information in the corresponding working coordinate system. Finally, the machining guide point sequence { BXj}j=1->nSends the corresponding operation to the robot to execute. By aligning sequences of machining guide points { BXj}j=1->nThe machining path can be obtained by using a curve interpolation algorithm, so that the machining operation is guided.
Therefore, the processing steps can reduce the calculation amount, reduce the calculation complexity, accelerate the processing speed, reduce the calculation time, meet the requirements of real-time processing, reduce the requirements on the performance of software and hardware, save the cost, reduce the development difficulty and meet the requirements on a high-speed large-scale production mode.
Further, in the step S40, setting the j to be from 1 to n in sequence, and performing a surface fitting algorithm to obtain the local point cloud S3D-jFitting to curved surfaces SS3D-jThen searching and extracting the point cloud S vertical to the local point by using the local search algorithm3D-jThe plane of the bounding box and the corresponding curved surface SS3D-jThe machining path point SX at the junction ofjThereby obtaining the machining path point sequence { SXj}j=1->n
Using fitted surfaces SS3D-jCan filter out local point cloud S3D-jData redundancy of (2) to make local point cloud S3D-jThe data are homogenized, the measurement deviation caused by a measurement system is reduced, and the fluctuation of the data is eliminated. Particularly, a smooth machining path can be generated by NURBS curve fitting.
Further, the step S20 is to register the two-dimensional image IRGBAnd corresponding said registration depth data IDAre merged to generate fused data IRGB-DUsing said fused data I as input by a pre-trained image segmentation modelRGB-DDividing the region S corresponding to the target to be processedRGB-DAccording to said region SRGB-DFrom the registered depth data IDExtracting 3D point cloud data S of the target to be processed3D
Using said fusion data IRGB-DThe area S corresponding to the target to be processed can be effectively dividedRGB-DPrecision and accuracy of, and is largeThe robustness and stability of the segmentation are greatly enhanced.
Further, the training samples required for the pre-training of the image segmentation model of the step S20 are obtained by acquiring the registered two-dimensional image I containing the object to be processed of the same kind using the RGB-D composite sensorRGBAnd corresponding said registration depth data IDAnd (4) generating.
Using the RGB-D composite sensor with local grid space placement, a large number of the registered two-dimensional images I of the object to be processed can be obtainedRGBAnd corresponding said registration depth data IDThen, the training sample can be made by labeling; and then training the image segmentation model based on the deep learning framework, and finely adjusting related parameters in the training process until the accuracy of the model reaches an expected value. By the processing steps, a large amount of training sample data can be obtained very efficiently, so that the requirements of the image segmentation model based on the deep learning framework on precision and robustness are ensured.
Further, the conversion matrix BET of the step S50 is generated by a previously calibrated Denavit-Hartenberg (D-H) parameter.
The purpose of calibrating the Denavit-Hartenberg (D-H) parameters by using a laser tracker is to improve the overall accuracy of the robot vision guidance algorithm and to ensure that the machining path point SX is fixedjConversion to a machining guide point BXjThe method is accurate, and has the characteristics of high processing speed, maturity, reliability and easy engineering realization.
The robot vision guiding method based on RGB-D data fusion in the first embodiment of the robot vision guiding method based on RGB-D data fusion of the present invention may be implemented by the robot vision guiding device based on RGB-D data fusion provided in the first embodiment of the robot vision guiding device based on RGB-D data fusion of the present invention.
Referring to fig. 2, fig. 2 is a robot vision guiding device 1 based on RGB-D data fusion according to a first embodiment of the robot vision guiding device based on RGB-D data fusion of the present invention, where the device 1 includes:
a processing target data acquisition module 10, configured to acquire an RGB two-dimensional image and depth data including a target to be processed through an RGB-D composite sensor, and perform registration processing on the RGB two-dimensional image and the depth data according to preset registration parameters of the RGB-D composite sensor, so as to generate a registration two-dimensional image I corresponding to the RGB two-dimensional imageRGBAnd registered depth data I corresponding to said depth dataD
Referring to fig. 3, the RGB-D composite sensor is placed on top of the robot arm D40 with the RGB camera D20 in the middle of the RGB-D composite vision sensor, and the color image data is compressed before being transferred to the computer to ensure speed in analyzing the RGB data. The sensors D10 and D30 on the left and right sides of the RGB-D composite vision sensor are responsible for emitting and receiving infrared rays, respectively: firstly, infrared rays are emitted to an object O10 to be processed through an infrared emitter D10 on the left side, and the infrared rays have high randomness, so that light spots formed by the reflection of the infrared rays at any two different positions in space are different, and a three-dimensional 'light code' is formed for the environment; then, an infrared image in the visual field is collected through the infrared receiver D30 on the right side; finally, a series of complex calculations are carried out on the infrared image by using the parameters of the RGB-D composite vision sensor, and the depth data in the visual field can be obtained.
A processing target recognition module 20 for using a pre-trained image segmentation model to extract the registered two-dimensional image I from the image as inputRGBDividing the region S corresponding to the target to be processedRGBAccording to said region SRGBFrom the registered depth data IDExtracting 3D point cloud data S of the target to be processed3D
The image segmentation model based on the deep learning framework realizes a neural network model for semantic segmentation of the target to be processed, and has the characteristics of high precision, high processing speed and real-time processing.
A processing object segmentation module 30 for segmenting the 3D point cloud data S from3DExtracting significant feature point sequence { KXi}i=1->mUsing the significant feature point sequence { KXi}i=1->m(ii) assigning the 3D point cloud data S for a cut-off point3DDivided into local point cloud sets { S3D-j}j=1->nSaid KXiIs a significant feature point, i is the corresponding significant feature point KXiThe value range of i is [1, m ]]M is the significant feature point KXiTotal number of said S3D-jIs a local point cloud, j is the local point cloud S3D-jThe value range of j is [1, n ]]N is the local point cloud S3D-jThe total number of (a).
The KXiIs a coordinate vector of the corresponding salient feature point, said S3D-jA set of all points comprising the local point cloud.
A processing path point obtaining module 40 for sequentially setting the j from 1 to n, searching and extracting the point cloud S perpendicular to the local point by using a local search algorithm3D-jAnd the plane of the bounding box of (A) and the corresponding local point cloud S3D-jMachining path point SX at the boundary of (1)jThereby obtaining the machining path point sequence { SXj}j=1->n
The processing target segmentation module 30 divides the 3D point cloud data S of the target to be processed3DWith a sequence of salient feature points { KXi}i=1->mDividing the boundary point into local point cloud sets (S)3D-j}j=1->nThe machining path point acquiring module 40 may simply and effectively acquire the corresponding machining path point sequence { SX }j}j=1->n. Especially for very complex said 3D point cloud data S3DMore beneficial to processing path point sequence { SXj}j=1->nAnd stability and robustness are greatly enhanced.
A machining guide point switching module 50 for sequentially setting the j from 1 to n, using a tool comprising switching the machining path point SX by a switching matrix BETjConversion to a machining guide point BXjSo as to process the path point sequence { SXj}j=1->nConversion into a sequence of machining guide points { BXj}j=1->n
Machining path point SXjThe position coordinate information in the coordinate system corresponding to the RGB-D composite vision sensor needs to be converted into the position coordinate information in the corresponding working coordinate system. Finally, the machining guide point sequence { BXj}j=1->nAnd sending the data to the robot to execute corresponding operation. By aligning sequences of machining guide points { BXj}j=1->nThe machining path can be obtained by using a curve interpolation algorithm, so that the machining operation is guided.
Therefore, the module can reduce the calculation amount, reduce the calculation complexity, accelerate the processing speed, reduce the calculation time, meet the requirements of real-time processing, reduce the requirements on the performance of software and hardware, save the cost, reduce the development difficulty and meet the requirements on a high-speed large-scale production mode.
Further, the processing path point obtaining module 40 is configured to set j to be from 1 to n in sequence, and apply a surface fitting algorithm to the local point cloud S3D-jFitting to curved surfaces SS3D-jThen searching and extracting the point cloud S vertical to the local point by using the local search algorithm3D-jThe plane of the bounding box and the corresponding curved surface SS3D-jThe machining path point SX at the junction ofjThereby obtaining the machining path point sequence { SXj}j=1->n
Using fitted surfaces SS3D-jCan filter out local point cloud S3D-jData redundancy of (2) to make local point cloud S3D-jThe data are homogenized, the measurement deviation caused by a measurement system is reduced, and the fluctuation of the data is eliminated. Particularly, a smooth machining path can be generated by NURBS curve fitting.
Further, the processing object identification module 20 is used for registering the two-dimensional image IRGBAnd corresponding said registration depth data IDAre merged to generate fused data IRGB-DUsing a pre-trained image segmentation modelFor the input of said fusion data IRGB-DDividing the region S corresponding to the target to be processedRGB-DAccording to said region SRGB-DFrom the registered depth data IDExtracting 3D point cloud data S of the target to be processed3D
Using said fusion data IRGB-DThe area S corresponding to the target to be processed can be effectively dividedRGB-DAnd greatly enhances the robustness and stability of the segmentation.
Further, the training samples required for the pre-training of the image segmentation model of the processing target recognition module 20 are obtained by acquiring the registered two-dimensional image I containing the target to be processed of the same kind using the RGB-D composite sensorRGBAnd corresponding said registration depth data IDAnd (4) generating.
Using the RGB-D composite sensor with local grid space placement, a large number of the registered two-dimensional images I of the object to be processed can be obtainedRGBAnd corresponding said registration depth data IDThen, the training sample can be made by labeling; and then training the image segmentation model based on the deep learning framework, and finely adjusting related parameters in the training process until the accuracy of the model reaches an expected value. By the processing steps, a large amount of training sample data can be obtained very efficiently, so that the requirements of the image segmentation model based on the deep learning framework on precision and robustness are ensured.
Further, the conversion matrix BET of the machining guide point conversion module 50 is generated by a previously calibrated Denavit-Hartenberg (D-H) parameter.
The purpose of calibrating the Denavit-Hartenberg (D-H) parameters by using a laser tracker is to improve the overall accuracy of the robot vision guidance algorithm and to ensure that the machining path point SX is fixedjConversion to a machining guide point BXjThe method is accurate, and has the characteristics of high processing speed, maturity, reliability and easy engineering realization.
It should be noted that, in this document, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising an … …" does not exclude the presence of other like elements in a process, method, article, or apparatus that comprises the element.
The above-mentioned serial numbers of the embodiments of the present invention are merely for description and do not represent the merits of the embodiments.
It will be apparent to those skilled in the art that the above-described block units or steps of the present invention may be implemented by a general purpose computing device, or alternatively, they may be implemented by program code executable by a computing device, so that they may be stored in a storage device and executed by a computing device, and in some cases, the steps shown or described may be performed in a different order than here, or they may be separately fabricated into individual integrated circuit modules, or multiple ones of them may be fabricated into a single integrated circuit module. Thus, the present invention is not limited to any specific combination of hardware and software.
Through the above description of the embodiments, those skilled in the art will clearly understand that the method of the above embodiments can be implemented by software plus a necessary general hardware platform, and certainly can also be implemented by hardware, but in many cases, the former is a better implementation manner. Based on such understanding, the technical solutions of the present invention may be embodied in the form of a software product, which is stored in a storage medium (such as ROM/RAM, magnetic disk, optical disk) and includes instructions for enabling a terminal device (such as a mobile phone, a computer, a server, an air conditioner, or a network device) to execute the method according to the embodiments of the present invention.
The above description is only a preferred embodiment of the present invention, and not intended to limit the scope of the present invention, and all modifications of equivalent structures and equivalent processes, which are made by using the contents of the present specification and the accompanying drawings, or directly or indirectly applied to other related technical fields, are included in the scope of the present invention.

Claims (8)

1. The robot vision guiding method based on RGB-D data fusion is characterized by comprising the following steps:
step 1, acquiring an RGB two-dimensional image and depth data containing a target to be processed through an RGB-D composite sensor, and registering the RGB two-dimensional image and the depth data according to preset registration parameters of the RGB-D composite sensor to generate a registered two-dimensional image I corresponding to the RGB two-dimensional imageRGBAnd registered depth data I corresponding to said depth dataD
Step 2, using the image segmentation model trained in advance to obtain the registered two-dimensional image I as inputRGBA region S corresponding to the target to be processed is divided from the centerRGBAccording to said region SRGBFrom the registered depth data IDExtracting 3D point cloud data S of the target to be processed3D
Step 3, from the 3D point cloud data S3DExtracting significant feature point sequence { KXi}i=1->mUsing the significant feature point sequence { KXi}i=1->m(ii) assigning the 3D point cloud data S for a cut-off point3DDivided into local point cloud sets { S3D-j}j=1->n,KXiIs a significant feature point, i is the corresponding significant feature point KXiThe value range of i is [1, m ]]M is the significant feature point KXiTotal number of said S3D-jIs a local point cloud, j is the local point cloud S3D-jThe value range of j is [1, n ]]N is the local point cloud S3D-jThe total number of (c);
step 4, sequentially setting the j from 1 to n, and carrying out a surface fitting algorithm on the local point cloud S3D-jFitting to curved surfaces SS3D-jAnd then by makingSearching and extracting the point cloud S perpendicular to the local point by using a local search algorithm3D-jAnd the plane of the bounding box and the corresponding curved surface SS3D-jMachining path point SX at the boundary of (1)jThereby obtaining the machining path point sequence { SXj}j=1->n
Step 5, setting the j to be from 1 to n in sequence, and using a conversion matrix BET to convert the processing path point SXjConversion to a machining guide point BXjSo as to process the path point sequence { SXj}j=1->nConversion into a sequence of machining guide points { BXj}j=1->n
2. The RGB-D data fusion-based robotic visual guidance method of claim 1,
the step 2 of registering the two-dimensional images IRGBAnd corresponding said registration depth data IDAre merged to generate fused data IRGB-DUsing said fused data I as input by a pre-trained image segmentation modelRGB-DDividing the region S corresponding to the target to be processedRGB-DAccording to said region SRGB-DFrom the registered depth data IDExtracting 3D point cloud data S of the target to be processed3D
3. The RGB-D data fusion-based robot vision guiding method as claimed in claim 1, wherein the training samples required for the pre-training of the image segmentation model of the step 2 are registered two-dimensional images I including the objects to be processed of the same kind by using the RGB-D composite sensorRGBAnd corresponding said registration depth data IDAnd (4) generating.
4. The RGB-D data fusion-based robot vision guidance method of claim 1, wherein the conversion matrix BET of the step 5 is generated by a previously calibrated Denavit-Hartenberg (D-H) parameter.
5. Robot vision guiding device based on RGB-D data fusion, characterized by includes:
the processing target data acquisition module is used for acquiring an RGB two-dimensional image and depth data containing a target to be processed through an RGB-D composite sensor, and registering the RGB two-dimensional image and the depth data according to preset registration parameters of the RGB-D composite sensor so as to generate a registered two-dimensional image I corresponding to the RGB two-dimensional imageRGBAnd registered depth data I corresponding to said depth dataD
A processing target recognition module for using a pre-trained image segmentation model to extract the registered two-dimensional image I from the image as inputRGBA region S corresponding to the target to be processed is divided from the centerRGBAccording to said region SRGBFrom the registered depth data IDExtracting 3D point cloud data S of the target to be processed3D
A processing target segmentation module for segmenting the 3D point cloud data S3DExtracting significant feature point sequence { KXi}i=1->mUsing the significant feature point sequence { KXi}i=1->m(ii) assigning the 3D point cloud data S for a cut-off point3DDivided into local point cloud sets { S3D-j}j=1->n,KXiIs a significant feature point, i is the corresponding significant feature point KXiThe value range of i is [1, m ]]M is the significant feature point KXiTotal number of said S3D-jIs a local point cloud, j is the local point cloud S3D-jThe value range of j is [1, n ]]N is the local point cloud S3D-jThe total number of (c);
a processing path point acquisition module used for setting the j from 1 to n in sequence and carrying out local point cloud S through a curved surface fitting algorithm3D-jFitting to curved surfaces SS3D-jThen searching and extracting the point cloud S vertical to the local point by using a local search algorithm3D-jAnd the plane of the bounding box and the corresponding curved surface SS3D-jThe path of the junction ofPoint SXjThereby obtaining the machining path point sequence { SXj}j=1->n
A machining guide point conversion module for setting the j from 1 to n in sequence, the machining path point SX being obtained by converting the matrix BETjConversion to a machining guide point BXjSo as to process the path point sequence { SXj}j=1->nConversion into a sequence of machining guide points { BXj}j=1->n
6. The RGB-D data fusion-based robot vision guiding device of claim 5,
the processing target identification module is used for registering the two-dimensional image IRGBAnd corresponding said registration depth data IDAre merged to generate fused data IRGB-DUsing said fused data I as input by a pre-trained image segmentation modelRGB-DDividing the region S corresponding to the target to be processedRGB-DAccording to said region SRGB-DFrom the registered depth data IDExtracting 3D point cloud data S of the target to be processed3D
7. The RGB-D data fusion-based robot vision guiding device according to claim 5, wherein training samples required for the pre-training of the image segmentation model in the processing target recognition module are obtained by acquiring a registered two-dimensional image I containing the objects to be processed of the same kind using the RGB-D composite sensorRGBAnd corresponding said registration depth data IDAnd (4) generating.
8. The RGB-D data fusion based robot vision guidance apparatus of claim 5, wherein the conversion matrix BET among the processing guidance point conversion modules is generated by a precalibrated Denavit-Hartenberg (D-H) parameter.
CN201911196215.4A 2019-11-29 2019-11-29 Robot vision guiding method and device based on RGB-D data fusion Active CN110992410B (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN201911196215.4A CN110992410B (en) 2019-11-29 2019-11-29 Robot vision guiding method and device based on RGB-D data fusion
PCT/CN2020/101335 WO2021103558A1 (en) 2019-11-29 2020-07-10 Rgb-d data fusion-based robot vision guiding method and apparatus

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201911196215.4A CN110992410B (en) 2019-11-29 2019-11-29 Robot vision guiding method and device based on RGB-D data fusion

Publications (2)

Publication Number Publication Date
CN110992410A CN110992410A (en) 2020-04-10
CN110992410B true CN110992410B (en) 2020-10-30

Family

ID=70088010

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911196215.4A Active CN110992410B (en) 2019-11-29 2019-11-29 Robot vision guiding method and device based on RGB-D data fusion

Country Status (2)

Country Link
CN (1) CN110992410B (en)
WO (1) WO2021103558A1 (en)

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110992410B (en) * 2019-11-29 2020-10-30 广东技术师范大学 Robot vision guiding method and device based on RGB-D data fusion
CN113103226A (en) * 2021-03-08 2021-07-13 同济大学 Visual guide robot system for ceramic biscuit processing and manufacturing
CN116652951B (en) * 2023-06-08 2024-04-05 广州鑫帅机电设备有限公司 Method and device for positioning robot vision in unstructured large working space

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105665970A (en) * 2016-03-01 2016-06-15 中国科学院自动化研究所 System and method for automatic generation for path points of welding robot
CN109541997A (en) * 2018-11-08 2019-03-29 东南大学 It is a kind of towards the quick, intelligent programmed method of plane/almost plane workpiece spray robot

Family Cites Families (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104063894B (en) * 2014-06-13 2017-02-22 中国科学院深圳先进技术研究院 Point cloud three-dimensional model reestablishing method and system
CN104324861B (en) * 2014-08-12 2016-06-15 清华大学 A kind of multiparameter time-varying Control During Paint Spraying by Robot method
CN104856720B (en) * 2015-05-07 2017-08-08 东北电力大学 A kind of robot assisted ultrasonic scanning system based on RGB D sensors
WO2017096299A1 (en) * 2015-12-04 2017-06-08 Autodesk, Inc. Keypoint-based point-pair-feature for scalable automatic global registration of large rgb-d scans
US10150213B1 (en) * 2016-07-27 2018-12-11 X Development Llc Guide placement by a robotic device
CN106251353A (en) * 2016-08-01 2016-12-21 上海交通大学 Weak texture workpiece and the recognition detection method and system of three-dimensional pose thereof
JP6911777B2 (en) * 2018-01-23 2021-07-28 トヨタ自動車株式会社 Motion trajectory generator
CN110014426B (en) * 2019-03-21 2020-10-02 同济大学 Method for grabbing symmetrically-shaped workpieces at high precision by using low-precision depth camera
CN110992410B (en) * 2019-11-29 2020-10-30 广东技术师范大学 Robot vision guiding method and device based on RGB-D data fusion

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105665970A (en) * 2016-03-01 2016-06-15 中国科学院自动化研究所 System and method for automatic generation for path points of welding robot
CN109541997A (en) * 2018-11-08 2019-03-29 东南大学 It is a kind of towards the quick, intelligent programmed method of plane/almost plane workpiece spray robot

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
Amodal Detection of 3D Objects:Inferring 3D Bounding Boxes from 2D Ones in RGB-Depth Images;Zhuo Deng,and etc;《2017 IEEE Conference on Computer Vision and Pattern Recognition》;20171109;第398-406页 *
基于RGB-D信息的移动机器人SLAM和路径规划方法研究与实现;李元等;《智能***学报》;20180630;第13卷(第3期);第445-451页 *

Also Published As

Publication number Publication date
CN110992410A (en) 2020-04-10
WO2021103558A1 (en) 2021-06-03

Similar Documents

Publication Publication Date Title
CN111216124B (en) Robot vision guiding method and device based on integration of global vision and local vision
EP3629057B1 (en) Method and apparatus for calibrating relative pose and medium
CN110426051B (en) Lane line drawing method and device and storage medium
CN110992410B (en) Robot vision guiding method and device based on RGB-D data fusion
CN109658504B (en) Map data annotation method, device, equipment and storage medium
CN111179324B (en) Object six-degree-of-freedom pose estimation method based on color and depth information fusion
CN111089569B (en) Large box body measuring method based on monocular vision
CN109993793B (en) Visual positioning method and device
EP2026279B1 (en) Method and system for aligning three-dimensional surfaces
CN110176078B (en) Method and device for labeling training set data
CN109927036A (en) A kind of method and system of 3D vision guidance manipulator crawl
CN105729468A (en) Enhanced robot workbench based on multiple depth cameras
CN111476841B (en) Point cloud and image-based identification and positioning method and system
CN110930442B (en) Method and device for determining positions of key points in robot hand-eye calibration based on calibration block
CN107545247B (en) Stereo cognition method based on binocular recognition
Boughorbel et al. Laser ranging and video imaging for bin picking
KR20230003803A (en) Automatic calibration through vector matching of the LiDAR coordinate system and the camera coordinate system
CN109635640B (en) Traffic light identification method, device and equipment based on point cloud and storage medium
CN114863129A (en) Instrument numerical analysis method, device, equipment and storage medium
CN113724329A (en) Object attitude estimation method, system and medium fusing plane and stereo information
CN116188540A (en) Target identification and pose estimation method based on point cloud information
CN112232272B (en) Pedestrian recognition method by fusing laser and visual image sensor
CN112785647A (en) Three-eye stereo image detection method and system
CN112562008A (en) Target point matching method in local binocular vision measurement
CN111354031A (en) 3D vision guidance system based on deep learning

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
TR01 Transfer of patent right
TR01 Transfer of patent right

Effective date of registration: 20210526

Address after: 528225 room 417, block a, Nanhai industrial think tank City Phase I, Taoyuan Road, software park, Shishan town, Nanhai District, Foshan City, Guangdong Province (residence declaration)

Patentee after: FOSHAN DILE VISION TECHNOLOGY Co.,Ltd.

Address before: No. 293, Zhongshan Avenue West, Tianhe District, Guangzhou, Guangdong 510000

Patentee before: GUANGDONG POLYTECHNIC NORMAL University