CN112396655A - Point cloud data-based ship target 6D pose estimation method - Google Patents

Point cloud data-based ship target 6D pose estimation method Download PDF

Info

Publication number
CN112396655A
CN112396655A CN202011290504.3A CN202011290504A CN112396655A CN 112396655 A CN112396655 A CN 112396655A CN 202011290504 A CN202011290504 A CN 202011290504A CN 112396655 A CN112396655 A CN 112396655A
Authority
CN
China
Prior art keywords
point
target
dimensional
points
point cloud
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202011290504.3A
Other languages
Chinese (zh)
Other versions
CN112396655B (en
Inventor
苏丽
宋浩
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Harbin Engineering University
Original Assignee
Harbin Engineering University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Harbin Engineering University filed Critical Harbin Engineering University
Priority to CN202011290504.3A priority Critical patent/CN112396655B/en
Publication of CN112396655A publication Critical patent/CN112396655A/en
Application granted granted Critical
Publication of CN112396655B publication Critical patent/CN112396655B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • G06T7/73Determining position or orientation of objects or cameras using feature-based methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Engineering & Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Biomedical Technology (AREA)
  • General Health & Medical Sciences (AREA)
  • Evolutionary Biology (AREA)
  • Health & Medical Sciences (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Image Analysis (AREA)
  • Image Processing (AREA)

Abstract

The invention discloses a point cloud data-based ship target 6D pose estimation method, which comprises the following steps of: step 1: acquiring a ship point cloud data set of a marine scene, wherein a data set label comprises a target category, a three-dimensional coordinate of a target, a three-dimensional size of the target and a three-dimensional pose of the target; step 2: constructing a neural network, and extracting point-by-point cloud features by adopting PointNet + + to obtain point-by-point high-dimensional features; and step 3: generating a 3D boundary frame proposal by a bottom-up scheme, generating a real segmentation mask based on the 3D boundary frame, segmenting foreground points and simultaneously generating a boundary frame proposal with angle information from segmentation points for the input of RCNN; and 4, step 4: and (4) carrying out proposal optimization based on the proposal obtained in the step (3) and the foreground segmentation characteristics and spatial characteristics so as to output the final classification and 3D frame and posture angle. The method realizes the pose estimation effect of the three-dimensional target by adopting an end-to-end learning mode, and improves the instantaneity of pose estimation.

Description

Point cloud data-based ship target 6D pose estimation method
Technical Field
The invention relates to a point cloud data-based ship target 6D pose estimation method, and relates to the field of point clouds, ship target pose estimation of offshore scenes, deep learning and neural networks.
Background
Pose estimation plays a very important role in the field of computer vision. The method has great application in the aspects of estimating the pose of the robot by using the vision sensor for control, robot navigation, augmented reality and the like. The pose estimation problem of an object is to determine the spatial position of the object in 3D space, and the angle of rotation of the object about coordinate axes, Yaw (Yaw) about the Z-axis, Pitch (Pitch) about the Y-axis, and Roll (Roll) about the X-axis. In recent years, methods for 6D pose estimation can be divided into four broad categories, namely a model corresponding point-based method, a template matching-based method, a voting-based method, and a regression-based method. The method based on the model corresponding points mainly aims at objects with rich textures, and the method based on the template matching mainly aims at images with weak textures or no textures. The specificity of the marine scene for a ship target, such as the effect of marine surface lighting, and the variability of weather affects image quality. The traditional three-dimensional attitude estimation algorithm can only detect a single target generally and is time-consuming.
Disclosure of Invention
In view of the prior art, the technical problem to be solved by the invention is to provide a ship target 6D pose estimation method based on point cloud data, by learning the relation between the features of point cloud information and the 6D pose, and by improving on the basis of target detection based on deep learning, the 6D pose of a target is further regressed.
In order to solve the technical problem, the invention provides a ship target 6D pose estimation method based on point cloud data, which comprises the following steps:
step 1: acquiring a ship point cloud data set of a marine scene, wherein a data set label comprises a target category, a three-dimensional coordinate of a target, a three-dimensional size of the target and a three-dimensional pose of the target;
step 2: constructing a neural network, and extracting point-by-point cloud features by adopting PointNet + + to obtain point-by-point high-dimensional features;
and step 3: generating a 3D boundary frame proposal by a bottom-up scheme, generating a real segmentation mask based on the 3D boundary frame, segmenting foreground points and simultaneously generating a boundary frame proposal with angle information from segmentation points for the input of RCNN;
and 4, step 4: and (4) carrying out proposal optimization based on the proposal obtained in the step (3) and the foreground segmentation characteristics and spatial characteristics so as to output the final classification and 3D frame and posture angle.
The invention also includes:
1. in the step 2, pointwise point cloud feature extraction is carried out by adopting PointNet + +, and the obtained point-wise high-dimensional features are specifically as follows: the feature extraction of the point set comprises three parts, including Sampling layer, Grouping layer and Pointnet layer, wherein the Sampling algorithm of the Sampling layer uses an iteration farthest point Sampling method, a series of points are selected from the input point cloud, the center of a local area is defined, then a local neighborhood is constructed, points are searched within a given distance, then feature extraction is carried out by using a full connection layer, finally, pooling operation is carried out to obtain high-level features, the number of the original point sets is sampled from the point sets, and the high-dimensional features of the point sets one by one are obtained.
2. Generating a real segmentation mask based on the 3D boundary frame in the step 3, segmenting foreground points and generating a boundary frame proposal with angle information from the segmentation points specifically: classifying each point for two times, scoring foreground and background, normalizing to 0-1 by a sigmod function, considering that points with scores higher than a threshold are foreground points, segmenting foreground points and generating boundary frame proposals with angle information from segmented points, taking the foreground points as centers, wherein the total number is N, generating initial proposals on each point by using regression values and set average sizes, the size is (batch _ size N,9), 9-dimensional vectors are [ x, y, z, h, w, l, rx, ry, rz ], namely the center position, length, width and height of a target and rotation angles in the xyz direction respectively, sorting according to the classified scores, finding out 512 boundary frames in front of each batch by using non-maximum values, returning the boundary frames, the angle information and confidence scores, adopting a cross loss function as a loss function of a binary network in the segmentation stage, and predicting the initial proposals, where x, z and yaw angle ry are calculated based on the bin's loss function regression, the dimensional information (h, w, L) and rotation angles (rx, rz) are calculated using smoothed L1 losses.
The invention has the beneficial effects that: the ship sensing link is one of paths for intelligent ship development, and the identification of the position and the posture of a ship target is an indispensable link for ship sensing or ship tracking identification. The sensing precision can be enhanced by adding the attitude estimation of the ship target on the basis of target detection. The traditional three-dimensional attitude estimation algorithm can only detect a single target generally and consumes a lot of time, so the algorithm adopts a neural network based on deep learning and adopts a learning mode to train a priori data set, thereby greatly improving the real-time performance of the algorithm and being capable of estimating the attitude of a complex scene. The method realizes the pose estimation effect of the three-dimensional target by adopting an end-to-end learning mode, and improves the instantaneity of pose estimation.
Drawings
FIG. 1 is a schematic diagram of a generator in a network;
FIG. 2 is a schematic diagram of visualization of an algorithm pose estimation result.
Detailed Description
The following further describes the embodiments of the present invention with reference to the drawings.
The purpose of the invention is realized as follows:
1. acquiring a data set
A point cloud data set is obtained through manual production, and the ship point cloud data of the offshore scene required by the data set can be ship targets under the conditions of single ship, multiple ships, shielding and the like. Meanwhile, the point cloud data of each scene should contain a corresponding ship position and posture label, and the structure of the label comprises three angles, namely a yaw angle, a roll angle and a pitch angle corresponding to the ship target, of the category, the three-dimensional coordinate of the target, the three-dimensional size of the target and the three-dimensional position and posture of the target.
2. Constructing neural networks
The algorithm is based on a 3D target detection network PointRCNN, and the target of the 3D target detection algorithm is only output to a 3D frame, and network angle branches are added on the basis to estimate the 6D pose of the target. The network mainly comprises two parts: point cloud feature extraction based on PointNet + + and improvement based on 3D target detection network PointRCNN. The PointNet + + network serves as a backbone network and is responsible for feature extraction from input point cloud data.
The improved 6D pose estimation network based on PointRCNN comprises two stages, wherein the first stage aims to generate a 3D boundary frame proposal in a bottom-up scheme, obtain a three-dimensional pose angle of a target at the same time on the basis of the 3D boundary frame proposal, generate a real segmentation mask based on the boundary frame, segment foreground points and generate a small number of boundary frame proposals from segmentation points at the same time. Such a strategy avoids the use of a large number of anchor boxes throughout the three-dimensional space. And in the second stage, carrying out standard frame optimization with angles. After the bounding box and angle proposals are generated, the point representation from the first stage is pooled using a point cloud area pooling operation. Unlike the method of directly estimating global frame coordinates, the pooled points are converted to canonical coordinates and combined with the point features and the first stage segmentation mask to complete coordinate optimization. This strategy takes full advantage of the information provided by the segmentation and proposal sub-network of the first stage.
3. 6D pose estimation network main steps
(1) Generation of 3D bezel and angle proposals
And the main network adopts PointNet + + to extract point-by-point cloud characteristics. Inputting point cloud data of N x 3, and for each point cloud input of N x 3, extracting the characteristics of a point set, wherein the characteristics of the point set comprise three parts, namely a Sampling layer, a Grouping layer and a Pointnet layer. The sampling algorithm of the sampling layer uses iterative farthest point sampling method iterative Fast Point Sampling (FPS). A series of points is selected in the input point cloud, thereby defining the center of the local region. And then constructing a local neighborhood, searching points within a given distance, extracting features by using a full-link layer, performing pooling operation to obtain high-grade features, and upsampling the point set to the number of the original point set to obtain point-by-point high-dimensional features in the input point set.
Objects in a 3D scene are naturally separated and do not overlap with each other. The segmentation masks for all three-dimensional objects can be obtained directly by 3D bounding box annotation, i.e. the 3D points within the 3D box are considered foreground points. And (3) carrying out point-by-point characteristic after the processing of the backbone network, wherein the foreground prediction branch is used for carrying out classified prediction on each point, and the real segmentation mask of the point cloud is determined by the three-dimensional frame.
And predicting branches of a 3D frame and a three-dimensional angle while predicting the branches of the foreground, and adopting a full connection layer. Due to the fact that the scale change of the surrounding frame coordinate and the yaw angle of the ship target is large, the change scale of the roll angle and the pitch angle is usually small, and the thought regression calculation based on Bin is adopted for the horizontal direction coordinate x, z and the yaw angle of the surrounding frame. Specifically, the surrounding area of each foreground point is divided into a series of discrete bins along the X-axis and Z-axis, with Bin-based cross-entropy loss classification and residual regression in both directions rather than direct smooth L1 loss regression. In most offshore scenes, the horizontal direction scale change of a ship target is large, but the central coordinate scale change in the vertical direction is usually small, the pitch angle and the roll angle of the attitude angle change in a very small scale range, and the roll angle, the pitch angle and the roll angle of the ship can be obtained by directly performing regression by using the loss of smooth L1 through the algorithm.
(2) Pooling point cloud area
After the three-dimensional bounding box scheme is obtained, the position and the direction of the box are optimized according to the previously generated frame and the angle branch. And pooling each point and the characteristics thereof according to the position of each three-dimensional frame. The points and their features within the bounding box after slight enlargement will be retained. The segmentation mask is then used to distinguish between foreground and background points within a slightly enlarged box. Proposals without interior points will be eliminated.
(3) Canonical 3D bounding box and angle regression proposal optimization
The pooled point sets and their associated features are fed into the second stage subnetwork to optimize the position, angle and confidence of the foreground object of the border. Combining the characteristics, obtaining high-dimensional characteristics through a Sampling layer provided by 3 PointNet + +, and then predicting and outputting coordinate size and attitude angle information of the target by using classification and regression branches.
The loss function is as follows:
Figure BDA0002783698640000041
Figure BDA0002783698640000042
Figure BDA0002783698640000043
in which the loss function is based on bin
Figure BDA0002783698640000044
Figure BDA0002783698640000045
Residual loss function
Figure BDA0002783698640000046
Comprises the following steps:
Figure BDA0002783698640000047
wherein
Figure BDA0002783698640000048
Figure BDA0002783698640000049
B is the proposal set of the first stage, BposFor the positive proposal of regression, probiFor confidence, bleiAre corresponding labels, FclsA confidence for the prediction is calculated for the cross entropy loss function.
Pose estimation (position estimation) plays a very important role in the field of computer vision. The method has great application in the aspects of estimating the pose of the robot by using the vision sensor for control, robot navigation, augmented reality and the like. The pose estimation problem of an object is to determine the spatial position of the object in 3D space, and the angle of rotation of the object about coordinate axes, Yaw (Yaw) about the Z-axis, Pitch (Pitch) about the Y-axis, and Roll (Roll) about the X-axis.
With reference to fig. 1 and 2, the steps of the present invention are as follows:
step 1, preparing a position and pose data set of a ship scene, wherein a data set label comprises a target type, a three-dimensional coordinate of a target, a three-dimensional size of the target and a three-dimensional position and pose of the target, namely three angles of a yaw angle, a roll angle and a pitch angle corresponding to the ship target.
And 2, point cloud feature extraction based on PointNet + +, wherein the feature extraction part mainly comprises three parts of feature extraction of a point Set consisting of Set Abstraction sub-networks, input point clouds are arranged, a farthest point is searched based on a sampling algorithm iterative farthest point sampling method (FPS), and the center of a local area is defined after a series of farthest points are selected from the input point clouds. And then constructing a local neighborhood, searching points within a given distance, extracting features by using a full-link layer, performing pooling operation to obtain high-grade features, up-sampling the point set to the number of the original point sets to obtain point-by-point high-dimensional features in the input point set, and passing the input point cloud data through a backbone network to obtain the point-by-point high-dimensional features.
And 3, an RPN stage, namely generating a 3D boundary frame proposal by a bottom-up scheme, generating a real segmentation mask based on the 3D boundary frame, namely performing two classifications on each point, scoring the foreground and the background, normalizing to 0-1 by a sigmod function, considering that the point with the score higher than a threshold value is a foreground point, segmenting foreground points, generating a small amount of boundary frame proposals with angle information from the segmented points, taking the foreground points (the total number of the foreground points is N) as the center, and generating an initial proposal with the size of (batch _ size N,9) and 9-dimensional vectors of [ x, y, z, h, w, l, rx, ry, rz ] on each point by using a regression value and a set average size, namely the center position, the length, the width and the rotation angle of the target in the xyz direction. And then sorting according to the classified scores, finding out the first 512 bounding boxes of each batch by using a non-maximum value, returning bounding boxes, angle information and a confidence score for inputting in an RCNN stage, and predicting an initial proposal in a segmentation stage by using a cross entropy loss function as a loss function of a binary network, wherein x, z and a yaw angle ry are calculated based on the regression of the loss function of bin, and size information (h, w, L) and rotation angles (rx, rz) are calculated by using a smooth L1 loss.
And 4, an RCNN stage, aiming at carrying out proposal optimization based on a small number of proposals obtained in the RPN stage and the foreground segmentation characteristics and spatial characteristics so as to output final classification and 3D frame and attitude angles. According to the proposal of the RPN stage, calculating the IOU between the ROI and the truth value, wherein the IOU is more than 0.55, and the truth value is divided into the proposal of prediction for fine adjustment. And connecting and recombining the segmented mask with the point cloud coordinates and the high-dimensional features, obtaining high-level features of the combined feature vectors from a Set Abstraction sub-network of PointNet + +, and predicting by a classification layer and a regression layer network, wherein the classification layer is cross entropy loss, and the regression layer is bin-based loss and smooth L1 loss.
Example (b):
1. and making a data set containing a target type, a three-dimensional coordinate of the target, a three-dimensional size of the target and three postures of the target, namely three angles of a yaw angle, a roll angle and a pitch angle corresponding to the ship target, for training.
2. And constructing a neural network.
And the main network adopts PointNet + + to extract point-by-point cloud characteristics. The feature extraction of the point set comprises three parts, namely a Sampling layer, a Grouping layer and a Pointnet layer. The sampling algorithm of the sampling layer uses iterative farthest point sampling method iterative Fast Point Sampling (FPS). A series of points is selected in the input point cloud, thereby defining the center of the local region. And then constructing a local neighborhood, searching points within a given distance, extracting features by using a full-connection layer, performing pooling operation to obtain high-level features, up-sampling the point set to the number of the original point set to obtain high-dimensional features point by point in the input point set, and inputting the high-dimensional features with the conversion from n x 3 to n x 128.
3. 6D pose estimation network
The overall network structure is shown in fig. 1. The point-by-point characteristics processed by the backbone network, the high-dimensional characteristics pass through three subsequent structures, namely a foreground prediction branch, a frame regression branch and an angle regression branch. And (4) carrying out classified prediction on each point by using the foreground prediction branch, wherein the real segmentation mask of the point cloud is determined by the three-dimensional frame. And predicting branches of a 3D frame and a three-dimensional angle while predicting the branches of the foreground, and adopting a full connection layer. Due to the fact that the scale change of the surrounding frame coordinate and the yaw angle of the ship target is large, the change scale of the roll angle and the pitch angle is usually small, and the thought regression calculation based on Bin is adopted for the horizontal direction coordinate x, z and the yaw angle of the surrounding frame. Specifically, the surrounding area of each foreground point is divided into a series of discrete bins along the X-axis and Z-axis, with Bin-based cross-entropy loss classification and residual regression in both directions rather than direct smooth L1 loss regression. In most offshore scenes, the horizontal direction scale change of a ship target is large, but the central coordinate scale change in the vertical direction is usually small, the pitch angle and the roll angle of the attitude angle change in a very small scale range, and the algorithm can obtain an accurate value by directly performing regression by using the loss of smooth L1.
After the three-dimensional bounding box scheme is obtained, the position and the direction of the box are optimized according to the previously generated frame and the angle branch. And pooling each point and the characteristics thereof according to the position of each three-dimensional frame. The points and their features within the bounding box after slight enlargement will be retained. The segmentation mask is then used to distinguish between foreground and background points within a slightly enlarged box. Proposals without interior points will be eliminated.
The pooled point sets and their associated features are fed into the second stage subnetwork to optimize the position, angle and confidence of the foreground object of the border. Combining the characteristics, obtaining high-dimensional characteristics through a Sampling layer proposed by 3 PointNet + +, then predicting the coordinate size and the posture angle information of an output target by using classification and regression branches, and displaying the output information as a visualization result shown in FIG. 2.
The loss function is as follows:
Figure BDA0002783698640000061
Figure BDA0002783698640000062
Figure BDA0002783698640000063
in which the loss function is based on bin
Figure BDA0002783698640000064
Figure BDA0002783698640000071
Residual loss function
Figure BDA0002783698640000072
Comprises the following steps:
Figure BDA0002783698640000073
wherein
Figure BDA0002783698640000074
Figure BDA0002783698640000075
B is the proposal set of the first stage, BposFor the positive proposal of regression, probiFor confidence, bleiAre corresponding labels, FclsA confidence for the prediction is calculated for the cross entropy loss function.

Claims (3)

1. A ship target 6D pose estimation method based on point cloud data is characterized by comprising the following steps:
step 1: acquiring a ship point cloud data set of a marine scene, wherein a data set label comprises a target category, a three-dimensional coordinate of a target, a three-dimensional size of the target and a three-dimensional pose of the target;
step 2: constructing a neural network, and extracting point-by-point cloud features by adopting PointNet + + to obtain point-by-point high-dimensional features;
and step 3: generating a 3D boundary frame proposal by a bottom-up scheme, generating a real segmentation mask based on the 3D boundary frame, segmenting foreground points and simultaneously generating a boundary frame proposal with angle information from segmentation points for the input of RCNN;
and 4, step 4: and (4) carrying out proposal optimization based on the proposal obtained in the step (3) and the foreground segmentation characteristics and spatial characteristics so as to output the final classification and 3D frame and posture angle.
2. The point cloud data-based ship target 6D pose estimation method according to claim 1, characterized in that: step 2, point-by-point cloud feature extraction is carried out by adopting PointNet + +, and the obtained point-by-point high-dimensional features are specifically as follows: the feature extraction of the point set comprises three parts, including Sampling layer, Grouping layer and Pointnet layer, wherein the Sampling algorithm of the Sampling layer uses an iteration farthest point Sampling method, a series of points are selected from the input point cloud, the center of a local area is defined, then a local neighborhood is constructed, points are searched within a given distance, then feature extraction is carried out by using a full connection layer, finally, pooling operation is carried out to obtain high-level features, the number of the original point sets is sampled from the point sets, and the high-dimensional features of the point sets one by one are obtained.
3. The point cloud data-based ship target 6D pose estimation method according to claim 1 or 2, wherein: step 3, generating a real segmentation mask based on the 3D boundary frame, segmenting foreground points and generating a boundary frame proposal with angle information from the segmented points specifically comprises the following steps: classifying each point for two times, scoring foreground and background, normalizing to 0-1 by a sigmod function, considering that points with scores higher than a threshold are foreground points, segmenting foreground points and generating boundary frame proposals with angle information from segmented points, taking the foreground points as centers, wherein the total number is N, generating initial proposals on each point by using regression values and set average sizes, the size is (batch _ size N,9), 9-dimensional vectors are [ x, y, z, h, w, l, rx, ry, rz ], namely the center position, length, width and height of a target and rotation angles in the xyz direction respectively, sorting according to the classified scores, finding out 512 boundary frames in front of each batch by using non-maximum values, returning the boundary frames, the angle information and confidence scores, adopting a cross loss function as a loss function of a binary network in the segmentation stage, and predicting the initial proposals, where x, z and yaw angle ry are calculated based on the bin's loss function regression, the dimensional information (h, w, L) and rotation angles (rx, rz) are calculated using smoothed L1 losses.
CN202011290504.3A 2020-11-18 2020-11-18 Point cloud data-based ship target 6D pose estimation method Active CN112396655B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011290504.3A CN112396655B (en) 2020-11-18 2020-11-18 Point cloud data-based ship target 6D pose estimation method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011290504.3A CN112396655B (en) 2020-11-18 2020-11-18 Point cloud data-based ship target 6D pose estimation method

Publications (2)

Publication Number Publication Date
CN112396655A true CN112396655A (en) 2021-02-23
CN112396655B CN112396655B (en) 2023-01-03

Family

ID=74606473

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011290504.3A Active CN112396655B (en) 2020-11-18 2020-11-18 Point cloud data-based ship target 6D pose estimation method

Country Status (1)

Country Link
CN (1) CN112396655B (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113298163A (en) * 2021-05-31 2021-08-24 国网湖北省电力有限公司黄石供电公司 Target identification monitoring method based on LiDAR point cloud data
CN114972968A (en) * 2022-05-19 2022-08-30 长春市大众物流装配有限责任公司 Tray identification and pose estimation method based on multiple neural networks
WO2022188882A1 (en) * 2021-03-11 2022-09-15 先临三维科技股份有限公司 3-dimension instance segmentation method and apparatus, device and computer-readable storage medium

Citations (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107644121A (en) * 2017-08-18 2018-01-30 昆明理工大学 The reverse three-dimensionalreconstruction and body modeling method of a kind of ground surface material skeleton structure
CN109064514A (en) * 2018-07-03 2018-12-21 北京航空航天大学 A kind of six-freedom degree pose algorithm for estimating returned based on subpoint coordinate
CN109086683A (en) * 2018-07-11 2018-12-25 清华大学 A kind of manpower posture homing method and system based on cloud semantically enhancement
CN109801337A (en) * 2019-01-21 2019-05-24 同济大学 A kind of 6D position and orientation estimation method of Case-based Reasoning segmentation network and iteration optimization
US20190304134A1 (en) * 2018-03-27 2019-10-03 J. William Mauchly Multiview Estimation of 6D Pose
CN110473284A (en) * 2019-07-29 2019-11-19 电子科技大学 A kind of moving object method for reconstructing three-dimensional model based on deep learning
CN110533721A (en) * 2019-08-27 2019-12-03 杭州师范大学 A kind of indoor objects object 6D Attitude estimation method based on enhancing self-encoding encoder
CN110930452A (en) * 2019-10-23 2020-03-27 同济大学 Object pose estimation method based on self-supervision learning and template matching
CN110930454A (en) * 2019-11-01 2020-03-27 北京航空航天大学 Six-degree-of-freedom pose estimation algorithm based on boundary box outer key point positioning
CN111080693A (en) * 2019-11-22 2020-04-28 天津大学 Robot autonomous classification grabbing method based on YOLOv3
CN111126269A (en) * 2019-12-24 2020-05-08 京东数字科技控股有限公司 Three-dimensional target detection method, device and storage medium
CN111179324A (en) * 2019-12-30 2020-05-19 同济大学 Object six-degree-of-freedom pose estimation method based on color and depth information fusion
US20200160033A1 (en) * 2018-11-15 2020-05-21 Toyota Research Institute, Inc. System and method for lifting 3d representations from monocular images
CN111259934A (en) * 2020-01-09 2020-06-09 清华大学深圳国际研究生院 Stacked object 6D pose estimation method and device based on deep learning
CN111275758A (en) * 2020-01-15 2020-06-12 深圳市微埃智能科技有限公司 Hybrid 3D visual positioning method and device, computer equipment and storage medium
CN111368733A (en) * 2020-03-04 2020-07-03 电子科技大学 Three-dimensional hand posture estimation method based on label distribution learning, storage medium and terminal
US20200311956A1 (en) * 2019-03-25 2020-10-01 Dishcraft Robotics, Inc. Automated Manipulation Of Transparent Vessels
CN111862201A (en) * 2020-07-17 2020-10-30 北京航空航天大学 Deep learning-based spatial non-cooperative target relative pose estimation method
CN111915677A (en) * 2020-07-08 2020-11-10 哈尔滨工程大学 Ship pose estimation method based on three-dimensional point cloud characteristics

Patent Citations (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107644121A (en) * 2017-08-18 2018-01-30 昆明理工大学 The reverse three-dimensionalreconstruction and body modeling method of a kind of ground surface material skeleton structure
US20190304134A1 (en) * 2018-03-27 2019-10-03 J. William Mauchly Multiview Estimation of 6D Pose
CN109064514A (en) * 2018-07-03 2018-12-21 北京航空航天大学 A kind of six-freedom degree pose algorithm for estimating returned based on subpoint coordinate
CN109086683A (en) * 2018-07-11 2018-12-25 清华大学 A kind of manpower posture homing method and system based on cloud semantically enhancement
US20200160033A1 (en) * 2018-11-15 2020-05-21 Toyota Research Institute, Inc. System and method for lifting 3d representations from monocular images
CN109801337A (en) * 2019-01-21 2019-05-24 同济大学 A kind of 6D position and orientation estimation method of Case-based Reasoning segmentation network and iteration optimization
US20200311956A1 (en) * 2019-03-25 2020-10-01 Dishcraft Robotics, Inc. Automated Manipulation Of Transparent Vessels
CN110473284A (en) * 2019-07-29 2019-11-19 电子科技大学 A kind of moving object method for reconstructing three-dimensional model based on deep learning
CN110533721A (en) * 2019-08-27 2019-12-03 杭州师范大学 A kind of indoor objects object 6D Attitude estimation method based on enhancing self-encoding encoder
CN110930452A (en) * 2019-10-23 2020-03-27 同济大学 Object pose estimation method based on self-supervision learning and template matching
CN110930454A (en) * 2019-11-01 2020-03-27 北京航空航天大学 Six-degree-of-freedom pose estimation algorithm based on boundary box outer key point positioning
CN111080693A (en) * 2019-11-22 2020-04-28 天津大学 Robot autonomous classification grabbing method based on YOLOv3
CN111126269A (en) * 2019-12-24 2020-05-08 京东数字科技控股有限公司 Three-dimensional target detection method, device and storage medium
CN111179324A (en) * 2019-12-30 2020-05-19 同济大学 Object six-degree-of-freedom pose estimation method based on color and depth information fusion
CN111259934A (en) * 2020-01-09 2020-06-09 清华大学深圳国际研究生院 Stacked object 6D pose estimation method and device based on deep learning
CN111275758A (en) * 2020-01-15 2020-06-12 深圳市微埃智能科技有限公司 Hybrid 3D visual positioning method and device, computer equipment and storage medium
CN111368733A (en) * 2020-03-04 2020-07-03 电子科技大学 Three-dimensional hand posture estimation method based on label distribution learning, storage medium and terminal
CN111915677A (en) * 2020-07-08 2020-11-10 哈尔滨工程大学 Ship pose estimation method based on three-dimensional point cloud characteristics
CN111862201A (en) * 2020-07-17 2020-10-30 北京航空航天大学 Deep learning-based spatial non-cooperative target relative pose estimation method

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
W. ZHANG 等: ""Pose Estimation by Key Points Registration in Point Cloud"", 《2019 3RD INTERNATIONAL SYMPOSIUM ON AUTONOMOUS SYSTEMS (ISAS)》 *
李运川 等: ""一种用于点云配准的改进迭代最近点算法 "", 《软件导刊》 *

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2022188882A1 (en) * 2021-03-11 2022-09-15 先临三维科技股份有限公司 3-dimension instance segmentation method and apparatus, device and computer-readable storage medium
CN113298163A (en) * 2021-05-31 2021-08-24 国网湖北省电力有限公司黄石供电公司 Target identification monitoring method based on LiDAR point cloud data
CN114972968A (en) * 2022-05-19 2022-08-30 长春市大众物流装配有限责任公司 Tray identification and pose estimation method based on multiple neural networks

Also Published As

Publication number Publication date
CN112396655B (en) 2023-01-03

Similar Documents

Publication Publication Date Title
CN111798475B (en) Indoor environment 3D semantic map construction method based on point cloud deep learning
CN108648233B (en) Target identification and capture positioning method based on deep learning
CN111339903B (en) Multi-person human body posture estimation method
CN106682598B (en) Multi-pose face feature point detection method based on cascade regression
CN112396655B (en) Point cloud data-based ship target 6D pose estimation method
CN110852182B (en) Depth video human body behavior recognition method based on three-dimensional space time sequence modeling
CN111899172A (en) Vehicle target detection method oriented to remote sensing application scene
CN112288857A (en) Robot semantic map object recognition method based on deep learning
CN110688965B (en) IPT simulation training gesture recognition method based on binocular vision
CN111161317A (en) Single-target tracking method based on multiple networks
CN107527054B (en) Automatic foreground extraction method based on multi-view fusion
CN110334584B (en) Gesture recognition method based on regional full convolution network
CN112784736A (en) Multi-mode feature fusion character interaction behavior recognition method
CN113408584B (en) RGB-D multi-modal feature fusion 3D target detection method
CN110751097B (en) Semi-supervised three-dimensional point cloud gesture key point detection method
CN110533720A (en) Semantic SLAM system and method based on joint constraint
CN110533716B (en) Semantic SLAM system and method based on 3D constraint
CN113744311A (en) Twin neural network moving target tracking method based on full-connection attention module
Wang et al. An overview of 3d object detection
CN116052222A (en) Cattle face recognition method for naturally collecting cattle face image
CN113705579A (en) Automatic image annotation method driven by visual saliency
Sun et al. IRDCLNet: Instance segmentation of ship images based on interference reduction and dynamic contour learning in foggy scenes
CN108830222A (en) A kind of micro- expression recognition method based on informedness and representative Active Learning
Liu et al. PolishNet-2d and PolishNet-3d: Deep learning-based workpiece recognition
CN114299339A (en) Three-dimensional point cloud model classification method and system based on regional correlation modeling

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant