CN110400349B - Robot navigation tracking recovery method in small scene based on random forest - Google Patents

Robot navigation tracking recovery method in small scene based on random forest Download PDF

Info

Publication number
CN110400349B
CN110400349B CN201910593421.2A CN201910593421A CN110400349B CN 110400349 B CN110400349 B CN 110400349B CN 201910593421 A CN201910593421 A CN 201910593421A CN 110400349 B CN110400349 B CN 110400349B
Authority
CN
China
Prior art keywords
training
random
layer
images
nodes
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related
Application number
CN201910593421.2A
Other languages
Chinese (zh)
Other versions
CN110400349A (en
Inventor
何建军
汪泽睿
文青勇
齐文琴
裴雨听
郝思宇
何广平
王丽萍
王艺蓉
廖敏
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Chengdu Univeristy of Technology
Original Assignee
Chengdu Univeristy of Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Chengdu Univeristy of Technology filed Critical Chengdu Univeristy of Technology
Priority to CN201910593421.2A priority Critical patent/CN110400349B/en
Publication of CN110400349A publication Critical patent/CN110400349A/en
Application granted granted Critical
Publication of CN110400349B publication Critical patent/CN110400349B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/243Classification techniques relating to the number of classes
    • G06F18/24323Tree-organised classifiers
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • G06T7/73Determining position or orientation of objects or cameras using feature-based methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • General Engineering & Computer Science (AREA)
  • Image Processing (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a robot navigation tracking recovery method under a small scene based on random forests, which comprises the following steps of (1) selecting a scene to be tracked, and shooting the scene for multiple times by using an RGB-D camera; (2) obtaining a transformation matrix of each group of images by using a three-dimensional reconstruction algorithm to obtain M transformation matrices; (3) obtaining a training label and training characteristics; (4) training a random forest model T by using the training labels and the training characteristics; (5) if the robot fails to track, shooting the scene for 1 time by adopting an RGB-D camera; (6) randomly sampling a plurality of pixel points from an RGB image of the group of images, calculating random characteristics corresponding to each pixel point, sending all the random characteristics into T for training, and outputting 3D world coordinates corresponding to all the random characteristics; and finally, determining the self pose according to the obtained transformation matrix, and completing navigation tracking recovery.

Description

Robot navigation tracking recovery method in small scene based on random forest
Technical Field
The invention relates to a robot tracking recovery method, in particular to a robot navigation tracking recovery method in a small scene based on a random forest.
Background
With the development of the robot technology, the automatic path finding and mapping functions of the robot play an important role in the robot technology, and in the process, the robot needs to know the position of the robot relative to the real world and the posture of the robot, so that the tracking problem of the robot is always a hotspot problem in robotics.
With the development of vslam (visual Simultaneous Localization and mapping) technology, it becomes very important to restore robot tracking after the robot tracking fails, and for this reason, many tracking restoration methods have been developed. The key point of tracking recovery is to calculate the correct 6D pose of the robot after the tracking fails, and after the correct 6D position is obtained, the robot can know the correct pose of the robot relative to the real world, so that the tracking is recovered.
The method based on local features proposed by S.Se, D.G.Lowe et al has the idea that images with known poses and image feature points with unknown poses are extracted firstly, descriptors are calculated, and pose calculation and optimization are performed by utilizing a P3P algorithm and a RANSAC algorithm.
The keyframe-based methods proposed by g.klein and d.murray infer an assumed pose by querying the similarity of the known image and the keyframe, which is fast, but when the similarity of the unknown image and the known image is very different, the assumed pose may be much different from the true pose.
PoseNet proposed by Alex Kendall, Matthew Grimes and Robert Cipola of Cambridge university utilizes deep learning and utilizes an RGB image to directly predict the image pose, but the method cannot meet the requirement of real-time tracking recovery due to high computational complexity, long time consumption and low precision.
Disclosure of Invention
The invention aims to solve the problems, the training features and the training labels are calculated by using an RGB-D (red, green and blue) -graph, the random forest is trained to predict the coordinates of a 3D scene, and tracking is restored through the coordinate relation, so that the robot navigation tracking restoration method under the small scene based on the random forest is simple in restoration, and high in time period and precision.
In order to achieve the purpose, the technical scheme adopted by the invention is as follows: a robot navigation tracking recovery method in a small scene based on random forests comprises the following steps:
(1) selecting a scene to be tracked, and randomly shooting the scene for M times by using an RGB-D camera to obtain M groups of images, wherein each group of images comprises an RGB image and a depth image;
(2) obtaining a transformation matrix of each group of images by using a three-dimensional reconstruction algorithm to obtain M transformation matrices;
(3) obtaining a training label and training characteristics;
(31) randomly sampling a plurality of pixel points from an RGB image of a first group of images, calculating a 3D world coordinate and random characteristics corresponding to each pixel point, wherein the 3D world coordinate and the random characteristics of the pixel points are in one-to-one correspondence;
(32) obtaining randomly sampled pixel points, 3D world coordinates and random characteristics corresponding to the pixel points by using the remaining M-1 groups of images according to the method in the step (31), using the 3D world coordinates as a training label of the model, and using the random characteristics as training characteristics of the model;
(4) training a random forest model T by using the training labels and the training characteristics;
(41) sequencing all training features, randomly selecting a plurality of training features from the sequenced training features to form a random set, presetting the number of layers adopting a tree balance method as A layers, and presetting the depth threshold of the number as B layers, wherein A is less than B;
(42) acquiring splitting parameters of nodes from a1 st layer to an A th layer by adopting a tree balance method, and acquiring splitting parameters of nodes from an A +1 th layer to a B th layer by adopting a minimum spatial variance method;
(43) forming a random forest model T by the split nodes of each layer;
(5) the robot starts tracking, if the tracking fails, an RGB-D camera is adopted to shoot a scene for 1 time, and a group of images are obtained, wherein the group of images comprise an RGB image and a depth image;
(6) randomly sampling a plurality of pixel points from an RGB image of the group of images, calculating random characteristics corresponding to each pixel point, sending all the random characteristics into T, and outputting 3D world coordinates corresponding to all the random characteristics;
(7) obtaining a prediction transformation matrix according to the 3D world coordinates obtained in the step (6);
(8) and (5) the robot acquires the prediction transformation matrix in the step (7), determines the self pose, and completes navigation tracking recovery.
Preferably, the method comprises the following steps: in the step (2), a Kinectfusion three-dimensional reconstruction method is adopted for each group of images to obtain a transformation matrix.
Preferably, the method comprises the following steps: the step (31) is specifically as follows:
selecting a pixel point, calculating a 3D world coordinate X of the pixel point by adopting the following formula, and calculating a world coordinate W of the pixel point according to the 3D world coordinate X1
X=K-1*P*D(P);
W1=HX;
Wherein K is the internal reference matrix of the camera, K-1Expressing an inverse matrix of an internal reference matrix, wherein P is a vector of the pixel point, D (P) is a depth value in a depth map corresponding to the pixel point, and H is a transformation matrix corresponding to an RGB map where the pixel point is located;
calculating the random characteristic F of the pixel point by adopting the following formulap,
Figure BDA0002116840250000041
Wherein, c1And c2Representing two random channels of 3 image channels in an RGB image, I (P, c)1) Representing the channel c at vector P1And δ represents a 2D offset on the pixel coordinate.
Preferably, the method comprises the following steps: in the step (42):
the method for obtaining the splitting parameters of the nodes from the 1 st layer to the A th layer by the tree balancing method specifically comprises the following steps:
A1. constructing nodes of a first layer;
calculating the balance parameter Q of each training feature in the random set by adopting the following formulabSelecting the smallest QbThe corresponding training characteristics are used as splitting parameters of the nodes of the first layer;
Figure BDA0002116840250000042
in the formula, SLThe number of training features to the left of the current training feature, SRThe number of training features to the right of the current training feature;
A2. respectively forming a data set by using the left and right training features used as the training features of the splitting parameter of the node at the upper layer, and finding out the minimum Q in each data set by using the same method of A1bTaking the corresponding training characteristics as splitting parameters of the next layer of nodes;
A3. repeating the step A2 until a node of layer A is found;
the method for obtaining the splitting parameters of the nodes from the A +1 th layer to the B th layer by the minimum spatial variance method specifically comprises the following steps:
B1. respectively forming training characteristics at two ends of all splitting nodes on the A layer into a data set, and calculating the spatial variance Q of each training characteristic in the data set by adopting the following formulavFind Q in the data setvThe minimum training characteristic is used as a splitting parameter of the node of the A +1 layer;
Figure BDA0002116840250000051
wherein
Figure BDA0002116840250000052
Where n denotes the node index of the tree, m denotes the computed training labels, S denotes a randomly selected set of marked pixel points (p, m), S denotes a random selection of the set of marked pixel points (p, m)nRepresenting the complete set at node n, L representing the left sub-tree, R representing the right sub-tree,
Figure BDA0002116840250000053
representing a set of left or right sub-trees,
Figure BDA0002116840250000054
represents the mean value of m in S;
B2. according to the method of B1, the splitting parameter from the A +2 layer node to the B layer node is calculated.
Preferably, the method comprises the following steps: in step (7), the specific method for obtaining a prediction transformation matrix is as follows:
and calculating the 3D world coordinates by using a Kabsch method to obtain a pose matrix, and optimizing the pose matrix by using a RANSAC algorithm to finally obtain a prediction transformation matrix.
Compared with the prior art, the invention has the advantages that: the calculation of feature points and descriptors is not needed, the matching of pictures is not needed, and only a random forest model needs to be trained in advance. In the invention, the method for constructing the random forest model is very special, the random forest model is constructed by adopting a tree balance method and a minimum spatial variance method together, and in the tree balance method, Q is obtained according to a formula and a method of the random forest modelbFinally, the method can shorten the time in actual use, and compared with a deep learning method, the method is higher in accuracy and shorter in training and testing time.
Drawings
FIG. 1 is a flow chart of the present invention;
FIG. 2 is a comparison graph of the angle truth value and the evaluation value in example 2;
FIG. 3 is a comparison graph of the shift truth value and the estimation value in example 2;
FIG. 4 is a graph of the angle truth and estimation error of FIG. 2;
FIG. 5 is a graph of the shift values and estimation error of FIG. 4.
Detailed Description
The invention will be further explained with reference to the drawings.
Example 1: referring to fig. 1, a method for recovering robot navigation tracking in a small scene based on a random forest comprises the following steps:
(1) selecting a scene to be tracked, and randomly shooting the scene for M times by using an RGB-D camera to obtain M groups of images, wherein each group of images comprises an RGB image and a depth image;
(2) obtaining a transformation matrix of each group of images by using a three-dimensional reconstruction algorithm to obtain M transformation matrices;
(3) obtaining a training label and training characteristics;
(31) randomly sampling a plurality of pixel points from an RGB image of a first group of images, calculating a 3D world coordinate and random characteristics corresponding to each pixel point, wherein the 3D world coordinate and the random characteristics of the pixel points are in one-to-one correspondence;
(32) obtaining randomly sampled pixel points, 3D world coordinates and random characteristics corresponding to the pixel points by using the remaining M-1 groups of images according to the method in the step (31), using the 3D world coordinates as a training label of the model, and using the random characteristics as training characteristics of the model;
(4) training a random forest model T by using the training labels and the training characteristics;
(41) sequencing all training features, randomly selecting a plurality of training features from the sequenced training features to form a random set, presetting the number of layers adopting a tree balance method as A layers, and presetting the depth threshold of the number as B layers, wherein A is less than B;
(42) acquiring splitting parameters of nodes from a1 st layer to an A th layer by adopting a tree balance method, and acquiring splitting parameters of nodes from an A +1 th layer to a B th layer by adopting a minimum spatial variance method;
(43) forming a random forest model T by the split nodes of each layer;
(5) the robot starts tracking, if the tracking fails, an RGB-D camera is adopted to shoot a scene for 1 time, and a group of images are obtained, wherein the group of images comprise an RGB image and a depth image;
(6) randomly sampling a plurality of pixel points from an RGB image of the group of images, calculating random characteristics corresponding to each pixel point, sending all the random characteristics into T, and outputting 3D world coordinates corresponding to all the random characteristics;
(7) obtaining a prediction transformation matrix according to the 3D world coordinates obtained in the step (6);
(8) and (5) the robot acquires the prediction transformation matrix in the step (7), determines the self pose, and completes navigation tracking recovery.
In this embodiment: in the step (2), a Kinectfusion three-dimensional reconstruction method is adopted for each group of images to obtain a transformation matrix.
The step (31) is specifically as follows:
selecting a pixel point, calculating a 3D world coordinate X of the pixel point by adopting the following formula, and calculating a world coordinate W of the pixel point according to the 3D world coordinate X1
X=K-1*P*D(P);
W1=HX;
Wherein K is the internal reference matrix of the camera, K-1Expressing an inverse matrix of an internal reference matrix, wherein P is a vector of the pixel point, D (P) is a depth value in a depth map corresponding to the pixel point, and H is a transformation matrix corresponding to an RGB map where the pixel point is located;
calculating the random characteristic F of the pixel point by adopting the following formulap
Figure BDA0002116840250000071
Wherein, c1And c2Representing two random channels of 3 image channels in an RGB image, I (P, c)1) Representing the channel c at vector P1And δ represents a 2D offset on the pixel coordinate.
In the step (42):
the method for obtaining the splitting parameters of the nodes from the 1 st layer to the A th layer by the tree balancing method specifically comprises the following steps:
A1. constructing nodes of a first layer;
calculating the balance parameter Q of each training feature in the random set by adopting the following formulabSelecting the smallest QbThe corresponding training characteristics are used as splitting parameters of the nodes of the first layer;
Figure BDA0002116840250000081
in the formula, SLThe number of training features to the left of the current training feature, SRThe number of training features to the right of the current training feature;
A2. respectively forming a data set by using the left and right training features used as the training features of the splitting parameter of the node at the upper layer, and finding out the minimum Q in each data set by using the same method of A1bTaking the corresponding training characteristics as splitting parameters of the next layer of nodes;
A3. repeating the step A2 until a node of layer A is found;
the method for obtaining the splitting parameters of the nodes from the A +1 th layer to the B th layer by the minimum spatial variance method specifically comprises the following steps:
B1. respectively forming training characteristics at two ends of all splitting nodes on the A layer into a data set, and calculating the spatial variance Q of each training characteristic in the data set by adopting the following formulavFind Q in the data setvThe minimum training characteristic is used as a splitting parameter of the node of the A +1 layer;
Figure BDA0002116840250000082
wherein
Figure BDA0002116840250000083
Where n denotes the node index of the tree, m denotes the computed training labels, S denotes a randomly selected set of marked pixel points (p, m), S denotes a random selection of the set of marked pixel points (p, m)nRepresenting the complete set at node n, L representing the left sub-tree, R representing the right sub-tree,
Figure BDA0002116840250000091
representing a set of left or right sub-trees,
Figure BDA0002116840250000092
represents the mean value of m in S;
B2. according to the method of B1, the splitting parameter from the A +2 layer node to the B layer node is calculated.
In step (7), the specific method for obtaining a prediction transformation matrix is as follows:
and calculating the 3D world coordinates by using a Kabsch method to obtain a pose matrix, and optimizing the pose matrix by using a RANSAC algorithm to finally obtain a prediction transformation matrix.
In the invention, all coordinates adopt a homogeneous coordinate notation; in addition, in the embodiment, a strategy of calculating all the pixel points is not adopted, only the random features of the randomly selected pixel points are calculated, and the calculated random features and the training labels are in one-to-one correspondence and combined into a queue for training a random forest.
This embodiment mentions a balance parameter QbThe method has the advantages that whether the subtrees are balanced or not can be deduced without really calculating the number of the nodes of the left subtrees and the right subtrees, the calculation time complexity is reduced, the training time is shortened, moreover, when the left subtrees and the right subtrees are approximately balanced, the data part over-fitting or under-fitting caused by overlarge depth difference of any subtrees can be avoided, finally, the random feature with the minimum calculation result is selected as the splitting parameter of the node splitting, and the point worthy of explanation is that the number of the nodes of the left subtrees and the right subtrees is larger than the set minimum number of the nodes.
In this embodiment, the method of minimizing spatial variance is used from the a +1 th layer to the B th layer, and when the number of layers is greater than a, we use this method to obtain the classification parameters. And the random forest generation method is a traditional random forest generation method by adopting a method of minimizing spatial variance to generate random forests. The invention improves the method, only uses the method from A +1 layer to B layer, and uses the tree balance method in the rest.
In this example, to better illustrate the tree balancing method, we take the following example:
assuming that a plurality of training features are obtained in the step (41), randomly extracting 10 training features for sorting, and presetting the number of layers adopting a tree balance method as 2 layers, when the number of layers is less than or equal to 2, adopting the tree balance method, and adopting a minimum spatial variance method for the rest layers. The training characteristics are shown in table 1 below:
table 1: randomly extracting 10 training feature ordered tables
1 2 3 4 5 6 7 8 9 10
(1) By using QbThe calculation formula of (2) calculates 10 training features to find out Q of the random feature 6bAt a minimum, 6 is used as the splitting parameter, the left data set is 1, 2, 3, 4, 5, the right data set is 8, 9, 10, the left data set is for the left sub-tree, the right data set is for the right sub-tree, and the first layer is complete because there is only one node.
(2) And (5) building a second layer: because the number of layers is also two, the left and right subtrees respectively use the feature sets distributed to themselves by the previous layer to perform the same operation, so that each node is selected, and if 3 selected by the left subtree and 8 selected by the right subtree are assumed, 4 sets are obtained, and are respectively distributed to the left and right subtrees of the node, and then the third layer is established.
(3) And a third layer: the number of layers is more than 2, so that the method of minimizing the space variance is used for finding the QvAt a minimum of QvThe corresponding training features are used as splitting parameters.
This is repeated until the threshold B is reached.
Example 2: referring to fig. 1 to 5, a method for recovering robot navigation tracking in a small scene based on a random forest according to the present embodiment includes the following steps:
(1) selecting a scene to be tracked, and randomly shooting the scene for M times by using an RGB-D camera to obtain M groups of images, wherein each group of images comprises an RGB image and a depth image; in the embodiment, an indoor image set kitchen in a public data set 4-Scenes of Stanford university is directly adopted, the size of an image is 640 × 480, a training picture 550 comprises an RGB-D picture, a test picture comprises 520 RGB-D pictures, and the indoor image set kitchen contains information of a transformation matrix and serves as a true value;
(2) (3) and (4) obtaining a random forest model T in the same way as the embodiment 1;
(5) and the robot starts tracking, and if the tracking fails, the RGB-D camera is adopted to shoot the scene for 1 time to obtain a group of images, wherein the group of images comprises an RGB image and a depth image. Here, to explain the effect of the present embodiment, we adopt the following means:
randomly extracting a plurality of test pictures from the indoor image set kitchen, sending the test pictures into the random forest model T obtained in the step (4) for testing, outputting a prediction transformation matrix of each group of pictures, and taking the prediction transformation matrix as an estimated value of an image transformation matrix;
and (3) analyzing an experimental result:
for comparison, the transformation matrix is analyzed into 6 values in the pitch angle, yaw angle, roll angle, X direction, Y direction and Z direction. FIG. 2 is a comparison graph of the true values and estimated values of the angles calculated by 520 groups of pictures in example 2 of the present invention, and FIG. 3 is a comparison graph of the true values and estimated values of the shifts calculated by 520 groups of pictures in example 2 of the present invention; since the similarity is particularly high, we supplement fig. 4 and fig. 5, in which fig. 4 is a graph of the angle true value and the estimation error in fig. 2, and fig. 5 is a graph of the shift true value and the estimation error in fig. 4.
As can be seen from fig. 2 to 5, the overlap ratio between the estimated value curve and the true value curve is high, which indicates that the similarity between the estimated pose and the true pose is high, the angle error value is mostly within the range of-5 ° to +5 °, the displacement error is mostly within the range of-0.05 m to +0.05m, the difference between the estimated value and the true value is small, the error range is also small, which indicates that the algorithm has better robustness, the output result of the algorithm operation is stable, and the estimated value is used for recovering the tracking, so that a better effect can be obtained.
Example 3: compared with other classical algorithms, the method has obvious application effect.
Referring to table 2, the angle is 5 degrees, the displacement is 5cm, the test result of the method reaches 92.7% of accuracy, and the effect is obvious.
TABLE 2 comparison of the results of the different processes
Figure BDA0002116840250000121
In table 2, the method based on the keyframes uses a global keyframe matching method, the method based on the local features respectively uses two image features, namely ORB features and SIFT features, and the angle error is greater than 5 ° and the displacement error is greater than 50cm by using the deep learning method, so that the angle error is not listed in the table.
The above description is only for the purpose of illustrating the preferred embodiments of the present invention and is not to be construed as limiting the invention, and any modifications, equivalents and improvements made within the spirit and principle of the present invention are intended to be included within the scope of the present invention.

Claims (5)

1. A robot navigation tracking recovery method in a small scene based on random forests is characterized by comprising the following steps: the method comprises the following steps:
(1) selecting a scene to be tracked, and randomly shooting the scene for M times by using an RGB-D camera to obtain M groups of images, wherein each group of images comprises an RGB image and a depth image;
(2) obtaining a transformation matrix of each group of images by using a three-dimensional reconstruction algorithm to obtain M transformation matrices;
(3) obtaining a training label and training characteristics;
(31) randomly sampling a plurality of pixel points from an RGB image of a first group of images, calculating a 3D world coordinate and random characteristics corresponding to each pixel point, wherein the 3D world coordinate and the random characteristics of the pixel points are in one-to-one correspondence;
(32) obtaining randomly sampled pixel points, 3D world coordinates and random characteristics corresponding to the pixel points by using the remaining M-1 groups of images according to the method in the step (31), using the 3D world coordinates as a training label of the model, and using the random characteristics as training characteristics of the model;
(4) training a random forest model T by using the training labels and the training characteristics;
(41) sequencing all training features, randomly selecting a plurality of training features from the sequenced training features to form a random set, presetting the number of layers adopting a tree balance method as A layers, and presetting the depth threshold of the number as B layers, wherein A is less than B;
(42) acquiring splitting parameters of nodes from a1 st layer to an A th layer by adopting a tree balance method, and acquiring splitting parameters of nodes from an A +1 th layer to a B th layer by adopting a minimum spatial variance method;
(43) forming a random forest model T by the split nodes of each layer;
(5) the robot starts tracking, if the tracking fails, an RGB-D camera is adopted to shoot a scene for 1 time, and a group of images are obtained, wherein the group of images comprise an RGB image and a depth image;
(6) randomly sampling a plurality of pixel points from an RGB image of the group of images, calculating random characteristics corresponding to each pixel point, sending all the random characteristics into T, and outputting 3D world coordinates corresponding to all the random characteristics;
(7) obtaining a prediction transformation matrix according to the 3D world coordinates obtained in the step (6);
(8) and (5) the robot acquires the prediction transformation matrix in the step (7), determines the self pose, and completes navigation tracking recovery.
2. The method for recovering the robot navigation tracking under the small scene based on the random forest as claimed in claim 1, wherein: in the step (2), a Kinectfusion three-dimensional reconstruction method is adopted for each group of images to obtain a transformation matrix.
3. The method for recovering the robot navigation tracking under the small scene based on the random forest as claimed in claim 1, wherein: the step (31) is specifically as follows:
selecting a pixel point, calculating a 3D world coordinate X of the pixel point by adopting the following formula, and calculating a world coordinate W of the pixel point according to the 3D world coordinate X1
X=K-1*P*D(P);
W1=HX;
Wherein K is the internal reference matrix of the camera, K-1Expressing an inverse matrix of an internal reference matrix, wherein P is a vector of the pixel point, D (P) is a depth value in a depth map corresponding to the pixel point, and H is a transformation matrix corresponding to an RGB map where the pixel point is located;
calculating the random characteristic F of the pixel point by adopting the following formulap
Figure FDA0002116840240000021
Wherein, c1And c2Representing two random channels of 3 image channels in an RGB image, I (P, c)1) Representing the channel c at vector P1And δ represents a 2D offset on the pixel coordinate.
4. The method for recovering the robot navigation tracking under the small scene based on the random forest as claimed in claim 1, wherein: in the step (42):
the method for obtaining the splitting parameters of the nodes from the 1 st layer to the A th layer by the tree balancing method specifically comprises the following steps:
A1. constructing nodes of a first layer;
calculating the balance parameter Q of each training feature in the random set by adopting the following formulabSelecting the smallest QbCorresponding training features are used as the first layerSplitting parameters of the node of (1);
Figure FDA0002116840240000031
in the formula, SLThe number of training features to the left of the current training feature, SRThe number of training features to the right of the current training feature;
A2. respectively forming a data set by using the left and right training features used as the training features of the splitting parameter of the node at the upper layer, and finding out the minimum Q in each data set by using the same method of A1bTaking the corresponding training characteristics as splitting parameters of the next layer of nodes;
A3. repeating the step A2 until a node of layer A is found;
the method for obtaining the splitting parameters of the nodes from the A +1 th layer to the B th layer by the minimum spatial variance method specifically comprises the following steps:
B1. respectively forming training characteristics at two ends of all splitting nodes on the A layer into a data set, and calculating the spatial variance Q of each training characteristic in the data set by adopting the following formulavFind Q in the data setvThe minimum training characteristic is used as a splitting parameter of the node of the A +1 layer;
Figure FDA0002116840240000032
wherein
Figure FDA0002116840240000033
Where n denotes the node index of the tree, m denotes the computed training labels, S denotes a randomly selected set of marked pixel points (p, m), S denotes a random selection of the set of marked pixel points (p, m)nRepresenting the complete set at node n, L representing the left sub-tree, R representing the right sub-tree,
Figure FDA0002116840240000041
representing a set of left or right sub-trees,
Figure FDA0002116840240000042
represents the mean value of m in S;
B2. according to the method of B1, the splitting parameter from the A +2 layer node to the B layer node is calculated.
5. The method for recovering the robot navigation tracking under the small scene based on the random forest as claimed in claim 1, wherein: in step (7), the specific method for obtaining a prediction transformation matrix is as follows:
and calculating the 3D world coordinates by using a Kabsch method to obtain a pose matrix, and optimizing the pose matrix by using a RANSAC algorithm to finally obtain a prediction transformation matrix.
CN201910593421.2A 2019-07-03 2019-07-03 Robot navigation tracking recovery method in small scene based on random forest Expired - Fee Related CN110400349B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910593421.2A CN110400349B (en) 2019-07-03 2019-07-03 Robot navigation tracking recovery method in small scene based on random forest

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910593421.2A CN110400349B (en) 2019-07-03 2019-07-03 Robot navigation tracking recovery method in small scene based on random forest

Publications (2)

Publication Number Publication Date
CN110400349A CN110400349A (en) 2019-11-01
CN110400349B true CN110400349B (en) 2022-04-15

Family

ID=68322729

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910593421.2A Expired - Fee Related CN110400349B (en) 2019-07-03 2019-07-03 Robot navigation tracking recovery method in small scene based on random forest

Country Status (1)

Country Link
CN (1) CN110400349B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113255490A (en) * 2021-05-15 2021-08-13 成都理工大学 Unsupervised pedestrian re-identification method based on k-means clustering and merging

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108898623A (en) * 2018-05-24 2018-11-27 北京飞搜科技有限公司 Method for tracking target and equipment

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9940553B2 (en) * 2013-02-22 2018-04-10 Microsoft Technology Licensing, Llc Camera/object pose from predicted coordinates
US9626766B2 (en) * 2014-02-28 2017-04-18 Microsoft Technology Licensing, Llc Depth sensing using an RGB camera
US9613298B2 (en) * 2014-06-02 2017-04-04 Microsoft Technology Licensing, Llc Tracking using sensor data

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108898623A (en) * 2018-05-24 2018-11-27 北京飞搜科技有限公司 Method for tracking target and equipment

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
基于树结构分层随机森林在非约束环境下的头部姿态估计;刘袁缘等;《电子与信息学报》;20150315(第03期);第39-47页全文 *
基于深度优先随机森林分类器的目标检测;马娟娟等;《中国惯性技术学报》;20180815(第04期);第104-109页全文 *

Also Published As

Publication number Publication date
CN110400349A (en) 2019-11-01

Similar Documents

Publication Publication Date Title
CN109387204B (en) Mobile robot synchronous positioning and composition method facing indoor dynamic environment
Wang et al. Matchformer: Interleaving attention in transformers for feature matching
Shotton et al. Scene coordinate regression forests for camera relocalization in RGB-D images
CN108038420B (en) Human behavior recognition method based on depth video
CN114782691A (en) Robot target identification and motion detection method based on deep learning, storage medium and equipment
CN111862213A (en) Positioning method and device, electronic equipment and computer readable storage medium
Blanton et al. Extending absolute pose regression to multiple scenes
CN108171249B (en) RGBD data-based local descriptor learning method
KR20030062313A (en) Image conversion and encoding techniques
CN111583340B (en) Method for reducing monocular camera pose estimation error rate based on convolutional neural network
Meng et al. Exploiting points and lines in regression forests for RGB-D camera relocalization
WO2011131029A1 (en) Method for detecting similar units based on outline belt graph
CN114842033B (en) Image processing method for intelligent AR equipment
CN113361542A (en) Local feature extraction method based on deep learning
CN115761734A (en) Object pose estimation method based on template matching and probability distribution
CN111797692A (en) Depth image gesture estimation method based on semi-supervised learning
CN112465021A (en) Pose track estimation method based on image frame interpolation method
CN116188825A (en) Efficient feature matching method based on parallel attention mechanism
CN113673354A (en) Human body key point detection method based on context information and combined embedding
CN110400349B (en) Robot navigation tracking recovery method in small scene based on random forest
CN114663880A (en) Three-dimensional target detection method based on multi-level cross-modal self-attention mechanism
Yu et al. Learning bipartite graph matching for robust visual localization
EP3995993A1 (en) Process for the extension of deep-learning based 2d-object-detectors for the 6d-pose-estimation
CN113011359A (en) Method for simultaneously detecting plane structure and generating plane description based on image and application
KR102186764B1 (en) Apparatus and method for estimating optical flow and disparity via cycle consistency

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20220415