WO2022077863A1 - Visual positioning method, and method for training related model, related apparatus, and device - Google Patents
Visual positioning method, and method for training related model, related apparatus, and device Download PDFInfo
- Publication number
- WO2022077863A1 WO2022077863A1 PCT/CN2021/082198 CN2021082198W WO2022077863A1 WO 2022077863 A1 WO2022077863 A1 WO 2022077863A1 CN 2021082198 W CN2021082198 W CN 2021082198W WO 2022077863 A1 WO2022077863 A1 WO 2022077863A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- matching
- point
- image
- value
- map
- Prior art date
Links
- 238000012549 training Methods 0.000 title claims abstract description 83
- 238000000034 method Methods 0.000 title claims abstract description 75
- 230000000007 visual effect Effects 0.000 title claims abstract description 56
- 238000012545 processing Methods 0.000 claims abstract description 39
- 238000000605 extraction Methods 0.000 claims description 81
- 208000037170 Delayed Emergence from Anesthesia Diseases 0.000 claims description 30
- 238000003384 imaging method Methods 0.000 claims description 9
- 238000005315 distribution function Methods 0.000 claims description 8
- 230000009466 transformation Effects 0.000 claims description 8
- 238000004590 computer program Methods 0.000 claims description 2
- 230000004807 localization Effects 0.000 claims description 2
- 230000009286 beneficial effect Effects 0.000 description 14
- 238000010276 construction Methods 0.000 description 11
- 238000010586 diagram Methods 0.000 description 10
- 238000005070 sampling Methods 0.000 description 4
- 230000008878 coupling Effects 0.000 description 3
- 238000010168 coupling process Methods 0.000 description 3
- 238000005859 coupling reaction Methods 0.000 description 3
- 238000010606 normalization Methods 0.000 description 3
- 238000004891 communication Methods 0.000 description 2
- 238000005457 optimization Methods 0.000 description 2
- 230000000750 progressive effect Effects 0.000 description 2
- 238000013459 approach Methods 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 230000001413 cellular effect Effects 0.000 description 1
- 238000006243 chemical reaction Methods 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 238000003062 neural network model Methods 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N20/00—Machine learning
- G06N20/20—Ensemble learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/20—Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
- G06F16/29—Geographical information databases
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/77—Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
- G06V10/774—Generating sets of training patterns; Bootstrap methods, e.g. bagging or boosting
Definitions
- the present disclosure relates to the technical field of computer vision, and in particular, to a visual positioning method, a training method for a related model, and related devices and equipment.
- Visual positioning can be divided into various ways according to the expression of map data.
- the structure-based method also known as the feature-based method, has received extensive attention due to its high accuracy and excellent generalization performance.
- the present disclosure provides a visual positioning method, a training method for a related model, and related devices and equipment.
- a first aspect of the present disclosure provides a training method for a matching prediction model, including: using sample images and map data to construct sample matching data, wherein the sample matching data includes several groups of point pairs and an actual matching value of each group of point pairs, The two points of each group of point pairs come from the sample image and map data respectively; use the matching prediction model to perform prediction processing on several groups of point pairs to obtain the predicted matching value of the point pair; use the actual matching value and the predicted matching value to determine the matching prediction model loss value; use the loss value to adjust the parameters of the matching prediction model.
- the sample matching data is obtained by constructing the sample image and map data, and the sample matching data includes several groups of point pairs and the actual matching value of each group of point pairs.
- the two points of each group of point pairs come from the sample image and map data respectively, Therefore, the matching prediction model is used to perform prediction processing on several groups of point pairs, and the predicted matching value of the point pair is obtained, and then the actual matching value and the predicted matching value are used to determine the loss value of the matching prediction model, and the loss value is used to match the parameters of the prediction model. Therefore, the matching prediction model can be used to establish a matching relationship, so that the matching prediction model can be used to predict the matching value between point pairs in visual positioning, so the point pair with high matching value can be preferentially sampled based on the predicted matching value.
- the camera pose parameters of the image to be positioned can further improve the accuracy and immediacy of visual positioning.
- using the sample image and the map data to construct the sample matching data includes: obtaining several image points from the sample image, and obtaining several map points from the map data to form several groups of point pairs; wherein, the several groups of point pairs include at least one Matching point pairs that match between the image points and map points contained in the group; for each group of matching point pairs: use the pose parameters of the sample image to project the map points into the dimension to which the sample image belongs to obtain the projected points of the map points; And based on the difference between the image point and the projected point, the actual matching value of the matching point pair is determined.
- the several sets of point pairs include at least one set of matching between the included image points and the map points. Therefore, it is possible to generate samples for training the matching prediction model, and for each set of matching point pairs, use the pose parameters of the sample image to project the map points into the dimension to which the sample image belongs, and obtain the projected points of the map points. , so as to determine the actual matching value of the matching point pair based on the difference between the image point and the projection point, so that the matching prediction model can learn the geometric features of the matching point pair during the training process, which is beneficial to improve the accuracy of the matching prediction model.
- the several sets of point pairs include at least one set of non-matching points that do not match between the included image points and map points, and using sample images and map data to construct sample matching data further includes: converting the actual matching values of the non-matching point pairs into Set to default value
- several groups of point pairs include at least one group of non-matching point pairs that do not match between the included image points and map points, and different from the matching point pairs, the actual matching value of the non-matching point pairs is set to a preset value, thereby It can help to improve the robustness of the matching prediction model.
- acquiring several image points from the sample image and acquiring several map points from the map data to form several groups of point pairs including: dividing the image points in the sample image into a first image point and a second image point, wherein , the first image point has a matching map point in the map data, and the second image point does not have a matching map point in the map data; for each first image point, assign a number of first map points from the map data , and take the first image point and each first map point as a first point pair, wherein the first map point includes a map point matching the first image point; and, for each second image point, A number of second map points are allocated from the map data, and the second image point and each second map point are respectively used as a second point pair; several groups of point pairs are extracted from the first point pair and the second point pair.
- the second image point does not have a matching map point in the image data.
- image point, and for the first image point assign a number of first map points from the map data, respectively take the first image point and each first map point as a first point pair, and the first map point includes and the first map point.
- using the pose parameter of the sample image to project the map point into the dimension to which the sample image belongs, and obtaining the projected point of the map point includes: calculating the pose parameter of the sample image based on the matching point pair; using the pose parameter to project the map point To the dimension to which the sample image belongs, the projected point of the map point is obtained.
- the matching point pairs to calculate the pose parameters of the sample image, and using the pose parameters to project the map points into the dimension to which the sample image belongs, the projected points of the map points can be obtained, which can help to improve the relationship between the projected points and the image points.
- the accuracy of the difference between them can be beneficial to improve the accuracy of the matching prediction model.
- determining the actual matching value of the matching point pair based on the difference between the image point and the projection point includes: using a preset probability distribution function to convert the difference into a probability density value as the actual matching value of the matching point pair.
- the preset probability distribution function to convert the difference into a probability density value as the actual matching value of the matching point pair, it can help to accurately describe the difference between the projection point and the image point, which can help improve the matching prediction. accuracy of the model.
- the sample matching data is a bipartite graph, and the bipartite graph includes several groups of point pairs and connecting edges connecting each group of point pairs, and the connecting edges are marked with the actual matching values of the corresponding point pairs;
- the matching prediction model includes the dimension corresponding to the sample image.
- the first point feature extraction sub-model, the second point feature extraction sub-model corresponding to the dimension to which the map data belongs, and the edge feature extraction sub-model; the matching prediction model is used to perform prediction processing on several groups of point pairs, and the prediction matching of point pairs is obtained.
- the value includes: using the first point feature extraction sub-model and the second point feature extraction sub-model to perform feature extraction on the bipartite graph to obtain the first feature and the second feature; using the edge feature extraction sub-model to extract the first feature and the second feature. Perform feature extraction to obtain a third feature; use the third feature to obtain the predicted matching value of the point pair corresponding to the connecting edge.
- the matching prediction model can more effectively perceive the spatial geometric structure of the matching, which can help to improve the accuracy of the matching prediction model.
- the structure of the first point feature extraction sub-model and the second point feature extraction sub-model is any of the following: including at least one residual block, including at least one residual block and at least one spatial transformation network; and/or, edge
- the feature extraction submodel includes at least one residual block.
- the structure of the first point feature extraction sub-model and the second point feature extraction sub-model to any one of the following: including at least one residual block, including at least one residual block and at least one spatial transformation network, and
- the edge feature extraction sub-model is set to include at least one residual block, so it can facilitate the optimization of the matching prediction model and improve the accuracy of the matching prediction model.
- several sets of point pairs include at least one set of matching point pairs that match between the included image points and map points and at least one set of non-matching point pairs that do not match between the included image points and map points; using actual matching value and predicted matching value, and determining the loss value of the matching prediction model includes: using the predicted matching value and the actual matching value of the matching point pair to determine the first loss value of the matching prediction model; and using the predicted matching value and actual matching value of the non-matching point pair.
- the matching value is used to determine the second loss value of the matching prediction model; the first loss value and the second loss value are weighted to obtain the loss value of the matching prediction model.
- the first loss value of the matching prediction model is determined by using the predicted matching value and the actual matching value of the matching point pair
- the second loss value of the matching prediction model is determined using the predicted matching value and the actual loss value of the non-matching point pair.
- the method before determining the first loss value of the matching prediction model by using the predicted matching value and the actual matching value of the matching point pair, the method further includes: separately counting the first number of matching point pairs and the second number of non-matching point pairs; Using the predicted matching value and the actual matching value of the matching point pair to determine the first loss value of the matching prediction model includes: using the difference between the predicted matching value and the actual matching value of the matching point pair, and the first number, determining the first loss value of the matching prediction model. Loss value; using the predicted matching value and the actual matching value of the unmatched point pair to determine the second loss value of the matching prediction model includes: using the difference between the predicted matching value and the actual matching value of the unmatched point pair, and the second loss value of the matching prediction model. quantity to determine the second loss value.
- the first loss is determined by using the difference between the predicted matching value and the actual matching value of the matching point pair, and the first number value, and use the difference between the predicted matching value and the actual matching value of the non-matching point pair, and the second quantity, to determine the second loss value, which can help to improve the accuracy of the loss value of the matching prediction model, which can be beneficial to Improve the accuracy of matching prediction models.
- the dimension to which the sample image belongs is 2-dimensional or 3-dimensional
- the dimension to which the map data belongs is 2-dimensional or 3-dimensional
- the 3-3-dimensional matching prediction model can improve the applicable scope of the matching prediction model.
- a second aspect of the present disclosure provides a visual positioning method, comprising: constructing matching data to be identified by using an image to be positioned and map data, wherein the matching data to be identified includes several sets of point pairs, and two points of each set of point pairs are respectively From the image and map data to be positioned; use the matching prediction model to perform prediction processing on several groups of point pairs to obtain the predicted matching value of the point pair; based on the predicted matching value of the point pair, determine the pose parameters of the camera device of the image to be positioned.
- the to-be-identified matching data is constructed, and the to-be-identified matching data includes several sets of point pairs, and the two points of each set of point pairs are respectively from the to-be-located image and the map data, thereby using the matching prediction model.
- Predictive processing is performed on several groups of point pairs to obtain the predicted matching values of the point pairs, and then based on the predicted matching values of the point pairs, the pose parameters of the camera device of the image to be positioned are determined, which improves the accuracy and immediacy of visual positioning.
- determining the pose parameters of the imaging device of the image to be positioned based on the predicted matching values of the point pairs includes: sorting several groups of point pairs in descending order of the predicted matching values; using the previously preset number of groups of point pairs , and determine the pose parameters of the imaging device of the image to be positioned.
- the matching prediction model is obtained by using the training method of the matching prediction model in the first aspect.
- a third aspect of the present disclosure provides a training device for matching prediction models, including: a sample construction module, a prediction processing module, a loss determination module, and a parameter adjustment module.
- the sample construction module is used for using sample images and map data to construct sample matching data.
- the sample matching data includes several groups of point pairs and the actual matching values of each group of point pairs, and the two points of each group of point pairs come from the sample image and map data respectively;
- the prediction processing module is used to use the matching prediction model to match several groups of points.
- the loss determination module is used to use the actual matching value and the predicted matching value to determine the loss value of the matching prediction model;
- the parameter adjustment module is used to use the loss value to adjust the parameters of the matching prediction model .
- a fourth aspect of the present disclosure provides a visual positioning device, comprising: a data construction module, a prediction processing module, and a parameter determination module, where the data construction module is used to construct matching data to be identified by using images to be positioned and map data, wherein the to-be-identified matching data is The matching data includes several groups of point pairs, and the two points of each group of point pairs are respectively from the image to be located and the map data; the prediction processing module is used to perform prediction processing on several groups of point pairs by using the matching prediction model to obtain the predicted matching value of the point pair. ; The parameter determination module is used to determine the pose parameters of the camera device of the image to be positioned based on the predicted matching value of the point pair.
- a fifth aspect of the present disclosure provides an electronic device, including a memory and a processor coupled to each other, where the processor is configured to execute program instructions stored in the memory, so as to implement the training method for a matching prediction model in the first aspect, or to implement The visual positioning method in the above second aspect.
- a sixth aspect of the present disclosure provides a computer-readable storage medium on which program instructions are stored, and when the program instructions are executed by a processor, implement the training method for a matching prediction model in the first aspect above, or implement the second aspect above. visual positioning method.
- a sixth aspect of the present disclosure provides a computer program, comprising computer-readable codes, which, when the computer-readable codes are executed in an electronic device and executed by a processor in the electronic device, realize the above-mentioned first aspect
- the above scheme can use the matching prediction model to establish a matching relationship, so that the matching prediction model can be used to predict the matching value between point pairs in visual positioning, so the point pair with high matching value can be preferentially sampled based on the predicted matching value, and establish The matching relationship can be beneficial to improve the accuracy and immediacy of visual positioning.
- FIG. 1 is a schematic flowchart of an embodiment of a training method for a matching prediction model of the present disclosure
- FIG. 2 is a state schematic diagram of an embodiment of a training method for a matching prediction model of the present disclosure
- step S11 in FIG. 1 is a schematic flowchart of an embodiment of step S11 in FIG. 1;
- FIG. 4 is a schematic flowchart of an embodiment of step S111 in FIG. 3;
- FIG. 5 is a schematic flowchart of an embodiment of the visual positioning method of the present disclosure.
- FIG. 6 is a schematic diagram of a framework of an embodiment of a training device for matching prediction models of the present disclosure
- FIG. 7 is a schematic frame diagram of an embodiment of the visual positioning device of the present disclosure.
- FIG. 8 is a schematic diagram of a framework of an embodiment of an electronic device of the present disclosure.
- FIG. 9 is a schematic diagram of a framework of an embodiment of a computer-readable storage medium of the present disclosure.
- system and “network” are often used interchangeably herein.
- the term “and/or” in this article is only an association relationship to describe the associated objects, indicating that there can be three kinds of relationships, for example, A and/or B, it can mean that A exists alone, A and B exist at the same time, and A and B exist independently B these three cases.
- the character "/” in this document generally indicates that the related objects are an “or” relationship.
- “multiple” herein means two or more than two.
- FIG. 1 is a schematic flowchart of an embodiment of a training method for a matching prediction model of the present disclosure.
- the training method of the matching prediction model may include the following steps:
- Step S11 Using the sample image and map data to construct sample matching data.
- the sample matching data includes several groups of point pairs and actual matching values of each group of point pairs, and the two points of each group of point pairs are respectively from the sample image and the map data.
- the map data may be constructed from sample images.
- the dimension to which the sample image belongs may be 2-dimensional or 3-dimensional, and the dimension to which the map data belongs may be 2-dimensional or 3-dimensional, which is not limited herein.
- the sample image is a two-dimensional image
- the two-dimensional image can be processed by three-dimensional reconstruction methods such as SFM (Structure From Motion) to obtain map data such as a sparse point cloud model.
- the sample image can also include three-dimensional information.
- the sample image may also be an RGB-D image (ie, a color image and a depth image), which is not limited herein.
- the map data may be composed of a simple two-dimensional image, a three-dimensional point cloud map, or a combination of a two-dimensional image and a three-dimensional point cloud, which is not limited here.
- the execution body of the training method for matching prediction models may be a training device for matching prediction models, which is described as a training device hereinafter; for example, the training method for matching prediction models may be performed by a terminal device or a server or other processing device.
- the terminal device may be a user equipment (User Equipment, UE), a mobile device, a user terminal, a terminal, a cellular phone, a cordless phone, a personal digital assistant (Personal Digital Assistant, PDA), a handheld device, a computing device, a vehicle-mounted device , wearable devices, etc.
- the training method of the matching prediction model may be implemented by the processor calling computer-readable instructions stored in the memory.
- the sample matching data may be a bipartite graph.
- a bipartite graph also known as a bipartite graph, is an undirected graph composed of a point set and an edge set, and the point set can be divided into two mutually disjoint sub-graphs. The two points associated with each edge in the edge set belong to the two disjoint subsets.
- the sample matching data when it is a bipartite graph, it includes several groups of point pairs and connecting edges connecting each group of point pairs, and the connecting edges are marked with the actual matching value of the corresponding point pair, which is used to describe the matching degree of the corresponding point pair, for example , the actual matching value can be a value between 0 and 1; here, when the actual matching value is 0.1, it can indicate that the matching degree between the corresponding point pairs is low, and the points from the sample image in the point pairs are the same as the points from the map data.
- FIG. 2 is a schematic state diagram of an embodiment of the training method of the matching prediction model of the present disclosure.
- the left side is the sample matching data represented by the bipartite graph, and the upper side and the lower side of the bipartite graph
- the two are mutually disjoint point sets, and the points connecting the two point sets are connected edges, and the connected edges are marked with actual matching values (not shown).
- the training device may further perform data enhancement on the sample matching data.
- the training device may randomly rotate the coordinates of the three-dimensional points in the sample matching data to three axes; or, it may also perform normalization processing on the three-dimensional points in the sample matching data, which is not limited here.
- Step S12 Use the matching prediction model to perform prediction processing on several groups of point pairs to obtain the predicted matching values of the point pairs.
- the matching prediction model may include a first point feature extraction sub-model corresponding to the dimension to which the sample image belongs, and a second sub-model corresponding to the dimension to which the map data belongs. Point feature extraction sub-model, and edge feature extraction sub-model.
- the matching prediction model obtained by training can be used For two-dimensional-three-dimensional matching prediction; or, when the sample image is a three-dimensional image and the map data includes a three-dimensional point cloud, the first point feature extraction sub-model and the second point feature extraction sub-model are three-dimensional point feature extraction sub-models, then training
- the obtained matching prediction model can be used for 3D-3D matching prediction; or, when the sample image is a 2D image and the map data includes a 3D point cloud, the first point feature extraction sub-model is a 2D point feature extraction sub-model, the second The point feature extraction sub-model is a three-dimensional point feature extraction sub-model, and the matching prediction model obtained by training can be used for 2D-3D matching prediction; here, the matching prediction model can be set according to the actual application, which is not limited here
- the training device may use the first point feature extraction sub-model and the second point feature extraction sub-model to perform feature extraction on the bipartite graph to obtain the first feature and the second feature; and then use the edge feature extraction sub-model to perform feature extraction on the second feature
- the first feature and the second feature are extracted to obtain the third feature; the third feature is used to obtain the predicted matching value of the corresponding point of the connecting edge; as shown in Figure 2, the predicted matching value corresponding to each connecting edge in the bipartite graph is shown.
- each residual block may be included, for example, one residual block, 2 residual blocks, 3 residual blocks, etc.
- each residual block (resblock) consists of multiple basic blocks (base blocks), and each basic block (base block) consists of a layer of 1*1 It consists of a convolutional layer, a batch normalization layer, and a context normalization layer.
- At least one residual block (resblock) and at least one spatial transformation network may be included, for example , 1 residual block, 2 residual blocks, 3 residual blocks, etc., which are not limited here.
- the number of spatial transformation networks can be one or two.
- the spatial transformation networks can be located at the beginning and end of the model, which is not limited here.
- the edge feature extraction sub-model may include at least one residual block, for example, one residual block, two residual blocks, three residual blocks, etc., which are not limited here, and the structure of the residual block (resblock) can be Referring to the structure in the foregoing implementation scenario, details are not repeated here.
- Step S13 Determine the loss value of the matching prediction model by using the actual matching value and the predicted matching value.
- the training device may count the difference between the actual matching value and the predicted matching value, so as to determine the loss value of the matching prediction model.
- the training device can count the sum of the differences between the predicted matching values of all point pairs and their actual matching values, and then use the sum and the number of all point pairs to obtain the average of the predicted matching values of all point pairs, as matching The loss value of the prediction model.
- several sets of matching point pairs may include at least one set of matching point pairs that match between the included image points and map points, that is, the image points and map points included in the matching point pairs are the same in space.
- One point, several sets of matching point pairs may also include at least one set of non-matching point pairs that do not match between the image points and map points contained in the non-matching point pairs, that is, the image points and map points contained in the non-matching point pairs are different points in space , then the training device can use the predicted matching value w * and the actual matching value w of the matching point pair to determine the first loss value L pos (w, w * ) of the matching prediction model, and use the predicted matching value w of the non-matching point pair * and the actual matching value w, determine the second loss value L neg (w, w * ) of the matching prediction model, so that by comparing the first loss value L pos (w, w * ) and the second loss value L neg (w, w * ) is weighted
- L(w, w * ) represents the loss value of the matching prediction model
- L pos (w, w * ) represents the first loss value corresponding to the matching point pair
- L neg (w, w * ) represents the second loss value corresponding to the non-matching point pair
- ⁇ and ⁇ represent the weight of the first loss value L pos (w,w * ) and the weight of the second loss value L neg (w,w * ), respectively.
- the training device may also count the first number
- L pos (w, w * ) represents the first loss value
- represents the first quantity
- w, w * represent the actual matching value and the predicted matching value of the matching point pair, respectively.
- the training device can also use the difference between the predicted matching value and the actual matching value of the non-matching point pair, as well as the second quantity, to determine the second loss value, see formula (3):
- L neg (w, w * ) represents the second loss value
- represents the second quantity
- w, w * represent the actual matching value and the predicted matching value of the non-matching point pair, respectively; in addition, , the actual matching value w of the non-matching point pair can also be uniformly set to a preset value (for example, 0).
- Step S14 Using the loss value, adjust the parameters of the matching prediction model.
- the training device may adopt methods such as Stochastic Gradient Descent (SGD), Batch Gradient Descent (BGD), Mini-Batch Gradient Descent (MBGD), etc.
- the loss value adjusts the parameters of the matching prediction model; among them, batch gradient descent refers to using all samples for parameter update in each iteration; stochastic gradient descent refers to using one sample for parameter update in each iteration ; Mini-batch gradient descent refers to using a batch of samples to update parameters in each iteration, which will not be repeated here.
- a training end condition may also be set, and when the training end condition is satisfied, the training device may end the training of the matching prediction model.
- the training end condition may include: the loss value is less than a preset loss threshold, and the loss value is no longer reduced; the current training times reaches a preset times threshold (eg, 500 times, 1000 times, etc.), which is not limited here.
- the sample matching data is obtained by constructing the sample image and the map data, and the sample matching data includes several groups of point pairs and the actual matching value of each group of point pairs, and the two points of each group of point pairs come from the sample image and the map data respectively.
- the matching prediction model is used to perform prediction processing on several groups of point pairs, and the predicted matching value of the point pair is obtained, and then the actual matching value and the predicted matching value are used to determine the loss value of the matching prediction model, and the loss value is used to determine the matching prediction model.
- the matching prediction model can be used to establish a matching relationship, so that the matching prediction model can be used to predict the matching value between point pairs in visual positioning, so the point pair with high matching value can be preferentially sampled based on the predicted matching value. This can help to improve the accuracy and immediacy of visual positioning.
- FIG. 3 is a schematic flowchart of an embodiment of step S11 in FIG. 1 .
- the training device can construct the sample matching data through the following steps:
- Step S111 Acquire several image points from the sample image, and acquire several map points from the map data to form several groups of point pairs.
- Several sets of point pairs include at least one set of matched point pairs that match between the included image points and map points; that is, at least one set of the included image points and map points in the several sets of point pairs corresponds to the same in space.
- a matching pair of dots Taking the sample image as a two-dimensional image and the map data as a sparse point cloud model obtained by SFM reconstruction as an example, several groups of point pairs contain at least one triangulated point and the triangulated point corresponds to the sparse point cloud model. three-dimensional point.
- the several groups of point pairs may further include at least one group of non-matching point pairs that do not match between the included image points and map points, that is, the several groups of point pairs may further include at least one group of non-matching point pairs.
- the included image points and map points correspond to unmatched pairs of points at different points in space.
- several groups of point pairs can also include untriangulated points and any point in the sparse point cloud model to form a sparse point cloud model.
- a set of non-matching point pairs can add noise to the sample matching data, thereby improving the robustness of the matching prediction model.
- FIG. 4 is a schematic flowchart of an embodiment of step S111 in FIG. 3 .
- the training device can obtain several sets of point pairs through the following steps:
- Step S41 Divide the image points in the sample image into a first image point and a second image point.
- the first image point has a matching map point in the map data
- the second image point does not have a matching map point in the map data.
- the first image point can be a triangulated feature point in the sample image
- the second image point can be a non-triangulated point in the sample image. Triangulated feature points; in other application scenarios, it can be deduced by analogy, which is not limited here.
- the image points in the sample image are feature points of the sample image.
- the coordinates of the feature points can also be converted to a normalized plane.
- Step S42 For each first image point, assign a number of first map points from the map data, and use the first image point and each first map point as a first point pair, wherein, among the first map points Contains map points that match the first image point.
- a number of first map points are allocated from the map data, and the first image point and each first map point are respectively regarded as a first point pair, and the first map point includes a The map point to which the image point matches.
- the number of first map points allocated to each first image point may be the same or different.
- a number of first image points may be randomly selected from the first image points obtained by division, and the first image points obtained by extraction may be assigned from map data. The steps of using a plurality of first map points and respectively using the first image point and each first map point as a first point pair are not limited herein.
- N points may be randomly selected from the first image points obtained by division, and for each of the N first image points obtained by the extraction, K points are randomly allocated from the map data The first map point, and the randomly assigned K first map points include map points that match the first image point.
- Step S43 For each second image point, assign a number of second map points from the map data, and use the second image point and each second map point as a second point pair.
- a number of second map points are allocated from the map data, and the second image point and each second map point are respectively regarded as a second point pair.
- the number of second map points allocated to each second image point may be the same or different.
- a number of second image points may be randomly selected from the second image points obtained by division, and the second image points obtained by extraction may be assigned from map data.
- the steps of using a plurality of second map points and using the second image point and each second map point as a second point pair are not limited herein.
- M points may be randomly selected from the divided second image points, and K points may be randomly allocated from the map data for each of the M second image points obtained by the extraction Second map point.
- each first point pair and each second point pair are matching point pairs
- traverse each first point pair and each second point pair and use the first identifier
- An identifier eg, 1 marks matching point pairs
- a second identifier eg, 0
- steps S42 and S43 can be performed in sequence, for example, step S42 is performed first, and then step S43 is performed; or, step S43 is performed first, and then step S42 is performed; This is not limited.
- Step S44 Extracting several groups of point pairs from the first point pair and the second point pair.
- several groups of point pairs may be obtained by randomly extracting from the first point pair and the second point pair, as a sample matching data.
- the first point pair and the second point pair may also be randomly selected several times to obtain several sample matching data.
- multiple sample images and map data can also be obtained, and the above steps can be repeatedly performed for each sample image and map data to obtain multiple sample matching data, which can increase the number of samples and help improve matching. Predictive model accuracy.
- Step S112 for each set of matching point pairs: using the pose parameters of the sample image to project the map point into the dimension to which the sample image belongs to obtain the projected point of the map point; and determine the matching based on the difference between the image point and the projected point The actual match value of the point pair.
- the map point can be projected into the dimension to which the sample image belongs by using the pose parameters of the corresponding sample image to obtain the projected point of the map point.
- the training device can use the pose parameters to reproject the three-dimensional points to obtain their projected points.
- a preset probability distribution function can be used to convert the difference between the image point and its projected point into a probability density value, which is used as the actual matching value of the matching point pair.
- the preset probability distribution function may be a standard Gaussian distribution function, so that the difference in the value range from negative infinity to positive infinity can be converted into a corresponding probability density value, and the greater the absolute value of the difference, The smaller the corresponding probability density value, the lower the matching degree of the corresponding point pair, the smaller the absolute value of the difference, the smaller the corresponding probability density value, the higher the matching degree of the corresponding point pair, when the absolute value of the difference is 0 , the corresponding probability density value is the largest.
- the training device before using the pose parameters to project the map points to the dimension to which the sample images belong, can also calculate the pose parameters of the sample images based on the matched point pairs; here, BA (Bundle Adjustment) can be used to calculate The pose parameter is used to project the map point into the dimension to which the sample image belongs to obtain the projected point of the map point.
- BA Breast Adjustment
- the actual matching value of the non-matching point pair may also be set to a preset value, for example, the actual matching value of the non-matching point pair is set to 0.
- several sets of point pairs are formed by acquiring several image points from the sample image and several map points from the map data, and the several sets of point pairs include at least one set of the image points and the map.
- the matching point pairs matched between the points can generate samples for training the matching prediction model, and for each set of matching point pairs, the map points are projected to the dimension of the sample image by using the pose parameters of the sample image to obtain a map.
- the actual matching value of the matching point pair is determined based on the difference between the image point and the projected point, so the matching prediction model can learn the geometric features of the matching point pair during the training process, which is conducive to improving the matching prediction model. accuracy.
- FIG. 5 is a schematic flowchart of an embodiment of the visual positioning method of the present disclosure.
- the visual positioning method may include the following steps:
- Step S51 Using the image to be located and the map data to construct matching data to be identified.
- the matching data to be identified includes several groups of point pairs, and the two points of each group of point pairs respectively come from the image to be located and the map data.
- the dimension to which the image to be located and the map data belong may be 2-dimensional or 3-dimensional, which is not limited herein.
- the image to be positioned may be a two-dimensional image, or the image to be positioned may also be an RGB-D image, which is not limited here;
- the map data may be composed of a simple two-dimensional image, or may be composed of a three-dimensional point cloud map , or a combination of a two-dimensional image and a three-dimensional point cloud, which is not limited here.
- Step S52 Use the matching prediction model to perform prediction processing on several groups of point pairs to obtain the predicted matching values of the point pairs.
- the matching prediction model is a neural network model trained in advance through sample matching data.
- the matching prediction model may be obtained by training through the steps in any of the foregoing embodiments of the matching prediction model training method, wherein the training steps may refer to the steps in the foregoing embodiments, which will not be repeated here.
- the matching prediction model By using the matching prediction model to perform prediction processing on several groups of point pairs, the predicted matching values of the point pairs in the matching data to be identified can be obtained.
- the matching data to be identified is a bipartite graph, and the bipartite graph includes several groups of point pairs and connecting edges connecting each group of point pairs, and the matching prediction model includes a first point feature extraction corresponding to the dimension to which the image to be located belongs.
- the steps in the foregoing embodiments which will not be repeated here.
- Step S53 Determine the pose parameters of the imaging device of the image to be positioned based on the predicted matching value of the point pair.
- the point pair with a relatively high predicted matching value can be preferentially used to determine the pose parameters of the imaging device of the image to be positioned.
- the PnP (Perspective-n-Point) problem can be constructed by using n point pairs with relatively high predicted matching values, so as to solve the PnP problem by means such as EPnP (Efficient PnP), and then obtain the to-be-located problem.
- the pose parameters of the camera device of the image In another implementation scenario, several sets of point pairs may also be sorted in descending order of predicted matching values, and the previously preset number of sets of point pairs may be used to determine the pose parameters of the camera device of the image to be positioned.
- the first preset number can be set according to the actual situation, for example, point pairs whose predicted matching value is not 0 among the sorted groups of point pairs are used as the first preset number of point pairs; If the predicted matching value in the group point pair is greater than the lower limit value, the point pair is used as the first preset number of group point pairs; the first preset number can be set according to the actual application, which is not limited here.
- a method such as PROSAC (PROgressive SAmple Consensus, Progressive Consistent Sampling) can also be used to process the sorted point pairs to obtain the pose parameters of the camera device of the image to be positioned.
- the pose parameters of the camera device of the image to be positioned may include 6 degrees of freedom (DoF) of the camera device in the map coordinate system to which the map data belongs, including: pose, That is, the coordinates, and the deflection yaw (pitch angle) around the x-axis, the deflection pitch (yaw angle) around the y-axis, and the deflection roll (roll angle) around the z-axis.
- DoF degrees of freedom
- the matching data to be identified is constructed by using the image to be located and the map data, and the matching data to be identified includes several groups of point pairs, and the two points of each group of point pairs are respectively from the image to be located and the map data, so as to use the matching prediction
- the model performs prediction processing on several groups of point pairs to obtain the predicted matching value of the point pair, and then determines the pose parameters of the camera device of the image to be positioned based on the predicted matching value of the point pair, so the matching prediction model can be used in visual positioning to predict Establishing a matching relationship based on matching values between point pairs can help improve the accuracy and immediacy of visual positioning.
- FIG. 6 is a schematic diagram of a framework of an embodiment of a training apparatus 60 for matching prediction models of the present disclosure.
- the training device 60 for matching the prediction model includes a sample construction part 61, a prediction processing part 62, a loss determination part 63 and a parameter adjustment part 64, and the sample construction part 61 is configured to use the sample image and map data to construct sample matching data, wherein the sample matching The data includes several groups of point pairs and the actual matching value of each group of point pairs, and the two points of each group of point pairs are respectively from the sample image and the map data;
- the prediction processing part 62 is configured to use the matching prediction model to perform prediction processing on the several groups of point pairs , to obtain the predicted matching value of the point pair;
- the loss determination part 63 is configured to use the actual matching value and the predicted matching value to determine the loss value of the matching prediction model;
- the parameter adjustment part 64 is configured to use the loss value to adjust the parameters of the matching prediction model.
- the above solution can use the matching prediction model to establish a matching relationship, so that the matching prediction model can be used to predict the matching value between the point pairs in the visual positioning, so the point pair with high matching value can be preferentially sampled based on the predicted matching value, and then can It is beneficial to improve the accuracy and immediacy of visual positioning.
- the sample construction section 61 includes a point pair acquisition subsection configured to acquire several image points from the sample image and several map points from the map data to form several sets of point pairs; wherein the several sets of points The pair includes at least one set of matching point pairs that match between the included image points and map points, and the sample construction section 61 includes a first matching value determination subsection configured to, for each set of matching point pairs: use the pose parameters of the sample image Project the map point into the dimension to which the sample image belongs to obtain the projected point of the map point; and determine the actual matching value of the matching point pair based on the difference between the image point and the projected point.
- several groups of point pairs are formed by acquiring several image points from the sample image and several map points from the map data, and the several groups of point pairs include at least one group of the image points and the map.
- the matching point pairs matched between the points can generate samples for training the matching prediction model, and for each group of matching point pairs, the map points are projected to the dimension of the sample image by using the pose parameters of the sample image to obtain a map.
- the actual matching value of the matching point pair is determined based on the difference between the image point and the projected point, so the matching prediction model can learn the geometric features of the matching point pair during the training process, which is beneficial to improve the matching prediction model. accuracy.
- the sets of point pairs include at least one set of non-matching point pairs that do not match between the included image points and map points
- the sample construction section 61 includes a second matching value determination subsection configured to convert the non-matching points The actual match value of the point pair is set to the preset value.
- several groups of point pairs include at least one group of non-matching point pairs that do not match between the included image points and map points, and different from the matching point pairs, the actual matching value of the non-matching point pairs is set to a preset value. Set the value, which can help to improve the robustness of the matching prediction model.
- the point pair acquisition subsection includes an image point division section configured to divide the image points in the sample image into a first image point and a second image point, wherein the first image point exists in the map data with The matching map point, the second image point does not have a matching map point in the map data, the point pair acquisition subsection includes a first point pair acquisition section, and is configured to allocate a number of points from the map data for each first image point.
- the first map point, and the first image point and each first map point are respectively regarded as a first point pair, wherein the first map point includes a map point matching the first image point, and the point pair acquisition subsection includes The second point pair acquisition part is configured to allocate a number of second map points from the map data for each second image point, and respectively use the second image point and each second map point as a second point pair, point
- the pair acquisition subsection includes a point pair extraction section configured to extract several groups of point pairs from the first point pair and the second point pair.
- the image points in the sample image are divided into first image points and second image points, and the first image point has a matching map point in the map, and the second image point does not exist in the image data.
- There is an image point that matches it and for the first image point, a number of first map points are allocated from the map data, and the first image point and each first map point are respectively regarded as a first point pair, and the first map point contains map points matching the first image point, and for each second image point, assigns a number of second map points from the map data, and takes the second image point and each second map point as a second point respectively pair, and extract several groups of point pairs from the first point pair and the second point pair, so that several groups of point pairs that are abundant and include non-matching point pairs and matching point pairs can be constructed to be used for training matching prediction models. , so it can help to improve the accuracy of the matching prediction model.
- the first matching value determination subsection includes a pose calculation section configured to calculate pose parameters of the sample image based on the matching point pairs, and the first matching value determination subsection includes a projection section configured to utilize the pose The parameter projects the map point into the dimension to which the sample image belongs to obtain the projected point of the map point.
- the matching point pair to calculate the pose parameters of the sample image, and using the pose parameters to project the map points into the dimension to which the sample image belongs, the projection points of the map points are obtained, which can help to improve the projection.
- the accuracy of the difference between the point and the image point can be beneficial to improve the accuracy of the matching prediction model.
- the first matching value determination subsection includes a probability density conversion section configured to convert the difference into a probability density value using a preset probability distribution function as the actual matching value of the matching point pair
- the sample matching data is a bipartite graph, and the bipartite graph includes several groups of point pairs and connecting edges connecting each group of point pairs, and the connecting edges are marked with the actual matching values of the corresponding point pairs;
- the matching prediction model includes matching with the sample image.
- the first point feature extraction sub-model corresponding to the dimension to which the map data belongs, the second point feature extraction sub-model and the edge feature extraction sub-model corresponding to the dimension to which the map data belongs, the prediction processing part 62 includes a point feature extraction sub-part, which is configured to use the The first point feature extraction sub-model and the second point feature extraction sub-model perform feature extraction on the bipartite graph to obtain the first feature and the second feature.
- the prediction processing part 62 includes an edge feature extraction sub-part, which is configured to extract the sub-model by using the edge feature. Feature extraction is performed on the first feature and the second feature to obtain a third feature.
- the prediction processing part 62 includes a prediction sub-part configured to use the third feature to obtain the predicted matching value of the point pair corresponding to the connecting edge.
- the matching prediction model can more effectively perceive the spatial geometric structure of the matching, thereby improving the accuracy of the matching prediction model.
- the structure of the first point feature extraction sub-model and the second point feature extraction sub-model is any of the following: including at least one residual block, including at least one residual block and at least one spatial transformation network; and /or, the edge feature extraction sub-model includes at least one residual block.
- the structures of the first point feature extraction sub-model and the second point feature extraction sub-model are set to any one of the following: including at least one residual block, including at least one residual block and at least one spatial transformation network, and the edge feature extraction sub-model is set to include at least one residual block, so it can facilitate the optimization of the matching prediction model and improve the accuracy of the matching prediction model.
- the sets of point pairs include at least one set of matched point pairs that match between the included image points and map points and at least one set of non-matched point pairs that do not match between the included image points and map points
- the loss determination section 63 includes a first loss determination subsection configured to use the predicted matching value and the actual matching value of the matching point pair to determine a first loss value matching the prediction model
- the loss determination section 63 includes a second loss determination subsection, is configured to use the predicted matching value and the actual matching value of the non-matching point pair to determine the second loss value of the matching prediction model
- the loss determination section 63 includes a loss weighting subsection configured to weight the first loss value and the second loss value Process to get the loss value that matches the prediction model.
- the first loss value of the matching prediction model is determined by using the predicted matching value and the actual matching value of the matching point pair, and the predicted matching value and the actual loss value of the non-matching point pair are used to determine the matching prediction model.
- the second loss value so that the first loss value and the second loss value are weighted to obtain the loss value of the matching prediction model, so it can help the matching prediction model to effectively perceive the matching spatial geometry, thereby improving the matching prediction model. accuracy.
- the loss determination part 63 further includes a quantity statistics subsection, configured to count the first number of matched point pairs and the second number of non-matched point pairs, respectively; the first loss determination subsection is configured to use matching The difference between the predicted matching value and the actual matching value of the point pair, and the first number, determine the first loss value; the second loss determining subsection is configured to use the predicted matching value and the actual matching value of the non-matching point pair The difference between , and the second quantity, determines the second loss value.
- a quantity statistics subsection configured to count the first number of matched point pairs and the second number of non-matched point pairs, respectively; the first loss determination subsection is configured to use matching The difference between the predicted matching value and the actual matching value of the point pair, and the first number, determine the first loss value; the second loss determining subsection is configured to use the predicted matching value and the actual matching value of the non-matching point pair The difference between , and the second quantity, determines the second loss value.
- the difference between the predicted matching value and the actual matching value of the matching point pair, and the first number are used, Determining the first loss value, and using the difference between the predicted matching value and the actual matching value of the unmatched point pair, and the second quantity, determining the second loss value can help to improve the accuracy of the loss value of the matching prediction model, Thus, the accuracy of the matching prediction model can be improved.
- the dimension to which the sample image belongs is 2D or 3D
- the dimension to which the map data belongs is 2D or 3D
- a matching prediction model for 2-2D can be trained, or a matching prediction model for 2-3D can be trained, or a matching prediction model can be trained.
- the matching prediction model for 3D-3D is obtained by training, so that the applicable range of the matching prediction model can be improved.
- FIG. 7 is a schematic diagram of a frame of an embodiment of a visual positioning device 70 of the present disclosure.
- the visual positioning device 70 includes a data construction part 71, a prediction processing part 72 and a parameter determination part 73.
- the data construction part 71 is configured to use the image to be located and the map data to construct matching data to be identified, wherein the matching data to be identified includes several groups of points Right, the two points of each group of point pairs are respectively from the image to be located and the map data;
- the prediction processing part 72 is configured to use the matching prediction model to perform prediction processing on several groups of point pairs to obtain the predicted matching values of the point pairs;
- the parameter determination part 73 It is configured to determine the pose parameters of the imaging device of the image to be positioned based on the predicted matching value of the point pair.
- the matching prediction model can be used to establish the matching relationship, so that the matching prediction model can be used to predict the matching value between the point pairs in the visual positioning to establish the matching relationship, which can help to improve the accuracy and immediacy of the visual positioning.
- the parameter determination section 73 includes a point pair sorting subsection configured to sort several groups of point pairs in descending order of predicted matching values, and the parameter determination section 73 further includes a parameter determination subsection configured as Using the previously preset number of point pairs, the pose parameters of the imaging device of the image to be positioned are determined.
- a "part" may be a part of a circuit, a part of a processor, a part of a program or software, etc., of course, a unit, a module or a non-modularity.
- the matching prediction model is obtained by training the matching prediction model training device in any of the above-mentioned embodiments of the matching prediction model training device.
- performing visual positioning through the matching prediction model obtained by the training device for matching prediction models in any of the above-mentioned embodiments of the training device for matching prediction models can improve the accuracy and immediacy of visual positioning.
- FIG. 8 is a schematic diagram of a framework of an embodiment of an electronic device 80 of the present disclosure.
- the electronic device 80 includes a mutually coupled memory 81 and a processor 82, and the processor 82 is configured to execute program instructions stored in the memory 81 to implement the steps in any of the above-mentioned embodiments of the training method for matching the prediction model, or to implement any of the above-mentioned training methods. Steps in an embodiment of a visual positioning method.
- the electronic device 80 may include, but is not limited to, mobile devices such as a mobile phone and a matching computer, which are not limited herein.
- the processor 82 is configured to control itself and the memory 81 to implement the steps in any of the above-mentioned embodiments of the matching prediction model training method, or to implement the steps in any of the above-mentioned embodiments of the visual positioning method.
- the processor 82 may also be referred to as a CPU (Central Processing Unit, central processing unit).
- the processor 82 may be an integrated circuit chip with signal processing capability.
- the processor 82 may also be a general-purpose processor, a digital signal processor (Digital Signal Processor, DSP), an application specific integrated circuit (Application Specific Integrated Circuit, ASIC), a field programmable gate array (Field-Programmable Gate Array, FPGA) or other Programmable logic devices, discrete gate or transistor logic devices, discrete hardware components.
- a general purpose processor may be a microprocessor or the processor may be any conventional processor or the like.
- the processor 82 may be jointly implemented by an integrated circuit chip.
- the above solution can use the matching prediction model to establish a matching relationship, so that the matching prediction model can be used in the visual positioning to predict the matching value between the point pairs, so the point pair with the high matching value can be preferentially sampled based on the predicted matching value, and then the matching value can be sampled. It is beneficial to improve the accuracy and immediacy of visual positioning.
- FIG. 9 is a schematic diagram of a framework of an embodiment of the disclosed computer-readable storage medium 90 .
- the computer-readable storage medium 90 stores program instructions 901 that can be executed by the processor, and the program instructions 901 are used to implement the steps in any of the above-mentioned embodiments of the training method for matching prediction models, or to implement any of the above-mentioned embodiments of the visual positioning method. A step of.
- the above solution can use the matching prediction model to establish a matching relationship, so that the matching prediction model can be used in the visual positioning to predict the matching value between the point pairs, so the point pair with the high matching value can be preferentially sampled based on the predicted matching value, and then the matching value can be sampled. It is beneficial to improve the accuracy and immediacy of visual positioning.
- the disclosed method and apparatus may be implemented in other manners.
- the device implementations described above are only illustrative.
- the division of modules or units is only a logical function division. In actual implementation, there may be other divisions.
- units or components may be combined or integrated. to another system, or some features can be ignored, or not implemented.
- the shown or discussed mutual coupling or direct coupling or communication connection may be through some interfaces, indirect coupling or communication connection of devices or units, which may be in electrical, mechanical or other forms.
- Units described as separate components may or may not be physically separated, and components shown as units may or may not be physical units, that is, may be located in one place, or may be distributed over network units. Some or all of the units may be selected according to actual needs to achieve the purpose of the solution in this implementation manner.
- each functional unit in each embodiment of the present disclosure may be integrated into one processing unit, or each unit may exist physically alone, or two or more units may be integrated into one unit.
- the above-mentioned integrated units may be implemented in the form of hardware, or may be implemented in the form of software functional units.
- the integrated unit if implemented as a software functional unit and sold or used as a stand-alone product, may be stored in a computer-readable storage medium.
- the technical solutions of the present disclosure can be embodied in the form of software products in essence, or the part that contributes to the prior art, or all or part of the technical solutions, and the computer software product is stored in a storage medium , including several instructions for causing a computer device (which may be a personal computer, a server, or a network device, etc.) or a processor (processor) to execute all or part of the steps of the methods of the various embodiments of the present disclosure.
- the aforementioned storage medium includes: U disk, mobile hard disk, Read-Only Memory (ROM, Read-Only Memory), Random Access Memory (RAM, Random Access Memory), magnetic disk or optical disk and other media that can store program codes .
- the matching prediction model can be used to establish a matching relationship, so that the matching prediction model can be used to predict the matching value between point pairs in visual positioning, so that the point pair with high matching value can be preferentially sampled based on the predicted matching value. , and establish a matching relationship, which can help to improve the accuracy and immediacy of visual positioning.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- General Engineering & Computer Science (AREA)
- Software Systems (AREA)
- Artificial Intelligence (AREA)
- Evolutionary Computation (AREA)
- Computing Systems (AREA)
- Databases & Information Systems (AREA)
- Mathematical Physics (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Biophysics (AREA)
- Molecular Biology (AREA)
- Computational Linguistics (AREA)
- Biomedical Technology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Medical Informatics (AREA)
- Remote Sensing (AREA)
- Multimedia (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Evolutionary Biology (AREA)
- Image Analysis (AREA)
Abstract
Description
Claims (18)
- 一种匹配预测模型的训练方法,包括:A training method for matching predictive models, including:利用样本图像和地图数据,构建样本匹配数据,其中,所述样本匹配数据包括若干组点对以及每组点对的实际匹配值,每组点对的两个点分别来自所述样本图像和所述地图数据;Using the sample image and map data, construct sample matching data, wherein the sample matching data includes several groups of point pairs and the actual matching value of each group of point pairs, and the two points of each group of point pairs come from the sample image and the the map data;利用匹配预测模型对所述若干组点对进行预测处理,得到所述点对的预测匹配值;Use a matching prediction model to perform prediction processing on the several groups of point pairs to obtain the predicted matching values of the point pairs;利用所述实际匹配值和所述预测匹配值,确定所述匹配预测模型的损失值;Using the actual matching value and the predicted matching value, determining the loss value of the matching prediction model;利用所述损失值,调整所述匹配预测模型的参数。Using the loss value, the parameters of the matching prediction model are adjusted.
- 根据权利要求1所述的训练方法,其中,所述利用样本图像和地图数据,构建样本匹配数据包括:The training method according to claim 1, wherein, using sample images and map data to construct sample matching data comprises:从所述样本图像中获取若干图像点,以及从所述地图数据中获取若干地图点,以组成若干组点对;其中,所述若干组点对包括至少一组所包含的图像点和地图点之间匹配的匹配点对;Several image points are obtained from the sample image, and several map points are obtained from the map data to form several sets of point pairs; wherein the several sets of point pairs include at least one set of included image points and map points Matching point pairs that match between;对于每组所述匹配点对:利用所述样本图像的位姿参数将所述地图点投影至所述样本图像所属的维度中,得到所述地图点的投影点;并基于所述图像点和所述投影点之间的差异,确定所述匹配点对的实际匹配值。For each set of matched point pairs: project the map point into the dimension to which the sample image belongs by using the pose parameters of the sample image to obtain the projected point of the map point; and based on the image point and The difference between the projected points determines the actual matching value of the matching point pair.
- 根据权利要求2所述的训练方法,其中,所述若干组点对包括至少一组所包含的图像点和地图点之间不匹配的非匹配点对,所述利用样本图像和地图数据,构建样本匹配数据还包括:The training method according to claim 2, wherein the several groups of point pairs include at least one group of non-matching point pairs that do not match between the included image points and map points, and the sample image and map data are used to construct the Sample match data also includes:将所述非匹配点对的实际匹配值设置为预设数值。The actual matching value of the non-matching point pair is set as a preset value.
- 根据权利要求2或3所述的训练方法,其中,所述从所述样本图像中获取若干图像点,以及从所述地图数据中获取若干地图点,以组成若干组点对,包括:The training method according to claim 2 or 3, wherein the acquiring several image points from the sample image and acquiring several map points from the map data to form several groups of point pairs, including:将所述样本图像中的图像点划分为第一图像点和第二图像点,其中,所述第一图像点在所述地图数据中存在与其匹配的所述地图点,所述第二图像点在所述地图数据中不存在与其匹配的所述地图点;Dividing the image points in the sample image into a first image point and a second image point, wherein the first image point has the map point matching it in the map data, and the second image point there is no matching said map point in said map data;对于每一所述第一图像点,从所述地图数据中分配若干第一地图点,并分别将所述第一图像点与每一所述第一地图点作为一第一点对,其中,所述第一地图点中包含与所述第一图像点匹配的所述地图点;以及,For each of the first image points, a number of first map points are allocated from the map data, and the first image point and each of the first map points are respectively used as a first point pair, wherein, The first map point includes the map point that matches the first image point; and,对于每一所述第二图像点,从所述地图数据中分配若干第二地图点,并分别将所述第二图像点与每一所述第二地图点作为一第二点对;for each of the second image points, assigning a plurality of second map points from the map data, and respectively using the second image point and each of the second map points as a second point pair;从所述第一点对和所述第二点对中抽取得到若干组点对。Several sets of point pairs are extracted from the first point pair and the second point pair.
- 根据权利要求2至4任一项所述的训练方法,其中,所述利用所述样本图像的位姿参数将所述地图点投影至所述样本图像所属的维度中,得到所述地图点的投影点包括:The training method according to any one of claims 2 to 4, wherein the map point is projected into the dimension to which the sample image belongs by using the pose parameters of the sample image to obtain the map point's Projection points include:基于所述匹配点对,计算所述样本图像的位姿参数;Based on the matched point pair, calculate the pose parameter of the sample image;利用所述位姿参数将所述地图点投影至所述样本图像所属的维度中,得到所述地图点的投影点;Using the pose parameter to project the map point into the dimension to which the sample image belongs, to obtain the projected point of the map point;和/或,所述基于所述图像点和所述投影点之间的差异,确定所述匹配点对的实际匹配值包括:And/or, determining the actual matching value of the matching point pair based on the difference between the image point and the projection point includes:利用预设概率分布函数将所述差异转换为概率密度值,作为所述匹配点对的实际匹配值。The difference is converted into a probability density value using a preset probability distribution function as the actual matching value of the matching point pair.
- 根据权利要求1至5任一项所述的训练方法,其中,所述样本匹配数据为二分图,所述二分图包括若干组点对和连接每组点对的连接边,且所述连接边标注有对应所述点对的实际匹配值;所述匹配预测模型包括与所述样本图像所属的维度对应的第一点特征提取子模型、与所述地图数据所属的维度对应的第二点特征提取子模型以及边特征提取子模型;The training method according to any one of claims 1 to 5, wherein the sample matching data is a bipartite graph, and the bipartite graph includes several groups of point pairs and connecting edges connecting each group of point pairs, and the connecting edges The actual matching value corresponding to the point pair is marked; the matching prediction model includes a first point feature extraction sub-model corresponding to the dimension to which the sample image belongs, and a second point feature corresponding to the dimension to which the map data belongs Extract sub-models and edge feature extraction sub-models;所述利用匹配预测模型对所述若干组点对进行预测处理,得到所述点对的预测匹配值包括:The performing prediction processing on the several groups of point pairs by using the matching prediction model, and obtaining the predicted matching values of the point pairs includes:分别利用所述第一点特征提取子模型和所述第二点特征提取子模型对所述二分图进行特征提取,得到第一特征和第二特征;Using the first point feature extraction sub-model and the second point feature extraction sub-model to perform feature extraction on the bipartite graph, respectively, to obtain the first feature and the second feature;利用所述边特征提取子模型对所述第一特征和所述第二特征进行特征提取,得到第三特征;Using the edge feature extraction sub-model to perform feature extraction on the first feature and the second feature to obtain a third feature;利用所述第三特征,得到所述连接边对应的点对的预测匹配值。Using the third feature, the predicted matching value of the point pair corresponding to the connecting edge is obtained.
- 根据权利要求6所述的训练方法,其中,所述第一点特征提取子模型和所述第二点特征提取子模型的结构为以下任一种:包括至少一个残差块,包括至少一个残差块和至少一个空间变换网络;The training method according to claim 6, wherein the structure of the first point feature extraction sub-model and the second point feature extraction sub-model is any of the following: including at least one residual block, including at least one residual block difference blocks and at least one spatial transformation network;和/或,所述边特征提取子模型包括至少一个残差块。And/or, the edge feature extraction sub-model includes at least one residual block.
- 根据权利要求1至7任一项所述的训练方法,其中,所述若干组点对包括至少一组所包含的图像点和地图点之间匹配的匹配点对和至少一组所包含的图像点和地图点之间不匹配的非匹配点对;The training method according to any one of claims 1 to 7, wherein the several sets of point pairs include at least one set of matched point pairs matched between the included image points and map points and at least one set of included images Non-matching point pairs that do not match between points and map points;所述利用所述实际匹配值和所述预测匹配值,确定所述匹配预测模型的损失值包括:The determining the loss value of the matching prediction model by using the actual matching value and the predicted matching value includes:利用所述匹配点对的所述预测匹配值和所述实际匹配值,确定所述匹配预测模型的第一损失值;Using the predicted matching value and the actual matching value of the matching point pair to determine a first loss value of the matching prediction model;并利用所述非匹配点对的所述预测匹配值和所述实际匹配值,确定所述匹配预测模型的第二损失值;and using the predicted matching value and the actual matching value of the non-matching point pair to determine the second loss value of the matching prediction model;对所述第一损失值和所述第二损失值进行加权处理,得到所述匹配预测模型的损失值。The first loss value and the second loss value are weighted to obtain the loss value of the matching prediction model.
- 根据权利要求8所述的训练方法,其中,所述利用所述匹配点对的所述预测匹配值和所述实际匹配值,确定所述匹配预测模型的第一损失值之前,所述方法还包括:The training method according to claim 8, wherein before the first loss value of the matching prediction model is determined by using the predicted matching value and the actual matching value of the matching point pair, the method further include:分别统计所述匹配点对的第一数量,以及所述非匹配点对的第二数量;respectively count the first number of the matching point pairs and the second number of the non-matching point pairs;所述利用所述匹配点对的所述预测匹配值和所述实际匹配值,确定所述匹配预测模型的第一损失值包括:The determining the first loss value of the matching prediction model by using the predicted matching value and the actual matching value of the matching point pair includes:利用所述匹配点对的所述预测匹配值和所述实际匹配值之间的差值,以及所述第一数量,确定所述第一损失值;Using the difference between the predicted matching value and the actual matching value of the matching point pair, and the first number, determining the first loss value;所述利用所述非匹配点对的所述预测匹配值和所述实际匹配值,确定所述匹配预测模型的第二损失值包括:Using the predicted matching value and the actual matching value of the non-matching point pair to determine the second loss value of the matching prediction model includes:利用所述非匹配点对的所述预测匹配值和所述实际匹配值之间的差值,以及所述第二数量,确定所述第二损失值。The second loss value is determined using the difference between the predicted match value and the actual match value for the pair of non-matching points, and the second number.
- 根据权利要求1至9任一项所述的训练方法,其中,所述样本图像所属的维度为2维或3维,所述地图数据所属的维度为2维或3维。The training method according to any one of claims 1 to 9, wherein the dimension to which the sample image belongs is 2-dimensional or 3-dimensional, and the dimension to which the map data belongs is 2-dimensional or 3-dimensional.
- 一种视觉定位方法,包括:A visual positioning method comprising:利用待定位图像和地图数据,构建待识别匹配数据,其中,所述待识别匹配数据包括若干组点对,每组点对的两个点分别来自所述待定位图像和所述地图数据;Using the to-be-located image and the map data, the to-be-identified matching data is constructed, wherein the to-be-identified matching data includes several groups of point pairs, and the two points of each group of point pairs are respectively from the to-be-located image and the map data;利用匹配预测模型对所述若干组点对进行预测处理,得到所述点对的预测匹配值;Use a matching prediction model to perform prediction processing on the several groups of point pairs to obtain the predicted matching values of the point pairs;基于所述点对的预测匹配值,确定所述待定位图像的摄像器件的位姿参数。Based on the predicted matching value of the point pair, the pose parameter of the imaging device of the image to be positioned is determined.
- 根据权利要求11所述的视觉定位方法,其中,所述基于所述点对的预测匹配值,确定所述待定位图像的摄像器件的位姿参数,包括:The visual positioning method according to claim 11, wherein the determining the pose parameters of the camera device of the image to be positioned based on the predicted matching value of the point pair comprises:将所述若干组点对按照所述预测匹配值从高到低的顺序进行排序;sorting the groups of point pairs in descending order of the predicted matching values;利用前预设数量组所述点对,确定所述待定位图像的摄像器件的位姿参数。The pose parameters of the imaging device of the to-be-located image are determined by using the previously preset number of sets of the point pairs.
- 根据权利要求11或12所述的视觉定位方法,其中,所述匹配预测模型是利用权利要求1至10任一项所述的匹配预测模型的训练方法得到的。The visual positioning method according to claim 11 or 12, wherein the matching prediction model is obtained by using the training method of the matching prediction model according to any one of claims 1 to 10.
- 一种匹配预测模型的训练装置,包括:A training device for matching a prediction model, comprising:样本构建模块,用于利用样本图像和地图数据,构建样本匹配数据,其中,所述样本匹配数据包括若干组点对以及每组点对的实际匹配值,每组点对的两个点分别来自所述样本图像和所述地图数据;The sample building module is used to construct sample matching data by using sample images and map data, wherein the sample matching data includes several groups of point pairs and the actual matching value of each group of point pairs, and the two points of each group of point pairs come from the sample image and the map data;预测处理部分,配置为利用匹配预测模型对所述若干组点对进行预测处理,得到所述点对的预测匹配值;a prediction processing part, configured to perform prediction processing on the several groups of point pairs by using a matching prediction model to obtain the predicted matching values of the point pairs;损失确定部分,配置为利用所述实际匹配值和所述预测匹配值,确定所述匹配预测模型的损失值;a loss determination part configured to use the actual matching value and the predicted matching value to determine the loss value of the matching prediction model;参数调整部分,配置为利用所述损失值,调整所述匹配预测模型的参数。The parameter adjustment part is configured to use the loss value to adjust the parameters of the matching prediction model.
- 一种视觉定位装置,其中,A visual positioning device, wherein,待定位图像和地图数据,构建待识别匹配数据,其中,所述待识别匹配数据包括若干组点对,每组点对的两个点分别来自所述待定位图像和所述地图数据;The to-be-located image and map data, and the to-be-identified matching data is constructed, wherein the to-be-identified matching data includes several sets of point pairs, and the two points of each set of point pairs are respectively from the to-be-located image and the map data;预测处理部分,配置为利用匹配预测模型对所述若干组点对进行预测处理,得到所述点对的预测匹配值;a prediction processing part, configured to perform prediction processing on the several groups of point pairs by using a matching prediction model to obtain the predicted matching values of the point pairs;参数确定部分,配置为基于所述点对的预测匹配值,确定所述待定位图像的摄像器件的位姿参数。The parameter determination part is configured to determine the pose parameter of the imaging device of the to-be-positioned image based on the predicted matching value of the point pair.
- 一种电子设备,包括相互耦接的存储器和处理器,所述处理器用于执行所述存储器中存储的程序指令,以实现权利要求1至10任一项所述的匹配预测模型的训练方法,或权利要求11至13任一项所述的视觉定位方法。An electronic device, comprising a mutually coupled memory and a processor, the processor is configured to execute program instructions stored in the memory to implement the training method for a matching prediction model according to any one of claims 1 to 10, Or the visual positioning method according to any one of claims 11 to 13.
- 一种计算机可读存储介质,其上存储有程序指令,所述程序指令被处理器执行时实现权利要求1至10任一项所述的匹配预测模型的训练方法,或权利要求11至13任一项所述的视觉定位方法。A computer-readable storage medium on which program instructions are stored, and when the program instructions are executed by a processor, the training method of the matching prediction model according to any one of claims 1 to 10, or any one of claims 11 to 13 is realized. A described visual localization method.
- 一种计算机程序,包括计算机可读代码,在所述计算机可读代码在电子设备中运行,被所述电子设备中的处理器执行的情况下,实现权利要求1至10任一项所述的匹配预测模型的训练方法,或权利要求11至13任一项所述的视觉定位方法。A computer program, comprising computer-readable codes, in the case that the computer-readable codes are executed in an electronic device and executed by a processor in the electronic device, to implement the method described in any one of claims 1 to 10 The training method of the matching prediction model, or the visual positioning method according to any one of claims 11 to 13.
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020227003201A KR20220051162A (en) | 2020-10-16 | 2021-03-22 | Visual positioning methods, training methods for related models, and related devices and devices |
JP2021578181A JP7280393B2 (en) | 2020-10-16 | 2021-03-22 | Visual positioning method, related model training method and related device and equipment |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202011110569.5 | 2020-10-16 | ||
CN202011110569.5A CN112328715B (en) | 2020-10-16 | 2020-10-16 | Visual positioning method, training method of related model, related device and equipment |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2022077863A1 true WO2022077863A1 (en) | 2022-04-21 |
Family
ID=74313967
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2021/082198 WO2022077863A1 (en) | 2020-10-16 | 2021-03-22 | Visual positioning method, and method for training related model, related apparatus, and device |
Country Status (5)
Country | Link |
---|---|
JP (1) | JP7280393B2 (en) |
KR (1) | KR20220051162A (en) |
CN (1) | CN112328715B (en) |
TW (1) | TW202217662A (en) |
WO (1) | WO2022077863A1 (en) |
Families Citing this family (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112328715B (en) * | 2020-10-16 | 2022-06-03 | 浙江商汤科技开发有限公司 | Visual positioning method, training method of related model, related device and equipment |
CN113240656B (en) * | 2021-05-24 | 2023-04-07 | 浙江商汤科技开发有限公司 | Visual positioning method and related device and equipment |
CN113822916B (en) * | 2021-08-17 | 2023-09-15 | 北京大学 | Image matching method, device, equipment and readable storage medium |
CN116529590A (en) * | 2021-11-30 | 2023-08-01 | 宁德时代新能源科技股份有限公司 | Machine vision detection method, detection device and detection system thereof |
CN114998600B (en) * | 2022-06-17 | 2023-07-25 | 北京百度网讯科技有限公司 | Image processing method, training method, device, equipment and medium for model |
CN117351306B (en) * | 2023-12-04 | 2024-03-22 | 齐鲁空天信息研究院 | Training method, determining method and device for three-dimensional point cloud projection pose solver |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20090142029A1 (en) * | 2007-12-03 | 2009-06-04 | Institute For Information Industry | Motion transition method and system for dynamic images |
CN110274598A (en) * | 2019-06-24 | 2019-09-24 | 西安工业大学 | A kind of robot monocular vision robust location estimation method |
CN111260726A (en) * | 2020-02-07 | 2020-06-09 | 北京三快在线科技有限公司 | Visual positioning method and device |
CN111627050A (en) * | 2020-07-27 | 2020-09-04 | 杭州雄迈集成电路技术股份有限公司 | Training method and device for target tracking model |
CN112328715A (en) * | 2020-10-16 | 2021-02-05 | 浙江商汤科技开发有限公司 | Visual positioning method, training method of related model, related device and equipment |
Family Cites Families (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102236798B (en) * | 2011-08-01 | 2012-12-05 | 清华大学 | Image matching method and device |
US9989969B2 (en) * | 2015-01-19 | 2018-06-05 | The Regents Of The University Of Michigan | Visual localization within LIDAR maps |
US9934587B2 (en) * | 2016-06-30 | 2018-04-03 | Daqri, Llc | Deep image localization |
CN107967457B (en) * | 2017-11-27 | 2024-03-19 | 全球能源互联网研究院有限公司 | Site identification and relative positioning method and system adapting to visual characteristic change |
CN109658445A (en) * | 2018-12-14 | 2019-04-19 | 北京旷视科技有限公司 | Network training method, increment build drawing method, localization method, device and equipment |
CN110009722A (en) | 2019-04-16 | 2019-07-12 | 成都四方伟业软件股份有限公司 | Three-dimensional rebuilding method and device |
CN110095752B (en) * | 2019-05-07 | 2021-08-10 | 百度在线网络技术(北京)有限公司 | Positioning method, apparatus, device and medium |
CN110473259A (en) * | 2019-07-31 | 2019-11-19 | 深圳市商汤科技有限公司 | Pose determines method and device, electronic equipment and storage medium |
CN111508019A (en) * | 2020-03-11 | 2020-08-07 | 上海商汤智能科技有限公司 | Target detection method, training method of model thereof, and related device and equipment |
CN111476251A (en) | 2020-03-26 | 2020-07-31 | 中国人民解放军战略支援部队信息工程大学 | Remote sensing image matching method and device |
CN111414968B (en) | 2020-03-26 | 2022-05-03 | 西南交通大学 | Multi-mode remote sensing image matching method based on convolutional neural network characteristic diagram |
CN111538855B (en) * | 2020-04-29 | 2024-03-08 | 浙江商汤科技开发有限公司 | Visual positioning method and device, electronic equipment and storage medium |
CN111627065B (en) * | 2020-05-15 | 2023-06-20 | Oppo广东移动通信有限公司 | Visual positioning method and device and storage medium |
CN111652929A (en) * | 2020-06-03 | 2020-09-11 | 全球能源互联网研究院有限公司 | Visual feature identification and positioning method and system |
-
2020
- 2020-10-16 CN CN202011110569.5A patent/CN112328715B/en active Active
-
2021
- 2021-03-22 JP JP2021578181A patent/JP7280393B2/en active Active
- 2021-03-22 KR KR1020227003201A patent/KR20220051162A/en unknown
- 2021-03-22 WO PCT/CN2021/082198 patent/WO2022077863A1/en active Application Filing
- 2021-08-30 TW TW110132126A patent/TW202217662A/en unknown
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20090142029A1 (en) * | 2007-12-03 | 2009-06-04 | Institute For Information Industry | Motion transition method and system for dynamic images |
CN110274598A (en) * | 2019-06-24 | 2019-09-24 | 西安工业大学 | A kind of robot monocular vision robust location estimation method |
CN111260726A (en) * | 2020-02-07 | 2020-06-09 | 北京三快在线科技有限公司 | Visual positioning method and device |
CN111627050A (en) * | 2020-07-27 | 2020-09-04 | 杭州雄迈集成电路技术股份有限公司 | Training method and device for target tracking model |
CN112328715A (en) * | 2020-10-16 | 2021-02-05 | 浙江商汤科技开发有限公司 | Visual positioning method, training method of related model, related device and equipment |
Also Published As
Publication number | Publication date |
---|---|
TW202217662A (en) | 2022-05-01 |
JP7280393B2 (en) | 2023-05-23 |
KR20220051162A (en) | 2022-04-26 |
JP2023502819A (en) | 2023-01-26 |
CN112328715A (en) | 2021-02-05 |
CN112328715B (en) | 2022-06-03 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2022077863A1 (en) | Visual positioning method, and method for training related model, related apparatus, and device | |
CN110322453B (en) | 3D point cloud semantic segmentation method based on position attention and auxiliary network | |
CN108764041B (en) | Face recognition method for lower shielding face image | |
CN110276768B (en) | Image segmentation method, image segmentation device, image segmentation apparatus, and medium | |
CN111291768B (en) | Image feature matching method and device, equipment and storage medium | |
CN116188805B (en) | Image content analysis method and device for massive images and image information network | |
CN114186632B (en) | Method, device, equipment and storage medium for training key point detection model | |
CN111310821B (en) | Multi-view feature fusion method, system, computer equipment and storage medium | |
CN105654483A (en) | Three-dimensional point cloud full-automatic registration method | |
CN111831844A (en) | Image retrieval method, image retrieval device, image retrieval apparatus, and medium | |
CN113298870B (en) | Object posture tracking method and device, terminal equipment and storage medium | |
CN112364881B (en) | Advanced sampling consistency image matching method | |
CN116503575A (en) | Method for graphically generating high-quality building model and electronic equipment | |
CN114332125A (en) | Point cloud reconstruction method and device, electronic equipment and storage medium | |
WO2021051562A1 (en) | Facial feature point positioning method and apparatus, computing device, and storage medium | |
CN111597367B (en) | Three-dimensional model retrieval method based on view and hash algorithm | |
CN110060290B (en) | Binocular parallax calculation method based on 3D convolutional neural network | |
CN111862176A (en) | Three-dimensional oral cavity point cloud orthodontic front and back accurate registration method based on palatine fold | |
CN111192302A (en) | Feature matching method based on motion smoothness and RANSAC algorithm | |
CN111383353A (en) | Fractured bone model registration method based on Gaussian mixture model and contour descriptor | |
KR20230098058A (en) | Three-dimensional data augmentation method, model training and detection method, device, and autonomous vehicle | |
CN113487713B (en) | Point cloud feature extraction method and device and electronic equipment | |
CN114154222A (en) | Image segmentation method for BIM (building information modeling), electronic equipment and storage medium | |
CN111145081B (en) | Three-dimensional model view projection method and system based on spatial volume characteristics | |
Tuyet et al. | Adaptive content-based medical image retrieval based on local features extraction in shearlet domain |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
ENP | Entry into the national phase |
Ref document number: 2021578181 Country of ref document: JP Kind code of ref document: A |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 21878919 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 21878919 Country of ref document: EP Kind code of ref document: A1 |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 21878919 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 081123) |