CN112562081A - Visual map construction method for visual layered positioning - Google Patents
Visual map construction method for visual layered positioning Download PDFInfo
- Publication number
- CN112562081A CN112562081A CN202110175262.1A CN202110175262A CN112562081A CN 112562081 A CN112562081 A CN 112562081A CN 202110175262 A CN202110175262 A CN 202110175262A CN 112562081 A CN112562081 A CN 112562081A
- Authority
- CN
- China
- Prior art keywords
- frame
- visual
- superpoint
- points
- key
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T17/00—Three dimensional [3D] modelling, e.g. data description of 3D objects
- G06T17/05—Geographic models
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/70—Determining position or orientation of objects or cameras
- G06T7/73—Determining position or orientation of objects or cameras using feature-based methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2200/00—Indexing scheme for image data processing or generation, in general
- G06T2200/04—Indexing scheme for image data processing or generation, in general involving 3D image data
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Software Systems (AREA)
- General Physics & Mathematics (AREA)
- General Health & Medical Sciences (AREA)
- Artificial Intelligence (AREA)
- Computational Linguistics (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- Biomedical Technology (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- General Engineering & Computer Science (AREA)
- Biophysics (AREA)
- Mathematical Physics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Health & Medical Sciences (AREA)
- Geometry (AREA)
- Remote Sensing (AREA)
- Computer Graphics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Image Analysis (AREA)
Abstract
The invention discloses a visual map construction method for visual layered positioning, which comprises the following steps: 1) sequentially acquiring binocular image frame data collected by a binocular camera and determining a motion track; 2) respectively extracting a NetVLAD global descriptor, a SuperPoint characteristic point and a local descriptor from the image frame; 3) finding out the matching key points of the characteristic points of each frame of image and restoring the 3D position of each characteristic point in an incremental manner; 4) determining the optimal co-view key frame of each frame according to the 2D observation of the 3D feature points of the frame; 5) and generating visual map information finally used for visual layered positioning. The visual map construction method provided by the invention combines deep learning characteristics, the descriptive performance of the visual map is enhanced, and the generated map for visual positioning contains multi-level description information and can be used for visual global robust positioning with consistent coordinate systems.
Description
Technical Field
The invention relates to the technical field of computer vision, in particular to a visual map construction method for visual layered positioning.
Background
In recent years, with the continuous development of computer vision, SLAM technology is widely used in various fields, virtual reality, augmented reality, robotics, unmanned planes, and the like. With the continuous development of computer hardware, the real-time processing of visual information becomes possible, and the real-time positioning and map construction by using the visual information greatly reduces the price of intelligent robot products while improving the information acquisition amount.
However, most visual SLAM systems focus on online localization estimation, do not have localization function under global coordinate system, and most SLAM system outputs do not include representation of visual maps. The traditional method is based on matching of a bag-of-words model and traditional feature points, and has limited environment adaptation capability and low positioning success rate.
Disclosure of Invention
In view of the above, the invention provides a visual map construction method for visual layered positioning, which combines with deep learning domain knowledge and visual map construction method and expression form, and can construct a global consistent map and be used for visual-based layered global consistent positioning.
The invention adopts the following technical scheme: a visual map construction method for visual layered positioning comprises the following steps:
(1) sequentially acquiring binocular image frames acquired by the binocular camera and recording the motion trail of the binocular camera to obtain a binocular image sequence containing the motion trailWherein, in the step (A),is shown asThe left frame image to which the frame corresponds,is shown asThe right frame image to which the frame corresponds,is an external parameter of the binocular camera,is as followsThe pose of the frame left camera relative to the world coordinate system, and N is the frame number of the binocular image; using the first frame as a key frame, and aiming at the binocular image sequenceSelecting subsequent key frames to form a key frame sequence,Selecting a binocular image frame number N;
(2) respectively extracting NetVLAD global descriptors from key framesSuperPoint feature pointsAnd corresponding local descriptorsAnd according to SuperPoint feature pointsThe response values are sorted from high to low, the first 2000 feature points and the corresponding local descriptors are reserved, and for the secondObtaining two triple description information by one key frame(ii) a The NetVLAD global descriptorIs composed ofFeature vectors of dimensions, the local descriptorsIs composed ofA vector of dimensions;
(3) finding out SuperPoint feature points of key framesAnd incrementally restoring the 3D position of each SuperPoint feature point;
(4) Traversing each of the steps (3)Statistical observation of 3D positionAll key frames of the points are sorted in descending order, and the first 5 key frames are taken as the optimal common-view key frame of the key frames(ii) a And finally, corresponding the key frame sequence and each SuperPoint characteristic point to a 3D positionAs visual map information.
Further, the selection of the subsequent key frame needs to satisfy the following condition: the Euclidean distance between adjacent left frame images or right frame images is larger than 0.3 meter, and the rotation angle of the adjacent left frame images or right frame images is larger than 3 degrees.
Further, the step (3) comprises the following sub-steps:
(3.1) describing information for a certain tripleSuperPoint feature points in (1)Selecting the characteristic points which can observe SuperPoint simultaneouslyCandidate frame set ofWherein A is selected from the key frames, satisfying: pose of keyframes in the set of candidate frames CAndthe movement translation distance of the rotating shaft is less than 10m, and the rotating angle is less than 45 degrees;
(3.2) traverse each frame in the candidate frame set C, pairThe local descriptor corresponding to each feature point inCalculatingAndand obtaining two nearest local descriptorsAndsatisfy the following requirementsThen local descriptorCorresponding toAnd SuperPoint feature pointsFor mutually matching points, SuperPoint characteristic points are obtained after traversal is finishedSet of all match points in candidate frame set CSuperPoint feature pointsCorresponding 3D positionThe associated matching information isEstablishing a constraint equation to solve according to the internal reference K of the binocular cameraThe constraint equation is:
(3.3) traversing each SuperPoint feature pointAnd obtaining the 3D position corresponding to each SuperPoint characteristic point.
Compared with the prior art, the invention has the beneficial effects that: the method combines an image global descriptor NetVLAD method for image retrieval and a local image feature point and descriptor SuperPoint method based on a convolutional neural network, the combination of the two enhances the descriptive performance of a visual map, the visual localization is decoupled into global and local localization, and the generated map for visual localization comprises multi-level description information, wherein the global information comprises the size information of the map, the number of feature points and the corresponding 3D point positions, the number of key frames and NetVLAD descriptors corresponding to each frame, the local information comprises the posture of each frame key frame, the position of the feature points in each key frame and the corresponding SuperPoint descriptors, and the optimal co-view key frame index corresponding to each key frame, and the two are combined and can be used for visual global robust localization with consistent coordinate systems. On one hand, the problem that the traditional SLAM system cannot realize map multiplexing can be solved, and on the other hand, the robustness of global positioning can be improved.
Drawings
FIG. 1 is a flow chart of a visual map construction method for visual layered positioning according to the present invention;
FIG. 2 is a schematic diagram of multi-view observation recovery 3D feature points according to the present invention;
FIG. 3 is a schematic diagram of a visual map containing information according to the present invention;
FIG. 4 is a schematic view of the layered global positioning based on the visual map according to the present invention.
Detailed Description
The principles and aspects of the present invention will be further explained with reference to the drawings and the detailed description, it being understood that the illustrated embodiments are only some examples and the specific embodiments described herein are merely illustrative of the relevant invention and are not intended to limit the invention.
Fig. 1 is a schematic flow chart of a visual map construction method for visual layered positioning according to the present invention, where the visual map construction method includes the following steps:
(1) installing a binocular camera in a robot body, moving the robot and starting to acquire data, sequentially acquiring binocular image frames acquired by the binocular camera and recording the motion trail of the binocular camera to obtain a binocular image sequence containing the motion trailWherein, in the step (A),is shown asThe left frame image to which the frame corresponds,is shown asThe right frame image to which the frame corresponds,is an external parameter of the binocular camera,is as followsThe pose of the frame left camera relative to the world coordinate system, and N is the frame number of the binocular image; using the first frame as a key frame, and aiming at the binocular image sequenceSelecting subsequent key frames to form a key frame sequence,Selecting a binocular image frame number N; the selection of the subsequent key frame needs to satisfy the following conditions: the Euclidean distance between adjacent left frame images or right frame images is larger than 0.3 meter, and the rotation angle of the adjacent left frame images or right frame images is larger than 3 degrees.
(2) Extracting NetVLAD global descriptor based on convolution neural network for key frameThe NetVLAD global descriptorIs described asExtracting SuperPoint characteristic points of the key frame based on the convolution neural network by using the characteristic vector of the dimensionAnd corresponding local descriptorsThe local descriptorIs described asA vector of dimensions; and according to SuperPoint characteristic pointsThe response values are sorted from high to low, the first 2000 feature points and the corresponding local descriptors are reserved, and for the secondObtaining two triple description information by one key frame。
(3) In order to find out SuperPoint characteristic points of key framesAnd incrementally restoring the 3D position of each SuperPoint feature pointIt is necessary to find out the sequence of key framesThe same pairWith observed 2D point locations, as shown in fig. 2, comprising the sub-steps of:
(3.1) describing information for a certain tripleSuperPoint feature points in (1)Selecting the characteristic points which can observe SuperPoint simultaneouslyCandidate frame set ofWherein A is selected from the key frames, satisfying: pose of keyframes in the set of candidate frames CAndthe movement translation distance of the rotating shaft is less than 10m, and the rotating angle is less than 45 degrees;
(3.2) traverse each frame in the candidate frame set C, pairThe local descriptor corresponding to each feature point inCalculatingAndand obtaining two nearest local descriptorsAndsatisfy the following requirementsThen local descriptorCorresponding toAnd SuperPoint feature pointsFor mutually matching points, SuperPoint characteristic points are obtained after traversal is finishedSet of all match points in candidate frame set C Corresponding 3D point sets are(ii) a SuperPoint feature pointsCorresponding 3D positionThe associated matching information isEstablishing a constraint equation to solve according to the internal reference K of the binocular cameraThe constraint equation is:
(3.3) traversing each SuperPoint feature pointObtaining a 3D position corresponding to each SuperPoint characteristic point;
(3.4) the wholeAnd traversing each feature point of the sequence to obtain a corresponding 3D value of the feature point. And finally obtaining basic information of the whole map:
(4) corresponding 3D position for each acquired SuperPoint feature pointWill be observed by multiple frames, traverseEach of which isStatistical observation of 3D positionAll key frames of the points are sorted in descending order, and the first 5 key frames are taken as the optimal common-view key frame of the key frames(ii) a And finally, corresponding the key frame sequence and each SuperPoint characteristic point to a 3D positionAs a visual groundThe map information, as shown in fig. 3, the final map information, which is exemplified by the left camera of the binocular camera, may be represented as map basic information and feature information, where the basic information includes map size information, the number of key frames, the number of feature points, and the number of 3D points; the feature information comprises key frames, key frame global descriptors, common view key frames, feature points, feature point descriptors and feature point corresponding sets of 3D points.
The visual map information obtained by the visual map construction method is stored and loaded on the robot, the robot is restarted and placed in the environment for positioning, and as shown in fig. 4, firstly, the characteristics of the collected image are extracted: the system comprises a NetVLAD global descriptor, a SuperPoint characteristic point and a descriptor. Comparing the NetVLAD global descriptor of the current frame with the NetVLAD set of the global descriptors in the map visual information, calculating the Euclidean distance of the NetVLAD global descriptor to obtain a key frame closest to the map, finishing the first-layer positioning of the hierarchical positioning, namely the global coarse positioning part, and indicating that the frame is to be positioned near the closest key frame and is coarse positioning. And finding out 5 optimal co-viewing key frames corresponding to the key frames, performing feature matching on the current frame and the closest key frame and the 5 optimal co-viewing key frames corresponding to the current frame by the method in the step (3.2), finally obtaining 3D-2D matching according to data association in the step (3.3), finally obtaining 6DoF postures according to PnP solution, and finishing the fine positioning function of layered positioning.
Table 1 gives, for example, a comparison of the visual mapping method of the present invention with some prior disclosed SLAM methods. The sparse and dense evaluation criteria in table 1 are the number of map 3D points, if the map is constructed by using only a small number of feature points, such as 2000 feature points selected by the scheme, the map is sparse, and if all pixel points in the selected image are constructed, the map is considered dense. In table 1, the coarse positioning mode DBoW is a bag-of-words model, which requires pre-training a dictionary of features for image retrieval, and finally outputs a word vector for coarse positioning for each image.
According to the comparison in table 1, in the prior art scheme, map reuse is basically not considered, which means that the corresponding method cannot perform globally consistent positioning, and the positioning result of each time is related to the initial position of the acquired data.
TABLE 1 comparison of the Performance of the location method of the present invention with that of the prior art method
In terms of robustness, table 2 compares the VINS-Mono scheme closest to the present invention in the collected data set. Because the positioning precision of the invention is related to the camera pose of the image construction, the rough positioning result can only be compared with the VINS-Mono, the total frame number of 7 sequences is counted, and the frame number of the image which passes through the same place and is collected for at least two times in the sequence is counted and recorded as the number of the loop frame. The 7 sequences of images contain challenging factors such as motion blur, occlusion, scene lighting changes and the like. As can be seen from the results in table 2, the present invention has a great improvement in the success rate and robustness of coarse positioning, which is a front step of fine positioning, and all of which have great improvements on the whole positioning system.
TABLE 2 comparison of the coarse positioning scheme of the present invention with the VINS-Mono method
Therefore, the visual map construction method of the invention can solve the problem that the traditional SLAM system can not realize map multiplexing on one hand, and can improve the robustness of global positioning on the other hand.
The above are merely preferred embodiments of the present invention; the scope of the invention is not limited thereto. Any person skilled in the art should be able to cover the technical scope of the present invention by equivalent or modified solutions and modifications within the technical scope of the present invention.
Claims (3)
1. A visual map construction method for visual layered positioning is characterized in that: the method comprises the following steps:
(1) sequentially acquiring binocular image frames acquired by the binocular camera and recording the motion trail of the binocular camera to obtain a binocular image sequence containing the motion trailWherein, in the step (A),is shown asThe left frame image to which the frame corresponds,is shown asThe right frame image to which the frame corresponds,is an external parameter of the binocular camera,is as followsThe pose of the frame left camera relative to the world coordinate system, and N is the frame number of the binocular image; using the first frame as a key frame, and aiming at the binocular image sequenceSelecting subsequent key frames to form a key frame sequence,Selecting a binocular image frame number N;
(2) respectively extracting NetVLAD global descriptors from key framesSuperPoint feature pointsAnd corresponding local descriptorsAnd according to SuperPoint feature pointsThe response values are sorted from high to low, the first 2000 feature points and the corresponding local descriptors are reserved, and for the secondObtaining two triple description information by one key frame(ii) a The NetVLAD global descriptorIs composed ofFeature vectors of dimensions, the local descriptorsIs composed ofA vector of dimensions;
(3) finding out SuperPoint feature points of key framesAnd incrementally restoring the 3D position of each SuperPoint feature point;
(4) Traversing each of the steps (3)Statistical observation of 3D positionAll key frames of the points are sorted in descending order, and the first 5 key frames are taken as the optimal common-view key frame of the key frames(ii) a And finally, corresponding the key frame sequence and each SuperPoint characteristic point to a 3D positionAs visual map information.
2. The visual map construction method for visual layered positioning according to claim 1,
the selection of the subsequent key frame needs to satisfy the following conditions: the Euclidean distance between adjacent left frame images or right frame images is larger than 0.3 meter, and the rotation angle of the adjacent left frame images or right frame images is larger than 3 degrees.
3. A visual mapping method for visual layered positioning according to claim 1, wherein step (3) comprises the sub-steps of:
(3.1) describing information for a certain tripleSuperPoint feature points in (1)Selecting the characteristic points which can observe SuperPoint simultaneouslyCandidate frame set ofWherein A is selected from the key frames, satisfying: pose of keyframes in the set of candidate frames CAndthe movement translation distance of the rotating shaft is less than 10m, and the rotating angle is less than 45 degrees;
(3.2) traverse each frame in the candidate frame set C, pairThe local descriptor corresponding to each feature point inCalculatingAndand obtaining two nearest local descriptorsAndsatisfy the following requirementsThen local descriptorCorresponding toAnd SuperPoint feature pointsFor mutually matching points, SuperPoint characteristic points are obtained after traversal is finishedSet of all match points in candidate frame set CSuperPoint feature pointsCorresponding 3D positionThe associated matching information isEstablishing a constraint equation to solve according to the internal reference K of the binocular cameraThe constraint equation is:
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110175262.1A CN112562081B (en) | 2021-02-07 | 2021-02-07 | Visual map construction method for visual layered positioning |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110175262.1A CN112562081B (en) | 2021-02-07 | 2021-02-07 | Visual map construction method for visual layered positioning |
Publications (2)
Publication Number | Publication Date |
---|---|
CN112562081A true CN112562081A (en) | 2021-03-26 |
CN112562081B CN112562081B (en) | 2021-05-11 |
Family
ID=75035905
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202110175262.1A Active CN112562081B (en) | 2021-02-07 | 2021-02-07 | Visual map construction method for visual layered positioning |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN112562081B (en) |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113781563A (en) * | 2021-09-14 | 2021-12-10 | 中国民航大学 | Mobile robot loop detection method based on deep learning |
CN114639006A (en) * | 2022-03-15 | 2022-06-17 | 北京理工大学 | Loop detection method and device and electronic equipment |
CN114674328A (en) * | 2022-03-31 | 2022-06-28 | 北京百度网讯科技有限公司 | Map generation method, map generation device, electronic device, storage medium, and vehicle |
CN114694013A (en) * | 2022-04-11 | 2022-07-01 | 北京理工大学 | Distributed multi-machine cooperative vision SLAM method and system |
CN115049731A (en) * | 2022-06-17 | 2022-09-13 | 感知信息科技(浙江)有限责任公司 | Visual mapping and positioning method based on binocular camera |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110070615A (en) * | 2019-04-12 | 2019-07-30 | 北京理工大学 | A kind of panoramic vision SLAM method based on polyphaser collaboration |
CN111292420A (en) * | 2020-02-28 | 2020-06-16 | 北京百度网讯科技有限公司 | Method and device for constructing map |
CN111652934A (en) * | 2020-05-12 | 2020-09-11 | Oppo广东移动通信有限公司 | Positioning method, map construction method, device, equipment and storage medium |
CN111768498A (en) * | 2020-07-09 | 2020-10-13 | 中国科学院自动化研究所 | Visual positioning method and system based on dense semantic three-dimensional map and mixed features |
-
2021
- 2021-02-07 CN CN202110175262.1A patent/CN112562081B/en active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110070615A (en) * | 2019-04-12 | 2019-07-30 | 北京理工大学 | A kind of panoramic vision SLAM method based on polyphaser collaboration |
CN111292420A (en) * | 2020-02-28 | 2020-06-16 | 北京百度网讯科技有限公司 | Method and device for constructing map |
CN111652934A (en) * | 2020-05-12 | 2020-09-11 | Oppo广东移动通信有限公司 | Positioning method, map construction method, device, equipment and storage medium |
CN111768498A (en) * | 2020-07-09 | 2020-10-13 | 中国科学院自动化研究所 | Visual positioning method and system based on dense semantic three-dimensional map and mixed features |
Non-Patent Citations (2)
Title |
---|
XIAO HAN等: "SuperPointVO A Lightweight Visual Odometrybased on CNN Feature Extraction", 《IEEE》 * |
唐灿 等: "图像特征检测与匹配方法研究综述", 《南京信息工程大学学报》 * |
Cited By (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113781563A (en) * | 2021-09-14 | 2021-12-10 | 中国民航大学 | Mobile robot loop detection method based on deep learning |
CN113781563B (en) * | 2021-09-14 | 2023-10-24 | 中国民航大学 | Mobile robot loop detection method based on deep learning |
CN114639006A (en) * | 2022-03-15 | 2022-06-17 | 北京理工大学 | Loop detection method and device and electronic equipment |
CN114639006B (en) * | 2022-03-15 | 2023-09-26 | 北京理工大学 | Loop detection method and device and electronic equipment |
CN114674328A (en) * | 2022-03-31 | 2022-06-28 | 北京百度网讯科技有限公司 | Map generation method, map generation device, electronic device, storage medium, and vehicle |
CN114694013A (en) * | 2022-04-11 | 2022-07-01 | 北京理工大学 | Distributed multi-machine cooperative vision SLAM method and system |
CN114694013B (en) * | 2022-04-11 | 2022-11-15 | 北京理工大学 | Distributed multi-machine cooperative vision SLAM method and system |
CN115049731A (en) * | 2022-06-17 | 2022-09-13 | 感知信息科技(浙江)有限责任公司 | Visual mapping and positioning method based on binocular camera |
Also Published As
Publication number | Publication date |
---|---|
CN112562081B (en) | 2021-05-11 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN112562081B (en) | Visual map construction method for visual layered positioning | |
CN109166149B (en) | Positioning and three-dimensional line frame structure reconstruction method and system integrating binocular camera and IMU | |
CN103577793B (en) | Gesture identification method and device | |
Li et al. | Object detection in the context of mobile augmented reality | |
CN107016319B (en) | Feature point positioning method and device | |
CN110555408B (en) | Single-camera real-time three-dimensional human body posture detection method based on self-adaptive mapping relation | |
CN112598775B (en) | Multi-view generation method based on contrast learning | |
CN109272577B (en) | Kinect-based visual SLAM method | |
Laga | A survey on deep learning architectures for image-based depth reconstruction | |
CN110119768B (en) | Visual information fusion system and method for vehicle positioning | |
CN110942476A (en) | Improved three-dimensional point cloud registration method and system based on two-dimensional image guidance and readable storage medium | |
CN112419497A (en) | Monocular vision-based SLAM method combining feature method and direct method | |
Dharmasiri et al. | MO-SLAM: Multi object slam with run-time object discovery through duplicates | |
CN115147599A (en) | Object six-degree-of-freedom pose estimation method for multi-geometric feature learning of occlusion and truncation scenes | |
Wang et al. | Joint head pose and facial landmark regression from depth images | |
CN113592015B (en) | Method and device for positioning and training feature matching network | |
CN111402331A (en) | Robot repositioning method based on visual word bag and laser matching | |
CN112329662B (en) | Multi-view saliency estimation method based on unsupervised learning | |
Shin et al. | Loop closure detection in simultaneous localization and mapping using descriptor from generative adversarial network | |
CN111626417A (en) | Closed loop detection method based on unsupervised deep learning | |
CN110070626B (en) | Three-dimensional object retrieval method based on multi-view classification | |
Lu et al. | Model and exemplar-based robust head pose tracking under occlusion and varying expression | |
CN115496859A (en) | Three-dimensional scene motion trend estimation method based on scattered point cloud cross attention learning | |
Song et al. | ConcatNet: A deep architecture of concatenation-assisted network for dense facial landmark alignment | |
Ma et al. | Capsule-based regression tracking via background inpainting |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |