CN109341703A - A kind of complete period uses the vision SLAM algorithm of CNNs feature detection - Google Patents

A kind of complete period uses the vision SLAM algorithm of CNNs feature detection Download PDF

Info

Publication number
CN109341703A
CN109341703A CN201811087509.9A CN201811087509A CN109341703A CN 109341703 A CN109341703 A CN 109341703A CN 201811087509 A CN201811087509 A CN 201811087509A CN 109341703 A CN109341703 A CN 109341703A
Authority
CN
China
Prior art keywords
training
feature
cnns
detection
image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201811087509.9A
Other languages
Chinese (zh)
Other versions
CN109341703B (en
Inventor
赵永嘉
张宁
雷小永
戴树岭
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beihang University
Original Assignee
Beihang University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beihang University filed Critical Beihang University
Priority to CN201811087509.9A priority Critical patent/CN109341703B/en
Publication of CN109341703A publication Critical patent/CN109341703A/en
Application granted granted Critical
Publication of CN109341703B publication Critical patent/CN109341703B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G01MEASURING; TESTING
    • G01CMEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
    • G01C21/00Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
    • G01C21/26Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 specially adapted for navigation in a road network
    • G01C21/28Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 specially adapted for navigation in a road network with correlation of data from several navigational instruments
    • G01C21/30Map- or contour-matching
    • G01C21/32Structuring or formatting of map data
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/22Matching criteria, e.g. proximity measures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches

Landscapes

  • Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Remote Sensing (AREA)
  • Radar, Positioning & Navigation (AREA)
  • General Physics & Mathematics (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Evolutionary Computation (AREA)
  • Evolutionary Biology (AREA)
  • General Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Automation & Control Theory (AREA)
  • Image Analysis (AREA)

Abstract

The present invention discloses a kind of vision SLAM algorithm that the complete period uses CNNs feature to detect, first in front end, pre-training is carried out to raw image data with unsupervised model, the joint of movement and depth is indicated that drawn game portion speed and direction change by the CNN network architecture followed by the data of pre-training to associate, executes visual odometry;Last execution route prediction.The present invention also uses OverFeat neural network model to carry out winding detection, for eliminating front end bring accumulated error, constructs a vision slam framework based on deep learning.Time and spatial continuity filter are constructed simultaneously, matching result is verified, improves matching accuracy rate, eliminates error hiding.The present invention has big advantage and potentiality in terms of improving visual odometry precision and closed loop detection accuracy.

Description

A kind of complete period uses the vision SLAM algorithm of CNNs feature detection
Technical field
The technical field with map structuring (slam) algorithm is positioned while the invention belongs in computer vision, is specifically related to And a kind of complete period uses the vision SLAM algorithm of CNNs feature detection.
Background technique
SLAM (Simultaneous Localization and Mapping) Chinese be " while positioning and map Building ".SLAM is a fascinating research field, has in robot, navigation and other many applications and widely answers With.Vision SLAM is essentially related to estimate camera motion according to visual sensor information and is attempted to construct the ground of ambient enviroment Figure, such as the sequence frame from one or more cameras.The research means of current SLAM problem are mainly by robot sheet Sensor is installed to estimate the characteristic information of robot body motion information and circumstances not known on body, is merged using information It realizes to the accurate estimation of robot pose and the spatial modeling of scene.Although the sensor that SLAM is used has laser and vision Equal multiple types, but its treatment process generally comprises 3 parts: front-end vision odometer, rear end optimization and closed loop detection.
Typical vision SLAM algorithm is to estimate pose of camera as main target, 3D is reconstructed by multiple view geometry theory Figure.To improve data processing speed, partial visual SLAM algorithm extracts sparse characteristics of image first, by between characteristic point Visual odometry and closed loop detection are realized in matching, such as special based on SIFT (scale invariant feature transform) The vision SLAM [13] of the sign and vision SLAM for being based on ORB (oriented FAST and rotated BRIEF) feature.SIFT Rely on its preferable robustness and preferably separating capacity and quick processing speed with ORB feature, in the field vision SLAM It is widely applied.The sparse characteristics of image of engineer currently has many limitations, and it is special on the one hand how to design sparse image Sign optimally indicates that image information is still the unsolved major issue of computer vision field, on the other hand sparse characteristics of image In terms of reply illumination variation, dynamic object movement, camera parameters change and lack texture or texture is single Still there is more challenge.Traditional visual odometry (Visual Odometry, VO) essentially relates to be estimated according to visual information Meter movement, such as the sequence frame from one or more cameras.One common feature of most number attribute in these work is They estimate visual token by critical point detection and tracking and camera geometry.
In recent years, the method based on study many fields of computer vision show desirable as a result, Defect present in traditional vision slam algorithm can be overcome, and (sparse characteristics of image is in reply illumination variation, dynamic object fortune Dynamic, camera parameters, which change and lack texture or the single environment etc. of texture, more difficulty).As convolutional Neural net Model as network (CNNs) is proved to highly effective in various visual tasks, such as classifies and positions, estimation of Depth etc..Nothing Supervise the ability that feature learning model display is indicated by partial transformation in multiplication interactive learning data.Research shows that by nothing Data after monitor model pre-training reapply in CNNs network, preferable can cross noise filtering and prevent " over-fitting ".
Visual odometry (VO) is responsible for the initial value of estimation track and map.VO principle is the connection that only considered consecutive frame picture System.Error inevitably will be with time integral, so that cumulative errors occurs in whole system, the result estimated for a long time will not Reliably, in other words, we can not construct globally consistent track and map.Winding detection module can provide one except consecutive frame A bit after constraint more remote.The key of winding detection, be how effectively to detect camera by the same place this Part thing is related to the correctness of the estimation and map of our estimations under long-time.Therefore, winding detection is to entire SLAM system The promotion of precision and robustness be obviously.
And this winding detection based on appearance is substantially a task of image similarity match.By scheming to two As carrying out characteristic matching, to verify whether as same place.The conventional method of winding detection is that " bag of words " is utilized to generate word Allusion quotation does characteristic matching.Now, CNNs shows optimal performance in various classification tasks.Existing road sign test result Show that the further feature of the CNN from different levels executes more preferably than SIFT always in terms of descriptors match, shows SIFT Or SURF may no longer be the preferred descriptor of matching task.Therefore, our invention by CNN in terms of image classification go out The inspiration for the feasibility evidence that color table is existing and they are in terms of characteristic matching.The bag of words method to discard tradition, with being based on CNNs using depth learning technology as the stratification image characteristic extracting method of representative, to do winding detection.Deep learning algorithm It is the recognizer of computer vision field mainstream, relies on the stratification mark sheet of multilayer neural network study image Show, compared with traditional recognition method, the accuracy rate of higher feature extraction and position identification may be implemented.
Summary of the invention
In view of the above-mentioned problems, the present invention proposes that a kind of complete period uses the vision SLAM algorithm of CNNs feature detection, it is one Kind is all handled using convolutional neural networks in the front end (VO) that SLAM algorithm is run and winding detection, realizes a complete period The SLAM system of deep learning algorithm.
Complete period of the invention uses the vision SLAM algorithm of CNNs feature detection, comprising the following steps:
Step 1: using the environmental information around binocular camera scanning;A part in collected multiple video flowings is made For training dataset, a part is used as test data set.
Step 2: training dataset carries out pre-training in the video flowing acquired by the method for synchronous detection to step 1.
Step 3: being changed using the part that convolutional neural networks training obtains speed and changed with the part in direction, to execute view Feel odometer.
Step 4: the part of the speed obtained using step 3, which is changed, changes information with the part in direction, recovers camera Motion path.
Step 5: carrying out closed loop detection using convolutional neural networks, eliminate the accumulated error of path prediction.
The invention has the advantages that:
1, the complete period of the invention uses the vision SLAM algorithm of CNNs feature detection, and convolutional neural networks are used for front end, Compared to conventional front-end algorithm, the side based on study substitutes cumbersome formula and calculates, and extracts and matches without manual features, it appears succinct Intuitively, and on-line operation fast speed.
2, the complete period of the invention uses the vision SLAM algorithm of CNNs feature detection, the utilization bag of words to discard tradition into The method of row closed loop detection obtains the effect of relatively good position identification by accurate characteristic matching.
3, the complete period of the invention uses the vision SLAM algorithm of CNNs feature detection, can pass through neural network learning image In profound feature, discrimination can achieve higher level.Compared to Conventional visual slam algorithm, closed loop inspection can be improved Survey accuracy rate, expression image information it is more abundant, have more strong robustness to environmental changes such as illumination, seasons, can to 2 frame images it Between similitude calculated, to realize more succinct visual odometry, and complete using database pre-training neural network While at characteristic Design, the design of matched classifier can also be synchronously completed.
Detailed description of the invention
Fig. 1 is the vision SLAM algorithm overall flow figure for using the detection of CNNs feature in the complete period of the invention;
Fig. 2 is the complete period of the invention using closed loop detection method flow chart in the vision SLAM algorithm of CNNs feature detection.
Specific embodiment
The present invention is described in further detail below in conjunction with the accompanying drawings.
Complete period of the invention uses the vision SLAM algorithm of CNNs feature detection, as shown in Figure 1, comprising the following steps:
Step 1: scanning the environmental information of surrounding;
Using binocular camera along being moved in square area, ambient image information acquisition is carried out to real scene, is obtained Video flowing be transmitted in host computer in real time.Binocular camera mobile ring number is 1~2 circle, constitutes closed loop, facilitates subsequent close Compensation of the ring detection to cumulative errors.Repeat the above process, by a part of conduct in collected multiple video flowings Training dataset, a part are used as test data set.
Step 2: training dataset carries out pre-training in the video flowing acquired by the method for synchronous detection to step 1.
It indicates to obtain the combination of camera motion and image depth information, is adopted using unsupervised learning model (SAE-D) Pre-training is carried out to training dataset with stochastic gradient descent training.It is the model of a single layer based on synchronous SAE-D, allows Feature is obtained from training dataset by part and Hebbian type study.
SAE-D model concentrates training, the size of each binocular block on the local binocular block cut out immediately in training data For 16*16*5 (space*space*time), the characteristic information of movement and depth joint expression is obtained;It then will movement and depth The characteristic information that joint indicates goes albefaction back to image space, completes the pre-training to training dataset.Instruction after pre-training Practice the first layer that data set is used to initialize CNN (convolutional neural networks).
Step 3: executing visual odometry using convolutional neural networks (CNN) training.
Convolutional neural networks (CNN) are a kind of learning models based on supervision.CNN is trained to by partial-depth and movement Indicate associated with the localized variation in speed and direction, to learn to execute visual odometry.By CNN framework the fortune of acquisition Dynamic and depth joint expression is trained associated with desired label (direction and speed change).
The feature that SAE-D model training obtains in step 2 is inputted to the CNN first layer of two same architectures in the present invention, Initialization for two CNN first layers;The part of two CNN difference output speeds changes to be changed with the part in direction, and is had It is associated with desired label that the part of speed is changed the part change with direction respectively by two CNN remainders.
Entire CNN network shares 6 layers, and first layer is the convolutional layer of 5*5, respectively left images learning characteristic.Following Two layers of feature extracted to two convolutional layers in left and right carry out element multiplication.The convolutional layer of third layer 1*1, the 4th layer is pond Layer, layer 5 is full articulamentum, and last output layer is Softmax layers.
The input of above-mentioned two CNN is 5 frame subsequences, and target output is local velocity and the vector that direction changes It indicates.The speed of part and direction, which change the effect of information and precision, to be evaluated and tested by binocular data set KITTI.
Step 4: the prediction of camera motion path
For entire video flowing, the speed of the every 5 frame subsequence obtained using step 3 and direction change information, Ji Keli The scattered motion path for recovering camera.
Step 5: carrying out closed loop detection using CNN, eliminate the accumulated error of path prediction;
Local speed obtained in step 4 and direction change information can not completely precisely, and there is a certain error.With The continuous accumulation of error, predicted path can constantly increase with the gap in true path from origin-to-destination.So needing Subsequent closed loop detection eliminates these cumulative errors, reduces the gap of predicted path and true path.It is this Algorithm includes two parts: carrying out feature extraction using convolutional neural networks;Location matches are assumed by comparing characteristic response Carry out spatio-temporal filtering.
Equally handled using the algorithm based on CNN in closed loop detection, the CNN model being applied to it is aforementioned CNN model is different, is independently used in visual odometry and closed loop detection, pre- with the camera motion path that this removal process 4 obtains The cumulative errors of survey realize from main closed loop, method particularly includes:
Image characteristics extraction is carried out using the convolutional neural networks of pre-training.The present invention uses overfeat convolutional Neural net Network carries out image characteristics extraction.
Overfeat convolutional neural networks carry out pre-training on 2012 data set of ImageNet, and the data set is by 1,200,000 Open image and 1000 classes compositions.Overfeat convolutional neural networks include that five convolution stages and three are fully connected the stage. Two convolution stages of beginning are by a convolutional layer, a maximum pond layer and rectification (ReLU) non-linear layer composition.The Three and the Volume Four product stage by convolutional layer, zero padding layer and ReLU non-linear layer composition.5th stage include a convolutional layer, one A zero padding layer, one ReLU layers and one Maxpooling layers.Finally, the 6th and the 7th be fully connected the stage include one it is complete Full articulamentum and one ReLU layer, and it includes the output layer for being only fully connected layer that the 8th stage, which was,.Entire convolutional neural networks are total Share 21 layers.
When image I is input into network, it generates a series of layering activation.L is used in the present inventionk(I), k= 1 ..., 21 indicate the corresponding output of given input picture I kth layer.The feature vector L of each layer of outputk(I) every in One be all image I deep learning indicate;Execution position is compared to the corresponding feature vector of different images by these Set identification.The network is capable of handling the image that any size is equal to or more than 231 × 231 pixels, therefore overfeat convolution mind Input through network is all using being sized as the image of 256 × 256 pixels.
It is good using pre-training as a result, by using the training dataset acquired in step 1 and test data set as input Overfeat convolutional neural networks carry out feature extraction.
Step 6: characteristic matching generates hybrid matrix and carries out space-time expending detection.
As shown in Fig. 2, being extracted by overfeat convolutional neural networks to from the picture that each test data is concentrated Feature concentrate the feature proposed to be matched with from each training data.
To compare performance difference of the feature of every tomographic image in overfeat convolutional neural networks in scene Recognition, into one Step utilizes each layer latent structure hybrid matrix:
Mk(i, j)=d (Lk(Ii),Lk(Ij)), i=1 ..., R, j=1 ..., T
Wherein, IiRepresent the image input of the i-th frame training data concentration, IjThe image for representing jth frame test data concentration is defeated Enter, Lk(Ii) represent and IiCorresponding kth layer output, Mk(i, j) represents the Europe between kth layer training sample i and test sample j Family name's distance describes matching degree between the two;R and T respectively represent the number of training image and the number of test image. The averaged feature vector that the above-mentioned each column of hybrid matrix store between jth frame test image and all training images is poor.
In order to find strongest location matches it is assumed that having the member of minimum feature vector difference in search hybrid matrix each column Element.
For location matches possible in hybrid matrix it is assumed that further constructing spatial continuity filter and Time Continuous Property filter carry out comprehensive verification, improve matching accuracy rate.Meanwhile the characteristic performance come is trained to each layer network and is visited Rope, it is preferable that discovery network middle layer feature describes similar for visual angle image matching effect, and in after layer for scene visual angle Variation has stronger adaptability and robustness.
There are accurately location matches, so that it may the cumulative errors caused by the visual odometry of not winding compensate, Construct globally consistent track.

Claims (5)

1. the vision SLAM algorithm that a kind of complete period uses the detection of CNNs feature, comprising the following steps:
Step 1: using the environmental information around binocular camera scanning;By a part in collected multiple video flowings as instruction Practice data set, a part is used as test data set;
Step 2: training dataset carries out pre-training in the video flowing acquired by the method for synchronous detection to step 1;
Step 3: being changed using the part that convolutional neural networks training obtains speed and changed with the part in direction, to execute in vision Journey meter;
Step 4: the part of the speed obtained using step 3, which is changed, changes information with the part in direction, recovers the movement of camera Path;
Step 5: carrying out closed loop detection using convolutional neural networks, eliminate the accumulated error of path prediction.
2. a kind of complete period uses the vision SLAM algorithm of CNNs feature detection as described in claim 1, it is characterised in that: step Using binocular camera along moving in annular region in 1, mobile ring number is 1~2 circle, constitutes closed loop.
3. a kind of complete period uses the vision SLAM algorithm of CNNs feature detection as described in claim 1, it is characterised in that: step Pre-training is carried out to training dataset using stochastic gradient descent training using unsupervised learning model in 2, is moved and deep The characteristic information that movement and depth joint indicate then is gone albefaction back to image space by the characteristic information that degree joint indicates.
4. a kind of complete period uses the vision SLAM algorithm of CNNs feature detection as described in claim 1, it is characterised in that: step In 3, the feature that step 2 training obtains is inputted to the CNN first layer of two same architectures, by CNN framework the movement of acquisition Joint expression with depth is trained associated with desired label.
5. a kind of complete period uses the vision SLAM algorithm of CNNs feature detection as described in claim 1, it is characterised in that: step In 5, firstly, by using the training dataset acquired in step 1 and test data set as input, using in imagenet number According to upper pre-training is collected, good overfeat convolutional neural networks carry out feature extraction;Then, it is concentrated to from each test data Picture in the feature extracted concentrate the feature proposed to be matched with from each training data;
Utilize each layer latent structure hybrid matrix of overfeat convolutional neural networks:
Mk(i, j)=d (Lk(Ii),Lk(Ij)), i=1 ..., R, j=1 ..., T
Wherein, IiRepresent the image input of the i-th frame training data concentration, IjThe image input of jth frame test data concentration is represented, Lk(Ii) represent and IiCorresponding kth layer output, Mk(i, j) represent the Euclidean between kth layer training sample i and test sample j away from From describing matching degree between the two;R and T respectively represent the number of training image and the number of test image;It is above-mentioned The averaged feature vector that each column of hybrid matrix store between jth frame test image and all training images is poor;
Search for the element in hybrid matrix each column with minimum feature vector difference;
CN201811087509.9A 2018-09-18 2018-09-18 Visual SLAM algorithm adopting CNNs characteristic detection in full period Active CN109341703B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201811087509.9A CN109341703B (en) 2018-09-18 2018-09-18 Visual SLAM algorithm adopting CNNs characteristic detection in full period

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201811087509.9A CN109341703B (en) 2018-09-18 2018-09-18 Visual SLAM algorithm adopting CNNs characteristic detection in full period

Publications (2)

Publication Number Publication Date
CN109341703A true CN109341703A (en) 2019-02-15
CN109341703B CN109341703B (en) 2022-07-01

Family

ID=65305452

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201811087509.9A Active CN109341703B (en) 2018-09-18 2018-09-18 Visual SLAM algorithm adopting CNNs characteristic detection in full period

Country Status (1)

Country Link
CN (1) CN109341703B (en)

Cited By (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109840598A (en) * 2019-04-29 2019-06-04 深兰人工智能芯片研究院(江苏)有限公司 A kind of method for building up and device of deep learning network model
CN110146099A (en) * 2019-05-31 2019-08-20 西安工程大学 A kind of synchronous superposition method based on deep learning
CN110296705A (en) * 2019-06-28 2019-10-01 苏州瑞久智能科技有限公司 A kind of vision SLAM winding detection method based on learning distance metric
CN110487274A (en) * 2019-07-30 2019-11-22 中国科学院空间应用工程与技术中心 SLAM method, system, navigation vehicle and storage medium for weak texture scene
CN110659619A (en) * 2019-09-27 2020-01-07 昆明理工大学 Depth space-time information-based correlation filtering tracking method
CN110738128A (en) * 2019-09-19 2020-01-31 天津大学 repeated video detection method based on deep learning
CN111144550A (en) * 2019-12-27 2020-05-12 中国科学院半导体研究所 Simplex deep neural network model based on homologous continuity and construction method
CN111243021A (en) * 2020-01-06 2020-06-05 武汉理工大学 Vehicle-mounted visual positioning method and system based on multiple combined cameras and storage medium
CN111241986A (en) * 2020-01-08 2020-06-05 电子科技大学 Visual SLAM closed loop detection method based on end-to-end relationship network
CN111753789A (en) * 2020-07-01 2020-10-09 重庆邮电大学 Robot vision SLAM closed loop detection method based on stack type combined self-encoder
CN113066152A (en) * 2021-03-18 2021-07-02 内蒙古工业大学 AGV map construction method and system
CN110399821B (en) * 2019-07-17 2023-05-30 上海师范大学 Customer satisfaction acquisition method based on facial expression recognition

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106651830A (en) * 2016-09-28 2017-05-10 华南理工大学 Image quality test method based on parallel convolutional neural network
CN106780631A (en) * 2017-01-11 2017-05-31 山东大学 A kind of robot closed loop detection method based on deep learning
CN107563430A (en) * 2017-08-28 2018-01-09 昆明理工大学 A kind of convolutional neural networks algorithm optimization method based on sparse autocoder and gray scale correlation fractal dimension
CN107808132A (en) * 2017-10-23 2018-03-16 重庆邮电大学 A kind of scene image classification method for merging topic model
CN107944386A (en) * 2017-11-22 2018-04-20 天津大学 Visual scene recognition methods based on convolutional neural networks

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106651830A (en) * 2016-09-28 2017-05-10 华南理工大学 Image quality test method based on parallel convolutional neural network
CN106780631A (en) * 2017-01-11 2017-05-31 山东大学 A kind of robot closed loop detection method based on deep learning
CN107563430A (en) * 2017-08-28 2018-01-09 昆明理工大学 A kind of convolutional neural networks algorithm optimization method based on sparse autocoder and gray scale correlation fractal dimension
CN107808132A (en) * 2017-10-23 2018-03-16 重庆邮电大学 A kind of scene image classification method for merging topic model
CN107944386A (en) * 2017-11-22 2018-04-20 天津大学 Visual scene recognition methods based on convolutional neural networks

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
何元烈等: "基于精简卷积神经网络的快速闭环检测方法", 《计算机工程》 *

Cited By (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109840598B (en) * 2019-04-29 2019-08-09 深兰人工智能芯片研究院(江苏)有限公司 A kind of method for building up and device of deep learning network model
CN109840598A (en) * 2019-04-29 2019-06-04 深兰人工智能芯片研究院(江苏)有限公司 A kind of method for building up and device of deep learning network model
CN110146099A (en) * 2019-05-31 2019-08-20 西安工程大学 A kind of synchronous superposition method based on deep learning
CN110296705A (en) * 2019-06-28 2019-10-01 苏州瑞久智能科技有限公司 A kind of vision SLAM winding detection method based on learning distance metric
CN110296705B (en) * 2019-06-28 2022-01-25 苏州瑞久智能科技有限公司 Visual SLAM loop detection method based on distance metric learning
CN110399821B (en) * 2019-07-17 2023-05-30 上海师范大学 Customer satisfaction acquisition method based on facial expression recognition
CN110487274B (en) * 2019-07-30 2021-01-29 中国科学院空间应用工程与技术中心 SLAM method and system for weak texture scene, navigation vehicle and storage medium
CN110487274A (en) * 2019-07-30 2019-11-22 中国科学院空间应用工程与技术中心 SLAM method, system, navigation vehicle and storage medium for weak texture scene
CN110738128A (en) * 2019-09-19 2020-01-31 天津大学 repeated video detection method based on deep learning
CN110659619A (en) * 2019-09-27 2020-01-07 昆明理工大学 Depth space-time information-based correlation filtering tracking method
CN111144550A (en) * 2019-12-27 2020-05-12 中国科学院半导体研究所 Simplex deep neural network model based on homologous continuity and construction method
CN111243021A (en) * 2020-01-06 2020-06-05 武汉理工大学 Vehicle-mounted visual positioning method and system based on multiple combined cameras and storage medium
CN111241986B (en) * 2020-01-08 2021-03-30 电子科技大学 Visual SLAM closed loop detection method based on end-to-end relationship network
CN111241986A (en) * 2020-01-08 2020-06-05 电子科技大学 Visual SLAM closed loop detection method based on end-to-end relationship network
CN111753789A (en) * 2020-07-01 2020-10-09 重庆邮电大学 Robot vision SLAM closed loop detection method based on stack type combined self-encoder
CN113066152A (en) * 2021-03-18 2021-07-02 内蒙古工业大学 AGV map construction method and system
CN113066152B (en) * 2021-03-18 2022-05-27 内蒙古工业大学 AGV map construction method and system

Also Published As

Publication number Publication date
CN109341703B (en) 2022-07-01

Similar Documents

Publication Publication Date Title
CN109341703A (en) A kind of complete period uses the vision SLAM algorithm of CNNs feature detection
CN110956651B (en) Terrain semantic perception method based on fusion of vision and vibrotactile sense
Fang et al. 3d-siamrpn: An end-to-end learning method for real-time 3d single object tracking using raw point cloud
CN111311666B (en) Monocular vision odometer method integrating edge features and deep learning
CN109800689B (en) Target tracking method based on space-time feature fusion learning
CN110084850B (en) Dynamic scene visual positioning method based on image semantic segmentation
CN110378281A (en) Group Activity recognition method based on pseudo- 3D convolutional neural networks
CN108921107B (en) Pedestrian re-identification method based on sequencing loss and Simese network
CN110135249B (en) Human behavior identification method based on time attention mechanism and LSTM (least Square TM)
CN109816689A (en) A kind of motion target tracking method that multilayer convolution feature adaptively merges
CN109559320A (en) Realize that vision SLAM semanteme builds the method and system of figure function based on empty convolution deep neural network
CN106023211A (en) Robot image positioning method and system base on deep learning
CN108805149A (en) A kind of winding detection method and device of visual synchronization positioning and map structuring
CN111161315A (en) Multi-target tracking method and system based on graph neural network
Wen et al. Hybrid semi-dense 3D semantic-topological mapping from stereo visual-inertial odometry SLAM with loop closure detection
Wei et al. GMSK-SLAM: a new RGB-D SLAM method with dynamic areas detection towards dynamic environments
Ma et al. Line-based stereo SLAM by junction matching and vanishing point alignment
Yuan et al. A novel approach to image-sequence-based mobile robot place recognition
Li et al. Multi-modal neural feature fusion for automatic driving through perception-aware path planning
Zhu et al. DualDA-Net: Dual-head rectification for cross-domain object detection of remote sensing
Feng et al. Effective venue image retrieval using robust feature extraction and model constrained matching for mobile robot localization
CN112967317B (en) Visual odometry method based on convolutional neural network architecture in dynamic environment
CN112699954B (en) Closed loop detection method based on deep learning and bag-of-word model
Ding et al. Robust indoor slam based on pedestrian recognition by using rgb-d camera
Lee et al. Camera pose estimation using voxel-based features for autonomous vehicle localization tracking

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant