CN110599489A - Target space positioning method - Google Patents

Target space positioning method Download PDF

Info

Publication number
CN110599489A
CN110599489A CN201910792381.4A CN201910792381A CN110599489A CN 110599489 A CN110599489 A CN 110599489A CN 201910792381 A CN201910792381 A CN 201910792381A CN 110599489 A CN110599489 A CN 110599489A
Authority
CN
China
Prior art keywords
target
pixel point
positioning
dimensional coordinate
coordinate set
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201910792381.4A
Other languages
Chinese (zh)
Inventor
韩守东
夏晨斐
陈国荣
刘巾英
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Huazhong University of Science and Technology
Original Assignee
Huazhong University of Science and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Huazhong University of Science and Technology filed Critical Huazhong University of Science and Technology
Priority to CN201910792381.4A priority Critical patent/CN110599489A/en
Publication of CN110599489A publication Critical patent/CN110599489A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G01MEASURING; TESTING
    • G01CMEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
    • G01C11/00Photogrammetry or videogrammetry, e.g. stereogrammetry; Photographic surveying
    • G01C11/04Interpretation of pictures
    • G01C11/06Interpretation of pictures by comparison of two or more pictures of the same area
    • G01C11/08Interpretation of pictures by comparison of two or more pictures of the same area the pictures not being supported in the same relative position as when they were taken
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20212Image combination
    • G06T2207/20221Image fusion; Image merging

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Theoretical Computer Science (AREA)
  • Multimedia (AREA)
  • Radar, Positioning & Navigation (AREA)
  • Remote Sensing (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a target space positioning method, which comprises the following steps: simultaneously collecting two target images with different visual angles, and positioning through a binocular visual space to obtain a three-dimensional coordinate set of each pixel point in one image; classifying and regressing the target in the image based on example segmentation to obtain a target binary mask set; and mapping and fusing pixel point coordinates based on the three-dimensional coordinate set of each pixel point and the target binary mask set to obtain a target three-dimensional coordinate in the image, so as to realize target space positioning. Based on binocular vision space positioning, acquiring sparse three-dimensional coordinates for describing real scale and space positioning information of a target; and performing monocular instance segmentation on the interested specific class target by adopting a deep learning method, and accurately defining the semantic attribute of the pixel. And finally, based on the coupling of the three-dimensional coordinates and the example segmentation result, positioning the target space under the connection relation of the pixel coordinates, so that the three-dimensional coordinates which are scattered sparsely are densified, and the positioning accuracy is improved.

Description

Target space positioning method
Technical Field
The invention belongs to the field of target space positioning, and particularly relates to a target space positioning method.
Background
With the continuous development of production and life, the position information of the target is concerned by more and more fields, and the target space positioning has wide application in many application scenes, such as factory dangerous area alarm, automatic driving obstacle prediction, aerospace position and attitude estimation and the like.
The existing target space positioning means mainly include three types, namely hardware auxiliary methods, traditional geometric methods and depth estimation methods based on deep learning. The hardware auxiliary method mainly refers to the steps of analyzing and calculating a positioning target through transmitting/receiving active signals by means of radio frequency identification, particle filtering, WI-FI, Bluetooth, radar, multi-sensor fusion and the like. The traditional geometric method is a method for acquiring three-dimensional coordinates of a target in geometric modes such as camera calibration or optical instrument measurement. The deep estimation method based on deep learning mainly refers to a method for carrying out deep estimation after collecting samples containing deep information and training by using a deep learning network.
However, the three methods have the defects that signal transmission is easily interfered and even shielded, point clouds are sparse and difficult to be utilized in a centralized manner, scene mobility is poor, and real scale information cannot be restored. Therefore, how to overcome the above defects to improve the accuracy of target spatial positioning is a technical problem to be solved in the field.
Disclosure of Invention
The invention provides a target space positioning method, which is used for solving the technical problem that the existing target space positioning method cannot overcome inherent defects and further causes insufficient space positioning precision due to independent adoption of a single type method such as a geometric method, a deep learning method and the like.
The technical scheme for solving the technical problems is as follows: a method of spatial localization of an object, comprising:
s1, simultaneously collecting two target images with different visual angles, and positioning through a binocular visual space to obtain a three-dimensional coordinate set of each pixel point in one of the images;
s2, carrying out classification regression on the target in the image based on example segmentation to obtain a target binary mask set;
and S3, mapping and fusing pixel point coordinates based on the pixel point three-dimensional coordinate sets and the target binary mask set to obtain target three-dimensional coordinates in the image, and realizing target space positioning.
The invention has the beneficial effects that: based on binocular vision space positioning, acquiring sparse three-dimensional coordinates (actual distance values relative to an original point) for describing real scale and space positioning information of a target; by adopting a deep learning method, monocular instance segmentation is carried out on the interested specific class target, the semantic attribute of the pixel is accurately defined, the target contour is finer, and the huge deviation possibly caused by introducing a large number of non-target pixel points into the target position indicated by a rectangular frame is avoided. And finally, based on the coupling of the three-dimensional coordinates and the example segmentation result, carrying out target space positioning under the connection relation of the pixel coordinates, so that the sparsely-dispersed three-dimensional coordinates are densified to position the target, and the accuracy of the target space positioning is improved. The invention improves the space positioning effect of the traditional visual positioning method, and can quickly and accurately realize the three-dimensional accurate positioning of the target under the real space scene based on the real space size and combined with deep learning.
On the basis of the technical scheme, the invention can be further improved as follows.
Further, in S1, the simultaneously acquiring two target images at different viewing angles specifically includes: and simultaneously acquiring two target images with different visual angles by adopting a calibrated binocular camera or two monocular cameras.
The invention has the further beneficial effects that: by adopting a binocular camera or two monocular cameras, the similarity (including the image size) of the acquired target images of two different visual angles can be maximized, so that the accuracy of the three-dimensional coordinates is improved.
Further, in S1, the binocular vision space positioning specifically includes:
and performing image correction, stereo matching and depth recovery on the two target images with different visual angles by adopting a visual principle.
Further, in S2, the performing classification regression on the target in the image specifically includes:
and performing classification regression on the targets in the image by adopting a Mask R-CNN algorithm.
The invention has the further beneficial effects that: and performing monocular instance segmentation on the interested specific class target by adopting a Mask R-CNN algorithm, and taking the segmented object as foreground/background prior information, thereby being beneficial to realizing the classification of pixels inside and outside the target contour and accurately defining the attribute of the image pixel.
Further, the S2 includes:
s2.1, carrying out classification regression on the target in the image based on example segmentation to obtain a target contour and a target binary mask set of a delineating region of the target contour;
s2.2, reducing the target contour through a convolution kernel to obtain a new target contour and a target binary mask set of a delineating area of the new target contour.
The invention has the further beneficial effects that: in order to eliminate non-target errors introduced by the target contour edge, pixel corrosion is introduced, edge error optimization is carried out, and target precision is provided.
Further, in S2.2, the narrowing the target contour specifically includes:
and adopting a convolution kernel to scan all the pixel points in sequence, carrying out AND or NOT calculation on binary masks of all the pixel points covered by the convolution kernel when any pixel point is scanned, and updating the binary masks of all the covered pixel points according to the calculation result so as to reduce the target contour.
The invention has the further beneficial effects that: the convolution kernel scans and operates among elements in the convolution kernel, and the value of each element in the convolution kernel is updated, so that the aim of reducing the target contour is fulfilled, the corrosion effect is realized, the calculation amount is small, and the method is convenient and quick.
Further, the S3 includes:
s3.1, extracting a target three-dimensional coordinate set corresponding to the target pixel point coordinate set from the target pixel point three-dimensional coordinate set corresponding to the target binary mask set;
and S3.2, performing mean filtering on the target three-dimensional coordinate set to obtain a target average three-dimensional coordinate, and realizing target space positioning.
The invention has the further beneficial effects that: aiming at the problems that a large amount of noise is easily introduced due to the fact that a three-dimensional coordinate set of each pixel point is obtained only by binocular visual space positioning, and the three-dimensional coordinates are sparse and difficult to use in a centralized mode, a target pixel point coordinate set obtained by example segmentation is fused, fine outlines and semantic classification of targets can be obtained, a pixel-level optimization algorithm for the targets is added beneficially, and positioning accuracy is further improved beneficially.
Further, before S3.1, the S3 further includes:
and uniformly and randomly sampling a target pixel point coordinate set corresponding to the target binary mask set to obtain a new target pixel point coordinate set, and executing the S3.1.
The invention has the further beneficial effects that: through even random sampling, can practice thrift computational resource, improve the positioning speed, the operation of even random sampling has the error optimization effect simultaneously, and very big degree reduces marginal error, further improves the location rate of accuracy.
Further, before S3.1 and after S3.2, the S3 further includes:
and based on the new target three-dimensional coordinate set, carrying out background point filtering on the target pixel point coordinate set to obtain a new target three-dimensional coordinate set.
The invention has the further beneficial effects that: and the error optimization is carried out by utilizing the background pixel point filtering, so that the edge error and the background interference can be greatly reduced, and the positioning accuracy is further improved.
The invention also provides a storage medium, wherein the storage medium stores instructions, and when the instructions are read by a computer, the computer is enabled to execute any one of the target space positioning methods.
Drawings
Fig. 1 is a flowchart of a target space positioning method according to an embodiment of the present invention;
FIG. 2 is a comparison graph of pixel corrosion effects provided by an embodiment of the present invention;
fig. 3 is a network frame diagram of target space positioning based on fusion of three-dimensional coordinates of all pixel points and a target binary mask set according to an embodiment of the present invention;
fig. 4 is a flowchart illustrating a coupling process between a three-dimensional coordinate and a target binary mask set according to an embodiment of the present invention.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention more apparent, the present invention is described in further detail below with reference to the accompanying drawings and embodiments. It should be understood that the specific embodiments described herein are merely illustrative of the invention and are not intended to limit the invention. In addition, the technical features involved in the embodiments of the present invention described below may be combined with each other as long as they do not conflict with each other.
Example one
A method 100 for spatial localization of an object, as shown in fig. 1, comprising:
step 110, simultaneously collecting two target images with different visual angles, and positioning through a binocular visual space to obtain a three-dimensional coordinate set of each pixel point in one of the images;
step 120, classifying and regressing the target in the image based on example segmentation to obtain a target binary mask set;
and step 130, mapping and fusing pixel point coordinates based on the pixel point three-dimensional coordinate sets and the target binary mask set to obtain target three-dimensional coordinates in the image, so as to realize target space positioning.
Acquiring sparse three-dimensional coordinates (actual distance values relative to an origin) for describing real dimensions of a target and space positioning information based on binocular vision space positioning; by adopting a deep learning method, monocular instance segmentation is carried out on the interested specific class target, the semantic attribute of the pixel is accurately defined, the target contour is finer, and the huge deviation possibly caused by introducing a large number of non-target pixel points into the target position indicated by a rectangular frame is avoided. And finally, positioning a target space based on the coupling of the three-dimensional coordinates and the example segmentation result, and particularly fusing the three-dimensional coordinates and the target binary mask set to densify the sparse three-dimensional coordinates and improve the accuracy of positioning the target space. The embodiment improves the space positioning effect of the traditional visual positioning method, and can quickly and accurately realize the three-dimensional accurate positioning of the target under the real space scene.
Preferably, in step 110, two target images with different viewing angles are acquired simultaneously, specifically, two target images with different viewing angles are acquired simultaneously by using a calibrated binocular camera or two monocular cameras.
For example, the experimental environment is an indoor laboratory scene, two Logitech/Robotic C920 network cameras with the resolution of 640 × 480 are used, the target category is designated as a pedestrian, and target images at the same time, the same scene and different viewing angles are obtained.
By adopting a binocular camera or two monocular cameras, the similarity (including the image size) of the acquired target images of two different visual angles can be maximized, so that the precision of three-dimensional coordinates is improved.
Preferably, in step 110, the binocular vision space positioning specifically includes: and (3) performing image correction, stereo matching and depth recovery on two target images with different visual angles by adopting a visual principle.
The stereo matching method can be a BM method, and three-dimensional coordinates are derived according to parallax.
Based on the captured images, 307200 sets of data were obtained, each set including X, Y, Z coordinate values with the left camera as the origin.
Preferably, in step 120, a classification regression is performed on the target in the image, specifically: and performing classification regression on the targets in the image by adopting a Mask R-CNN algorithm. The Mask R-CNN algorithm uses a Region generation Network (RPN) to find regions of Interest (RoI) in an input image, where each Region of Interest includes a detection target. And classifying and regressing the target by using a pre-training model, outputting the class and frame positioning information of the target, and generating and outputting binary mask information of the target by using a convolution network.
For example, a Mask R-CNN algorithm is adopted to perform example segmentation on the left image, and 640 × 480 binary masks of the pedestrian category are classified and regressed to output.
Preferably, step 120 includes:
step 121, classifying and regressing the target in the image based on example segmentation to obtain a target binary mask set of a target contour and a target delineated region;
and step 122, reducing the target contour through a convolution kernel to obtain a new target contour and a target binary mask set of the delineating area.
In order to eliminate non-target errors introduced by the target contour edge, pixel corrosion is introduced, edge error optimization is carried out, and target positioning precision is provided.
Preferably, in step 122, the narrowing of the target contour specifically includes: and sequentially scanning all the pixel points by adopting a convolution kernel, carrying out AND or NOT calculation on binary masks of all the pixel points covered by the convolution kernel when any pixel point is scanned, updating the binary masks of the pixel points in the covered area according to a calculation result, and reducing the target contour.
For example, in order to eliminate non-target errors introduced by an edge, pixel erosion is added to perform edge error optimization, and the specific structure of the edge error optimization is as follows: selecting a convolution kernel with the size of n multiplied by n and the value of 1 as a corrosion factor, and dividing the binary mask by using a corrosion factor scanning example. Supposing that the target binary mask is 1 and the non-target binary mask is 0, carrying out AND operation on corresponding positions in each scanning, taking the minimum value as a scanning result, and taking the value of each pixel point as at most n experienced by the pixel point2Minimum of the secondary scan results. And updating all pixel values after the corrosion factor scans the whole image to obtain a final mask of the example segmentation.
Specifically, convolution kernels with the sizes of 3 × 3, 5 × 5 and 10 × 10 and the values of 1 can be selected as the two-value mask for dividing the corrosion factor scanning example, and operation is performed on corresponding positions in each scanning, and the minimum value is taken as a scanning result. The value of each pixel point is the minimum value of the scanning result of at most 9, 25 or 100 times that the pixel point undergoes, and all pixel values are updated after the full image is scanned, so as to obtain an example target final mask, as shown in fig. 2, (a) in fig. 2 is that no corrosion target mask is introduced, (b) in fig. 2 is a 3 × 3 corrosion target mask, (c) in fig. 2 is a 5 × 5 corrosion target mask, and (d) in fig. 2 is a 10 × 10 corrosion target mask, and as the size of the convolution kernel increases, the target profile becomes smaller and smaller. Thus, the greater the erosion factor (i.e., convolution kernel size), the greater the pixel erosion, and the lower the likelihood of introducing edge errors.
Preferably, step 130 includes:
131, extracting a target three-dimensional coordinate set corresponding to the target pixel point coordinate set from the target pixel point three-dimensional coordinate sets according to the target pixel point coordinate sets corresponding to the target binary mask set;
and 132, performing mean filtering on the target three-dimensional coordinate set to obtain a target average three-dimensional coordinate, and realizing target space positioning.
An overall network framework fusing the three-dimensional coordinates and the example segmentation Mask results is shown in fig. 3, the three-dimensional coordinates of all pixel points in an image are derived through binocular visual space positioning (namely, the content shown by a left dotted line frame in fig. 3), an example final target binary Mask is obtained through a Mask R-CNN network and edge error optimization (such as contour corrosion), and finally, a target space three-dimensional positioning result is generated under fusion calculation (such as shown in fig. 4). It should be noted that the stereo image pair in fig. 3 is two target images with different viewing angles acquired simultaneously, and here, the target images can be divided into a left image and a right image. The left image is used here for instance segmentation, depending on the processing software. The "masks" in fig. 3 each represent a target binary mask set to which the target corresponds.
Based on the final binary mask of the example target, the mask is firstly analyzed and all pixel values are detected, and two-dimensional coordinates of pixel point images with pixel values not being 0(0 represents the binary mask of the background part, and 1 represents the binary mask of the target part) are stored. Then importing three-dimensional coordinates of all pixel points, specifically including three-dimensional coordinates of space points corresponding to all pixel points; and finally, extracting three-dimensional coordinates corresponding to the example target sample points, and performing mean value filtering calculation on the extracted three-dimensional coordinates to obtain a target positioning result.
Preferably, before step 131, step 130 further comprises: and uniformly and randomly sampling a target pixel point coordinate set corresponding to the target binary mask set to obtain a new target pixel point coordinate set, and executing the step 131.
In order to save computing resources and improve positioning speed, the extracted target pixel point coordinate set needs to be uniformly and randomly sampled, which specifically includes: a uniform randomly sampled network. The input of the network is a target binary mask set divided by an example, and the coordinates of all pixel points of a target can be obtained. If M is greater than or equal to N, then everyEach group of the pixel points is randomly sampled to be used as a sample point, and N sample points are obtained; if M is<N, each pixel point is a sample point, and M sample points are total, wherein M is the coordinate number of the target pixel point, and N is the sampling point number of the target pixel point
For example, based on the final target binary mask set of the example, there are 57297 target pixel points, the number of sampling points is set to 3000 according to the computing resources, each 19 pixel points are in one group, one pixel point is randomly sampled in each group as a sample point, there are 3000 sample points, and the sample points are uniformly distributed, so that the overall spatial position of the target can be reflected.
And an error optimization link is designed, uniform random sampling is introduced for error optimization, so that edge errors and background interference are greatly reduced, and the positioning accuracy is further improved.
Preferably, after step 131 and before step 132, step 130 further comprises: and based on the new target three-dimensional coordinate set, carrying out background point filtering on the target pixel point coordinate set to obtain the new target three-dimensional coordinate set.
In order to further improve the positioning accuracy, the extracted target pixel sample points need to be filtered by background points, so that the target three-dimensional coordinate set is ensured to introduce fewer non-target pixel point coordinates. The specific structure is a background filtering network, the input of the network is target pixel sample points (namely a new target pixel point coordinate set) which are uniformly and randomly sampled, three-dimensional coordinates of all the sample points are extracted, the sample point depths are clustered, and background type pixel points are filtered.
For example, based on 3000 sample points after uniform random sampling, three-dimensional coordinates of all sample points are extracted, and the sample point depths are clustered. In this embodiment, when the depth mean of one class is greater than twice the depth mean of another class and the number of members is less than one fifth of the depth mean of another class, the class is determined as a background class and filtered. In this embodiment, the real depth of the example target is 50cm, and the depth clustering of the sample points is two types: the mean depth value of the class A is 49.21cm, and the class A contains 2977 sample points; the mean depth of class B was 100000cm, containing 23 sample points. Therefore, the B class is judged to be the background class, the B class is filtered, an error optimization link is designed for the final positioning result of 49.21cm, error optimization is carried out by utilizing background pixel point filtering, edge errors and background interference are greatly reduced, and the positioning accuracy is further improved.
Example two
A storage medium having stored therein instructions, which when read by a computer, cause the computer to execute any one of the target space positioning methods described in the first embodiment above.
The related technical solution is the same as the first embodiment, and is not described herein again.
It will be understood by those skilled in the art that the foregoing is only a preferred embodiment of the present invention, and is not intended to limit the invention, and that any modification, equivalent replacement, or improvement made within the spirit and principle of the present invention should be included in the scope of the present invention.

Claims (10)

1. A method for spatially locating an object, comprising:
s1, simultaneously collecting two target images with different visual angles, and positioning through a binocular visual space to obtain a three-dimensional coordinate set of each pixel point in one of the images;
s2, carrying out classification regression on the target in the image based on example segmentation to obtain a target binary mask set;
and S3, mapping and fusing pixel point coordinates based on the pixel point three-dimensional coordinate sets and the target binary mask set to obtain target three-dimensional coordinates in the image, and realizing target space positioning.
2. The method according to claim 1, wherein in S1, the step of simultaneously acquiring two target images from different viewing angles includes: and simultaneously acquiring two target images with different visual angles by adopting a calibrated binocular camera or two monocular cameras.
3. The method for spatial localization of objects according to claim 1, wherein in S1, the binocular vision spatial localization specifically comprises:
and performing image correction, stereo matching and depth recovery on the two target images with different visual angles by adopting a visual principle.
4. The method for spatial localization of objects according to claim 1, wherein in S2, the classifying regression of the objects in the image is specifically:
and performing classification regression on the targets in the image by adopting a Mask R-CNN algorithm.
5. The method as claimed in claim 1, wherein said S2 includes:
s2.1, carrying out classification regression on the target in the image based on example segmentation to obtain a target contour and a target binary mask set of a delineating region of the target contour;
s2.2, reducing the target contour through a convolution kernel to obtain a new target contour and a target binary mask set of a delineating area of the new target contour.
6. The method according to claim 5, wherein in S2.2, the reducing the target contour specifically comprises:
and adopting a convolution kernel to scan all the pixel points in sequence, carrying out AND or NOT calculation on binary masks of all the pixel points covered by the convolution kernel when any pixel point is scanned, and updating the binary masks of all the covered pixel points according to the calculation result so as to reduce the target contour.
7. The method according to any one of claims 1 to 6, wherein said S3 includes:
s3.1, extracting a target three-dimensional coordinate set corresponding to the target pixel point coordinate set from the target pixel point three-dimensional coordinate set corresponding to the target binary mask set;
and S3.2, performing mean filtering on the target three-dimensional coordinate set to obtain a target average three-dimensional coordinate, and realizing target space positioning.
8. The method as claimed in claim 7, wherein said S3 further includes, before said S3.1:
and uniformly and randomly sampling a target pixel point coordinate set corresponding to the target binary mask set to obtain a new target pixel point coordinate set, and executing the S3.1.
9. The method as claimed in claim 7, wherein after s.1 and before S3.2, the step S3 further includes:
and based on the new target three-dimensional coordinate set, carrying out background point filtering on the target pixel point coordinate set to obtain a new target three-dimensional coordinate set.
10. A storage medium having stored thereon instructions which, when read by a computer, cause the computer to carry out a method of spatial localization of an object as claimed in any one of claims 1 to 9.
CN201910792381.4A 2019-08-26 2019-08-26 Target space positioning method Pending CN110599489A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910792381.4A CN110599489A (en) 2019-08-26 2019-08-26 Target space positioning method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910792381.4A CN110599489A (en) 2019-08-26 2019-08-26 Target space positioning method

Publications (1)

Publication Number Publication Date
CN110599489A true CN110599489A (en) 2019-12-20

Family

ID=68855590

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910792381.4A Pending CN110599489A (en) 2019-08-26 2019-08-26 Target space positioning method

Country Status (1)

Country Link
CN (1) CN110599489A (en)

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112163502A (en) * 2020-09-24 2021-01-01 电子科技大学 Visual positioning method under indoor dynamic scene
CN112184797A (en) * 2020-10-15 2021-01-05 广州计量检测技术研究院 Method for spatially positioning key part of kilogram group weight
CN112541936A (en) * 2020-12-09 2021-03-23 中国科学院自动化研究所 Method and system for determining visual information of operating space of actuating mechanism
CN112884841A (en) * 2021-04-14 2021-06-01 哈尔滨工业大学 Binocular vision positioning method based on semantic target
CN113267128A (en) * 2021-05-31 2021-08-17 西南石油大学 Binocular vision automatic side slope displacement monitoring method
CN113298702A (en) * 2021-06-23 2021-08-24 重庆科技学院 Reordering and dividing method based on large-size image pixel points
CN114359411A (en) * 2022-01-10 2022-04-15 杭州巨岩欣成科技有限公司 Method and device for detecting drowning prevention target of swimming pool, computer equipment and storage medium
WO2023098487A1 (en) * 2021-11-30 2023-06-08 西门子股份公司 Target detection method and apparatus, electronic device, and computer storage medium
CN116798056A (en) * 2023-08-28 2023-09-22 星汉智能科技股份有限公司 Form image positioning method, apparatus, device and computer readable storage medium
CN117541590A (en) * 2024-01-10 2024-02-09 腾讯科技(深圳)有限公司 Image processing method and device, storage medium and electronic equipment

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107883875A (en) * 2017-11-23 2018-04-06 哈尔滨工程大学 Autonomous type sea cucumber finishing device visual detection positioning device and vision-based detection localization method
CN108682039A (en) * 2018-04-28 2018-10-19 国网山西省电力公司电力科学研究院 A kind of binocular stereo vision measurement method
CN108876855A (en) * 2018-05-28 2018-11-23 哈尔滨工程大学 A kind of sea cucumber detection and binocular visual positioning method based on deep learning
CN109903507A (en) * 2019-03-04 2019-06-18 上海海事大学 A kind of fire disaster intelligent monitor system and method based on deep learning
CN110008915A (en) * 2019-04-11 2019-07-12 电子科技大学 The system and method for dense human body attitude estimation is carried out based on mask-RCNN
CN110060299A (en) * 2019-04-18 2019-07-26 中国测绘科学研究院 Danger source identifies and positions method in passway for transmitting electricity based on binocular vision technology

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107883875A (en) * 2017-11-23 2018-04-06 哈尔滨工程大学 Autonomous type sea cucumber finishing device visual detection positioning device and vision-based detection localization method
CN108682039A (en) * 2018-04-28 2018-10-19 国网山西省电力公司电力科学研究院 A kind of binocular stereo vision measurement method
CN108876855A (en) * 2018-05-28 2018-11-23 哈尔滨工程大学 A kind of sea cucumber detection and binocular visual positioning method based on deep learning
CN109903507A (en) * 2019-03-04 2019-06-18 上海海事大学 A kind of fire disaster intelligent monitor system and method based on deep learning
CN110008915A (en) * 2019-04-11 2019-07-12 电子科技大学 The system and method for dense human body attitude estimation is carried out based on mask-RCNN
CN110060299A (en) * 2019-04-18 2019-07-26 中国测绘科学研究院 Danger source identifies and positions method in passway for transmitting electricity based on binocular vision technology

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
吴保烁等: "双目立体视觉的测距技术研究", 《电视技术》 *
欧攀等: "基于Mask RCNN的目标识别与空间定位", 《计算机测量与控制》 *
马建设等: "基于轮廓提取与深度筛选的双目三维重构技术", 《计算机工程与科学》 *

Cited By (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112163502A (en) * 2020-09-24 2021-01-01 电子科技大学 Visual positioning method under indoor dynamic scene
CN112163502B (en) * 2020-09-24 2022-07-12 电子科技大学 Visual positioning method under indoor dynamic scene
CN112184797A (en) * 2020-10-15 2021-01-05 广州计量检测技术研究院 Method for spatially positioning key part of kilogram group weight
CN112184797B (en) * 2020-10-15 2023-04-07 广州计量检测技术研究院 Method for spatially positioning key part of kilogram group weight
CN112541936B (en) * 2020-12-09 2022-11-08 中国科学院自动化研究所 Method and system for determining visual information of operating space of actuating mechanism
CN112541936A (en) * 2020-12-09 2021-03-23 中国科学院自动化研究所 Method and system for determining visual information of operating space of actuating mechanism
CN112884841A (en) * 2021-04-14 2021-06-01 哈尔滨工业大学 Binocular vision positioning method based on semantic target
CN113267128A (en) * 2021-05-31 2021-08-17 西南石油大学 Binocular vision automatic side slope displacement monitoring method
CN113298702B (en) * 2021-06-23 2023-08-04 重庆科技学院 Reordering and segmentation method based on large-size image pixel points
CN113298702A (en) * 2021-06-23 2021-08-24 重庆科技学院 Reordering and dividing method based on large-size image pixel points
WO2023098487A1 (en) * 2021-11-30 2023-06-08 西门子股份公司 Target detection method and apparatus, electronic device, and computer storage medium
CN114359411B (en) * 2022-01-10 2022-08-09 杭州巨岩欣成科技有限公司 Method and device for detecting drowning prevention target of swimming pool, computer equipment and storage medium
CN114359411A (en) * 2022-01-10 2022-04-15 杭州巨岩欣成科技有限公司 Method and device for detecting drowning prevention target of swimming pool, computer equipment and storage medium
CN116798056A (en) * 2023-08-28 2023-09-22 星汉智能科技股份有限公司 Form image positioning method, apparatus, device and computer readable storage medium
CN116798056B (en) * 2023-08-28 2023-11-17 星汉智能科技股份有限公司 Form image positioning method, apparatus, device and computer readable storage medium
CN117541590A (en) * 2024-01-10 2024-02-09 腾讯科技(深圳)有限公司 Image processing method and device, storage medium and electronic equipment
CN117541590B (en) * 2024-01-10 2024-04-09 腾讯科技(深圳)有限公司 Image processing method and device, storage medium and electronic equipment

Similar Documents

Publication Publication Date Title
CN110599489A (en) Target space positioning method
US8199977B2 (en) System and method for extraction of features from a 3-D point cloud
CN111665842B (en) Indoor SLAM mapping method and system based on semantic information fusion
CN111563442A (en) Slam method and system for fusing point cloud and camera image data based on laser radar
CN111553859A (en) Laser radar point cloud reflection intensity completion method and system
CN109887021B (en) Cross-scale-based random walk stereo matching method
US7747106B2 (en) Method and system for filtering, registering, and matching 2.5D normal maps
US9576375B1 (en) Methods and systems for detecting moving objects in a sequence of image frames produced by sensors with inconsistent gain, offset, and dead pixels
CN109961506A (en) A kind of fusion improves the local scene three-dimensional reconstruction method of Census figure
CN114022830A (en) Target determination method and target determination device
JP2014523572A (en) Generating map data
CN112364865B (en) Method for detecting small moving target in complex scene
CN110910421A (en) Weak and small moving object detection method based on block characterization and variable neighborhood clustering
Cui et al. Dense depth-map estimation based on fusion of event camera and sparse LiDAR
CN115222884A (en) Space object analysis and modeling optimization method based on artificial intelligence
CN113822996B (en) Pose estimation method and device for robot, electronic device and storage medium
Sun et al. An improved binocular visual odometry algorithm based on the random sample consensus in visual navigation systems
US20210350595A1 (en) Methods and apparatus for generating point cloud histograms
CN113536959A (en) Dynamic obstacle detection method based on stereoscopic vision
CN117292076A (en) Dynamic three-dimensional reconstruction method and system for local operation scene of engineering machinery
CN116843754A (en) Visual positioning method and system based on multi-feature fusion
Liu et al. Outdoor camera calibration method for a GPS & camera based surveillance system
Arnaud et al. Partial linear gaussian models for tracking in image sequences using sequential monte carlo methods
CN112365600B (en) Three-dimensional object detection method
CN110399892B (en) Environmental feature extraction method and device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20191220