CN115272493B - Abnormal target detection method and device based on continuous time sequence point cloud superposition - Google Patents

Abnormal target detection method and device based on continuous time sequence point cloud superposition Download PDF

Info

Publication number
CN115272493B
CN115272493B CN202211145212.XA CN202211145212A CN115272493B CN 115272493 B CN115272493 B CN 115272493B CN 202211145212 A CN202211145212 A CN 202211145212A CN 115272493 B CN115272493 B CN 115272493B
Authority
CN
China
Prior art keywords
point
abnormal
point cloud
points
mapping
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202211145212.XA
Other languages
Chinese (zh)
Other versions
CN115272493A (en
Inventor
黄倩
刘云涛
朱永东
赵志峰
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhejiang Lab
Original Assignee
Zhejiang Lab
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhejiang Lab filed Critical Zhejiang Lab
Priority to CN202211145212.XA priority Critical patent/CN115272493B/en
Publication of CN115272493A publication Critical patent/CN115272493A/en
Application granted granted Critical
Publication of CN115272493B publication Critical patent/CN115272493B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/80Analysis of captured images to determine intrinsic or extrinsic camera parameters, i.e. camera calibration
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/02Affine transformations
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/762Arrangements for image or video recognition or understanding using pattern recognition or machine learning using clustering, e.g. of similar faces in social networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10028Range image; Depth image; 3D point clouds
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30244Camera pose

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Artificial Intelligence (AREA)
  • Health & Medical Sciences (AREA)
  • Computing Systems (AREA)
  • Databases & Information Systems (AREA)
  • Evolutionary Computation (AREA)
  • General Health & Medical Sciences (AREA)
  • Medical Informatics (AREA)
  • Software Systems (AREA)
  • Multimedia (AREA)
  • Optical Radar Systems And Details Thereof (AREA)

Abstract

The invention discloses an abnormal target detection method and device based on continuous time sequence point cloud superposition, which adopts a time sequence superposition method, utilizes continuous time sequence point cloud data frame mapping superposition to generate a background depth map, converts non-fixed and disordered point clouds into a fixed and ordered depth map, identifies abnormal point clouds based on the difference between the depth values of mapping points of the abnormal target point clouds and the depth values of the background depth map under corresponding coordinates, adds semantic category information to the abnormal point clouds by utilizing example region categories of the abnormal point clouds falling on a space-time alignment image, improves the anti-interference capability of the abnormal point cloud clustering by combining the space and semantic distances of the point clouds, forms accurate and independent point cloud target clusters, and calculates and generates detection information of the target clusters. According to the method, the background depth map is constructed by a continuous time sequence superposition method to detect the abnormal point cloud target, so that the problem of low accuracy of detection by directly adopting point cloud or images due to the fact that the size and the distance of the abnormal target are not fixed is solved.

Description

Abnormal target detection method and device based on continuous time sequence point cloud superposition
Technical Field
The invention relates to the technical field of intelligent perception, in particular to an abnormal target detection method and device based on continuous time sequence point cloud superposition.
Background
Along with the reduction of the cost of the sensor, more and more security monitoring scenes realize abnormity detection and alarm by installing the sensor. The system comprises a plurality of sensors, a monitoring center and a monitoring center, wherein abnormal intrusion targets such as pedestrians, non-motor vehicles and animals are monitored and identified through fusion perception calculation of the sensors in a park prevention and control range, and the system is an important application of realizing safety management of an unmanned monitoring area by using sensor equipment in a park. The low-cost solid-state laser radar is usually used for sensing the position of a target, but because the size and the distance of an abnormal target are not fixed, a deep learning method is directly adopted, detection and identification are carried out based on radar point cloud data, and the accuracy rate is low. Meanwhile, the detection range of the solid laser radar is wide, the formed point cloud space range is large, the arrangement is disordered, and the difficulty of detecting the abnormal point cloud target in the point cloud space range based on the background point cloud is high. On the other hand, the single camera cannot accurately identify the target position due to the fact that depth information cannot be obtained, and the accuracy of detection and identification of an abnormal target directly through point cloud or images is low due to the fact that the size of the abnormal target is not fixed and the position distance is not fixed.
Therefore, aiming at the problem that the solid-state laser radar and the camera cannot independently and accurately detect the abnormal target, the invention provides a continuous time sequence point cloud superposition method, which is used for constructing a background depth map to detect the abnormal point cloud target and realizing high-precision detection of the abnormal target.
Disclosure of Invention
The invention aims to provide an abnormal target detection method and device based on continuous time sequence point cloud superposition, which aims to overcome the defects of the prior art, and the abnormal target detection method and device based on continuous time sequence point cloud superposition are characterized in that a background depth map is generated by mapping and superposing continuous time sequence point cloud data frames, large-range and disordered point clouds are converted into a fixed-range and ordered depth map, then the abnormal target depth value and the depth value difference of the background depth map are utilized to identify abnormal point clouds, semantic category information is added to the abnormal point clouds by fusing image semantic information, and finally, the abnormal targets are clustered and identified by combining the space and the semantic distance of the point clouds, so that the problem of low accuracy in detection and identification directly by using the point clouds or images due to the problems of unfixed size and unfixed position distance of the abnormal targets is solved.
The purpose of the invention is realized by the following technical scheme: in a first aspect, the invention provides an abnormal target detection method based on continuous time sequence point cloud superposition, which comprises the following steps:
the method comprises the following steps: collecting a plurality of frames of point cloud data frames with continuous time sequence by a solid-state laser radar, mapping the point clouds in all the point cloud data frames to a depth map by using an affine transformation matrix from the point cloud data to image data, superposing the depth values of mapping points with the same mapping coordinates and calculating an average depth value, and updating the depth value of the corresponding coordinate of the depth map by using the obtained average depth value; repeating the step until the depth value before and after any coordinate of the depth map is updated does not change any more, wherein the updated depth map is the background depth map;
step two: acquiring a point cloud data frame of the solid-state laser radar in real time, mapping all point clouds in the point cloud data frame to a background depth map by using an affine transformation matrix from the point cloud data to image data, and judging that any point in the point clouds is a newly added abnormal point if the difference between the mapping point depth value and the depth value of the background depth map under the corresponding coordinate is greater than a threshold value;
step three: acquiring an image data frame which is aligned with a corresponding point cloud data frame in a time-space mode, segmenting all target examples in the image data frame by adopting a semantic segmentation method, respectively mapping the newly increased abnormal points to the image data frame, and adding semantic type information to the newly increased abnormal points according to the semantic type of an image target example area where the mapping points are located;
step four: clustering all the newly added abnormal points based on the space semantic joint distance between the points to form a cluster;
step five: and calculating the volume and the center point coordinates of each cluster, identifying the cluster with the volume larger than the threshold as an abnormal target, and generating the detection information of the abnormal target.
Further, the step one includes the following steps:
(1.1) defining a blank depth map, wherein the depth value of each coordinate in the blank depth map is initialized to be 0; the size of the blank depth map is consistent with that of an image shot by a camera aligned with the solid-state laser radar in time and space;
(1.2) acquiring an affine transformation matrix from the solid-state laser radar point cloud data to the camera image data aligned in time and space; the method comprises the following specific steps: controlling the time synchronization of data frames of a laser radar and a camera in a hardware line control mode, carrying out combined calibration on internal parameters of the camera and external parameters from a laser radar coordinate system to a camera coordinate system to obtain internal parameters and external parameters matrixes, and generating an affine transformation matrix from point cloud data to image data according to the obtained internal parameters and external parameters;
assuming a calibrated internal reference matrix of
Figure DEST_PATH_IMAGE001
The external reference matrix is
Figure 100002_DEST_PATH_IMAGE002
Affine transformation matrix of point cloud data to image data
Figure DEST_PATH_IMAGE003
Comprises the following steps:
Figure 100002_DEST_PATH_IMAGE004
wherein the internal reference matrix
Figure 81631DEST_PATH_IMAGE001
Dimension of 3*3, and extrinsic parameter matrix of
Figure 529930DEST_PATH_IMAGE002
Dimension of 3*4, affine transformation matrix
Figure 840825DEST_PATH_IMAGE003
Dimension size 3*4;
(1.3) setting the solid-state laser radar to be in a non-repetitive scanning mode, continuously collecting N frames of point cloud data frames in continuous time sequence according to a certain frequency, respectively mapping point clouds in all the point cloud data frames to a blank depth map by using an affine transformation matrix from the point cloud data to image data, overlapping the depth values of mapping points with the same mapping coordinates, and recording the overlapping times of the depth values; the method comprises the following specific steps: for any one point in N frames of continuous time sequence point cloud data frames, the coordinate of the point cloud in the point cloud is assumed to be
Figure DEST_PATH_IMAGE005
Mapping the point cloud data to the affine transformation matrix of the image data to the coordinates of the mapping points in the blank depth map as
Figure 100002_DEST_PATH_IMAGE006
Depth value of
Figure DEST_PATH_IMAGE007
Respectively, as follows:
Figure 100002_DEST_PATH_IMAGE008
wherein ceil means rounding up,
Figure DEST_PATH_IMAGE009
respectively representing floating point coordinate values of the point cloud mapped to the mapped points in the depth map,
Figure 100002_DEST_PATH_IMAGE010
is that
Figure DEST_PATH_IMAGE011
Divided by depth value
Figure 100002_DEST_PATH_IMAGE012
The integral coordinate value of the mapping point after the upward integration is carried out;
executing the mapping operation on the points in all the point cloud data frames, superposing the depth values of the mapping points with the same integer coordinate value in an adding mode, and recording the superposition times of the depth values under each coordinate;
(1.4) for each mapping point coordinate, calculating the average depth value of the coordinate by using the superposition depth value and the superposition times under the coordinate, and updating the depth value of the corresponding coordinate of the blank depth map by using the obtained average depth value; the method comprises the following specific steps: assuming that the depth value of the overlay under the coordinate of a certain mapping point is SumDepth, and the number of overlays is NumD, the average depth value depth of the coordinate is expressed as:
Figure DEST_PATH_IMAGE013
calculating an average depth value for all mapping point coordinates, and updating the depth value of the blank depth map at the corresponding coordinate by using the obtained average depth value;
and (1.5) repeating the step (1.3) and the step (1.4) until the depth value of any coordinate of the blank depth map is not changed before and after updating, wherein the blank depth map updated for the last time is the background depth map.
Further, in the second step, for any point in the point cloud, the coordinate of the point in the point cloud is assumed to be
Figure 100002_DEST_PATH_IMAGE014
The coordinates of the mapping points mapped into the background depth map are
Figure DEST_PATH_IMAGE015
Depth value of
Figure 100002_DEST_PATH_IMAGE016
Corresponding to the background depth map coordinates of
Figure DEST_PATH_IMAGE017
At a depth value of
Figure 100002_DEST_PATH_IMAGE018
If it satisfies
Figure DEST_PATH_IMAGE019
If the point is a new abnormal point, the point is judged to be a point in the abnormal target and not a point in the background; wherein
Figure 100002_DEST_PATH_IMAGE020
The empirical value is obtained by observing the difference in depth values between the abnormal target point and the background point.
Further, the third step includes the following steps:
(3.1) segmenting all target instances in the image data frame by adopting a Mask-RCNN-based semantic segmentation method;
(3.2) for any newly added abnormal point, assuming that the coordinate of the abnormal point in the point cloud is
Figure 100002_DEST_PATH_IMAGE021
Mapping the point cloud data to the coordinate of the mapping point in the image data frame by using the affine transformation matrix of the point cloud data to the image data
Figure DEST_PATH_IMAGE022
Is represented as follows:
Figure 100002_DEST_PATH_IMAGE023
wherein ceil means rounding up,
Figure 126049DEST_PATH_IMAGE022
integer coordinates representing mapping of point cloud data to mapping points in image dataThe value of the standard value is marked,
Figure DEST_PATH_IMAGE024
the depth value of the mapping point is obtained;
(3.3) hypothetical coordinates
Figure 487892DEST_PATH_IMAGE022
And a set of image coordinate points PixelCols contained in a certain target example, and satisfies
Figure 100002_DEST_PATH_IMAGE025
If the new abnormal point is added with the semantic category information, the new abnormal point is expressed as
Figure DEST_PATH_IMAGE026
Where cls is the semantic class of the target instance.
Further, the fourth step includes the following steps:
(4.1) searching and determining a core anomaly point; the method specifically comprises the following steps: set the space radius to
Figure DEST_PATH_IMAGE028
The minimum number of neighbors is minS; for any newly added abnormal point
Figure 100002_DEST_PATH_IMAGE029
Assuming its coordinates as
Figure 100002_DEST_PATH_IMAGE030
Semantic categories of
Figure DEST_PATH_IMAGE031
Then it is represented as
Figure 100002_DEST_PATH_IMAGE032
Go through the newly added abnormal point
Figure DEST_PATH_IMAGE033
Radius of space
Figure 294042DEST_PATH_IMAGE028
All new outliers in the range, for their spatial radius
Figure 92234DEST_PATH_IMAGE028
Any newly added abnormal point in the range
Figure 100002_DEST_PATH_IMAGE034
Assuming its coordinates as
Figure DEST_PATH_IMAGE035
Semantic categories of
Figure 100002_DEST_PATH_IMAGE036
Then it is represented as
Figure DEST_PATH_IMAGE037
If the space semantic joint distance between two points
Figure 100002_DEST_PATH_IMAGE038
Satisfies the following conditions:
Figure DEST_PATH_IMAGE039
the new abnormal point is considered
Figure 464703DEST_PATH_IMAGE034
Is newly added with abnormal points
Figure 415473DEST_PATH_IMAGE033
The neighbor of (2); if the abnormal point is newly added
Figure 939995DEST_PATH_IMAGE033
Radius of space of
Figure 225483DEST_PATH_IMAGE028
If the number of newly added abnormal points satisfying the above formula in the range is not less than minS, determining the abnormal points
Figure 493653DEST_PATH_IMAGE033
Is a core anomaly point, otherwise is a non-core anomaly point;
wherein Dis is the euclidean spatial distance between two points, and Sclass is the semantic distance between two points, which are respectively expressed as:
Figure 100002_DEST_PATH_IMAGE040
wherein
Figure DEST_PATH_IMAGE041
Is a weight of the spatial distance and is,
Figure 100002_DEST_PATH_IMAGE042
in order to be a semantic distance weight,
Figure DEST_PATH_IMAGE043
the distance threshold values are spatial semantic union distance threshold values which are empirical values;
executing the step (4.1) on all the newly added abnormal points until all the newly added abnormal points are confirmed to be core abnormal points or not; performing subsequent processing on the core abnormal point, and directly discarding the non-core abnormal point;
(4.2) clustering the core abnormal points to form a cluster; the method specifically comprises the following steps: starting from any one core anomaly point, the spatial radius of the core anomaly point is equal to that of the space
Figure 390940DEST_PATH_IMAGE028
The core abnormal points of the inner neighbors are clustered into a class, and the core abnormal points of the neighbors are continuously searched and clustered into a class from any core abnormal point of the neighbors until the core abnormal points of the neighbors cannot be found, the clustered core abnormal points are a cluster, and the cluster type is the semantic type of the initial core abnormal point; repeating the step (4.2) from any remaining core abnormal point until no new cluster is formed;
(4.3) the remaining non-clustered core outliers are outliers and are discarded directly.
Further, in step five, for each cluster, the spatial distance is maximal according to the spatial distanceCalculating the volume of the clustering cluster by using the two core abnormal points; the coordinates of two core abnormal points with the maximum space distance are assumed to be respectively
Figure 100002_DEST_PATH_IMAGE044
And
Figure DEST_PATH_IMAGE045
semantic categories are
Figure 100002_DEST_PATH_IMAGE046
Are then respectively represented as
Figure DEST_PATH_IMAGE047
Volume of the cluster
Figure 100002_DEST_PATH_IMAGE048
Expressed as:
Figure DEST_PATH_IMAGE049
coordinates of center point
Figure 100002_DEST_PATH_IMAGE050
Expressed as:
Figure DEST_PATH_IMAGE051
if it is
Figure 807402DEST_PATH_IMAGE048
Greater than a threshold value
Figure 100002_DEST_PATH_IMAGE052
Then the cluster is considered as an abnormal target, wherein
Figure 596498DEST_PATH_IMAGE052
If the upper limit value of the volume of the small target object forming the interference is an empirical value, the abnormal target detection information is as follows: the object class is
Figure DEST_PATH_IMAGE053
Distance solid state laser radar mounting origin
Figure 100002_DEST_PATH_IMAGE054
The included angle between the meter and the abscissa of the coordinate system of the solid-state laser radar is
Figure DEST_PATH_IMAGE055
In a second aspect, the invention provides an abnormal target detection device based on continuous time-series point cloud superposition, which comprises a memory and one or more processors, wherein the memory stores executable codes, and the processors are used for implementing the steps of the abnormal target detection method based on continuous time-series point cloud superposition when executing the executable codes.
In a third aspect, the present invention provides a computer-readable storage medium, on which a program is stored, which, when executed by a processor, implements the steps of the above-described abnormal object detection method based on continuous time-series point cloud overlay.
The invention has the beneficial effects that: the method solves the problem that the accuracy of detecting and identifying the abnormal target with the variable size and the variable distance is low by sensing the abnormal target by the solid-state laser radar and adopting a deep learning method. A background depth map based on point clouds is constructed through a continuous time sequence superposition method, abnormal point clouds are detected by utilizing the difference between the mapping depth value of the abnormal target point clouds and the depth value of the background depth map, and the clustering accuracy of the abnormal point clouds is improved by adding semantic category information to the abnormal point clouds, so that clustering clusters and detection information of the abnormal targets are accurately generated, and reliable technical support is provided for unmanned safety monitoring of a park.
Drawings
Fig. 1 is a flowchart of an abnormal target detection method based on continuous time sequence point cloud overlay according to the present invention.
FIG. 2 is a background depth map based on point cloud constructed by the continuous time sequence superposition method.
FIG. 3 is a graph of the effect of the point cloud containing abnormal objects of the present invention mapping to a background depth map.
FIG. 4 is a graph of the effect of the present invention of mapping a point cloud containing an anomalous target onto a spatio-temporally aligned image.
Fig. 5 is a structural diagram of an abnormal object detection apparatus based on continuous time-series point cloud overlay according to the present invention.
Detailed Description
The objects and effects of the present invention will become more apparent from the following detailed description of the present invention with reference to the accompanying drawings. It should be understood that the specific embodiments described herein are merely illustrative of the invention and are not intended to limit the invention.
As shown in fig. 1, the invention provides an abnormal target detection method based on continuous time sequence point cloud superposition, which is used for solving the problem of low reliability of abnormal target detection caused by low accuracy of detection and identification directly by a deep learning method when a solid-state laser radar or a camera is adopted to sense an abnormal target with a large volume and a non-fixed distance.
According to the method, a background depth map based on the point cloud is constructed through a continuous time sequence point cloud superposition method, the abnormal point cloud is detected by utilizing the difference between the mapping depth value of the abnormal point cloud and the depth value of the background depth map, and image semantic category information is added to the abnormal point cloud through introducing time synchronization image features, so that the information content of the point cloud is improved. The method for combining the space and the semantic distance of the point clouds can effectively improve the anti-interference capability of clustering abnormal point clouds by adopting a density clustering method, and form accurate and independent target clusters, thereby generating accurate detection information of targets.
The invention specifically comprises the following steps:
the method comprises the following steps: the method comprises the steps of setting a solid-state laser radar to be in a non-repetitive scanning mode, collecting a plurality of frames of point cloud data frames with continuous time sequence according to a certain frequency, mapping point clouds in all the point cloud data frames to a depth map by using an affine transformation matrix from the point cloud data to image data, superposing depth values of mapping points with the same mapping coordinates, calculating an average depth value, and updating the depth value of a coordinate corresponding to the depth map by using the obtained average depth value. The process is repeated until the depth value before and after any coordinate of the depth map is updated does not change any more, and the finally updated depth map is the background depth map.
And the number of the point cloud data frames of the continuous time sequence of the frames is determined according to the scanning characteristics of the selected solid-state laser radar equipment. In this embodiment, the model of the selected solid-state laser radar is voyadvia, the scanning mode is set to be a non-repetitive scanning mode, when the scanning period is 10HZ, the number of frames of continuously acquired point cloud data is 4, and when the number of updating times of the background image is 10, a stable background depth map can be constructed, that is, the depth value of any coordinate of the background depth map does not change any more before and after updating.
A solid-state laser radar and a camera which are installed on a roadside lamp post in an unattended monitoring area of a park are utilized to collect point cloud data and image data which are aligned in a space-time mode, for the collected point cloud data, a background depth map which is constructed by the continuous time sequence point cloud superposition method and is based on the point cloud is shown in the attached drawing 2, the size of the background depth map is consistent with that of the image data, the generated depth map mapping points are very dense, and the depth value of the background depth map cannot be changed by adding a point cloud time sequence data frame.
Further, the step one includes the following steps:
(1.1) defining a blank depth map with the size of W x H, and initializing the depth value at each coordinate in the blank depth map to 0. And the size of the blank depth map is consistent with the size of an image shot by a camera aligned with the solid-state laser radar in space-time.
And (1.2) acquiring an affine transformation matrix from the solid-state laser radar point cloud data to the camera image data aligned in time and space.
The method comprises the following specific steps: controlling the time synchronization of data frames of a laser radar and a camera by adopting a hardware line control mode, carrying out combined calibration on internal parameters of the camera and external parameters from a laser radar coordinate system to a camera coordinate system to obtain internal parameters and external parameter matrixes, and generating an affine transformation matrix from point cloud data to image data according to the obtained internal parameters and external parameter matrixes;
assuming a calibrated internal reference matrix of
Figure 730676DEST_PATH_IMAGE001
The external reference matrix is
Figure 100002_DEST_PATH_IMAGE056
Affine transformation matrix of point cloud data to image data
Figure DEST_PATH_IMAGE057
Comprises the following steps:
Figure 100002_DEST_PATH_IMAGE058
wherein the internal reference matrix
Figure 482469DEST_PATH_IMAGE001
Dimension of 3*3, and extrinsic parameter matrix of
Figure 99526DEST_PATH_IMAGE002
Dimension of 3*4, affine transformation matrix
Figure 94027DEST_PATH_IMAGE057
Dimension size 3*4;
and (1.3) setting the solid-state laser radar to be in a non-repeated scanning mode, continuously collecting N frames of point cloud data frames in continuous time sequence according to a certain frequency, respectively mapping point clouds in all the point cloud data frames to a blank depth map by using an affine transformation matrix from the point cloud data to image data, overlapping the depth values of mapping points with the same mapping coordinates, and recording the overlapping times of the depth values.
The method specifically comprises the following steps: for any one point in N frames of continuous time sequence point cloud data frames, the coordinate of the point cloud in the point cloud is assumed to be
Figure 704000DEST_PATH_IMAGE005
Mapping the point cloud data to the affine transformation matrix of the image data to the mapping points in the blank depth mapIs marked by
Figure DEST_PATH_IMAGE059
Depth value of
Figure 998715DEST_PATH_IMAGE007
Which may be expressed as follows:
Figure 100002_DEST_PATH_IMAGE060
where ceil denotes rounding up,
Figure 73157DEST_PATH_IMAGE009
respectively representing floating point coordinate values of the point cloud mapped to the mapped points in the depth map,
Figure 554954DEST_PATH_IMAGE010
is that
Figure 703039DEST_PATH_IMAGE011
Divided by depth value
Figure 586681DEST_PATH_IMAGE007
The integral coordinate value of the mapping point after the upward integration is carried out;
and performing the mapping operation on the points in all the point cloud data frames, superposing the depth values of the mapping points with the same integral coordinate values in an adding mode, and recording the superposition times of the depth values under all the coordinates.
And (1.4) for each mapping point coordinate, calculating the average depth value of the coordinate by using the superposition depth value and the superposition times under the coordinate, and updating the depth value of the corresponding coordinate of the blank depth map by using the obtained average depth value.
The method comprises the following specific steps: let the coordinates of a mapping point be
Figure DEST_PATH_IMAGE061
The superposition depth value is SumDepth, the superposition times is NumD, and the coordinate of the position is
Figure 76699DEST_PATH_IMAGE061
Can be expressed as:
Figure 45792DEST_PATH_IMAGE013
and calculating the average depth value of all mapping point coordinates, and updating the depth value of the blank depth map at the corresponding coordinate by using the obtained average depth value.
And (1.5) repeating the step (1.3) and the step (1.4) until the depth value of any coordinate of the blank depth map is not changed before and after updating, wherein the blank depth map updated for the last time is the background depth map.
The method for acquiring the affine transformation matrix from the point cloud data of the solid-state laser radar to the image data of the camera aligned in time and space comprises the steps of firstly utilizing the solid-state laser radar and the camera which are installed on a roadside lamp pole in a garden, controlling the time synchronization of data frames of the laser radar and the camera in a hardware line control mode, and respectively carrying out combined calibration on internal parameters of the camera and external parameters from a laser radar coordinate system to a camera coordinate system. The camera internal reference calibration is generally carried out by adopting a checkerboard, acquiring checkerboard data of a plurality of angles and distances by utilizing the characteristics of clear black and white checkerboard and easy angular point finding, generating a plurality of groups of two-dimensional image angular point coordinates and three-dimensional space angular point coordinates, and solving and generating internal reference parameters based on a least square method. The method comprises the steps of external parameter combined calibration from a solid-state laser radar coordinate system to a camera coordinate system, recording multi-section time-aligned image data and point cloud data by using an ROS tool through placing a plurality of white boards with different sizes and distances, and marking four corner points of the white boards for each frame of time-aligned image data and point cloud data. And correcting external parameter parameters for any group of corner points by adopting a BP neural network method through multiple iterations until the mapping deviation generated by the external parameter is stabilized within a threshold range. By utilizing the calibrated camera internal reference and the external reference from the laser radar coordinate system to the camera coordinate system, the three-dimensional point cloud can be mapped to a two-dimensional image, and the conversion from the point cloud space to the image space is realized.
Step two: the method comprises the steps of acquiring a point cloud data frame of the solid-state laser radar in real time, mapping all point clouds in the point cloud data frame to a background depth map by using an affine transformation matrix from point cloud data to image data, and judging that any point in the point clouds is a newly-increased abnormal point if the difference between the mapping point depth value of the point cloud data frame and the depth value of the background depth map under a corresponding coordinate is larger than a threshold value.
The point cloud data frame of the solid-state laser radar is obtained in real time, the obtaining period is 10HZ, namely, the point cloud data of the solid-state laser radar is collected once every 100 ms.
The method specifically comprises the following steps: for any point in the point cloud, the coordinate of the point in the point cloud is assumed to be
Figure 997568DEST_PATH_IMAGE014
The coordinates of the mapping points mapped into the background depth map are
Figure 100002_DEST_PATH_IMAGE062
Depth value of
Figure DEST_PATH_IMAGE063
Corresponding to the background depth map coordinates of
Figure 47301DEST_PATH_IMAGE017
At a depth value of
Figure 691909DEST_PATH_IMAGE018
If it satisfies
Figure 100002_DEST_PATH_IMAGE064
Then the point is determined to be a new outlier, i.e., the point is likely to be a point in the outlier target, not a point in the background, where
Figure 164610DEST_PATH_IMAGE020
The empirical value can be obtained by observing the difference between the depth values of the abnormal target point and the background point.
As shown in fig. 3, is a bagAnd mapping the point cloud containing the abnormal target to an effect graph of the background depth graph. It can be observed that after the point cloud of the abnormal target is mapped to the background depth map, occlusion occurs at a corresponding position of the background depth map, so that the depth value of the abnormal target and the depth value of the background depth map under the corresponding coordinate have a larger difference. From the observation, the depth values differ
Figure DEST_PATH_IMAGE065
When the value is set to 0.5 (unit: meter), abnormal point clouds can be distinguished.
Step three: the method comprises the steps of obtaining an image data frame which is aligned with a corresponding point cloud data frame in a time-space mode through a camera, segmenting all target examples in the corresponding image data frame by adopting a semantic segmentation method, respectively mapping newly increased abnormal points to the image data frame, and adding semantic type information to the newly increased abnormal points according to the semantic type of an image target example area where the mapping points are located. The method comprises the following specific steps:
and (3.1) segmenting all target instances in the corresponding image data frame by adopting a Mask-RCNN-based semantic segmentation method.
(3.2) for any newly added abnormal point, assuming the coordinate of the abnormal point in the point cloud to be
Figure 100002_DEST_PATH_IMAGE066
Mapping the point cloud data to the coordinates of the mapping points in the image data frame by using the affine transformation matrix of the point cloud data to the image data
Figure 716814DEST_PATH_IMAGE022
Can be expressed as follows:
Figure 43890DEST_PATH_IMAGE023
where ceil denotes rounding up,
Figure 111597DEST_PATH_IMAGE022
integer coordinate values representing mapping points of the point cloud data to image data,
Figure 55282DEST_PATH_IMAGE024
are mapped point depth values.
(3.3) hypothetical coordinates
Figure 348860DEST_PATH_IMAGE022
And a set of image coordinate points PixelCols contained in a certain target example, and satisfies
Figure 61601DEST_PATH_IMAGE025
If the new abnormal point is added with the semantic category information, the new abnormal point can be represented as
Figure DEST_PATH_IMAGE067
Where cls is the semantic class of the target instance.
As shown in fig. 4, is an effect diagram of a point cloud containing an abnormal object mapped onto a spatio-temporally aligned image. In the present embodiment, it is considered that the abnormal object includes a pedestrian, a non-motor vehicle, an animal, other movable obstacles, and the like. The image is segmented through a Mask-RCNN semantic segmentation method, instance objects such as backgrounds, pedestrians, non-motor vehicles, animals and other movable obstacles in the image are segmented, all instance areas (including the backgrounds) in an image data frame are endowed with semantic category information, and the backgrounds in the image comprise static category objects such as sky, green plants, road surfaces and buildings. It can be observed that abnormal targets such as pedestrians and other movable obstacles appear in fig. 4, and semantic category information can be added to the point cloud falling in the corresponding area according to the semantic category of the example target area corresponding to the image mapped by the point cloud of the abnormal target.
Step four: and for all the newly added abnormal points, designing a density clustering method based on the space semantic joint distance between the points, and clustering the newly added abnormal points to form a cluster.
Further, the fourth step includes the following steps:
(4.1) finding and determining core anomaly points. The method specifically comprises the following steps: set the radius of the space to
Figure 798744DEST_PATH_IMAGE028
The minimum number of neighbors is minS; for any newly added abnormal point
Figure 229725DEST_PATH_IMAGE029
Assuming its coordinates as
Figure 100002_DEST_PATH_IMAGE068
Semantic categories of
Figure DEST_PATH_IMAGE069
Then it is represented as
Figure 100002_DEST_PATH_IMAGE070
Traversing newly added abnormal points
Figure 638579DEST_PATH_IMAGE033
Radius of space
Figure 471406DEST_PATH_IMAGE028
All new outliers in the range, for their spatial radius
Figure 363138DEST_PATH_IMAGE028
Any newly added abnormal point in the range
Figure 15837DEST_PATH_IMAGE034
Assuming its coordinates as
Figure 916796DEST_PATH_IMAGE035
Semantic categories of
Figure 89283DEST_PATH_IMAGE036
Then it is represented as
Figure 417496DEST_PATH_IMAGE037
If the space semantic joint distance between two points
Figure 291911DEST_PATH_IMAGE038
Satisfies the following conditions:
Figure 996562DEST_PATH_IMAGE039
the new abnormal point is considered
Figure 272823DEST_PATH_IMAGE034
Is newly added with abnormal points
Figure 24134DEST_PATH_IMAGE033
The neighbor of (2); if an abnormal point is newly added
Figure 385845DEST_PATH_IMAGE033
Radius of space of
Figure 628608DEST_PATH_IMAGE028
If the number of newly added abnormal points satisfying the above formula in the range is not less than minS, the abnormal points are determined
Figure 759375DEST_PATH_IMAGE033
Is the core exception point, otherwise is the non-core exception point.
Where Dis is the euclidean spatial distance between two points, sclas is the semantic distance between two points, which can be expressed as:
Figure 429391DEST_PATH_IMAGE040
wherein
Figure DEST_PATH_IMAGE071
In order to be the spatial distance weight,
Figure 560289DEST_PATH_IMAGE042
in order to be a semantic distance weight,
Figure 341163DEST_PATH_IMAGE043
the distance threshold values are spatial semantic joint distance threshold values which are empirical values.
Executing the step (4.1) on all the newly added abnormal points until all the newly added abnormal points are confirmed to be core abnormal points or not; and performing subsequent processing on the core abnormal point, and directly discarding the non-core abnormal point.
And (4.2) clustering the core abnormal points and forming a cluster. The method specifically comprises the following steps: starting from any one core anomaly point, the spatial radius of the core anomaly point is equal to that of the space
Figure 592016DEST_PATH_IMAGE028
The core abnormal points of the inner neighbors are gathered into one class, and the core abnormal points of the neighbors are continuously searched and gathered into one class from any core abnormal point of the neighbors until the core abnormal points of the neighbors can not be found, the core abnormal points gathered together in the process are a cluster, and the cluster class is the semantic class of the initial core abnormal point. And (4.2) repeating the step from any remaining core abnormal point until no new cluster is formed.
(4.3) the remaining unclustered core outliers are outliers and are discarded directly.
The space radius is
Figure 432933DEST_PATH_IMAGE028
Set as 1 (unit: meter), the minimum neighbor number minS is set as 2, and the space semantic united distance between two points
Figure 100002_DEST_PATH_IMAGE072
Set to 1, parameter
Figure 549662DEST_PATH_IMAGE041
Figure 868648DEST_PATH_IMAGE042
The semantic categories of the two points are required to be considered while the spatial distances of the two points are considered, if the spatial distances of the two points are close but the semantic categories are inconsistent, the final spatial semantic union distance is smaller than a threshold value, and the final spatial semantic union distance does not belong to a cluster. Meanwhile, due to the fact that the number of the point clouds is large, when abnormal points are detected, the background point clouds are easy to be mistakenly detected as abnormal target point clouds, and semantic information is introduced, so that the abnormal target point clouds can be further detectedAnd step three, excluding false detection of background point clouds. Parameter(s)
Figure 239587DEST_PATH_IMAGE041
Figure 251405DEST_PATH_IMAGE042
The arrangement mode of (2) emphasizes considering the space distance, and points with close space distance are considered to be larger possibly to be the same cluster.
Step five: and calculating the volume and the center point coordinates of each cluster, identifying the cluster with the volume larger than the threshold as an abnormal target, and generating the position information of the abnormal target.
Specifically, for each cluster, the volume of the cluster is calculated according to the two core abnormal points with the largest spatial distance. The coordinates of two core abnormal points with the maximum space distance are respectively assumed to be
Figure 825737DEST_PATH_IMAGE044
And
Figure 948414DEST_PATH_IMAGE045
semantic categories are
Figure 173859DEST_PATH_IMAGE046
Are then respectively represented as
Figure 90999DEST_PATH_IMAGE047
Then the volume of the cluster
Figure 667474DEST_PATH_IMAGE048
Can be expressed as:
Figure 580460DEST_PATH_IMAGE049
coordinates of center point
Figure 660411DEST_PATH_IMAGE050
Can be expressed as:
Figure DEST_PATH_IMAGE073
if it is
Figure 545191DEST_PATH_IMAGE048
Greater than a threshold value
Figure 343382DEST_PATH_IMAGE052
Then the cluster is considered as an abnormal target, wherein
Figure 558594DEST_PATH_IMAGE052
If the upper limit value of the volume of the small target object forming the interference is an empirical value, the abnormal target detection information is as follows: the object class is
Figure 493052DEST_PATH_IMAGE053
Distance solid state laser radar mounting origin
Figure 283154DEST_PATH_IMAGE054
The included angle between the meter and the abscissa of the coordinate system of the solid-state laser radar is
Figure 568641DEST_PATH_IMAGE055
The volume threshold of the cluster
Figure 86079DEST_PATH_IMAGE052
According to empirical observation, the setting is 0.008 (unit:
Figure 100002_DEST_PATH_IMAGE074
). Because in daily scene, often can appear fallen leaves, rubbish, plastic bag, carton case etc. and produce the less object of interference and volume to the testing result, consequently need set up the threshold value volume for the cluster, avoid false retrieval, false retrieval.
Corresponding to the embodiment of the abnormal target detection method based on continuous time sequence point cloud superposition, the invention also provides an embodiment of an abnormal target detection device based on continuous time sequence point cloud superposition.
Referring to fig. 5, an abnormal target detection apparatus based on continuous time-series point cloud overlay according to an embodiment of the present invention includes a memory and one or more processors, where the memory stores executable codes, and the processors execute the executable codes to implement the abnormal target detection method based on continuous time-series point cloud overlay in the foregoing embodiments.
The embodiment of the abnormal target detection device based on continuous time-series point cloud superposition can be applied to any equipment with data processing capability, such as computers and other equipment or devices. The apparatus embodiments may be implemented by software, or by hardware, or by a combination of hardware and software. The software implementation is taken as an example, and as a logical device, the device is formed by reading corresponding computer program instructions in the nonvolatile memory into the memory for running through the processor of any device with data processing capability. From a hardware aspect, as shown in fig. 5, a hardware structure diagram of an arbitrary device with data processing capability where an abnormal object detection apparatus based on continuous time-sequence point cloud superposition is located according to the present invention is shown, except for the processor, the memory, the network interface, and the nonvolatile memory shown in fig. 5, in an embodiment, the arbitrary device with data processing capability where the apparatus is located may generally include other hardware according to an actual function of the arbitrary device with data processing capability, which is not described again.
The implementation process of the functions and actions of each unit in the above device is specifically described in the implementation process of the corresponding step in the above method, and is not described herein again.
For the device embodiments, since they substantially correspond to the method embodiments, reference may be made to the partial description of the method embodiments for relevant points. The above-described embodiments of the apparatus are merely illustrative, and the units described as separate parts may or may not be physically separate, and parts displayed as units may or may not be physical units, may be located in one place, or may be distributed on a plurality of network units. Some or all of the modules can be selected according to actual needs to achieve the purpose of the solution of the present invention. One of ordinary skill in the art can understand and implement it without inventive effort.
The embodiment of the invention also provides a computer-readable storage medium, on which a program is stored, and when the program is executed by a processor, the abnormal target detection method based on continuous time sequence point cloud superposition in the above embodiment is realized.
The computer readable storage medium may be an internal storage unit, such as a hard disk or a memory, of any data processing capability device described in any of the foregoing embodiments. The computer readable storage medium may also be any external storage device of a device with data processing capabilities, such as a plug-in hard disk, a Smart Media Card (SMC), an SD Card, a Flash memory Card (Flash Card), etc. provided on the device. Further, the computer readable storage medium may include both an internal storage unit and an external storage device of any data processing capable device. The computer-readable storage medium is used for storing the computer program and other programs and data required by the arbitrary data processing-capable device, and may also be used for temporarily storing data that has been output or is to be output.
The above-described embodiments are intended to illustrate rather than to limit the invention, and any modifications and variations of the present invention are within the spirit of the invention and the scope of the appended claims.

Claims (8)

1. An abnormal target detection method based on continuous time sequence point cloud superposition is characterized by comprising the following steps:
the method comprises the following steps: the method comprises the following steps of collecting a plurality of continuous time sequence point cloud data frames through a solid-state laser radar, and utilizing an affine transformation matrix from the point cloud data to image data, wherein the method specifically comprises the following steps: controlling the time synchronization of data frames of a laser radar and a camera by adopting a hardware line control mode, carrying out combined calibration on internal parameters of the camera and external parameters from a laser radar coordinate system to a camera coordinate system to obtain internal parameters and external parameter matrixes, and generating an affine transformation matrix from point cloud data to image data according to the obtained internal parameters and external parameter matrixes; mapping the point clouds in all the point cloud data frames to a depth map, overlapping the depth values of the mapping points with the same mapping coordinates, calculating an average depth value, and updating the depth value of the corresponding coordinate of the depth map by using the obtained average depth value; repeating the step until the depth value before and after any coordinate of the depth map is updated does not change any more, wherein the finally updated depth map is the background depth map;
step two: acquiring a point cloud data frame of the solid-state laser radar in real time, mapping all point clouds in the point cloud data frame to a background depth map by using an affine transformation matrix from the point cloud data to image data, and judging that any point in the point clouds is a newly added abnormal point if the difference between the mapping point depth value and the depth value of the background depth map under the corresponding coordinate is greater than a threshold value;
step three: acquiring an image data frame which is in time-space alignment with a corresponding point cloud data frame, segmenting all target instances in the image data frame by adopting a semantic segmentation method, respectively mapping newly increased abnormal points to the image data frame, and adding semantic type information to the newly increased abnormal points according to the semantic type of an image target instance region where the mapping points are located;
step four: searching and determining core outliers for all newly added outliers, clustering the core outliers based on the spatial semantic union distance between the points, forming cluster clusters, and directly discarding the remaining non-clustered core outliers as outliers;
step five: and for each cluster, calculating the volume and the center point coordinates of the cluster according to the two core abnormal points with the largest spatial distance, identifying the cluster with the volume larger than the threshold value as an abnormal target, and generating the detection information of the abnormal target.
2. The method according to claim 1, wherein the first step comprises the following steps:
(1.1) defining a blank depth map, wherein the depth value of each coordinate in the blank depth map is initialized to be 0; the size of the blank depth map is consistent with that of an image shot by a camera aligned with the solid-state laser radar in time and space;
(1.2) acquiring an affine transformation matrix from the solid-state laser radar point cloud data to the camera image data aligned in time and space; assuming a calibrated internal reference matrix of
Figure DEST_PATH_IMAGE002
The external reference matrix is
Figure DEST_PATH_IMAGE004
Affine transformation matrix of point cloud data to image data
Figure DEST_PATH_IMAGE006
Comprises the following steps:
Figure DEST_PATH_IMAGE008
wherein the internal reference matrix
Figure 945370DEST_PATH_IMAGE002
Dimension of 3*3, and extrinsic parameter matrix of
Figure 596931DEST_PATH_IMAGE004
Dimension of 3*4, affine transformation matrix
Figure 94777DEST_PATH_IMAGE006
Dimension size 3*4;
(1.3) setting the solid-state laser radar to be in a non-repetitive scanning mode, continuously acquiring N frames of point cloud data frames with continuous time sequence according to a certain frequency, respectively mapping point clouds in all the point cloud data frames to a blank depth map by utilizing an affine transformation matrix from the point cloud data to image data, and mapping the point clouds with the same mapping coordinateOverlapping the depth values of the points, and recording the overlapping times of the depth values; the method specifically comprises the following steps: for any point in N frames of collected continuous time sequence point cloud data frames, the coordinate of the point cloud in the point cloud is assumed to be
Figure DEST_PATH_IMAGE010
Mapping the point cloud data to the affine transformation matrix of the image data to the coordinates of the mapping points in the blank depth map as
Figure DEST_PATH_IMAGE012
Depth value of
Figure DEST_PATH_IMAGE014
Respectively, as follows:
Figure DEST_PATH_IMAGE016
wherein ceil means rounding up,
Figure DEST_PATH_IMAGE018
respectively representing floating point coordinate values of mapping points of the point cloud to the depth map,
Figure DEST_PATH_IMAGE020
is that
Figure DEST_PATH_IMAGE021
Divided by depth value
Figure 286724DEST_PATH_IMAGE014
The integral coordinate value of the mapping point after the upward integration is carried out;
executing the mapping operation on the points in all the point cloud data frames, superposing the depth values of the mapping points with the same integral coordinate values in an adding mode, and recording the superposition times of the depth values under all the coordinates;
(1.4) for each mapping point coordinate, calculating the average depth value of the coordinate by using the superposition depth value and the superposition times under the coordinate, and updating the depth value of the corresponding coordinate of the blank depth map by using the obtained average depth value; the method specifically comprises the following steps: assuming that the superposition depth value under the coordinate of a certain mapping point is SumDepth, and the superposition frequency is NumD, the average depth value depth of the coordinate is expressed as:
Figure DEST_PATH_IMAGE023
calculating an average depth value for all mapping point coordinates, and updating the depth value of the blank depth map at the corresponding coordinate by using the obtained average depth value;
and (1.5) repeating the step (1.3) and the step (1.4) until the depth value of any coordinate of the blank depth map is not changed before and after updating, wherein the blank depth map updated for the last time is the background depth map.
3. The method according to claim 1, wherein in the second step, for any point in the point cloud, the coordinate of the point cloud in the point cloud is assumed to be
Figure DEST_PATH_IMAGE025
The coordinates of the mapping points mapped into the background depth map are
Figure DEST_PATH_IMAGE027
Depth value of
Figure DEST_PATH_IMAGE029
Corresponding to the background depth map coordinates of
Figure DEST_PATH_IMAGE030
At a depth value of
Figure DEST_PATH_IMAGE032
If it satisfies
Figure DEST_PATH_IMAGE034
If the point is a new abnormal point, the point is judged to be a point in the abnormal target and not a point in the background; wherein
Figure DEST_PATH_IMAGE036
The empirical value is obtained by observing the difference in depth values between the abnormal target point and the background point.
4. The abnormal target detection method based on continuous time series point cloud superposition as claimed in claim 1, wherein the step three comprises the following steps:
(3.1) segmenting all target instances in the image data frame by adopting a Mask-RCNN-based semantic segmentation method;
(3.2) for any newly added abnormal point, assuming that the coordinate of the abnormal point in the point cloud is
Figure DEST_PATH_IMAGE038
Mapping the point cloud data to the coordinates of the mapping points in the image data frame by using the affine transformation matrix of the point cloud data to the image data
Figure DEST_PATH_IMAGE040
Is represented as follows:
Figure DEST_PATH_IMAGE042
wherein ceil means rounding up,
Figure 286384DEST_PATH_IMAGE040
integer coordinate values representing mapping points of the point cloud data to image data,
Figure DEST_PATH_IMAGE044
the depth value of the mapping point is obtained;
(3.3) hypothetical sittingSign board
Figure 46530DEST_PATH_IMAGE040
And a set of image coordinate points PixelCols contained in a certain target example, and satisfies
Figure DEST_PATH_IMAGE046
If the new abnormal point is added with the semantic category information, the new abnormal point is expressed as
Figure DEST_PATH_IMAGE048
Where cls is the semantic class of the target instance.
5. The abnormal target detection method based on continuous time series point cloud superposition as claimed in claim 1, wherein in the fourth step, the following steps are included:
(4.1) searching and determining a core anomaly point; the method specifically comprises the following steps: set the space radius to
Figure DEST_PATH_IMAGE050
The minimum number of neighbors is minS; for any newly added abnormal point
Figure DEST_PATH_IMAGE052
Assuming its coordinates as
Figure DEST_PATH_IMAGE054
Semantic categories of
Figure DEST_PATH_IMAGE056
Then it is represented as
Figure DEST_PATH_IMAGE058
Go through the newly added abnormal point
Figure 96919DEST_PATH_IMAGE052
Radius of space
Figure 764660DEST_PATH_IMAGE050
All new outliers in the range, for their spatial radius
Figure 417227DEST_PATH_IMAGE050
Any newly-added abnormal point in the range
Figure DEST_PATH_IMAGE060
Assuming its coordinates as
Figure DEST_PATH_IMAGE062
Semantic categories of
Figure DEST_PATH_IMAGE064
Then it is represented as
Figure DEST_PATH_IMAGE066
If the space-semantic joint distance between two points
Figure DEST_PATH_IMAGE068
Satisfies the following conditions:
Figure DEST_PATH_IMAGE070
the new abnormal point is considered
Figure 472908DEST_PATH_IMAGE060
Is newly added with abnormal points
Figure 696079DEST_PATH_IMAGE052
The neighbor of (2); if the abnormal point is newly added
Figure 167512DEST_PATH_IMAGE052
Radius of space of
Figure 409006DEST_PATH_IMAGE050
New abnormal point satisfying the above formula in rangeIf the number of the sub-points is greater than or equal to minS, determining an abnormal point
Figure 307692DEST_PATH_IMAGE052
Is a core anomaly point, otherwise is a non-core anomaly point;
wherein Dis is the euclidean spatial distance between two points, sclas is the semantic distance between two points, and is respectively expressed as:
Figure DEST_PATH_IMAGE072
wherein
Figure DEST_PATH_IMAGE074
In order to be the spatial distance weight,
Figure DEST_PATH_IMAGE076
in order to be a semantic distance weight,
Figure DEST_PATH_IMAGE078
the distance threshold values are spatial semantic union distance threshold values which are empirical values;
executing the step (4.1) on all the newly added abnormal points until all the newly added abnormal points are confirmed to be core abnormal points or not; performing subsequent processing on the core abnormal point, and directly discarding the non-core abnormal point;
(4.2) clustering the core abnormal points to form a cluster; the method comprises the following specific steps: starting from any one core anomaly point, the spatial radius of the core anomaly point is equal to that of the space
Figure 676881DEST_PATH_IMAGE050
Neighbor core abnormal points in the cluster are clustered, and the neighbor core abnormal points are continuously searched and clustered from any neighbor core abnormal point until the neighbor core abnormal points cannot be found, the clustered core abnormal points are a cluster, and the cluster category is the semantic category of the initial core abnormal point; and repeating the step (4.2) from any remaining core abnormal point until no new core abnormal point is formedClustering;
(4.3) the remaining non-clustered core outliers are outliers and are discarded directly.
6. The abnormal target detection method based on continuous time sequence point cloud superposition of claim 1, wherein in the fifth step, for each cluster, the volume of the cluster is calculated according to two core abnormal points with the largest spatial distance; the coordinates of two core abnormal points with the maximum space distance are respectively assumed to be
Figure DEST_PATH_IMAGE080
And
Figure DEST_PATH_IMAGE082
semantic categories are
Figure DEST_PATH_IMAGE084
Are then respectively represented as
Figure DEST_PATH_IMAGE086
Volume of the cluster
Figure DEST_PATH_IMAGE088
Expressed as:
Figure DEST_PATH_IMAGE090
coordinates of center point
Figure DEST_PATH_IMAGE092
Expressed as:
Figure DEST_PATH_IMAGE094
if it is
Figure 466851DEST_PATH_IMAGE088
Greater than a threshold value
Figure DEST_PATH_IMAGE096
Then the cluster is considered as an abnormal target, wherein
Figure DEST_PATH_IMAGE097
If the upper limit value of the volume of the small target object forming the interference is an empirical value, the abnormal target detection information is as follows: the object class is
Figure DEST_PATH_IMAGE098
Distance solid state laser radar mounting origin
Figure DEST_PATH_IMAGE100
The included angle between the meter and the abscissa of the coordinate system of the solid-state laser radar is
Figure DEST_PATH_IMAGE102
7. An abnormal target detection device based on continuous time-series point cloud superposition, comprising a memory and one or more processors, wherein the memory is stored with executable codes, and the processors are used for implementing the steps of the abnormal target detection method based on continuous time-series point cloud superposition according to any one of claims 1 to 6 when executing the executable codes.
8. A computer-readable storage medium, on which a program is stored, which, when being executed by a processor, carries out the steps of the method for detecting an anomalous target based on superposition of successive time-series point clouds according to any one of claims 1 to 6.
CN202211145212.XA 2022-09-20 2022-09-20 Abnormal target detection method and device based on continuous time sequence point cloud superposition Active CN115272493B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202211145212.XA CN115272493B (en) 2022-09-20 2022-09-20 Abnormal target detection method and device based on continuous time sequence point cloud superposition

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202211145212.XA CN115272493B (en) 2022-09-20 2022-09-20 Abnormal target detection method and device based on continuous time sequence point cloud superposition

Publications (2)

Publication Number Publication Date
CN115272493A CN115272493A (en) 2022-11-01
CN115272493B true CN115272493B (en) 2022-12-27

Family

ID=83757249

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202211145212.XA Active CN115272493B (en) 2022-09-20 2022-09-20 Abnormal target detection method and device based on continuous time sequence point cloud superposition

Country Status (1)

Country Link
CN (1) CN115272493B (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117037120B (en) * 2023-10-09 2024-02-09 之江实验室 Target perception method and device based on time sequence selection
CN118154588A (en) * 2024-05-09 2024-06-07 中铁七局集团第三工程有限公司 Large-diameter pressure steel pipe quality detection method and system based on contour extraction

Citations (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108389221A (en) * 2018-01-30 2018-08-10 深圳市菲森科技有限公司 The scan method and system of 3-D view
WO2018176440A1 (en) * 2017-04-01 2018-10-04 深圳市速腾聚创科技有限公司 Method for fusing point cloud and planar image, intelligent device and non-volatile computer-readable storage medium
CN109961440A (en) * 2019-03-11 2019-07-02 重庆邮电大学 A kind of three-dimensional laser radar point cloud Target Segmentation method based on depth map
CN111652179A (en) * 2020-06-15 2020-09-11 东风汽车股份有限公司 Semantic high-precision map construction and positioning method based on dotted line feature fusion laser
CN111798475A (en) * 2020-05-29 2020-10-20 浙江工业大学 Indoor environment 3D semantic map construction method based on point cloud deep learning
CN112348867A (en) * 2020-11-18 2021-02-09 南通市测绘院有限公司 Method and system for constructing city high-precision three-dimensional terrain based on LiDAR point cloud data
CN113111887A (en) * 2021-04-26 2021-07-13 河海大学常州校区 Semantic segmentation method and system based on information fusion of camera and laser radar
CN113128348A (en) * 2021-03-25 2021-07-16 西安电子科技大学 Laser radar target detection method and system fusing semantic information
CN113393514A (en) * 2021-06-11 2021-09-14 中国科学院自动化研究所 Three-dimensional disordered point cloud data processing method, system and equipment
CN114266960A (en) * 2021-12-01 2022-04-01 国网智能科技股份有限公司 Point cloud information and deep learning combined obstacle detection method
WO2022141912A1 (en) * 2021-01-01 2022-07-07 杜豫川 Vehicle-road collaboration-oriented sensing information fusion representation and target detection method
CN114724120A (en) * 2022-06-10 2022-07-08 东揽(南京)智能科技有限公司 Vehicle target detection method and system based on radar vision semantic segmentation adaptive fusion
CN114758504A (en) * 2022-06-13 2022-07-15 之江实验室 Online vehicle overspeed early warning method and system based on filtering correction
CN114782519A (en) * 2022-03-11 2022-07-22 陕西天视致远航空技术有限公司 Method, device and medium for positioning spherical or quasi-spherical object based on point cloud information
CN114862901A (en) * 2022-04-26 2022-08-05 青岛慧拓智能机器有限公司 Road-end multi-source sensor fusion target sensing method and system for surface mine
CN114937081A (en) * 2022-07-20 2022-08-23 之江实验室 Internet vehicle position estimation method and device based on independent non-uniform incremental sampling

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11676296B2 (en) * 2017-08-11 2023-06-13 Sri International Augmenting reality using semantic segmentation

Patent Citations (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2018176440A1 (en) * 2017-04-01 2018-10-04 深圳市速腾聚创科技有限公司 Method for fusing point cloud and planar image, intelligent device and non-volatile computer-readable storage medium
CN108389221A (en) * 2018-01-30 2018-08-10 深圳市菲森科技有限公司 The scan method and system of 3-D view
CN109961440A (en) * 2019-03-11 2019-07-02 重庆邮电大学 A kind of three-dimensional laser radar point cloud Target Segmentation method based on depth map
CN111798475A (en) * 2020-05-29 2020-10-20 浙江工业大学 Indoor environment 3D semantic map construction method based on point cloud deep learning
CN111652179A (en) * 2020-06-15 2020-09-11 东风汽车股份有限公司 Semantic high-precision map construction and positioning method based on dotted line feature fusion laser
CN112348867A (en) * 2020-11-18 2021-02-09 南通市测绘院有限公司 Method and system for constructing city high-precision three-dimensional terrain based on LiDAR point cloud data
WO2022141912A1 (en) * 2021-01-01 2022-07-07 杜豫川 Vehicle-road collaboration-oriented sensing information fusion representation and target detection method
CN113128348A (en) * 2021-03-25 2021-07-16 西安电子科技大学 Laser radar target detection method and system fusing semantic information
CN113111887A (en) * 2021-04-26 2021-07-13 河海大学常州校区 Semantic segmentation method and system based on information fusion of camera and laser radar
CN113393514A (en) * 2021-06-11 2021-09-14 中国科学院自动化研究所 Three-dimensional disordered point cloud data processing method, system and equipment
CN114266960A (en) * 2021-12-01 2022-04-01 国网智能科技股份有限公司 Point cloud information and deep learning combined obstacle detection method
CN114782519A (en) * 2022-03-11 2022-07-22 陕西天视致远航空技术有限公司 Method, device and medium for positioning spherical or quasi-spherical object based on point cloud information
CN114862901A (en) * 2022-04-26 2022-08-05 青岛慧拓智能机器有限公司 Road-end multi-source sensor fusion target sensing method and system for surface mine
CN114724120A (en) * 2022-06-10 2022-07-08 东揽(南京)智能科技有限公司 Vehicle target detection method and system based on radar vision semantic segmentation adaptive fusion
CN114758504A (en) * 2022-06-13 2022-07-15 之江实验室 Online vehicle overspeed early warning method and system based on filtering correction
CN114937081A (en) * 2022-07-20 2022-08-23 之江实验室 Internet vehicle position estimation method and device based on independent non-uniform incremental sampling

Non-Patent Citations (6)

* Cited by examiner, † Cited by third party
Title
Recognition of Key Targets of Locomotive Bottom Based on 3D Point Cloud Data;Qian Huang等;《2017 Far East NDT New Technology & Application Forum (FENDT)》;20181223;第149-153页 *
动态环境下融合语义信息的视觉SLAM算法研究;赵永杰;《中国优秀硕士学位论文全文数据库 信息科技辑》;20210115(第1期);第I140-758页 *
基于图像与点云融合的三维目标检测算法研究;岳文涛;《中国优秀硕士学位论文全文数据库 信息科技辑》;20220415(第4期);第I138-1116页 *
基于深度学习的图像目标检测算法综述;张婷婷等;《电信科学》;20200720(第07期);全文 *
基于视觉与激光雷达数据融合的多目标识别算法研究;吴庆祝;《万方数据知识服务平台》;20210907;全文 *
视觉与激光点云融合的深度图像获取方法;王东敏等;《军事交通学院学报》;20171025(第10期);全文 *

Also Published As

Publication number Publication date
CN115272493A (en) 2022-11-01

Similar Documents

Publication Publication Date Title
CN109034018B (en) Low-altitude small unmanned aerial vehicle obstacle sensing method based on binocular vision
CN115272493B (en) Abnormal target detection method and device based on continuous time sequence point cloud superposition
CN110163904B (en) Object labeling method, movement control method, device, equipment and storage medium
CN111753609A (en) Target identification method and device and camera
CN110470333B (en) Calibration method and device of sensor parameters, storage medium and electronic device
CN111507327B (en) Target detection method and device
CN115049700A (en) Target detection method and device
CN115376109B (en) Obstacle detection method, obstacle detection device, and storage medium
CN110992424B (en) Positioning method and system based on binocular vision
CN116310679A (en) Multi-sensor fusion target detection method, system, medium, equipment and terminal
CN114926808A (en) Target detection and tracking method based on sensor fusion
Muresan et al. Real-time object detection using a sparse 4-layer LIDAR
CN113012215A (en) Method, system and equipment for space positioning
CN112699711A (en) Lane line detection method, lane line detection device, storage medium, and electronic apparatus
CN114241448A (en) Method and device for obtaining heading angle of obstacle, electronic equipment and vehicle
CN114549542A (en) Visual semantic segmentation method, device and equipment
CN117475355A (en) Security early warning method and device based on monitoring video, equipment and storage medium
CN113189610A (en) Map-enhanced autonomous driving multi-target tracking method and related equipment
CN110864670B (en) Method and system for acquiring position of target obstacle
CN114648639B (en) Target vehicle detection method, system and device
WO2023283929A1 (en) Method and apparatus for calibrating external parameters of binocular camera
CN115240168A (en) Perception result obtaining method and device, computer equipment and storage medium
CN111372051B (en) Multi-camera linkage blind area detection method and device and electronic equipment
Dekkiche et al. Vehicles detection in stereo vision based on disparity map segmentation and objects classification
CN117392347B (en) Map construction method, device, computer equipment and readable storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant