CN106408940A - Microwave and video data fusion-based traffic detection method and device - Google Patents

Microwave and video data fusion-based traffic detection method and device Download PDF

Info

Publication number
CN106408940A
CN106408940A CN201610952272.0A CN201610952272A CN106408940A CN 106408940 A CN106408940 A CN 106408940A CN 201610952272 A CN201610952272 A CN 201610952272A CN 106408940 A CN106408940 A CN 106408940A
Authority
CN
China
Prior art keywords
video
data
sensor
target
information
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201610952272.0A
Other languages
Chinese (zh)
Other versions
CN106408940B (en
Inventor
张德锋
何抱
顾丹丹
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nanjing Hui Er Looks Intelligent Science And Technology Ltd
Original Assignee
Nanjing Hui Er Looks Intelligent Science And Technology Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nanjing Hui Er Looks Intelligent Science And Technology Ltd filed Critical Nanjing Hui Er Looks Intelligent Science And Technology Ltd
Priority to CN201610952272.0A priority Critical patent/CN106408940B/en
Publication of CN106408940A publication Critical patent/CN106408940A/en
Application granted granted Critical
Publication of CN106408940B publication Critical patent/CN106408940B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G08SIGNALLING
    • G08GTRAFFIC CONTROL SYSTEMS
    • G08G1/00Traffic control systems for road vehicles
    • G08G1/01Detecting movement of traffic to be counted or controlled
    • G08G1/0104Measuring and analyzing of parameters relative to traffic conditions
    • G08G1/0125Traffic data processing
    • GPHYSICS
    • G08SIGNALLING
    • G08GTRAFFIC CONTROL SYSTEMS
    • G08G1/00Traffic control systems for road vehicles
    • G08G1/01Detecting movement of traffic to be counted or controlled
    • G08G1/017Detecting movement of traffic to be counted or controlled identifying vehicles
    • G08G1/0175Detecting movement of traffic to be counted or controlled identifying vehicles by photographing vehicles, e.g. when violating traffic rules
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Landscapes

  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Chemical & Material Sciences (AREA)
  • Analytical Chemistry (AREA)
  • Radar Systems Or Details Thereof (AREA)
  • Traffic Control Systems (AREA)

Abstract

The invention discloses a microwave and video data fusion-based traffic detection device. The microwave and video data fusion-based traffic detection device comprises a video sensor, a microwave sensor, an A/D conversion module, a processor module, a network communication module, a traffic flow parameter data fusion module, a traffic flow management platform, a target tracking module and a traffic event and information management platform. The invention also discloses a microwave and video data fusion-based traffic detection method. According to the method, the device is adopted to realize information complementation and data fusion. With the device of the invention adopted, the reliability of a system can be improved, the accurate estimation of the position of a target can be provided, more accurate data can be obtained, and powerful parameters can be provided for relevant departments. According to the microwave and video data fusion-based traffic detection method, after the microwave sensor detects some traffic behaviors which are inputted into a database, the video sensor is controlled to take pictures, and then, whether obtained data are matched with the data of the database is analyzed, an alarm is given, and therefore, false alarm rate can be reduced, human and material resources can be reduced, and intelligent detection can be realized actually.

Description

The Vehicle Detection method and device being merged based on microwave and video data
Technical field
The invention belongs to intelligent transportation field, more particularly it relates to a kind of melted based on microwave and video data The Vehicle Detection method and device closed.
Background technology
Data fusion mainly has data level, feature level and three kinds of modes of decision level fusion.Pixel-based fusion refers to merging It is desirable to the sensing data being merged has the fusion of any information of the matching precision being accurate to a pixel in algorithm; Feature-based fusion refers to carry out feature extraction from the initial data that each sensor provides, and then merges these features;Decision-making Level fusion refers to that each sensing data source is all passed through and converted and obtain independent identity estimation before merging.
Data fusion process includes target detection, data association, tracking and identification, situation estimation and the conjunction of multisensor And.Data fusion is that the Incomplete information with regard to a certain environmental characteristic that multiple sensors and information source are provided is in addition comprehensive Close, to form relatively complete, consistent perception description, thus realizing more accurately identifying arbitration functions.Obtained by merging Than each input data single, more information, due to the collective effect of more multisensor, make the effectiveness of system be increased By force.
The design of multisensor syste equipment and the productivity improve so that sensor performance greatly improves, and how to process number Amount is huge, and miscellaneous information becomes the problem that multisensor syste first has to consider.Particularly have uncertain in information Property in the case of, be the process that data obtained to single-sensor or information are carried out with respect to single-sensor data processing, There may be partly imperfect or insecure information, Fusion can comprehensively utilize multisensor letter effectively Breath, such that it is able to obtaining more accurate, the complete information of detected target and environment to a great extent and conforming retouching State or understand.
Microwave remote sensor is a kind of radar installations for round-the-clock monitoring traffic.It can measure the microwave area of coverage The distance of target in domain, azimuth, speed, size etc., it is provided that the complete positional information of target and doppler information, by this A little measurement to realize the detection of the vehicle to multilane and pedestrian.When being detected, microwave remote sensor receives returning of reflection Ripple signal, carries out background suppression to echo-signal, extracts useful signal, is capable of detecting when telecommunication flow information, in target acquisition Play important function with tracking aspect.Shortcoming is can not intuitively to see the kinestate of target as video, and The judgement of the information such as the license plate number of vehicle, color.
Video frequency vehicle sensor is as video sensor using video camera, is a kind of based on video image analysis and calculating Machine vision technique road pavement runs the integrated system that vehicle is tested and analyzed, and the method real-time monitoring using Image Engineering is divided The traffic image of analysis input, can detect traffic dynamic behavior and various traffic data, including traffic flow, vehicle classification, account for There are rate, speed, queue length, license plate number, body color etc..Shortcoming is to be limited by live lighting condition, current image procossing Real-time is poor, and accuracy of detection is limited by whole system software and hardware.
Content of the invention
The invention discloses a kind of Vehicle Detection method and device, the method and dress being merged based on microwave and video data Put and can more accurately monitor traffic behavior state and statistics telecommunication flow information, it is achieved thereby that the mobilism of general communication system is excellent Change and run, effectively meet the transport need that the public constantly expands.
The technical solution adopted in the present invention is:
A kind of Vehicle Detection device being merged based on microwave and video data, described device includes video sensor, microwave Sensor, A/D modular converter, processor module, network communication module, traffic flow parameter data fusion module, traffic flow management Platform, target tracking module, traffic events and information management platform;
Described video sensor is connected with described A/D modular converter respectively with described microwave remote sensor, video sensor With the non-electric charge quantity signalling of the different characteristic of microwave remote sensor output, it is then passed through described A/D modular converter and converts them to energy Digital quantity by computer disposal;Described A/D modular converter is connected with described processor module, described processor module with described Network communication module connects, and described processor module enters to via the data that described A/D module processing is converted into digital quantity Row is processed, and filters some abnormal datas to obtain useful signal, useful signal is transmitted by described network communication module again;
Described network communication module is connected with described traffic flow parameter data fusion module, described target tracking module respectively Connect, useful signal is transferred to described traffic flow parameter data fusion module, described target following mould by described network communication module Block;
Described traffic flow parameter data fusion module and described traffic flow management platform connect, described traffic flow parameter data Fusion Module carries out space-time uniformity, feature extraction to useful signal, and based on data fusion being carried out to characteristic quantity by certain rule Calculate, finally export fusion results to described traffic flow management platform;
Described target tracking module and traffic events are connected with information management platform, and described target tracking module is to useful letter Number carry out space-time uniformity, feature extraction, and by certain rule, data fusion calculating is carried out to characteristic quantity, finally by fusion results Export to described information management platform.
A kind of Vehicle Detection method being merged based on microwave and video data, is comprised the steps of:
The first step:Detection, carries out background noise suppression respectively in two sensors detection zone, export traffic flow, put down All speed, occupation rate, queue length and other instant messages;
Second step:Initial data pretreatment, is standardized and carries out pretreatment to multigroup sensing data of input, full The requirement to amount of calculation and computation sequence of sufficient subsequent estimation and processor module;
Grubbs statistical method is adopted for abnormal data preprocess method;
3rd step:Space-time is calibrated, and calibrates time and the spatial reference point of unified each sensor, snaps to same in time Time reference, spatially it is transformed into the same coordinate system, set up coordinate corresponding relation so that the result after processing seems data It is the same that fusion treatment central station is gathered;If each sensor is independently asynchronous working over time and space, must enter The row time moves and coordinate transform, merges required unified time and spatial reference point to be formed;By to single sensor The position of acquisition is merged with the estimated information of identity category, obtains more accurate target location, state and identity category Estimation;
4th step:Basic dynamic traffic Parameter fusion, can detect from video sensor and microwave remote sensor simultaneously The basis such as the traffic flow on section, average speed, occupation rate, queue length traffic parameter carries out fusion treatment, draws more accurate Really reliable traffic flow parameter;The fusion results of this level are the inputs of next emerging system simultaneously;
5th step:Data association, differentiates whether the data in different time space is derived from same target, radar and video object Mated, real target can be defined as by successful match, processed by setting means it is impossible to the target of coupling is it is believed that can not be true Fixed target is it is impossible to exclude the possibility;Using the distance of target, orientation, relative velocity as parameter, calculate radar target and video The association angle value of target, when associating angle value and being more than the threshold value setting it is believed that mating;The correlation that same sensor is observed and predicted in succession Data carries out synthesis and state estimation, and with reference to the checking that data is modified of observing and predicting in other information source, each sensor is passed The point mark sent is associated, and keeps target is continuously followed the tracks of;
6th step:Target recognition and tracking;Form the spy of a N-dimensional according to a certain target characteristic that different sensors record Levy vector, often an one-dimensional independent characteristic representing target, is compared with consistent feature, so that it is determined that the classification of target. New data set is just merged by the end of scan with original data every time, and the observation according to sensor estimates target component, And estimate the position of target in prediction scanning next time with these;
7th step:Traffic behavior is estimated;Detections of radar, to target, exports three-dimensional coordinate, controls video monitoring output image, According to video coordinates model and radar and the position relationship of video, using 2 points of lowest distance value d of A, B as matching condition, make The information obtaining the same object that two sensors detect corresponds to, and is gone out for same target with match cognization from synchronous images. Which kind of the data set of all targets is compared with the behavioral pattern of previously determined possible situation, so that determined behavioral pattern and prison In viewed area, the state of all targets is mated most, is saved in traffic information platform by same for these information.
Preferably, in described second step, described had with Grubbs statistical method for abnormal data preprocess method Body is as follows:
Calculate each detection data Z of outputiAverage
Calculate standard deviation
Calculate Grubbs statistic
Give according to data volume n, significant level a=0.05, find out marginal value T of Grubbs statistic by look-up table (n a), is compared with T;It is small probability event according to P [T >=T (n, a)]=a, give up T >=T (n, data a).
Preferably, in described 3rd step, the establishment step of coordinate corresponding relation is as follows:
First, calculate the inner parameter of video sensor using calibration technique, set up video sensor coordinate model;
Secondly, according to the position relationship between video sensor coordinate model and microwave remote sensor and video sensor, Set up coordinate in video sensor acquired image plane for the target that microwave remote sensor under world coordinate system monitored Corresponding relation;
Finally, the information of microwave remote sensor just can be realized according to coordinate corresponding relation and video information is merged, real The 3D world coordinates that existing microwave remote sensor detects is converted into corresponding 2D image coordinate p ' (u ', v ') in video image, with abundant The positional information correspondence being monitored using microwave remote sensor is to video image.
Preferably, in described 4th step, for same object of observation, the result of different sensors output can not Same, in the case of there is no priori, take following methods to carry out data fusion:
Using adaptive optimal Weighted Fusion model, if the traffic flow data variance of two sensors is respectively σ1、σ2, institute True value to be estimated is X, and the measured value of each sensor is respectively X1、X2, they are independent each other, and are that the unbiased of X is estimated Meter;The weighter factor of each sensor is respectively W1、W2, then the measured value after fusionFor:
Wherein
The method can require no knowledge about any priori of this two detection measurement data, simply applies multisensor The detection data providing the minimum data fusion value of mean value error it is possible to merge.
Preferably, Fuzzy Synthetical Decision Model is adopted to construct a traffic events recognizer, step in described 7th step As follows:
A1, traffic behavior are estimated, set up model library, to the traffic abnormity state modeling generally occurring within, are easy to record Pattern match in behavioral pattern and data base;
A2, monitor in real time pavement state, carry out monitor in real time by microwave remote sensor 2 and video sensor 1;
A3, there is the judgement of situation in monitoring range by radar, if it has not, then return A2 proceeding in real time Monitoring pavement state, if it has, then enter next step;
A4, the three-dimensional coordinate of outgoing event target, the current synchronous images of video acquisition;Radar and video information merge, Three-dimensional coordinate is mapped to the radar detection coordinate in synchronous images, sends early warning information, and regarded by video sensor 1 Frequency gathers current synchronous images;
A5, radar and video information merge, and three-dimensional coordinate mapping is obtained the radar detection coordinate in synchronous images;
A6, in world coordinate system, set up the matching relationship of radar detection coordinate and image detection target, from synchronous images In identify event information;
The information such as the picture of A7, output traffic events type and event vehicle, license plate number are to traffic events and information management Management platform.
After technique scheme, in the present invention, video and radar complex are using composition radar-video multisensor system System, using message complementary sense, by Data fusion technique, becomes detection means of tracking that is separate and supplementing each other, Neng Gouti High system reliability can provide the accurate estimation to target location;Melted by the telecommunication flow information that each sensor detects Conjunction is processed, and obtains more precisely data, provides strong parameter for relevant department;Taken by video based on detections of radar Supplemented by card, traffic behavior state is estimated, event information is carried out with alarm and captures evidence obtaining;The present invention proposes to be based on The fusion method of microwave and video data, after carrying out pretreatment to the initial data of both sensors, through space-time Unified, draw standardized characteristic information;In data fusion module, by using the decision making level data fusion side based on weights Method, exports more accurate telecommunication flow information;After microwave remote sensor detects some traffic behaviors of typing in data base, control Whether video sensor is taken pictures, then be analyzed mating with data base, is reported to the police, and reduces false alarm rate, reduces manpower Material resources, are truly realized intellectualized detection.
Brief description
Fig. 1 is the schematic block diagram of the Vehicle Detection device that the present invention is merged based on microwave and video data;
Fig. 2 is microwave and the schematic block diagram of video data fusion method;
Fig. 3 is the schematic block diagram of data preprocessing method;
Fig. 4 be the traffic behavior method of estimation that merged based on microwave and video data show that frame is intended to.
Specific embodiment
Purpose, technical scheme and advantage for making the embodiment of the present invention are clearer, below in conjunction with the embodiment of the present invention In accompanying drawing, the technical scheme in the embodiment of the present invention is clearly and completely described.Following examples are only used for more clear Chu's ground explanation technical scheme, and can not be limited the scope of the invention with this.
A kind of Vehicle Detection device being merged based on microwave and video data, as shown in figure 1, described device includes video pass Sensor 1, microwave remote sensor 2, A/D modular converter 3, processor module 4, network communication module 5, traffic flow parameter data fusion mould Block 6, traffic flow management platform 7, target tracking module 8, traffic events and information management platform 9, video sensor 1 and microwave pass Sensor 2 is connected with A/D modular converter 3 respectively, the non electrical quantity of the different characteristic of video sensor 1 and microwave remote sensor 2 output Signal, being then passed through that A/D modular converter 3 converts them to can be by the digital quantity of computer disposal;A/D modular converter 3 and place Reason device module 4 connects, and processor module 4 is connected with network communication module 5, and processor module 4 is at via A/D modular converter 3 The data that reason is converted into digital quantity is processed, and filters some abnormal datas to obtain useful signal, useful signal is again by net Network communication module 5 is transmitted;Network communication module 5 respectively with traffic flow parameter data fusion module 6, target tracking module 8 It is connected, useful signal is transferred to traffic flow parameter data fusion module 6, target tracking module 8 by network communication module 5;Hand over Through-flow supplemental characteristic Fusion Module 6 and traffic flow management platform 7 connect, and traffic flow parameter data fusion module 6 is to useful signal Carry out space-time uniformity, feature extraction, and by certain rule, data fusion calculating is carried out to characteristic quantity, finally that fusion results are defeated Go out to traffic flow management platform 7;Target tracking module 8 and traffic events are connected with information management platform 9, target tracking module 8 Useful signal is carried out with space-time uniformity, feature extraction, and by certain rule, data fusion calculating is carried out to characteristic quantity, finally will Fusion results export to information management platform.
As shown in Fig. 2 Fig. 4, illustrate to based on the Vehicle Detection method that microwave and video data merge below, a kind of Flow process as shown in Figure 2 is included based on the Vehicle Detection method that microwave and video data merge, as follows:
The first step:Detection, carries out background noise suppression respectively in two sensors detection zone, export traffic flow, put down All speed, occupation rate, queue length and other instant messages;
Second step:Initial data pretreatment, is standardized and carries out pretreatment to multigroup sensing data of input, full The requirement to amount of calculation and computation sequence of sufficient subsequent estimation and processor module 4;
For abnormal data preprocess method Grubbs statistical method;
Circular is as follows:
Calculate each detection data Z of outputiAverage
Calculate standard deviation
Calculate Grubbs statistic
Give according to data volume n, significant level a=0.05, find out marginal value T of Grubbs statistic by look-up table (n a), is compared with T;It is small probability event according to P [T >=T (n, a)]=a, give up T >=T (n, data a);
3rd step:Space-time is calibrated, and calibrates time and the spatial reference point of unified each sensor, snaps to same in time Time reference, spatially it is transformed into the same coordinate system, set up coordinate corresponding relation so that the result after processing seems data It is the same that fusion treatment central station is gathered;If each sensor is independently asynchronous working over time and space, must enter The row time moves and coordinate transform, merges required unified time and spatial reference point to be formed;By to single sensor The position of acquisition is merged with the estimated information of identity category, obtains more accurate target location, state and identity category Estimation;
The establishment step of coordinate corresponding relation is:Calculate the inner parameter of video sensor 1 using calibration technique, foundation regards Video sensor 1 coordinate model;According between video sensor 1 coordinate model and microwave remote sensor 2 and video sensor 1 Position relationship, sets up the target that under world coordinate system, microwave remote sensor 2 is monitored and puts down in video sensor 1 acquired image Coordinate corresponding relation in face;Information and the video information of microwave remote sensor 2 just can be realized finally according to coordinate corresponding relation Merged, the 3D world coordinates realizing microwave remote sensor 2 detection is converted into corresponding 2D image coordinate p ' in video image (u ', v '), to make full use of the positional information correspondence that microwave remote sensor 2 monitors to video image;
4th step:Basic dynamic traffic Parameter fusion, can examine from video sensor 1 and microwave remote sensor 2 simultaneously Survey the basic traffic parameter such as the traffic flow on section, average speed, occupation rate, queue length and carry out fusion treatment, draw more Accurately and reliably traffic flow parameter.The fusion results of this level are the inputs of next emerging system simultaneously, this multi-level Emerging system structure design, be advantageously implemented multi-main-body cooperating information processing, the processing load of each processing center can be disperseed, Be conducive to improving system effectiveness.
For same object of observation, the result of different sensors output can be different, is not having the situation of priori Under, take following methods to carry out data fusion so that the detection data providing being capable of mean value error minimum;
Using adaptive optimal Weighted Fusion model, if the traffic flow data variance of two sensors is respectively σ1、σ2, institute True value to be estimated is X, and the measured value of each sensor is respectively X1, X2, they are independent each other, and are the unbiased of X Estimate;The weighter factor of each sensor is respectively W1、W2, then the measured value after fusionFor:
Wherein
The method can require no knowledge about any priori of this two detection measurement data, simply applies multisensor The detection data providing the minimum data fusion value of mean value error it is possible to merge;
5th step:Data association, differentiates whether the data in different time space is derived from same target, radar and video object Mated, real target can be defined as by successful match, processed by setting means it is impossible to the target of coupling is it is believed that can not be true Fixed target is it is impossible to exclude the possibility;Using the distance of target, orientation, relative velocity as parameter, calculate radar target and video The association angle value of target, when associating angle value and being more than the threshold value setting it is believed that mating;The correlation that same sensor is observed and predicted in succession Data carries out synthesis and state estimation, and with reference to the checking that data is modified of observing and predicting in other information source, each sensor is passed The point mark sent is associated, and keeps target is continuously followed the tracks of;
6th step:Target recognition and tracking;Form the spy of a N-dimensional according to a certain target characteristic that different sensors record Levy vector, often an one-dimensional independent characteristic representing target, is compared with consistent feature, so that it is determined that the classification of target. New data set is just merged by the end of scan with original data every time, and the observation according to sensor estimates target component, And estimate the position of target in prediction scanning next time with these;
7th step:Traffic behavior is estimated;Detections of radar, to target, exports three-dimensional coordinate, controls video monitoring output image, According to video coordinates model and radar and the position relationship of video, using 2 points of lowest distance value d of A, B as matching condition, make The information obtaining the same object that two sensors detect corresponds to, and goes out for same target with match cognization from synchronous images, Which kind of the data set of all targets is compared with the behavioral pattern of previously determined possible situation, so that determined behavioral pattern and prison In viewed area, the state of all targets is mated most, is saved in traffic information platform by same for these information;
Traffic events refer to incident such as vehicle traffic accident, scram, traffic jam etc., these things on road Part will cause traffic to be blocked when occurring, and will become relatively crowded at this, can be drawn the traffic parameter on basis by the 4th step Information, when occupation rate increases, speed reduces, and needs to determine whether there is event when density becomes big, needs to process in time.This step It is the traffic flow parameter after preliminary analysis merge, carry out demonstration and whether there is abnormal traffic event;
In the 7th step, traffic behavior recognition methodss are constructed using Fuzzy Synthetical Decision Model, step is as follows:
A1, traffic behavior are estimated, set up model library, to the traffic abnormity state modeling generally occurring within, are easy to record Pattern match in behavioral pattern and data base;
A2, monitor in real time pavement state, carry out monitor in real time by microwave remote sensor 2 and video sensor 1;
A3, there is the judgement of situation in monitoring range by radar, if it has not, then return A2 proceeding in real time Monitoring pavement state, if it has, then enter next step;
A4, the three-dimensional coordinate of outgoing event target, the current synchronous images of video acquisition;Radar and video information merge, Three-dimensional coordinate is mapped to the radar detection coordinate in synchronous images, sends early warning information, and regarded by video sensor 1 Frequency gathers current synchronous images;
A5, radar and video information merge, and three-dimensional coordinate mapping is obtained the radar detection coordinate in synchronous images;
A6, in world coordinate system, set up the matching relationship of radar detection coordinate and image detection target, from synchronous images In identify event information;
The information such as the picture of A7, output traffic events type and event vehicle, license plate number are to traffic events and information management Management platform 9.
The superiority of information fusion can be described as the robustness of computing, the coverage of expansion space and time, and increase is estimated The credibility of meter, improves detection performance, improves space resolution capability, make full use of resource and the scheduling system of multisensor, Limits play the utilization rate of resource and improve the survival ability of multisensor syste.
The present invention uses hierarchical fusion algorithm, in the pretreatment link of system, abnormal data is rejected, in traffic ginseng Number collection link and traffic behavior are estimated to carry out data fusion respectively, improve the robustness of system.This technological incorporation is accurately Multidimensional information, particularly in the case that information has uncertainty, with respect to single-sensor data processing be to single The process that the obtained data of sensor or information are carried out, it is understood that there may be partly imperfect or insecure information, multisensor Data fusion can comprehensively utilize multi-sensor information effectively, such that it is able to obtain detected target and environment to a great extent More accurate, complete information and conforming description or understanding.
In the present invention, video and radar complex, using constituting radar video multisensor syste, using message complementary sense, pass through Data fusion technique, becomes detection means of tracking that is separate and supplementing each other, it is possible to increase system reliability can provide Accurate estimation to target location;Fusion treatment is carried out by the telecommunication flow information that each sensor detects, obtains more accurate Ground data, provides strong parameter for relevant department;Supplemented by the evidence obtaining of video based on detections of radar, to traffic behavior shape State is estimated, event information is carried out with alarm and captures evidence obtaining;
The present invention proposes fusion method based on microwave and video data, by entering to the initial data of both sensors After row pretreatment, through space-time uniformity, draw standardized characteristic information;In data fusion module, by using based on power The decision making level data fusion method of value, exports more accurate telecommunication flow information;Microwave remote sensor 2 detects typing in data base Some traffic behaviors after, control video sensor 1 to be taken pictures, then be analyzed whether mating with data base, reported to the police, Reduce false alarm rate, reduce manpower and materials, be truly realized intellectualized detection.
Finally it should be noted that:The foregoing is only the preferred embodiments of the present invention, be not limited to the present invention, Although being described in detail to the present invention with reference to the foregoing embodiments, for a person skilled in the art, it still may be used To modify to the technical scheme described in foregoing embodiments, or equivalent is carried out to wherein some technical characteristics. All any modification, equivalent substitution and improvement within the spirit and principles in the present invention, made etc., should be included in the present invention's Within protection domain.

Claims (6)

1. a kind of Vehicle Detection device being merged based on microwave and video data it is characterised in that:Described device includes video and passes Sensor, microwave remote sensor, A/D modular converter, processor module, network communication module, traffic flow parameter data fusion module, friendship Through-flow management platform, target tracking module, traffic events and information management platform;
Described video sensor is connected with described A/D modular converter respectively with described microwave remote sensor, video sensor and micro- The non-electric charge quantity signalling of the different characteristic of wave sensor output, being then passed through that described A/D modular converter converts them to can be by counting The digital quantity that calculation machine is processed;Described A/D modular converter is connected with described processor module, described processor module and described network Communication module connects, and described processor module is converted at the data of digital quantity to via described A/D module processing Reason, filters some abnormal datas to obtain useful signal, useful signal is transmitted by described network communication module again;
Described network communication module is connected with described traffic flow parameter data fusion module, described target tracking module respectively, Useful signal is transferred to described traffic flow parameter data fusion module, described target tracking module by described network communication module;
Described traffic flow parameter data fusion module and described traffic flow management platform connect, described traffic flow parameter data fusion Module carries out space-time uniformity, feature extraction to useful signal, and carries out data fusion calculating by certain rule to characteristic quantity, Afterwards fusion results are exported to described traffic flow management platform;
Described target tracking module and traffic events are connected with information management platform, and described target tracking module is entered to useful signal Row space-time uniformity, feature extraction, and by certain rule, data fusion calculating is carried out to characteristic quantity, finally fusion results are exported To described information management platform.
2. a kind of Vehicle Detection method being merged based on microwave and video data is it is characterised in that comprise the steps of:
The first step:Detection, carries out background noise suppression, output traffic flow, averagely speed respectively in two sensors detection zone Degree, occupation rate, queue length and other instant messages;
Second step:Initial data pretreatment, is standardized and carries out pretreatment to multigroup sensing data of input, after satisfaction The continuous estimation and processor module requirement to amount of calculation and computation sequence;
Grubbs statistical method is adopted for abnormal data preprocess method;
3rd step:Space-time is calibrated, and calibrates time and the spatial reference point of unified each sensor, snaps to the same time in time Benchmark, spatially it is transformed into the same coordinate system, set up coordinate corresponding relation so that the result after processing seems data fusion It is the same that processing center station is gathered;If each sensor is independently asynchronous working over time and space, when must carry out Between move and coordinate transform, merge required unified time and spatial reference point to be formed;By obtaining to single sensor The estimated information of position and identity category merged, obtain more accurate the estimating of target location, state and identity category Meter;
4th step:Basic dynamic traffic Parameter fusion, can detect section from video sensor and microwave remote sensor simultaneously On traffic flow, average speed, occupation rate, the basic traffic parameter such as queue length carry out fusion treatment, draw and more accurately may be used The traffic flow parameter leaning on;The fusion results of this level are the inputs of next emerging system simultaneously;
5th step:Data association, differentiates whether the data in different time space is derived from same target, and radar is carried out with video object Coupling, can be defined as real target by successful match, process by setting means it is impossible to the target of coupling is it is believed that unascertainable Target is it is impossible to exclude the possibility;Using the distance of target, orientation, relative velocity as parameter, calculate radar target and video object Association angle value, when associate angle value be more than set threshold value when it is believed that coupling;The related data that same sensor is observed and predicted in succession Carry out synthesis and state estimation, and with reference to the checking that data is modified of observing and predicting in other information source, the transmission of each sensor is come Point mark be associated, keep target is continuously followed the tracks of;
6th step:Target recognition and tracking;According to a certain target characteristic that different sensors record formed the feature of a N-dimensional to Amount, often an one-dimensional independent characteristic representing target, is compared with consistent feature, so that it is determined that the classification of target.Every time New data set is just merged by the end of scan with original data, and the observation according to sensor estimates target component, is used in combination These estimate the position of target in prediction scanning next time;
7th step:Traffic behavior is estimated;Detections of radar, to target, exports three-dimensional coordinate, controls video monitoring output image, according to The position relationship of video coordinates model and radar and video, using 2 points of lowest distance value d of A, B as matching condition so that two The information of the same object that individual sensor detects corresponds to, and is gone out for same target with match cognization from synchronous images.By institute The data set having target compared with the behavioral pattern of previously determined possible situation, so that determined which kind of behavioral pattern and surveillance zone In domain, the state of all targets is mated most, is saved in traffic information platform by same for these information.
3. a kind of Vehicle Detection method being merged based on microwave and video data according to claim 2 it is characterised in that: In described second step, described specific as follows with Grubbs statistical method for abnormal data preprocess method:
Calculate each detection data Z of outputiAverage
Calculate standard deviation
Calculate Grubbs statistic
Given according to data volume n, significant level a=0.05, by look-up table find out Grubbs statistic marginal value T (n, A), it is compared with T;It is small probability event according to P [T >=T (n, a)]=a, give up T >=T (n, data a).
4. a kind of Vehicle Detection method being merged based on microwave and video data according to claim 2 it is characterised in that: In described 3rd step, the establishment step of coordinate corresponding relation is as follows:
First, calculate the inner parameter of video sensor using calibration technique, set up video sensor coordinate model;
Secondly, according to the position relationship between video sensor coordinate model and microwave remote sensor and video sensor, set up Coordinate pair in video sensor acquired image plane for the target that under world coordinate system, microwave remote sensor is monitored should Relation;
Finally, the information of microwave remote sensor just can be realized according to coordinate corresponding relation and video information is merged, realize micro- The 3D world coordinates that wave sensor detects is converted into corresponding 2D image coordinate p ' (u ', v ') in video image, to make full use of The positional information correspondence that microwave remote sensor monitors is to video image.
5. a kind of Vehicle Detection method being merged based on microwave and video data according to claim 2 it is characterised in that: In described 4th step, for same object of observation, the result of different sensors output can be different, is not having priori In the case of, take following methods to carry out data fusion:
Using adaptive optimal Weighted Fusion model, if the traffic flow data variance of two sensors is respectively σ1、σ2, estimate The true value of meter is X, and the measured value of each sensor is respectively X1、X2, they are independent each other, and are the unbiased esti-mator of X; The weighter factor of each sensor is respectively W1、W2, then the measured value after fusionFor:
Wherein
The detection data that the method application multisensor provides, merges and the minimum data fusion value of mean value error.
6. a kind of Vehicle Detection method being merged based on microwave and video data according to claim 2 it is characterised in that: Fuzzy Synthetical Decision Model is adopted to construct a traffic events recognizer in described 7th step, step is as follows:
A1, traffic behavior are estimated, set up model library, to the traffic abnormity state modeling generally occurring within, are easy to the behavior recording Pattern match in pattern and data base;
A2, monitor in real time pavement state, carry out monitor in real time by microwave remote sensor 2 and video sensor 1;
A3, there is the judgement of situation in monitoring range by radar, if it has not, then returning A2 to proceed monitor in real time Pavement state, if it has, then enter next step;
A4, the three-dimensional coordinate of outgoing event target, the current synchronous images of video acquisition;Radar and video information merge, by three Dimension coordinate is mapped to the radar detection coordinate in synchronous images, sends early warning information, and carries out video by video sensor 1 and adopt Collect current synchronous images;
A5, radar and video information merge, and three-dimensional coordinate mapping is obtained the radar detection coordinate in synchronous images;
A6, in world coordinate system, set up the matching relationship of radar detection coordinate and image detection target, know from synchronous images Other outgoing event information;
The information such as the picture of A7, output traffic events type and event vehicle, license plate number are to traffic events and information management management Platform.
CN201610952272.0A 2016-11-02 2016-11-02 Traffic detection method and device based on microwave and video data fusion Active CN106408940B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201610952272.0A CN106408940B (en) 2016-11-02 2016-11-02 Traffic detection method and device based on microwave and video data fusion

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201610952272.0A CN106408940B (en) 2016-11-02 2016-11-02 Traffic detection method and device based on microwave and video data fusion

Publications (2)

Publication Number Publication Date
CN106408940A true CN106408940A (en) 2017-02-15
CN106408940B CN106408940B (en) 2023-04-14

Family

ID=58014428

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201610952272.0A Active CN106408940B (en) 2016-11-02 2016-11-02 Traffic detection method and device based on microwave and video data fusion

Country Status (1)

Country Link
CN (1) CN106408940B (en)

Cited By (36)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107680054A (en) * 2017-09-26 2018-02-09 长春理工大学 Multisource image anastomosing method under haze environment
CN108734959A (en) * 2018-04-28 2018-11-02 扬州远铭光电有限公司 A kind of embedded vision train flow analysis method and system
CN108922188A (en) * 2018-07-24 2018-11-30 河北德冠隆电子科技有限公司 The four-dimensional outdoor scene traffic of radar tracking positioning perceives early warning monitoring management system
CN108961790A (en) * 2018-07-24 2018-12-07 河北德冠隆电子科技有限公司 Bad weather pre-warning management system and method based on four-dimensional outdoor scene traffic simulation
CN109343051A (en) * 2018-11-15 2019-02-15 众泰新能源汽车有限公司 A kind of multi-Sensor Information Fusion Approach driven for advanced auxiliary
CN109613537A (en) * 2019-01-16 2019-04-12 南京奥杰智能科技有限公司 A kind of hologram radar
CN109615866A (en) * 2019-01-16 2019-04-12 南京奥杰智能科技有限公司 Traffic monitoring system Internet-based
CN109671278A (en) * 2019-03-02 2019-04-23 安徽超远信息技术有限公司 A kind of bayonet precise positioning grasp shoot method and device based on multiple target radar
CN110126885A (en) * 2018-02-02 2019-08-16 保定市天河电子技术有限公司 A kind of railway circumference intrusion target monitoring method and system
CN110163270A (en) * 2019-05-10 2019-08-23 北京易控智驾科技有限公司 Multi-Sensor Information Fusion Approach and system
CN110444026A (en) * 2019-08-06 2019-11-12 北京万集科技股份有限公司 The triggering grasp shoot method and system of vehicle
CN110640737A (en) * 2018-11-07 2020-01-03 宁波赛朗科技有限公司 Industrial robot for measuring data fusion attitude
CN110796868A (en) * 2019-12-02 2020-02-14 江苏中路工程技术研究院有限公司 Video and microwave integrated traffic incident monitoring system and method
CN110865367A (en) * 2019-11-30 2020-03-06 山西禾源科技股份有限公司 Intelligent fusion method for radar video data
CN110969059A (en) * 2018-09-30 2020-04-07 长城汽车股份有限公司 Lane line identification method and system
CN111209327A (en) * 2020-01-14 2020-05-29 南京悠淼科技有限公司 Multi-sensor distributed sensing interconnection and edge fusion processing system and method
CN111477010A (en) * 2020-04-08 2020-07-31 图达通智能科技(苏州)有限公司 Device for intersection holographic sensing and control method thereof
CN112150799A (en) * 2020-08-19 2020-12-29 上海图丽信息技术有限公司 Method for collecting road vehicle traffic big data by fusing radar videos
CN112150797A (en) * 2020-08-19 2020-12-29 上海图丽信息技术有限公司 Traffic incident detection method fusing radar videos
CN112148769A (en) * 2020-09-15 2020-12-29 浙江大华技术股份有限公司 Data synchronization method, device, storage medium and electronic device
CN112509331A (en) * 2020-12-18 2021-03-16 芜湖易来达雷达科技有限公司 Verification system and verification method for traffic radar data
CN112532934A (en) * 2020-11-23 2021-03-19 国网山东省电力公司利津县供电公司 Multi-dimensional cooperative monitoring system
CN112731324A (en) * 2020-12-16 2021-04-30 中交第一公路勘察设计研究院有限公司 Multi-radar cross-regional networking multi-target tracking method for expressway
CN112837529A (en) * 2019-11-25 2021-05-25 斑马智行网络(香港)有限公司 Data processing method and system, acquisition device, processor and storage medium
CN113255708A (en) * 2020-02-10 2021-08-13 富士通株式会社 Data fusion method and device and data processing equipment
CN113393676A (en) * 2021-06-09 2021-09-14 东北林业大学 Traffic detection method and device based on unmanned aerial vehicle vision and millimeter wave radar
CN113412506A (en) * 2019-02-13 2021-09-17 日立安斯泰莫株式会社 Vehicle control device and electronic control system
CN113689691A (en) * 2020-05-18 2021-11-23 富士通株式会社 Traffic detection system
CN114530042A (en) * 2021-12-31 2022-05-24 威海南海数字产业研究院有限公司 Urban traffic brain monitoring system based on internet of things technology
CN114814720A (en) * 2022-06-20 2022-07-29 成都市克莱微波科技有限公司 Microwave direction finding device, system, method and storage medium
CN115278361A (en) * 2022-07-20 2022-11-01 重庆长安汽车股份有限公司 Driving video data extraction method, system, medium and electronic equipment
CN115376312A (en) * 2022-07-22 2022-11-22 交通运输部路网监测与应急处置中心 Road monitoring method and system based on radar and video fusion
CN111582130B (en) * 2020-04-30 2023-04-28 长安大学 Traffic behavior perception fusion system and method based on multi-source heterogeneous information
CN117636671A (en) * 2024-01-24 2024-03-01 四川君迪能源科技有限公司 Cooperation scheduling method and system for intelligent vehicle meeting of rural roads
US11934746B2 (en) 2017-10-25 2024-03-19 Ihi Corporation Information generation device
CN118013465A (en) * 2024-04-09 2024-05-10 微网优联科技(成都)有限公司 Non-motor vehicle identification method and system based on multi-sensor cooperation

Citations (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1940591A (en) * 2005-09-26 2007-04-04 通用汽车环球科技运作公司 System and method of target tracking using sensor fusion
CN101226689A (en) * 2008-02-03 2008-07-23 北京交通大学 Multi-sensor access device for acquisition of road traffic information and data fusion method thereof
WO2008142680A2 (en) * 2007-05-20 2008-11-27 Rafael Advanced Defense Systems Ltd Tracking and imaging data fusion
CN101318491A (en) * 2008-05-14 2008-12-10 合肥工业大学 Built-in integrated visual sensation auxiliary driving safety system
CN101655561A (en) * 2009-09-14 2010-02-24 南京莱斯信息技术股份有限公司 Federated Kalman filtering-based method for fusing multilateration data and radar data
CN101751782A (en) * 2009-12-30 2010-06-23 北京大学深圳研究生院 Crossroad traffic event automatic detection system based on multi-source information fusion
CN102542843A (en) * 2010-12-07 2012-07-04 比亚迪股份有限公司 Early warning method for preventing vehicle collision and device
CN102881162A (en) * 2012-09-29 2013-01-16 北京市交通信息中心 Data processing and fusion method for large-scale traffic information
CN103093625A (en) * 2013-01-09 2013-05-08 杭州师范大学 City road traffic condition real-time estimation method based on reliability verification
CN103116981A (en) * 2011-11-17 2013-05-22 无锡物联网产业研究院 Multi-sensor system and information fusion method
US20140195138A1 (en) * 2010-11-15 2014-07-10 Image Sensing Systems, Inc. Roadway sensing systems
CN104123837A (en) * 2013-04-28 2014-10-29 上海济祥智能交通科技有限公司 Interrupted flow travel time estimation method based on microwave and video data fusion
CN104200657A (en) * 2014-07-22 2014-12-10 杭州智诚惠通科技有限公司 Traffic flow parameter acquisition method based on video and sensor
CN104933879A (en) * 2014-03-19 2015-09-23 北京航天长峰科技工业集团有限公司 Traffic information collecting, inducing and publishing method based on Internet of Things
CN105015411A (en) * 2015-07-03 2015-11-04 河南工业技术研究院 Automobile microwave radar anti-collision early-warning method and system based on video fusion
CN105334514A (en) * 2015-10-19 2016-02-17 上海无线电设备研究所 Tramcar radar video compound early warning crashproof system and method
CN105427619A (en) * 2015-12-24 2016-03-23 上海新中新猎豹交通科技股份有限公司 Vehicle following distance automatic recording system and method
CN105807280A (en) * 2016-04-26 2016-07-27 南京鹏力***工程研究所 Echo fused target track association method based on track state estimation

Patent Citations (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1940591A (en) * 2005-09-26 2007-04-04 通用汽车环球科技运作公司 System and method of target tracking using sensor fusion
WO2008142680A2 (en) * 2007-05-20 2008-11-27 Rafael Advanced Defense Systems Ltd Tracking and imaging data fusion
CN101226689A (en) * 2008-02-03 2008-07-23 北京交通大学 Multi-sensor access device for acquisition of road traffic information and data fusion method thereof
CN101318491A (en) * 2008-05-14 2008-12-10 合肥工业大学 Built-in integrated visual sensation auxiliary driving safety system
CN101655561A (en) * 2009-09-14 2010-02-24 南京莱斯信息技术股份有限公司 Federated Kalman filtering-based method for fusing multilateration data and radar data
CN101751782A (en) * 2009-12-30 2010-06-23 北京大学深圳研究生院 Crossroad traffic event automatic detection system based on multi-source information fusion
US20140195138A1 (en) * 2010-11-15 2014-07-10 Image Sensing Systems, Inc. Roadway sensing systems
CN102542843A (en) * 2010-12-07 2012-07-04 比亚迪股份有限公司 Early warning method for preventing vehicle collision and device
CN103116981A (en) * 2011-11-17 2013-05-22 无锡物联网产业研究院 Multi-sensor system and information fusion method
CN102881162A (en) * 2012-09-29 2013-01-16 北京市交通信息中心 Data processing and fusion method for large-scale traffic information
CN103093625A (en) * 2013-01-09 2013-05-08 杭州师范大学 City road traffic condition real-time estimation method based on reliability verification
CN104123837A (en) * 2013-04-28 2014-10-29 上海济祥智能交通科技有限公司 Interrupted flow travel time estimation method based on microwave and video data fusion
CN104933879A (en) * 2014-03-19 2015-09-23 北京航天长峰科技工业集团有限公司 Traffic information collecting, inducing and publishing method based on Internet of Things
CN104200657A (en) * 2014-07-22 2014-12-10 杭州智诚惠通科技有限公司 Traffic flow parameter acquisition method based on video and sensor
CN105015411A (en) * 2015-07-03 2015-11-04 河南工业技术研究院 Automobile microwave radar anti-collision early-warning method and system based on video fusion
CN105334514A (en) * 2015-10-19 2016-02-17 上海无线电设备研究所 Tramcar radar video compound early warning crashproof system and method
CN105427619A (en) * 2015-12-24 2016-03-23 上海新中新猎豹交通科技股份有限公司 Vehicle following distance automatic recording system and method
CN105807280A (en) * 2016-04-26 2016-07-27 南京鹏力***工程研究所 Echo fused target track association method based on track state estimation

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
刘志强;程红星;王运霞;: "车辆防撞检测技术研究" *

Cited By (42)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107680054A (en) * 2017-09-26 2018-02-09 长春理工大学 Multisource image anastomosing method under haze environment
US11934746B2 (en) 2017-10-25 2024-03-19 Ihi Corporation Information generation device
CN110126885A (en) * 2018-02-02 2019-08-16 保定市天河电子技术有限公司 A kind of railway circumference intrusion target monitoring method and system
CN108734959A (en) * 2018-04-28 2018-11-02 扬州远铭光电有限公司 A kind of embedded vision train flow analysis method and system
CN108922188A (en) * 2018-07-24 2018-11-30 河北德冠隆电子科技有限公司 The four-dimensional outdoor scene traffic of radar tracking positioning perceives early warning monitoring management system
CN108961790A (en) * 2018-07-24 2018-12-07 河北德冠隆电子科技有限公司 Bad weather pre-warning management system and method based on four-dimensional outdoor scene traffic simulation
CN110969059A (en) * 2018-09-30 2020-04-07 长城汽车股份有限公司 Lane line identification method and system
CN110640737A (en) * 2018-11-07 2020-01-03 宁波赛朗科技有限公司 Industrial robot for measuring data fusion attitude
CN109343051A (en) * 2018-11-15 2019-02-15 众泰新能源汽车有限公司 A kind of multi-Sensor Information Fusion Approach driven for advanced auxiliary
CN109615866A (en) * 2019-01-16 2019-04-12 南京奥杰智能科技有限公司 Traffic monitoring system Internet-based
CN109613537A (en) * 2019-01-16 2019-04-12 南京奥杰智能科技有限公司 A kind of hologram radar
CN113412506B (en) * 2019-02-13 2023-06-13 日立安斯泰莫株式会社 Vehicle control device and electronic control system
CN113412506A (en) * 2019-02-13 2021-09-17 日立安斯泰莫株式会社 Vehicle control device and electronic control system
CN109671278A (en) * 2019-03-02 2019-04-23 安徽超远信息技术有限公司 A kind of bayonet precise positioning grasp shoot method and device based on multiple target radar
CN110163270A (en) * 2019-05-10 2019-08-23 北京易控智驾科技有限公司 Multi-Sensor Information Fusion Approach and system
CN110444026B (en) * 2019-08-06 2021-07-09 北京万集科技股份有限公司 Triggering snapshot method and system for vehicle
CN110444026A (en) * 2019-08-06 2019-11-12 北京万集科技股份有限公司 The triggering grasp shoot method and system of vehicle
CN112837529A (en) * 2019-11-25 2021-05-25 斑马智行网络(香港)有限公司 Data processing method and system, acquisition device, processor and storage medium
CN112837529B (en) * 2019-11-25 2022-12-02 斑马智行网络(香港)有限公司 Data processing method and system, acquisition device, processor and storage medium
CN110865367B (en) * 2019-11-30 2023-05-05 山西禾源科技股份有限公司 Intelligent radar video data fusion method
CN110865367A (en) * 2019-11-30 2020-03-06 山西禾源科技股份有限公司 Intelligent fusion method for radar video data
CN110796868A (en) * 2019-12-02 2020-02-14 江苏中路工程技术研究院有限公司 Video and microwave integrated traffic incident monitoring system and method
CN111209327A (en) * 2020-01-14 2020-05-29 南京悠淼科技有限公司 Multi-sensor distributed sensing interconnection and edge fusion processing system and method
CN113255708A (en) * 2020-02-10 2021-08-13 富士通株式会社 Data fusion method and device and data processing equipment
CN111477010A (en) * 2020-04-08 2020-07-31 图达通智能科技(苏州)有限公司 Device for intersection holographic sensing and control method thereof
CN111582130B (en) * 2020-04-30 2023-04-28 长安大学 Traffic behavior perception fusion system and method based on multi-source heterogeneous information
CN113689691A (en) * 2020-05-18 2021-11-23 富士通株式会社 Traffic detection system
CN112150797A (en) * 2020-08-19 2020-12-29 上海图丽信息技术有限公司 Traffic incident detection method fusing radar videos
CN112150799A (en) * 2020-08-19 2020-12-29 上海图丽信息技术有限公司 Method for collecting road vehicle traffic big data by fusing radar videos
CN112148769A (en) * 2020-09-15 2020-12-29 浙江大华技术股份有限公司 Data synchronization method, device, storage medium and electronic device
CN112532934A (en) * 2020-11-23 2021-03-19 国网山东省电力公司利津县供电公司 Multi-dimensional cooperative monitoring system
CN112731324A (en) * 2020-12-16 2021-04-30 中交第一公路勘察设计研究院有限公司 Multi-radar cross-regional networking multi-target tracking method for expressway
CN112509331A (en) * 2020-12-18 2021-03-16 芜湖易来达雷达科技有限公司 Verification system and verification method for traffic radar data
CN113393676A (en) * 2021-06-09 2021-09-14 东北林业大学 Traffic detection method and device based on unmanned aerial vehicle vision and millimeter wave radar
CN114530042A (en) * 2021-12-31 2022-05-24 威海南海数字产业研究院有限公司 Urban traffic brain monitoring system based on internet of things technology
CN114814720A (en) * 2022-06-20 2022-07-29 成都市克莱微波科技有限公司 Microwave direction finding device, system, method and storage medium
CN115278361A (en) * 2022-07-20 2022-11-01 重庆长安汽车股份有限公司 Driving video data extraction method, system, medium and electronic equipment
CN115278361B (en) * 2022-07-20 2023-08-01 重庆长安汽车股份有限公司 Driving video data extraction method, system, medium and electronic equipment
CN115376312A (en) * 2022-07-22 2022-11-22 交通运输部路网监测与应急处置中心 Road monitoring method and system based on radar and video fusion
CN117636671A (en) * 2024-01-24 2024-03-01 四川君迪能源科技有限公司 Cooperation scheduling method and system for intelligent vehicle meeting of rural roads
CN117636671B (en) * 2024-01-24 2024-04-30 四川君迪能源科技有限公司 Cooperation scheduling method and system for intelligent vehicle meeting of rural roads
CN118013465A (en) * 2024-04-09 2024-05-10 微网优联科技(成都)有限公司 Non-motor vehicle identification method and system based on multi-sensor cooperation

Also Published As

Publication number Publication date
CN106408940B (en) 2023-04-14

Similar Documents

Publication Publication Date Title
CN106408940A (en) Microwave and video data fusion-based traffic detection method and device
WO2020253308A1 (en) Human-machine interaction behavior security monitoring and forewarning method for underground belt transportation-related personnel
CN113255481B (en) Crowd state detection method based on unmanned patrol car
CN102811343B (en) Intelligent video monitoring system based on behavior recognition
Huang et al. Automatic moving object extraction through a real-world variable-bandwidth network for traffic monitoring systems
CN101552910B (en) Remnant detection device based on comprehensive computer vision
CN106203274A (en) Pedestrian's real-time detecting system and method in a kind of video monitoring
CA3094424A1 (en) Safety monitoring and early-warning method for man-machine interaction behavior of underground conveyor belt operator
Sun et al. Vehicle reidentification using multidetector fusion
JP5479907B2 (en) Network monitoring system
CN102254394A (en) Antitheft monitoring method for poles and towers in power transmission line based on video difference analysis
CN102163290A (en) Method for modeling abnormal events in multi-visual angle video monitoring based on temporal-spatial correlation information
CN109830078B (en) Intelligent behavior analysis method and intelligent behavior analysis equipment suitable for narrow space
CN113326719A (en) Method, equipment and system for target tracking
CN106846297A (en) Pedestrian's flow quantity detecting system and method based on laser radar
CN114333424B (en) Bridge prevents ship and hits monitoring early warning system
CN112183472A (en) Method for detecting whether test field personnel wear work clothes or not based on improved RetinaNet
CN113642403B (en) Crowd abnormal intelligent safety detection system based on edge calculation
CN110490150A (en) A kind of automatic auditing system of picture violating the regulations and method based on vehicle retrieval
Ua-Areemitr et al. Low-cost road traffic state estimation system using time-spatial image processing
CN117312801A (en) AI-based smart city monitoring system and method
CN115860144A (en) Machine learning system for anti-electricity-stealing site
CN114708544A (en) Intelligent violation monitoring helmet based on edge calculation and monitoring method thereof
CN110175521A (en) Method based on double camera linkage detection supervision indoor human body behavior
CN105471632B (en) A kind of detection method of autoregression line fault

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant