CN107749963A - A kind of source information that perceives merges video method more - Google Patents
A kind of source information that perceives merges video method more Download PDFInfo
- Publication number
- CN107749963A CN107749963A CN201710961702.XA CN201710961702A CN107749963A CN 107749963 A CN107749963 A CN 107749963A CN 201710961702 A CN201710961702 A CN 201710961702A CN 107749963 A CN107749963 A CN 107749963A
- Authority
- CN
- China
- Prior art keywords
- video
- data
- converged services
- frame
- picture
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/18—Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/70—Information retrieval; Database structures therefor; File system structures therefor of video data
- G06F16/71—Indexing; Data structures therefor; Storage structures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/70—Information retrieval; Database structures therefor; File system structures therefor of video data
- G06F16/73—Querying
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/76—Television signal recording
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/20—Adaptations for transmission via a GHz frequency band, e.g. via satellite
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Signal Processing (AREA)
- Data Mining & Analysis (AREA)
- Databases & Information Systems (AREA)
- General Engineering & Computer Science (AREA)
- Computational Linguistics (AREA)
- Astronomy & Astrophysics (AREA)
- Software Systems (AREA)
- Television Signal Processing For Recording (AREA)
Abstract
Source information fusion video method is perceived the invention discloses one kind, this method includes more:1) according to the hardware device in system architecture deployment system;2) converged services module obtains video flowing and frame of video picture from network hard disk video recorder;3) timing of converged services module obtains multi-source data by data access interface;4) converged services module is that frame of video picture generates pathname and filename, and preserves picture;5) converged services module is merged the data got and frame of video, and by the data storage after fusion in the local database;6) inquiry request that service module receives client is inquired about, returns to the picture for the condition that meets.The present invention is capable of the real-time acquisition of systematization, storage, fusion, retrieval data, the relevance established between data source, ensures global coherency, integrality, effectively improves the reaction speed and efficiency of production efficiency, safety precaution, and emergency processing.
Description
Technical field
The present invention relates to the fusion problem for perceiving source information in video monitoring system more, and in particular to Big Dipper receiver number
According to, the acquisition of the multi-source information such as Temperature Humidity Sensor data and video flowing, merge, storage, querying method, belong to security protection video prison
Control field.
Background technology
In the industrial production, each key area has all deployed video monitoring system, can obtain the video in scene
And image, improve the security and manageability of production scene.Generally, there be various environmental key-elements production scene, including when
Between, geography, noise, humiture, wind direction, come in and go out personnel identity information etc., these information can be obtained by various sensing equipments,
It is the information required for safety in production, security protection.But these systems are all the autonomous systems established one's own system, data source is all only
It is vertical to be stored in respective server, it is difficult to comprehensively utilize, carry out the feature of event that accurate description occurred and scene, it is difficult to support
Intelligentized production, scheduling, control, protection, this scattered, independent monitoring system and database are unfavorable for efficient, intelligent work
The needs of industry production management.
In order to improve video monitoring and safety in production, the efficiency of safety precaution and intelligence degree, it is necessary to will be above-mentioned more
Kind environmental information collects fusion, real-time acquisition, storage, the retrieval of systematization by the intelligent video monitoring system of integration
Data, the relevance established between data source, particularly ensure time, the accuracy in place, ensure global coherency, complete
Property, beneficial to alarm of pinpointing the problems, while situation can be analyzed, further do data mining, therefrom excavate valuable
Information, for improving the reaction speed and efficiency of production efficiency, safety precaution, and emergency processing.
The source information integration technologies that perceive come across in the 1970s, how that the data message of separate sources is whole studying more
It is combined, to obtain the result more accurate, more reliable than with single data.It is perceive level of the source information fusion by fusion more
It is divided into 3 kinds of pixel-based fusion, feature-based fusion and decision level fusion.Pixel-based fusion is the fusion of lowest level, directly to passing
The observation data of sensor carry out fusion treatment, are then based on fusion results and carry out feature extraction and judge decision-making.Feature-based fusion
Refer in the raw information that each sensor provides, extract one group of characteristic information first, form characteristic vector, and enter to target
Row classification or other before processings merge to each group information, sometimes referred to as middle rank fusion.Conventional method has clustering methodology,
Artificial neural network and K rank nearest neighbor methods etc..Decision level fusion is high-level fusion, is first based on itself institute by each sensor
The data of collection make decisions, and then complete the fusion treatment to local decision in fusion center.Common algorithms have Bayes to push away
Disconnected, expert system, D-S evidential reasonings etc..
The video frequency monitoring method of existing multisource data fusion is for particular source either in video encoder apparatus
Inside receive multi-source data.A kind of for example, video monitoring pedestrian identification side for multi-source big data fusion that Wuhan University proposes
Method (CN105357496A), the physical identity of pedestrian is identified by the push-to-talk handset number of control point.Observe on travel path
The number of recurrences of phone number on control point where specific suspicion pedestrian, calculates pedestrian, and phone number reappears probability, and two
The combined probability of person identifies the physical identity of pedestrian.Solve and only rely only on monitor video itself to identify monitoring objective identity
Problem of Failure in some cases.Zhongxing Microelectronci Co., Ltd., Beijing proposes a kind of video for supporting multisource data fusion
Method for hierarchically coding (CN106303538A), video is optimized and encode simultaneously together with the multi-source information of external sensor
Step combines, and source video is formed the classification code stream comprising video content and each object video relevant information.Solves existing video
Encoder apparatus can not receive external sensor information and operating efficiency is low, it is impossible to effectively give top priority to what is the most important information the problems such as.
The former invents and achievement is mainly for video monitoring pedestrian's identification of multisource data fusion, and main target is profit
With the number of recurrences of video monitoring and the push-to-talk handset number of control point to identifying rows people's identity.The latter invents and the main pin of achievement
Video spatial scalable coding method to supporting multisource data fusion, main target are that solve existing video encoder not possess outside reception
The problem of portion's Multiple Source Sensor informational function.
The content of the invention
The present invention proposes a kind of more perception source information fusion video frequency monitoring methods and system, system (the abbreviation multisource video
Monitoring system) by web camera, network hard disk video recorder, Big Dipper receiver, Temperature Humidity Sensor, converged services device, client
End composition.Converged services module is run on converged services device, inquires about two modules of service module.Converged services module is hard from network
Disk video recorder obtains video flowing, intercepts frame of video picture from video flowing at set time intervals, and pass through data-interface
Obtain the data of multiple data sources.After converged services module merges multi-source data with frame of video picture, according to set plan
Slightly, data are packed, are stored under locally specified catalogue, and establish information index.Inquiry service module is supported to come from client
The retrieval and inquisition at end.Multi-source awareness apparatus provides polytype number by data gateway or database for converged services device
According to the data-interface of converged services module is responsible for completing the collection of these data.
Specifically, method of the invention comprises the following steps (with reference to figure 1):
A. it is as follows according to the hardware device in system architecture deployment system, specific implementation step:
A1. described hardware device includes:Web camera, network hard disk video recorder, Temperature Humidity Sensor, the Big Dipper receive
Device, converged services device, client, it is connected to by interchanger in a LAN;
A2. two modules are run on converged services device:Converged services module is responsible for obtaining frame of video picture and multi-source data,
And fusion function;Inquiry service module is responsible for the request of customer in response end;
A3. Big Dipper receiver is connected with converged services device serial ports, from big-dipper satellite receive information, and issues converged services
Module;
A4. Temperature Humidity Sensor device is connected with converged services device serial ports, obtains ambient temperature and humidity data, and issue fusion
Service module;
A5. client includes but is not limited to PC computers, smart mobile phone, iPAD equipment, receives user's request, is taken from fusion
Business device obtains information needed and video;
A6. converged services module by unified data access interface and different pieces of information sources traffic and obtains data;
A7. service module is inquired about by http protocol and client communication, receives client request, and return to regarding for request
Frequency frame picture and the information inquired about;
B. converged services module obtains video flowing and frame of video picture from network hard disk video recorder;Specific implementation step is such as
Under:
B1. after converged services module starts, video flowing is asked to network hard disk video recorder;
B2. after network hard disk video recorder receives request, video flowing is sent to converged services module;
B3. converged services module is decoded it after receiving video flowing, and time driving mechanism or event are used per road video
Driving mechanism extracts frame of video picture in decoded video flowing;
Time driving mechanism described in B3.1 refers to, and starts a timer in converged services module, (or refers to every 1 second
Fixed other times interval) activate the operation for extracting frame of video picture;
Event-driven mechanism described in B3.2 refers to, in web camera, network hard disk video recorder, Big Dipper receiver, warm and humid
When specific event occurs in degree sensor, converged services device, the operation of activation extraction frame of video picture;
C. the timing of converged services module obtains multi-source data by data access interface, and specific implementation step is as follows:
C1. described data source includes but is not limited to Big Dipper receiver, Temperature Humidity Sensor;
C2. data are transmitted after big-dipper satellite receives Big Dipper data and give converged services mould by Big Dipper receiver
Block;Data include but is not limited to Big Dipper time, longitude and latitude, height above sea level etc.;
C3. converged services module obtains the data of packing according to time driving mechanism by data access interface from serial ports;
C4. Temperature Humidity Sensor device transmits data to converged services after environment gets data of the Temperature and Humidity module
Module;Data include temperature, humidity etc.;
C5. converged services module obtains the data of packing according to time driving mechanism by data access interface from serial ports;
D. converged services module is that frame of video picture generates pathname and filename, and preserves picture, specific implementation step
It is as follows:
D1. the step C2 Big Dipper times obtained and web camera mark generation pathname and filename are utilized;
D2. pathname describes the storage catalogue of frame of video picture, and its structure is divided into 6 grades:" year-month-day-when-point-network
Camera identification name ";Picture storage catalogue constructs in the way of extending step by step;Specific implementation step is as follows:
The D2.1 first class catalogues entitled time then;
D2.2 constructs second-level directory, the moon in second-level directory entitled higher level's directory name+this month under first class catalogue, according to month
Part;
D2.3 by that analogy, under every first class catalogue after two level, establishes next stage catalogue;
D2.4 afterbodies directory name is " web camera identification number ", is represented in certain year in such a month, and on such a day a certain point of some time
Zhong Li, the catalogue where the frame of video picture of the web camera;
D3. picture file is stored under the afterbody of above-mentioned bibliographic structure, and file name is:YMDHMS-IPCID;
It is expressed as the frame of video picture accessed by M S minute, IPCID second video cameras during the Y M D month, H day;
E. converged services module is merged the multi-source data got and frame of video, and by the data storage after fusion
In the local database, specific implementation step is as follows:
E1. converged services module will get frame of video picture, multi-source data, be merged according to set strategy;Merge plan
Slightly include but is not limited to be merged according to geographical position correlation, specific implementation step is:Same inspection place will be located at for the moment
The data and the pathname of frame of video picture storage and filename write-in one that Temperature Humidity Sensor, the Big Dipper receiver at quarter obtain
Individual buffering area, as a fused data bag;
E2 is stored in database using fused data bag as a record;Described database is including but not limited to MYSQL
Database;
E3. converged services module establishes information index, including but not limited to the time of frame of video picture, geographical position,
The foundation such as characteristic attribute, label are indexed, and index information is stored in into database;
F. the inquiry request that service module receives client is inquired about, returns to the picture for the condition that meets;Specific implementation step is such as
Under:
F1. inquire about service module and open nginx services;
F2. client sends querying condition and gives inquiry service module, the inquiry bar that can be inputted at client query interface
Part includes but is not limited to period, geographical position and other attribute conditions;
F3. inquire about service module and receive the inquiry request from client, obtained and corresponded in database according to querying condition
Path is locally stored, and the path is sent to client;
F4. after client receives path, frame of video picture is asked to inquiry service module according to the path;
F5. inquire about service module and receive the frame of video picture request from client, read institute from hard disk according to path
The frame of video picture of request, is sent to client;
F6. client receives the frame of video picture that inquiry service module is sent, and saves it under local directory, and show
On interface.
Brief description of the drawings
Fig. 1:A kind of flow chart for perceiving source information fusion video method more;
Fig. 2:Overall system framework figure;
Fig. 3:Query function timing diagram.
Embodiment
Below in conjunction with the accompanying drawings, one is entered to the present invention by illustrating that the multisource video monitoring system of certain unit is used as embodiment
Step description.
Specific implementation step is as follows:
1. according to the hardware device (such as Fig. 2) in system architecture deployment system, specific implementation step is as follows:
1.1 web cameras, network hard disk video recorder, Temperature Humidity Sensor, converged services device, client, pass through exchange
Machine is connected in a LAN;The IP of 4 web cameras is respectively 192.168.69.233,192.168.69.236,
192.168.69.237 192.168.69.242, identification number is respectively IPCID1, IPCID2, IPCID3, IPCID4.Network is hard
The IP of disk video recorder is that the IP of 192.168.69.231 converged services devices is 192.168.69.249, and client ip is
192.168.74.121;
Two modules are run on 1.2 converged services devices:Converged services module is responsible for obtaining frame of video picture and multi-source data,
And fusion function;Inquiry service module is responsible for the request of customer in response end;
Information is issued converged services by 1.3 Big Dipper receivers from big-dipper satellite receive information by converged services device serial ports
Module;
1.4 Temperature Humidity Sensors obtain the information such as humiture from environment, issue information by converged services device serial ports and melt
Close service module;
1.5 inquiry service modules receive client request, and return to regarding for request by http protocol and client communication
Frequency frame picture and the data inquired about;
2. converged services module obtains video flowing and frame of video picture from network hard disk video recorder;Specific implementation step is such as
Under:
2.1 converged services modules ask four road video flowings to network hard disk video recorder;A son is separately turned on per road video
Thread, for intercepting frame of video picture;
After 2.2 network hard disk video recorders receive request, video flowing is sent to converged services module, video stream format uses
H264 forms;
2.3 converged services modules decode it after receiving video flowing, start a timer in converged services module, often
Every the operation of 1 second activation extraction frame of video picture;
3. converged services module receives Big Dipper data from Big Dipper receiver, humiture information is obtained from Temperature Humidity Sensor;
3.1 Big Dipper receivers obtain time DateTime local on big-dipper satellite, longitude longitude, latitude
The information such as Latitude, height above sea level Height, converged services module is transferred data to 1HZ frequency by serial ports;
3.2 Temperature Humidity Sensors obtain the information such as temperature temperature, humidity humidity in environment, pass through string
Mouth transfers data to converged services module with 1HZ frequency;
4. converged services module, which is frame of video picture, generates pathname and filename, and preserves picture, specific implementation step
It is as follows:
4.1 times for assuming to transmit from Big Dipper receiver were 1 day 9 January in 2017:10:11, network hard disk video recorder transmits
The frame of video picture of interception is stored under local directory by four road video flowings, converged services module;
4.2 by " year-month-day-when-point-web camera identification number " extend step by step in the way of construct pathname, path
Name is as follows:
E:\\2017\201701\20170101\2017010109\201701010910\IPCID1\
E:\\2017\201701\20170101\2017010109\201701010910\IPCID2\
E:\\2017\201701\20170101\2017010109\201701010910\IPCID3\
E:\\2017\201701\20170101\2017010109\201701010910\IPCID4\
4.3 four IPC now obtained picture file is stored under four above-mentioned catalogues respectively, according to " YMDHMS-
IPCID " form construction filename, filename are as follows:
20170101091011-IPCID1.jpg
20170101091011-IPCID2.jpg
20170101091011-IPCID3.jpg
20170101091011-IPCID4.jpg
What 5. the data such as the time that converged services module transmits the Big Dipper, longitude and latitude, height above sea level and Temperature Humidity Sensor transmitted
In the routing information of the data such as temperature, humidity and picture deposit database db_M tb_IPC tables, four records are respectively:
6. inquiring about the inquiry request that service module receives client, the picture for the condition that meets is returned to, specific implementation step is such as
Under:
6.1 inquiry service modules open nginx services;
In query interface input inquiry condition, client sends querying condition and gives inquiry service module 6.2 users;
6.2.1 user is 1 day 00 January in 2017 in query interface input inquiry condition time:00:00 web camera mark
Know the picture for IPCID1;
6.2.2 inquiry request is sent to inquiry service module by client by http protocol;
Path is locally stored corresponding to being obtained according to querying condition in database in 6.3 inquiry service modules, and by the path
It is sent to client;
6.3.1 after inquiry service module receives client request, obtained according to querying condition in database and road is locally stored
Footpath, query statement are " select Path from tb_BeidouIPC where DateTime=' 20,170,101 00:00:
00 ' andIPCID=' IPCID1 ' ",
6.3.2 inquiring about the store path that service module inquires is
E:\\2017\201701\20170101\2017010100\201701010000\201701010000-IPCID1\
201701010000-IPCID1\20170101000000-IPCID1.jpg,;
6.3.3 the path is sent to client by inquiry service module;
After 6.4 clients receive path, frame of video picture is asked to inquiry service module according to the path;
6.5 clients receive the frame of video picture that inquiry service module is sent, and save it under local directory, and show
On interface;
It is finally noted that the purpose for publicizing and implementing example is that help further understands the present invention, but this area
Technical staff be appreciated that:Without departing from the spirit and scope of the invention and the appended claims, it is various to replace and repair
It is all possible for changing.Therefore, the present invention should not be limited to embodiment disclosure of that, and the scope of protection of present invention is to weigh
The scope that sharp claim defines is defined.
Claims (3)
1. a kind of source information that perceives merges video method more, its step includes:
A. it is as follows according to the hardware device in system architecture deployment system, specific implementation step:
A1. described hardware device includes:Web camera, network hard disk video recorder, Temperature Humidity Sensor, Big Dipper receiver,
Converged services device, client, it is connected to by interchanger in a LAN;
A2. two modules are run on converged services device:Converged services module is responsible for obtaining frame of video picture and multi-source data, and melts
Close function;Inquiry service module is responsible for the request of customer in response end;
A3. Big Dipper receiver is connected with converged services device serial ports, from big-dipper satellite receive information, and issues converged services mould
Block;
A4. Temperature Humidity Sensor device is connected with converged services device serial ports, obtains ambient temperature and humidity data, and issue converged services
Module;
A5. client includes but is not limited to PC computers, smart mobile phone, iPAD equipment, user's request is received, from converged services device
Obtain information needed and video;
A5. converged services module by unified data access interface and different pieces of information sources traffic and obtains data;
A6. service module is inquired about by http protocol and client communication, receives client request, and return to the frame of video of request
Picture and the information inquired about;
B. converged services module obtains video flowing and frame of video picture from network hard disk video recorder;Specific implementation step is as follows:
B1. after converged services module starts, video flowing is asked to network hard disk video recorder;
B2. after network hard disk video recorder receives request, video flowing is sent to converged services module;
B3. converged services module is decoded it after receiving video flowing, and time driving mechanism or event-driven are used per road video
Mechanism extracts frame of video picture in decoded video flowing;
C. the timing of converged services module obtains multi-source data by data access interface, and specific implementation step is as follows:
C1. described data source includes but is not limited to Big Dipper receiver, Temperature Humidity Sensor;
C2. data are transmitted after big-dipper satellite receives Big Dipper data and give converged services module by Big Dipper receiver;Number
According to including but not limited to Big Dipper time, longitude and latitude, height above sea level etc.;
C3. converged services module obtains the data of packing according to time driving mechanism by data access interface from serial ports;
C4. data are transmitted after environment gets data of the Temperature and Humidity module and give converged services module by Temperature Humidity Sensor device;
Data include temperature, humidity etc.;
C5. converged services module obtains mobile phone relevant information by data access interface according to time driving mechanism from database;
D. converged services module is that frame of video picture generates pathname and filename, and preserves picture, and specific implementation step is as follows:
D1. the step C2 Big Dipper times obtained and web camera mark generation pathname and filename are utilized;
D2. pathname describes the storage catalogue of frame of video picture, and its structure is divided into 6 grades:" year-month-day-when-point-network shooting
Machine identification name ";Picture storage catalogue constructs in the way of extending step by step;
D3. picture file is stored under the afterbody of above-mentioned bibliographic structure, and file name is:YMDHMS-IPCID;Represent
For the Y M D month, H day when M S minute, IPCID second video cameras accessed by frame of video picture;
E. converged services module is merged the multi-source data got and frame of video, and by the data storage after fusion at this
In ground database, specific implementation step is as follows:
E1. converged services module will get frame of video picture, multi-source data, be merged according to set strategy;Convergence strategy bag
Include but be not limited to be merged according to geographical position correlation, specific implementation step is:By positioned at same inspection place synchronization
Data, the pathname of frame of video picture storage and the filename that Temperature Humidity Sensor, Big Dipper receiver obtain write one and delayed
Area is rushed, as a fused data bag;
E2 is stored in database using fused data bag as a record;Described database is including but not limited to MYSQL data
Storehouse;
E3. converged services module establishes information index, including but not limited to the time of frame of video picture, geographical position, feature
The foundation such as attribute, label are indexed, and index information is stored in into database;
F. the inquiry request that service module receives client is inquired about, returns to the picture for the condition that meets;Specific implementation step is as follows:
F1. inquire about service module and open nginx services;
F2. client sends querying condition and gives inquiry service module, the querying condition bag that can be inputted at client query interface
Include but be not limited to the period, geographical position and other attribute conditions;
F3. inquire about service module and receive the inquiry request from client, according to corresponding to querying condition in database acquisition originally
Ground store path, and the path is sent to client;
F4. after client receives path, frame of video picture is asked to inquiry service module according to the path;
F5. inquire about service module and receive the frame of video picture request from client, read and asked from hard disk according to path
Frame of video picture, be sent to client;
F6. client receives the frame of video picture that inquiry service module is sent, and saves it under local directory, and be shown in boundary
On face.
2. the source information that perceives as claimed in claim 1 merges video method more, it is characterised in that converged services module, which receives, to be regarded
Frequency decodes it after flowing, and is extracted per road video using time driving mechanism or event-driven mechanism in decoded video flowing
Frame of video picture, is comprised the following steps that:
Time driving mechanism described in B3.1 refers to, and starts timer in converged services module, (or is specified every 1 second
Other times interval) activate the operation for extracting frame of video picture;
Event-driven mechanism described in B3.2 refers to, and is passed in web camera, network hard disk video recorder, Big Dipper receiver, humiture
When specific event occurs in sensor, converged services device, the operation of activation extraction frame of video picture.
3. the source information that perceives as claimed in claim 1 merges video method more, it is characterised in that pathname describes frame of video figure
The storage catalogue of piece, its structure are divided into 6 grades:" year-month-day-when-point-web camera identification name ";Picture storage catalogue according to
The mode extended step by step constructs, and comprises the following steps that:
The D2.1 first class catalogues entitled time then;
D2.2 constructs second-level directory, the month in second-level directory entitled higher level's directory name+this month under first class catalogue, according to month;
D2.3 by that analogy, under every first class catalogue after two level, establishes next stage catalogue;
D2.4 afterbodies directory name is " web camera identification number ", is represented in such a month, and on such a day a certain minute in some time in certain year
In, the catalogue where the frame of video picture of the web camera.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710961702.XA CN107749963A (en) | 2017-10-17 | 2017-10-17 | A kind of source information that perceives merges video method more |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710961702.XA CN107749963A (en) | 2017-10-17 | 2017-10-17 | A kind of source information that perceives merges video method more |
Publications (1)
Publication Number | Publication Date |
---|---|
CN107749963A true CN107749963A (en) | 2018-03-02 |
Family
ID=61253845
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201710961702.XA Pending CN107749963A (en) | 2017-10-17 | 2017-10-17 | A kind of source information that perceives merges video method more |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN107749963A (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111983367A (en) * | 2020-08-27 | 2020-11-24 | 西安苏试广博环境可靠性实验室有限公司 | Novel high-low temperature-low pressure screening test method and system |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103402044A (en) * | 2013-08-07 | 2013-11-20 | 重庆大学 | Target recognition and tracking system based on multi-source video integration |
CN104899261A (en) * | 2015-05-20 | 2015-09-09 | 杜晓通 | Device and method for constructing structured video image information |
CN204795392U (en) * | 2015-06-26 | 2015-11-18 | 山东大学 | Found equipment of different structure information structure ization of special equipment |
US20150380055A1 (en) * | 2012-09-12 | 2015-12-31 | Intel Corporation | Techniques for indexing video files |
CN105493502A (en) * | 2015-04-29 | 2016-04-13 | 北京旷视科技有限公司 | Video monitoring method, video monitoring system, and computer program product |
CN105630897A (en) * | 2015-12-18 | 2016-06-01 | 武汉大学 | Content-aware geographic video multilayer correlation method |
CN106162193A (en) * | 2015-04-08 | 2016-11-23 | 广东中星电子有限公司 | A kind of video data and the coded method of sensing data and device |
CN106303538A (en) * | 2016-08-16 | 2017-01-04 | 北京中星微电子有限公司 | A kind of video spatial scalable coded method supporting multisource data fusion and framework |
-
2017
- 2017-10-17 CN CN201710961702.XA patent/CN107749963A/en active Pending
Patent Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20150380055A1 (en) * | 2012-09-12 | 2015-12-31 | Intel Corporation | Techniques for indexing video files |
CN103402044A (en) * | 2013-08-07 | 2013-11-20 | 重庆大学 | Target recognition and tracking system based on multi-source video integration |
CN106162193A (en) * | 2015-04-08 | 2016-11-23 | 广东中星电子有限公司 | A kind of video data and the coded method of sensing data and device |
CN105493502A (en) * | 2015-04-29 | 2016-04-13 | 北京旷视科技有限公司 | Video monitoring method, video monitoring system, and computer program product |
CN104899261A (en) * | 2015-05-20 | 2015-09-09 | 杜晓通 | Device and method for constructing structured video image information |
CN204795392U (en) * | 2015-06-26 | 2015-11-18 | 山东大学 | Found equipment of different structure information structure ization of special equipment |
CN105630897A (en) * | 2015-12-18 | 2016-06-01 | 武汉大学 | Content-aware geographic video multilayer correlation method |
CN106303538A (en) * | 2016-08-16 | 2017-01-04 | 北京中星微电子有限公司 | A kind of video spatial scalable coded method supporting multisource data fusion and framework |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111983367A (en) * | 2020-08-27 | 2020-11-24 | 西安苏试广博环境可靠性实验室有限公司 | Novel high-low temperature-low pressure screening test method and system |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11328163B2 (en) | Methods and apparatus for automated surveillance systems | |
US10955586B2 (en) | Weather forecasting system and methods | |
US9740940B2 (en) | Event triggered location based participatory surveillance | |
US10399650B2 (en) | System for monitoring marine vessels and determining rendezvouses therebetween and related methods | |
Geraldes et al. | UAV-based situational awareness system using deep learning | |
US11367346B2 (en) | Digitizing and mapping the public space using collaborative networks of mobile agents and cloud nodes | |
CA2949353C (en) | Computer-implemented systems and methods of analyzing data in an ad-hoc network for predictive decision-making | |
Räty | Survey on contemporary remote surveillance systems for public safety | |
US10302769B2 (en) | System for monitoring marine vessels using fractal processing of aerial imagery and related methods | |
CA2824330C (en) | An integrated intelligent server based system and method/systems adapted to facilitate fail-safe integration and/or optimized utilization of various sensory inputs | |
CN102752574A (en) | Video monitoring system and method | |
US20180205444A1 (en) | System for monitoring marine vessels providing expected passenger determination features and related methods | |
US11727317B2 (en) | Systems and methods for coherent monitoring | |
US20230412769A1 (en) | Scalable Visual Computing System | |
CN109905423B (en) | Intelligent management system | |
Alabdulkarim et al. | Urban analytics in crowd management in the context of Hajj | |
CN117319609A (en) | Internet of things big data intelligent video monitoring system and method | |
CN102387346B (en) | Intelligent front end of manageable, findable and inspectable monitoring system | |
CN107749963A (en) | A kind of source information that perceives merges video method more | |
Miloudi et al. | Leveraging the power of integrated solutions of IoT and GIS | |
Gialampoukidis et al. | Multimodal data fusion of social media and satellite images for emergency response and decision-making | |
Li | A suvey on edge intelligent video surveillance with deep reinforcement learning | |
Liu et al. | A threefold similarity analysis of crowdsourcing feeds | |
Zhu et al. | Application of Video Surveillance Intelligent Analysis System Based on KNN Algorithm | |
Vishwakarma et al. | Empowering Smart Cities: A Comprehensive Edge Computing Framework for Enhanced IoT Situation Awareness |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
WD01 | Invention patent application deemed withdrawn after publication | ||
WD01 | Invention patent application deemed withdrawn after publication |
Application publication date: 20180302 |