WO2014092552A2 - Method for non-static foreground feature extraction and classification - Google Patents
Method for non-static foreground feature extraction and classification Download PDFInfo
- Publication number
- WO2014092552A2 WO2014092552A2 PCT/MY2013/000266 MY2013000266W WO2014092552A2 WO 2014092552 A2 WO2014092552 A2 WO 2014092552A2 MY 2013000266 W MY2013000266 W MY 2013000266W WO 2014092552 A2 WO2014092552 A2 WO 2014092552A2
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- clouds
- cloud
- feature points
- disparity
- frame
- Prior art date
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/20—Analysis of motion
- G06T7/292—Multi-camera tracking
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/194—Segmentation; Edge detection involving foreground-background segmentation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/20—Analysis of motion
- G06T7/215—Motion-based segmentation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/20—Analysis of motion
- G06T7/246—Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/20—Analysis of motion
- G06T7/285—Analysis of motion using a sequence of stereo image pairs
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10016—Video; Image sequence
- G06T2207/10021—Stereoscopic video; Stereoscopic image sequence
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30232—Surveillance
Definitions
- the present invention relates to image processing. More particularly, the present invention relates to a method for extracting and classifying non-static foreground features.
- US patent no. 7,729,512 provides a stereo image processing system and method for detecting moving objects.
- disparity threshold is used for classifying the feature points. This system require extensive searching and computing of all the individual the feature points and their disparity infomiation, which may lead to high processing power needed for real-time tracking.
- US Patent publication no. 2011/0134221 too suggests a system and method of processing stereo images for extracting moving object. Similarly, feature vectors are being extracted and matched to detect moving object. [0005] It is recognized that feature points/vectors are widely used for detecting moving object from a video stream. However, it is a challenge to provide a reliable system and method to process these feature points in an effective and efficient way to extract the moving object. It is also desired that the system and method are adapted with adaptive capability to refine the detection results in real time.
- a surveillance system having a pair of stereo camera capturing videos having a plurality of frames, each frames includes at least a left and a right images, wherein the stereo camera having an accelerometer attached thereto.
- the surveillance system comprises a video acquisition module for acquiring the videos captured by the pair of stereo cameras and acceleration information of the stereo camera; a feature point disparity computational unit adapted for extracting feature points in right and left images of a current frame and matching the feature points between the right and left images, wherein disparity value of each feature points is computed; a feature point tracking unit adapted for clustering the feature points on a current frame into one or more cloud based on their appearance and distance, and determining a correspondence between the clouds of the current frame and a previous frame, a feature point classification unit for classifying the clouds through estimating the statistical disparity distribution of each cloud to identify if the respective clouds are new clouds, existing clouds or unknown clouds; and an object extraction unit for merging the clouds that are determined to be a same object based on the tracking information and changes of disparity of the clouds over time.
- the surveillance system may further comprise an event rules unit; a display module for displaying overplayed information on live video; and a post detection module to trigger post event actions defined by the event rules unit.
- an event rules unit for processing videos captured through a stereo camera, the videos having a plurality of frames, each frames includes at least a left and a right images, wherein the stereo camera having an accelerometer attached thereto.
- the method comprises acquiring videos captured by the pair of stereo cameras and acceleration information of the stereo camera detected through the accelerometer; extracting feature points in the right and left images of a current frame; matching the feature points between the right and left images, wherein disparity value of each feature point is computed; clustering the feature points on a current frame into one or more clouds based on their appearance and distance; determining a correspondence between the clouds of the current frame and previous frame; classifying the clouds through estimating statistical disparity distribution of each cloud to identify if the respective clouds are new clouds, existing clouds or unknown clouds; and merging the clouds that are determined to be a same object based on the tracking information and changes of disparity of the clouds over time.
- the method further comprises filtering noises while matching the feature points between the left and the right images; computing distances between the matched features points; and quantizing the distance based on a current disparity value range. It is also possible that matching the feature points between the left and right image comprising searching area for fining a best matched feature point based on matching feature points detennined in the previous frame.
- searching area for finding the best matched feature point may comprise matching feature points within an entire image of a first frame; and matching feature points for all subsequent frames based on areas corresponding to feature points matched in a previous frame.
- clustering the feature points into the one or more clouds follows by the steps of tracking each cloud by determining the correspondences between clouds detected in current frame and in previous frames; determining a tracking status of each cloud detected in current frame, wherein the tracking status include new object, existing moving object, existing background and existing undetermined object; performing feature points matching between one of the left and right images of the current frame and previous frame; and redistribute the feature point clouds into different or same cloud based on the matching information.
- classifying feature point may further comprise estimating a statistical disparity distribution of each feature point cloud through mean, mpd or standard deviation; Storing the information when the cloud is determined as a new cloud; computing the change of disparity of each clouds that are tracked between a current frame and a previous frame; determining the acceleration information of the stereo camera for the instance frame.
- the cloud is determined to be a moving object when the rate of change is not substantially zero, and the cloud is determined to be a static object or background when the rate of change is substantially zero; and when the acceleration is not substantially zero, the rate of disparity change between a current frame and a previous frame is computed, and the cloud is determined to be a static object or background if the rate of change is increasing/decreasing, and the cloud is determined to be a moving object when the rate of disparity change is substantially zero.
- FIG. 1 illustrates a schematic diagram a surveillance system in accordance with one embodiment of the present invention
- FIG. 2 illustrates the process flow operationally carries out by a feature point disparity computational unit in accordance with one embodiment of the present invention
- FIG. 3 illustrates the process flow operationally carries out by a feature point tracking unit in accordance with one embodiment of the present invention
- FIG. 4 illustrates the process flow of a feature point classification unit in accordance with one embodiment of the present invention.
- FIG. 5 illustrates the process flow carries out by the moving object extraction unit in accordance with one embodiment of the present invention.
- the objective of this invention is to extract foreground feature points captured by a pair of non-static cameras and classify them into a moving pixel or static / background pixel.
- the resultant output can be used to delineate moving objects from background for further processing or event analysis.
- the feature point extraction and classification for non-static cameras is difficult as the computer vision is unable to differentiate the movement on the feature point either it is caused by the actual moving object or by the motion of the cameras.
- proposed invention is to solve the problem on how to differentiate between moving object and background when everything in the scene is non-static between the frames.
- FIG. 1 illustrates a block diagram of a surveillance system 100 in accordance with one embodiment of the present invention.
- the surveillance system 100 is adapted for processing surveillance videos captured through a pair of surveillance cameras 101.
- the pair of cameras is adapted for capturing stereo images/videos operationally.
- the pair of cameras shall herewith below refer to as stereo cameras.
- stereo camera can be a commercially available stereo camera or a camera assembly that comprise two suitable videos that are installed on a platform, or frames.
- the surveillance cameras 101 equipped or attached with an accelerometer 102 for detecting the movements of the respective surveillance cameras 101b.
- the surveillance system 100 comprises a video acquisition module 103, database 112, an image-processing module 104, a display module 110, and a post detection module 111.
- the video acquisition module 103 is connected to the pair of cameras 101 to acquire images or videos captured through the cameras.
- the acceleration information concerning the plurality of cameras 101 recorded through the accelerometers 102 is also transmitted to the video acquisition module 103.
- the captured videos/images and the acceleration information are stored on the database module 112, which are retrievable by the imaging-processing module 104 for processing anytime.
- the image-processing module 104 responsible for processing the input videos and acceleration information transmitted from the video acquisition module 103, and analyzing the same to extract and detects moving feature points from the images/videos. Once the feature points are being detected and extracted, they are output to the display module 110 as visual outputs, and to the post detection module 111 for triggering further processing or control as necessary .
- the image-processing module 104 further comprises a feature point disparity computational unit 105, a feature point tracking unit 106, a feature point classification unit 107, a moving object extraction unit 108, and an event rule unit 109.
- the feature point disparity computational unit 105 performs feature point matching between right and left images of the stereo videos at any instance of time, and subsequently estimating the disparity values for each of the matched points.
- the disparity values are used to track and classify the matched points through the feature point-tracking unit 106.
- the feature point tracking unit 106 tracks the matched feature points by processing subsequent stereo image frames to extract the movement information of each feature point.
- the feature points are tracked in a group of similar properties of the features.
- some feature points may also be represented by the same or substantially the same feature vector.
- the feature vector may be appearance such as e.g. color, edges etc. or the like. It can be challenging to find the correct correspondences between the multiple feature points having same or substantially the same feature vector, as opposed to finding the correspondence between the groups of feature points between two video frames.
- the group of feature points that shares similar properties shall herein referred to as "the cloud”. As the cloud of feature points are being tracked, instead of the individual feature point, the computational processing needed by the image processing module 104 and the information stored over time can be reduced.
- the internal intrinsic parameters of the stereo camera or the camera assembly are known. These internal intrinsic parameters provide distance between the two cameras (or imaging sensors) and the rotation of one camera with respect to the other.
- the two cameras are also horizontally aligned with a 3D point projected in a relatively same horizontal line on both images.
- the matching can be done by assuming that any point in one line in one image is correspond to one point in same corresponding line in the other image. Then with these point corresponding in hand, together with the intrinsic parameter known earlier, disparity value of this 3D point from the camera can be computed.
- each feature point is being classified as either moving paints or background points through the feature point classification unit 107.
- the classification is made up based on the information from the feature point tracking unit 106.
- the information includes tracking and disparity information. It is desired that the classification be done on each cloud.
- the moving object extraction unit 108 further group one or more clouds into a moving entity based on the feature point's classification result, tracking information and the disparity information.
- the event rule unit 109 applies an appropriate rule based on the extraction results.
- the event rule unit 109 applies an intrusion event rule when the moving object is detected to be an intruder.
- These rules are predefined and they may be stored in the database 112 for retrieval.
- the image-processing module 104 then output the video processing results to the display module 110 for visual output and when appropriate, the post detection module 111 is triggered to activate a corresponding alarm based on the processed results.
- Feature points that have almost the same or similar appearance to each other will be group as one cloud. For example, all the feature points that are extracted from the subject's shirt are presumably having the same or similar appearances and will therefore be group as one cloud. The appearance that includes color would have certain level of similarity.
- all feature points extracted from the pant part shall be group as the same cloud, which is a separate cloud from the cloud for the shirt.
- the motion of both clouds may form specific corresponding patterns and therefore can further be grouped into one cloud defining the subject.
- FIG. 2 illustrates the process flow operationally carries out by the feature point disparity computational unit 105 of FIG. 1 in accordance with one embodiment of the present invention.
- the feature point disparity unit 105 derives the disparity information of feature points between the two stereo (i.e. left and right images) of the current frame.
- the process flow includes extracting feature points in cun-ently left and right images in step 201; matching feature points in cunent left and right images in step 202; filtering noises in step 203; computing distance between the respective matched feature points in step 204; and quantizing the distance based on cuirent disparity value in step 205.
- the process starts with extracting feature points light and left images.
- salient points are extracted as the feature points in the images.
- These salient points can be corner and edges points that can be detected through any known techniques. For example, Harris corner point detector, scale-invariant feature transform (SIFT), speeded up robust feature (SURF) etc. can be utilized for extracting the desired salient points.
- Harris corner point detector scale-invariant feature transform (SIFT), speeded up robust feature (SURF) etc.
- SIFT scale-invariant feature transform
- SURF speeded up robust feature
- the feature points i.e. the salient points
- they are matched between right and left images of the same frame in the step 202.
- searching and matching can be simplified to within a single or one dimensional (ID) space because the horizontal line in one image can be presumed to be in parallel to the same horizontal line in the other image.
- searching for matching or corresponding feature points may requires searching in two-dimensional (2D) space, as two imaging means may not be perfectly aligned. It is more so when the imaging means are two independent cameras.
- One possible approach for example is to perform image rectification between right and left cameras before matching. Such image rectification for processing stereo images is well known in the art.
- Another desired approach that can be taken is to adaptively refine the 2D searching area for matching, of which no rectification is required. In this adaptive refinery process, previous matching information is used to determine the search area in the current frame.
- the each point on the whole horizontal lines or rows of the right image of the first frame will be considered as the searching area.
- the results will be analyzed for range of possible matching horizontal lines on the right image for searching through, given a point on certain horizontal line in left image is determined. Accordingly, when the next frame is to be searched, the searching area will be constraint within these ranges. That could effectively refine the number of horizontal lines to be search over time.
- the distance or disparity value of these matched feature points is computed.
- the distance or disparity value provides a relative distance of these points in the real world with respect to focal point of the surveillance camera.
- the distance or disparity values are quantized into discrete levels. Accordingly, the feature points extracted from the right/left image of the current frame can be outputted by the feature point disparity computational unit 105 along with the respective disparity values.
- FIG. 3 illustrates the process flow operationally carries out by the feature point tracking unit 106 of FIG. 1 in accordance with one embodiment of the present invention.
- the feature point tracking unit 106 receives input from the feature point disparity computational unit 105.
- the input includes either one of the stereo image (i.e. left or right image) of the current frame and previous frames, the extracted feature points from current and previous frames, and previous tracking information of feature points in previous frames.
- the output from the feature point tracking unit 106 provides the correspondences between feature points in current frame and previous frames.
- the process includes clustering the feature points in the current frame into clouds in step 301; tracking each cloud in step 302; determining tracking status of the clouds in step 303; matching feature point between the right/left image of the current frame and previous frame in step 304; and redistributing the feature points into the cloud in step 305.
- the feature point tracking unit 106 primarily forms and group the feature points into clouds and track the same accordingly.
- the process starts with clustering the feature points in current frame into clouds in the step 301.
- the clustering is done based on the similarity in appearance and proximity within the image of the current frame.
- the feature point tracking unit 106 tracks the clouds by determining correspondences of all the resultant clouds in current frame with respect to previous frame in the step 302.
- the correspondences are determined based on similarity of the clouds with reference to the appearance and shape of the feature point distribution.
- Various correspondences determination methods are widely available, and they can be desired for the present invention. Therefore, no f irther detail is provided herewith.
- a tracking status is being identified for each cloud.
- the tracking status includes new cloud, existing moving cloud, existing background cloud and existing unknown cloud.
- Current and previous clouds i.e. from current and previous frames, are referring to clouds in current and previous frame respectively.
- the feature point distributions i.e. into clouds
- the matching results that defines which points in the current frames is matched with the points in the previous frame can be used to redistribute feature point initially assigned to one cloud to another cloud in the step 305.
- FIG. 4 illustrates the process flow of the feature point classification unit
- the process is adapted for classifying each cloud detected by the feature point tracking unit 106. It classifies the clouds into moving objects, background/static objects or unknown objects.
- the feature point classification unit 107 receives inputs such as the clouds, tracking information and disparity information from the feature point tracking unit 106 for classification.
- the clouds are classified based on the disparity tracking information of each cloud. Assuming that the camera is moving in constant speed, then the rate of change of the cloud disparity information over subsequent frames would be substantially constant as well. For the cloud that is new, it will not be classified in current frame.
- the global disparity information for this cloud will be stored.
- the statistical information such as the mean or mod or the standard deviation of the disparity distribution with a cloud can be used to represent the global disparity information of the cloud.
- the process flow starts with estimating a statistical disparity distribution of each cloud (i.e. mean, mpd, standard deviation etc.) in current frame in step 401. Then each of the cloud will be examined whether it is new in step 402. If the cloud is new, the disparity information will be stored in step 403 and the process ends accordingly. Returning to the step 402, if the cloud is determined not being a new cloud, the cloud is further determined whether it is an existing cloud or an unknown cloud in step 404. If it is an existing cloud at the step 404, the feature point classification unit 107 computes the change in disparity of each cloud between the current and previous frame in step 405.
- each cloud i.e. mean, mpd, standard deviation etc.
- the feature point classification unit 107 further retrieves the acceleration information of the surveillance camera recorded by the accelerometer in step 406. If the acceleration is zero or substantially zero in the step 407, it provides that the surveillance camera is moving at a constant speed.
- the rate of change of disparity information of each cloud between the current frame and subsequent frames is computed in step 408. At least three frames would be required to compute the rate of change. This computation takes into consideration the changes of disparity information tracked over subsequent frames. If the rate of change is determined to be zero or substantially zero in the step 409, it can conclude that the cloud belongs to background or static object in step 413. If the rate of change is determined to be non-zero in the step 409, it an then conclude that the cloud is a moving object in step 410.
- the surveillance camera is determined to be accelerating, i.e. not zero acceleration, it can either be accelerating or decelerating.
- the feature point classification unit 107 further computes the rate of disparity change over subsequent frames in the step 411.
- the rate of disparity change of the cloud over the subsequent frames is in conformance with the movement of the surveillance camera in step 412, the cloud shall be classified as background or static object in the step 413. Otherwise, it is classified as moving object in the step 410.
- FIG. 5 illustrates the process flow carries out by the moving object extraction unit 108 of FIG. 1 in accordance with one embodiment of the present invention.
- the moving objection extraction unit 108 receives the clouds and the corresponding classifications from the feature point classification unit 107.
- step 501 it merges all the clouds determined belonging to the same object based on the tracking information and changes of disparity collected over certain time period.
- the processing units are able to effectively determine moving objects from a stereo surveillance camera, and track the same accordingly.
- one object may be defined by two or more independent clouds when the video is processed initially.
- a human who is weaiing red shirt and black pant may be defined by two different clouds since the object comprises two groups of different appearances.
- the two clouds After tracking these clouds over subsequent frame, it is expected that the two clouds that have similar tracking information, e.g. moving path, relative position between the two clouds over subsequent names and tracking disparity information.
- tracking information e.g. moving path
- the output provides the group of feature point clouds, which belong to the same moving object.
Landscapes
- Engineering & Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Multimedia (AREA)
- Image Analysis (AREA)
Abstract
The present invention provides a method for processing videos captured through a stereo camera, the videos having a plurality of frames, each frames includes at least a left and a right images. The method includes acquiring videos captured by the pair of stereo cameras (101) and acceleration information of the stereo camera (101) detected through the accelerometer (102); extracting feature points in the right and left images of a current frame; matching the feature points between the right and left images, wherein disparity value of each feature point is computed; clustering the feature points on a current frame into one or more clouds based on their appearance and distance; determining a correspondence between the clouds of the current frame and previous frame; classifying the clouds through estimating statistical disparity distribution of each cloud to identify if the respective clouds are new clouds, existing clouds or unknown clouds; and merging the clouds that are determined to be a same object based on the tracking information and changes of disparity of the clouds over time. A system carrying out the above method is also provided herein.
Description
Method For Non-Static Foreground Feature Extraction and
Classification
Field of the Invention
[0001] The present invention relates to image processing. More particularly, the present invention relates to a method for extracting and classifying non-static foreground features.
Background
[0002] The feature point extraction and classification for non-static cameras is difficult as the computer vision is unable to differentiate the movement on the feature point either it is caused by the actual moving object or by the motion of the cameras.
[0003] Many studies and research had been earned out extensively. US patent no. 7,729,512 provides a stereo image processing system and method for detecting moving objects. In this patent, disparity threshold is used for classifying the feature points. This system require extensive searching and computing of all the individual the feature points and their disparity infomiation, which may lead to high processing power needed for real-time tracking.
[0004] US Patent publication no. 2011/0134221 too suggests a system and method of processing stereo images for extracting moving object. Similarly, feature vectors are being extracted and matched to detect moving object. [0005] It is recognized that feature points/vectors are widely used for detecting moving object from a video stream. However, it is a challenge to provide a reliable
system and method to process these feature points in an effective and efficient way to extract the moving object. It is also desired that the system and method are adapted with adaptive capability to refine the detection results in real time.
Summary [0006] In one aspect of the present invention, there is provided a surveillance system having a pair of stereo camera capturing videos having a plurality of frames, each frames includes at least a left and a right images, wherein the stereo camera having an accelerometer attached thereto. The surveillance system comprises a video acquisition module for acquiring the videos captured by the pair of stereo cameras and acceleration information of the stereo camera; a feature point disparity computational unit adapted for extracting feature points in right and left images of a current frame and matching the feature points between the right and left images, wherein disparity value of each feature points is computed; a feature point tracking unit adapted for clustering the feature points on a current frame into one or more cloud based on their appearance and distance, and determining a correspondence between the clouds of the current frame and a previous frame, a feature point classification unit for classifying the clouds through estimating the statistical disparity distribution of each cloud to identify if the respective clouds are new clouds, existing clouds or unknown clouds; and an object extraction unit for merging the clouds that are determined to be a same object based on the tracking information and changes of disparity of the clouds over time.
[0007] In one embodiment, the surveillance system may further comprise an event rules unit; a display module for displaying overplayed information on live video; and a post detection module to trigger post event actions defined by the event rules unit.
[0008] In another aspect of the present invention, there is further provided a method for processing videos captured through a stereo camera, the videos having a plurality of frames, each frames includes at least a left and a right images, wherein the stereo camera having an accelerometer attached thereto. The method comprises acquiring videos captured by the pair of stereo cameras and acceleration information of the stereo camera detected through the accelerometer; extracting feature points in the right and left images of a current frame; matching the feature points between the right and left images, wherein disparity value of each feature point is computed; clustering the feature points on a current frame into one or more clouds based on their appearance and distance; determining a correspondence between the clouds of the current frame and previous frame; classifying the clouds through estimating statistical disparity distribution of each cloud to identify if the respective clouds are new clouds, existing clouds or unknown clouds; and merging the clouds that are determined to be a same object based on the tracking information and changes of disparity of the clouds over time.
[0009] In one embodiment, the method further comprises filtering noises while matching the feature points between the left and the right images; computing distances between the matched features points; and quantizing the distance based on a current disparity value range. It is also possible that matching the feature points between the left and right image comprising searching area for fining a best matched feature point based on matching feature points detennined in the previous frame.
[0010] In yet another embodiment, searching area for finding the best matched feature point may comprise matching feature points within an entire image of a first
frame; and matching feature points for all subsequent frames based on areas corresponding to feature points matched in a previous frame.
[0011] In yet a further embodiment, clustering the feature points into the one or more clouds follows by the steps of tracking each cloud by determining the correspondences between clouds detected in current frame and in previous frames; determining a tracking status of each cloud detected in current frame, wherein the tracking status include new object, existing moving object, existing background and existing undetermined object; performing feature points matching between one of the left and right images of the current frame and previous frame; and redistribute the feature point clouds into different or same cloud based on the matching information.
[0012] In a further embodiment, classifying feature point may further comprise estimating a statistical disparity distribution of each feature point cloud through mean, mpd or standard deviation; Storing the information when the cloud is determined as a new cloud; computing the change of disparity of each clouds that are tracked between a current frame and a previous frame; determining the acceleration information of the stereo camera for the instance frame. When the acceleration is substantially zero, the rate of disparity changes of each tracked cloud in the current frame over a subsequent frame is computed, the cloud is determined to be a moving object when the rate of change is not substantially zero, and the cloud is determined to be a static object or background when the rate of change is substantially zero; and when the acceleration is not substantially zero, the rate of disparity change between a current frame and a previous frame is computed, and the cloud is determined to be a static object or background if the rate of change is increasing/decreasing, and the cloud is determined to be a moving object when the rate of disparity change is substantially zero.
Brief Description of the Drawings
[0013] Preferred embodiments according to the present invention will now be described with reference to the figures accompanied herein, in which like reference numerals denote like elements; [0014] FIG. 1 illustrates a schematic diagram a surveillance system in accordance with one embodiment of the present invention;
[0015] FIG. 2 illustrates the process flow operationally carries out by a feature point disparity computational unit in accordance with one embodiment of the present invention; [0016] FIG. 3 illustrates the process flow operationally carries out by a feature point tracking unit in accordance with one embodiment of the present invention;
[0017] FIG. 4 illustrates the process flow of a feature point classification unit in accordance with one embodiment of the present invention; and
[0018] FIG. 5 illustrates the process flow carries out by the moving object extraction unit in accordance with one embodiment of the present invention.
Detailed Description
[0019] Embodiments of the present invention shall now be described in detail, with reference to the attached drawings. It is to be understood that no limitation of the scope of the invention is thereby intended, such alterations and further modifications in the illustrated device, and such further applications of the principles of the invention as
illustrated therein being contemplated as would normally occur to one skilled in the art to which the invention relates.
[0020] The objective of this invention is to extract foreground feature points captured by a pair of non-static cameras and classify them into a moving pixel or static / background pixel. The resultant output (sparse information) can be used to delineate moving objects from background for further processing or event analysis. The feature point extraction and classification for non-static cameras is difficult as the computer vision is unable to differentiate the movement on the feature point either it is caused by the actual moving object or by the motion of the cameras. Thus, proposed invention is to solve the problem on how to differentiate between moving object and background when everything in the scene is non-static between the frames.
[0021] FIG. 1 illustrates a block diagram of a surveillance system 100 in accordance with one embodiment of the present invention. The surveillance system 100 is adapted for processing surveillance videos captured through a pair of surveillance cameras 101. The pair of cameras is adapted for capturing stereo images/videos operationally. For the purpose of description below, the pair of cameras shall herewith below refer to as stereo cameras. Such stereo camera can be a commercially available stereo camera or a camera assembly that comprise two suitable videos that are installed on a platform, or frames. Preferably, the surveillance cameras 101 equipped or attached with an accelerometer 102 for detecting the movements of the respective surveillance cameras 101b. The surveillance system 100 comprises a video acquisition module 103, database 112, an image-processing module 104, a display module 110, and a post detection module 111. The video acquisition module 103 is connected to the pair of cameras 101 to acquire images or videos captured through the cameras. The
acceleration information concerning the plurality of cameras 101 recorded through the accelerometers 102 is also transmitted to the video acquisition module 103. The captured videos/images and the acceleration information are stored on the database module 112, which are retrievable by the imaging-processing module 104 for processing anytime. The image-processing module 104 responsible for processing the input videos and acceleration information transmitted from the video acquisition module 103, and analyzing the same to extract and detects moving feature points from the images/videos. Once the feature points are being detected and extracted, they are output to the display module 110 as visual outputs, and to the post detection module 111 for triggering further processing or control as necessary .
[0022] Still referring to FIG. 1, the image-processing module 104 further comprises a feature point disparity computational unit 105, a feature point tracking unit 106, a feature point classification unit 107, a moving object extraction unit 108, and an event rule unit 109. The feature point disparity computational unit 105 performs feature point matching between right and left images of the stereo videos at any instance of time, and subsequently estimating the disparity values for each of the matched points. The disparity values are used to track and classify the matched points through the feature point-tracking unit 106. The feature point tracking unit 106 tracks the matched feature points by processing subsequent stereo image frames to extract the movement information of each feature point. In accordance with the present embodiment, the feature points are tracked in a group of similar properties of the features. It is recognized that some feature points may also be represented by the same or substantially the same feature vector. The feature vector may be appearance such as e.g. color, edges etc. or the like. It can be challenging to find the correct
correspondences between the multiple feature points having same or substantially the same feature vector, as opposed to finding the correspondence between the groups of feature points between two video frames. For simplicity, the group of feature points that shares similar properties shall herein referred to as "the cloud". As the cloud of feature points are being tracked, instead of the individual feature point, the computational processing needed by the image processing module 104 and the information stored over time can be reduced.
[0023] When a stereo camera or a camera assembly that comprising two cameras disposed on a frame with a fix distance apart, the internal intrinsic parameters of the stereo camera or the camera assembly are known. These internal intrinsic parameters provide distance between the two cameras (or imaging sensors) and the rotation of one camera with respect to the other. Typically, the two cameras are also horizontally aligned with a 3D point projected in a relatively same horizontal line on both images. Thus, for any instant of two images taken by these two cameras, the matching can be done by assuming that any point in one line in one image is correspond to one point in same corresponding line in the other image. Then with these point corresponding in hand, together with the intrinsic parameter known earlier, disparity value of this 3D point from the camera can be computed.
[0024] Through the present invention however, it is cater for stereo camera or the camera assembly that is may not be properly calibrated or misaligned in the cause of operating the same. Accordingly, one line in an image may not always correspond to the same line in the other image. The proposed invention has therefore proposed the optimized matching process in light of the above.
[0025] Each feature point is being classified as either moving paints or background points through the feature point classification unit 107. The classification is made up based on the information from the feature point tracking unit 106. The information includes tracking and disparity information. It is desired that the classification be done on each cloud. The moving object extraction unit 108 further group one or more clouds into a moving entity based on the feature point's classification result, tracking information and the disparity information. Accordingly, the event rule unit 109 applies an appropriate rule based on the extraction results. For example, the event rule unit 109 applies an intrusion event rule when the moving object is detected to be an intruder. These rules are predefined and they may be stored in the database 112 for retrieval. The image-processing module 104 then output the video processing results to the display module 110 for visual output and when appropriate, the post detection module 111 is triggered to activate a corresponding alarm based on the processed results. [0026] Feature points that have almost the same or similar appearance to each other will be group as one cloud. For example, all the feature points that are extracted from the subject's shirt are presumably having the same or similar appearances and will therefore be group as one cloud. The appearance that includes color would have certain level of similarity. In the case where the same subject containing more than one colors, for example, wearing a pant with a color different from the shirt, then all feature points extracted from the pant part shall be group as the same cloud, which is a separate cloud from the cloud for the shirt. After monitoring this subject, i.e. formed by the two detected cloud, over subsequent frames, the motion of both clouds may form specific
corresponding patterns and therefore can further be grouped into one cloud defining the subject.
[0027] FIG. 2 illustrates the process flow operationally carries out by the feature point disparity computational unit 105 of FIG. 1 in accordance with one embodiment of the present invention. The feature point disparity unit 105 derives the disparity information of feature points between the two stereo (i.e. left and right images) of the current frame. Briefly, the process flow includes extracting feature points in cun-ently left and right images in step 201; matching feature points in cunent left and right images in step 202; filtering noises in step 203; computing distance between the respective matched feature points in step 204; and quantizing the distance based on cuirent disparity value in step 205.
[0028] In the step 201, the process starts with extracting feature points light and left images. In one embodiment of the present invention, salient points are extracted as the feature points in the images. These salient points can be corner and edges points that can be detected through any known techniques. For example, Harris corner point detector, scale-invariant feature transform (SIFT), speeded up robust feature (SURF) etc. can be utilized for extracting the desired salient points.
[0029] Once the feature points (i.e. the salient points) are extracted, they are matched between right and left images of the same frame in the step 202. On a stereo camera where two imaging means are aligned to be coplanar, searching and matching can be simplified to within a single or one dimensional (ID) space because the horizontal line in one image can be presumed to be in parallel to the same horizontal line in the other image. In some camera configurations, searching for matching or
corresponding feature points may requires searching in two-dimensional (2D) space, as two imaging means may not be perfectly aligned. It is more so when the imaging means are two independent cameras. One possible approach for example is to perform image rectification between right and left cameras before matching. Such image rectification for processing stereo images is well known in the art. Another desired approach that can be taken is to adaptively refine the 2D searching area for matching, of which no rectification is required. In this adaptive refinery process, previous matching information is used to determine the search area in the current frame.
[0030] In one illustrative example for matching feature points on the right image that correspond to one feature point on the left image, the each point on the whole horizontal lines or rows of the right image of the first frame will be considered as the searching area. Once the matching is done for all feature points, the results will be analyzed for range of possible matching horizontal lines on the right image for searching through, given a point on certain horizontal line in left image is determined. Accordingly, when the next frame is to be searched, the searching area will be constraint within these ranges. That could effectively refine the number of horizontal lines to be search over time.
[0031] Once the feature points matching between the right and left images of the current frame is done, matching noises are removed in the step 203. Methods such as RANSAC can be used in the step. In the step 204, the distance or disparity value of these matched feature points is computed. The distance or disparity value provides a relative distance of these points in the real world with respect to focal point of the surveillance camera. Further, in the step 205, the distance or disparity values are quantized into discrete levels. Accordingly, the feature points extracted from the
right/left image of the current frame can be outputted by the feature point disparity computational unit 105 along with the respective disparity values.
[0032] FIG. 3 illustrates the process flow operationally carries out by the feature point tracking unit 106 of FIG. 1 in accordance with one embodiment of the present invention. The feature point tracking unit 106 receives input from the feature point disparity computational unit 105. The input includes either one of the stereo image (i.e. left or right image) of the current frame and previous frames, the extracted feature points from current and previous frames, and previous tracking information of feature points in previous frames. The output from the feature point tracking unit 106 provides the correspondences between feature points in current frame and previous frames. Briefly, the process includes clustering the feature points in the current frame into clouds in step 301; tracking each cloud in step 302; determining tracking status of the clouds in step 303; matching feature point between the right/left image of the current frame and previous frame in step 304; and redistributing the feature points into the cloud in step 305.
[0033] The feature point tracking unit 106 primarily forms and group the feature points into clouds and track the same accordingly. The process starts with clustering the feature points in current frame into clouds in the step 301. The clustering is done based on the similarity in appearance and proximity within the image of the current frame. Once the feature points are clustered into clouds, the feature point tracking unit 106 tracks the clouds by determining correspondences of all the resultant clouds in current frame with respect to previous frame in the step 302. The correspondences are determined based on similarity of the clouds with reference to the appearance and shape of the feature point distribution. Various correspondences determination methods are
widely available, and they can be desired for the present invention. Therefore, no f irther detail is provided herewith. In the step 303, a tracking status is being identified for each cloud. The tracking status includes new cloud, existing moving cloud, existing background cloud and existing unknown cloud. A new cloud assigned to one that has no correspondence found between current cloud and any of previous clouds. Current and previous clouds, i.e. from current and previous frames, are referring to clouds in current and previous frame respectively. In the step 304, the feature point distributions (i.e. into clouds) are further refined by feature point matching within certain clouds. The matching results that defines which points in the current frames is matched with the points in the previous frame can be used to redistribute feature point initially assigned to one cloud to another cloud in the step 305.
[0034] FIG. 4 illustrates the process flow of the feature point classification unit
107 of FIG. 1 in accordance with one embodiment of the present invention. The process is adapted for classifying each cloud detected by the feature point tracking unit 106. It classifies the clouds into moving objects, background/static objects or unknown objects. The feature point classification unit 107 receives inputs such as the clouds, tracking information and disparity information from the feature point tracking unit 106 for classification. The clouds are classified based on the disparity tracking information of each cloud. Assuming that the camera is moving in constant speed, then the rate of change of the cloud disparity information over subsequent frames would be substantially constant as well. For the cloud that is new, it will not be classified in current frame. The global disparity information for this cloud will be stored. The statistical information such as the mean or mod or the standard deviation of the disparity
distribution with a cloud can be used to represent the global disparity information of the cloud.
[0035] With the above in mind, the process flow starts with estimating a statistical disparity distribution of each cloud (i.e. mean, mpd, standard deviation etc.) in current frame in step 401. Then each of the cloud will be examined whether it is new in step 402. If the cloud is new, the disparity information will be stored in step 403 and the process ends accordingly. Returning to the step 402, if the cloud is determined not being a new cloud, the cloud is further determined whether it is an existing cloud or an unknown cloud in step 404. If it is an existing cloud at the step 404, the feature point classification unit 107 computes the change in disparity of each cloud between the current and previous frame in step 405. Following that, the feature point classification unit 107 further retrieves the acceleration information of the surveillance camera recorded by the accelerometer in step 406. If the acceleration is zero or substantially zero in the step 407, it provides that the surveillance camera is moving at a constant speed. The rate of change of disparity information of each cloud between the current frame and subsequent frames is computed in step 408. At least three frames would be required to compute the rate of change. This computation takes into consideration the changes of disparity information tracked over subsequent frames. If the rate of change is determined to be zero or substantially zero in the step 409, it can conclude that the cloud belongs to background or static object in step 413. If the rate of change is determined to be non-zero in the step 409, it an then conclude that the cloud is a moving object in step 410. Returning to the step 407, the surveillance camera is determined to be accelerating, i.e. not zero acceleration, it can either be accelerating or decelerating. The feature point classification unit 107 further computes the rate of disparity change
over subsequent frames in the step 411. The rate of disparity change of the cloud over the subsequent frames is in conformance with the movement of the surveillance camera in step 412, the cloud shall be classified as background or static object in the step 413. Otherwise, it is classified as moving object in the step 410. [0036] FIG. 5 illustrates the process flow carries out by the moving object extraction unit 108 of FIG. 1 in accordance with one embodiment of the present invention. The moving objection extraction unit 108 receives the clouds and the corresponding classifications from the feature point classification unit 107. In step 501, it merges all the clouds determined belonging to the same object based on the tracking information and changes of disparity collected over certain time period. Through the series of tracking and grouping, the processing units are able to effectively determine moving objects from a stereo surveillance camera, and track the same accordingly.
[0037] In an illustrative example, one object may be defined by two or more independent clouds when the video is processed initially. For example, a human who is weaiing red shirt and black pant may be defined by two different clouds since the object comprises two groups of different appearances. After tracking these clouds over subsequent frame, it is expected that the two clouds that have similar tracking information, e.g. moving path, relative position between the two clouds over subsequent names and tracking disparity information. Thus these similaiities can be exploited to assign all these clouds to belong to a same object entity in the step 501. As the entire process illustrated above is performed on the clouds that are classified as moving object, the output provides the group of feature point clouds, which belong to the same moving object.
[0038] While specific embodiments have been described and illustrated, it is understood that many changes, modifications, variations, and combinations thereof could be made to the present invention without depaiting from the scope of the invention.
Claims
1. A surveillance system having a pair of stereo camera capturing videos having a plurality of frames, each frames includes at least a left and a right images, wherein the stereo camera (101) having an accelerometer (102) attached thereto, the surveillance system comprising:
a video acquisition module (103) for acquiring the videos captured by the pair of stereo cameras (101) and acceleration information of the stereo camera;
a feature point disparity computational unit (105) adapted for extracting feature points in right and left images of a current frame and matching the feature points between the right and left images, wherein disparity value of each feature points is computed;
a feature point tracking unit (106) adapted for clustering the feature points on a current frame into one or more cloud based on their appearance and distance, and determining a correspondence between the clouds of the current frame and a previous frame,
a feature point classification unit (107) for classifying the clouds through estimating the statistical disparity distribution of each cloud to identify if the respective clouds are new clouds, existing clouds or unknown clouds; and
an object extraction unit (108) for merging the clouds that are determined to be a same object based on the tracking information and changes of disparity of the clouds over time.
2. The surveillance system according to claim 1, further comprising:
an event rules unit (109);
a display module for displaying overplayed information on live video; and a post detection module to trigger post event actions (11 1) defined by the event rales unit (109).
3. A method for processing videos captured through a stereo camera, the videos having a plurality of frames, each frames includes at least a left and a right images, wherein the stereo camera (101 ) having an accelerometer attached thereto, the method comprising:
acquiring videos captured by the pair of stereo cameras (101) and acceleration information of the stereo camera (101) detected through the accelerometer (102);
extracting feature points in the right and left images of a current frame;
matching the feature points between the right and left images, wherein disparity value of each feature point is computed;
clustering the feature points on a current frame into one or more clouds based on their appearance and distance;
determining a correspondence between the clouds of the current frame and previous frame;
classifying the clouds through estimating statistical disparity distribution of each cloud to identify if the respective clouds are new clouds, existing clouds or unknown clouds; and
merging the clouds that are determined to be a same object based on the tracking information and changes of disparity of the clouds over time.
4. The method according to claim 3, further comprising:
filtering noises while matching the feature points between the left and the right images;
computing distances between the matched features points; and
quantizing the distance based on a current disparity value range.
5. The method according to claim 4, wherein matching the feature points between the left and right image comprising searching area for fining a best matched feature point based on matching feature points determined in the previous frame.
6. The method according to claim 5, wherein searching area for finding the best matched feature point comprising:
matching feature points within an entire image of a first frame; and
matching feature points for all subsequent frames based on areas corresponding to feature points matched in a previous frame.
7. The method according to claim 3, wherein clustering the feature points into the one or more clouds follows by the steps comprising:
tracking (302) each cloud by determining the correspondences between clouds detected in cuirent frame and in previous frames;
deteimining (303) a tracking status of each cloud detected in current frame, wherein the tracking status include new object, existing moving object, existing background and existing undetermined object;
perforrning feature points matching (304) between one of the left and right images of the current frame and previous frame; and
redistribute (305) the feature point clouds into different or same cloud based on the matching information.
8. The method according to claim 3, wherein classifying feature point further comprising:
estimating a statistical disparity distribution of each feature point cloud through mean, mpd or standard deviation;
storing the information when the cloud is determined as a new cloud;
computing the change of disparity of each clouds that are tracked between a current frame and a previous frame;
deteimining the acceleration information of the stereo camera for the instance frame,
wherein when the acceleration is substantially zero, the rate of disparity changes of each tracked cloud in the cun'ent frame over a subsequent frame is computed, the cloud is determined to be a moving object when the rate of change is not substantially zero, and the cloud is determined to be a static object or background when the rate of change is substantially zero; and when the acceleration is not substantially zero, the rate of dispaiity change between a cuirent frame and a previous frame is computed, and the cloud is determined to be a static object or background if the rate of change is increasing/decreasing, and the cloud is determined to be a moving object when the rate of disparity change is substantially zero.
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
MYPI2012005406A MY172143A (en) | 2012-12-13 | 2012-12-13 | Method for non-static foreground feature extraction and classification |
MYPI2012005406 | 2012-12-13 |
Publications (2)
Publication Number | Publication Date |
---|---|
WO2014092552A2 true WO2014092552A2 (en) | 2014-06-19 |
WO2014092552A3 WO2014092552A3 (en) | 2014-09-12 |
Family
ID=50179900
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/MY2013/000266 WO2014092552A2 (en) | 2012-12-13 | 2013-12-13 | Method for non-static foreground feature extraction and classification |
Country Status (2)
Country | Link |
---|---|
MY (1) | MY172143A (en) |
WO (1) | WO2014092552A2 (en) |
Cited By (50)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
GB2539387A (en) * | 2015-06-09 | 2016-12-21 | Omg Plc | Motion capture system |
WO2018097915A1 (en) * | 2016-11-22 | 2018-05-31 | Symbol Technologies, Llc | Dimensioning system for, and method of, dimensioning freight in motion along an unconstrained path in a venue |
US10140725B2 (en) | 2014-12-05 | 2018-11-27 | Symbol Technologies, Llc | Apparatus for and method of estimating dimensions of an object associated with a code in automatic response to reading the code |
US10145955B2 (en) | 2016-02-04 | 2018-12-04 | Symbol Technologies, Llc | Methods and systems for processing point-cloud data with a line scanner |
US10352689B2 (en) | 2016-01-28 | 2019-07-16 | Symbol Technologies, Llc | Methods and systems for high precision locationing with depth values |
US10354411B2 (en) | 2016-12-20 | 2019-07-16 | Symbol Technologies, Llc | Methods, systems and apparatus for segmenting objects |
US10521914B2 (en) | 2017-09-07 | 2019-12-31 | Symbol Technologies, Llc | Multi-sensor object recognition system and method |
US10572763B2 (en) | 2017-09-07 | 2020-02-25 | Symbol Technologies, Llc | Method and apparatus for support surface edge detection |
US10591918B2 (en) | 2017-05-01 | 2020-03-17 | Symbol Technologies, Llc | Fixed segmented lattice planning for a mobile automation apparatus |
US10663590B2 (en) | 2017-05-01 | 2020-05-26 | Symbol Technologies, Llc | Device and method for merging lidar data |
US10721451B2 (en) | 2016-03-23 | 2020-07-21 | Symbol Technologies, Llc | Arrangement for, and method of, loading freight into a shipping container |
US10726273B2 (en) | 2017-05-01 | 2020-07-28 | Symbol Technologies, Llc | Method and apparatus for shelf feature and object placement detection from shelf images |
US10731970B2 (en) | 2018-12-13 | 2020-08-04 | Zebra Technologies Corporation | Method, system and apparatus for support structure detection |
US10740911B2 (en) | 2018-04-05 | 2020-08-11 | Symbol Technologies, Llc | Method, system and apparatus for correcting translucency artifacts in data representing a support structure |
US10776661B2 (en) | 2016-08-19 | 2020-09-15 | Symbol Technologies, Llc | Methods, systems and apparatus for segmenting and dimensioning objects |
US10809078B2 (en) | 2018-04-05 | 2020-10-20 | Symbol Technologies, Llc | Method, system and apparatus for dynamic path generation |
US10823572B2 (en) | 2018-04-05 | 2020-11-03 | Symbol Technologies, Llc | Method, system and apparatus for generating navigational data |
US10832436B2 (en) | 2018-04-05 | 2020-11-10 | Symbol Technologies, Llc | Method, system and apparatus for recovering label positions |
US10949798B2 (en) | 2017-05-01 | 2021-03-16 | Symbol Technologies, Llc | Multimodal localization and mapping for a mobile automation apparatus |
US11003188B2 (en) | 2018-11-13 | 2021-05-11 | Zebra Technologies Corporation | Method, system and apparatus for obstacle handling in navigational path generation |
US11010920B2 (en) | 2018-10-05 | 2021-05-18 | Zebra Technologies Corporation | Method, system and apparatus for object detection in point clouds |
US11015938B2 (en) | 2018-12-12 | 2021-05-25 | Zebra Technologies Corporation | Method, system and apparatus for navigational assistance |
US11042161B2 (en) | 2016-11-16 | 2021-06-22 | Symbol Technologies, Llc | Navigation control method and apparatus in a mobile automation system |
US11080566B2 (en) | 2019-06-03 | 2021-08-03 | Zebra Technologies Corporation | Method, system and apparatus for gap detection in support structures with peg regions |
US11079240B2 (en) | 2018-12-07 | 2021-08-03 | Zebra Technologies Corporation | Method, system and apparatus for adaptive particle filter localization |
US11093896B2 (en) | 2017-05-01 | 2021-08-17 | Symbol Technologies, Llc | Product status detection system |
US11090811B2 (en) | 2018-11-13 | 2021-08-17 | Zebra Technologies Corporation | Method and apparatus for labeling of support structures |
US11100303B2 (en) | 2018-12-10 | 2021-08-24 | Zebra Technologies Corporation | Method, system and apparatus for auxiliary label detection and association |
US11107238B2 (en) | 2019-12-13 | 2021-08-31 | Zebra Technologies Corporation | Method, system and apparatus for detecting item facings |
US11151743B2 (en) | 2019-06-03 | 2021-10-19 | Zebra Technologies Corporation | Method, system and apparatus for end of aisle detection |
US11200677B2 (en) | 2019-06-03 | 2021-12-14 | Zebra Technologies Corporation | Method, system and apparatus for shelf edge detection |
US11327504B2 (en) | 2018-04-05 | 2022-05-10 | Symbol Technologies, Llc | Method, system and apparatus for mobile automation apparatus localization |
US11341663B2 (en) | 2019-06-03 | 2022-05-24 | Zebra Technologies Corporation | Method, system and apparatus for detecting support structure obstructions |
US11367092B2 (en) | 2017-05-01 | 2022-06-21 | Symbol Technologies, Llc | Method and apparatus for extracting and processing price text from an image set |
CN114708304A (en) * | 2022-06-06 | 2022-07-05 | 苏州浪潮智能科技有限公司 | Cross-camera multi-target tracking method, device, equipment and medium |
US11392891B2 (en) | 2020-11-03 | 2022-07-19 | Zebra Technologies Corporation | Item placement detection and optimization in material handling systems |
US11402846B2 (en) | 2019-06-03 | 2022-08-02 | Zebra Technologies Corporation | Method, system and apparatus for mitigating data capture light leakage |
US11416000B2 (en) | 2018-12-07 | 2022-08-16 | Zebra Technologies Corporation | Method and apparatus for navigational ray tracing |
US11449059B2 (en) | 2017-05-01 | 2022-09-20 | Symbol Technologies, Llc | Obstacle detection for a mobile automation apparatus |
US11450024B2 (en) | 2020-07-17 | 2022-09-20 | Zebra Technologies Corporation | Mixed depth object detection |
US11507103B2 (en) | 2019-12-04 | 2022-11-22 | Zebra Technologies Corporation | Method, system and apparatus for localization-based historical obstacle handling |
US11506483B2 (en) | 2018-10-05 | 2022-11-22 | Zebra Technologies Corporation | Method, system and apparatus for support structure depth determination |
US11592826B2 (en) | 2018-12-28 | 2023-02-28 | Zebra Technologies Corporation | Method, system and apparatus for dynamic loop closure in mapping trajectories |
US11593915B2 (en) | 2020-10-21 | 2023-02-28 | Zebra Technologies Corporation | Parallax-tolerant panoramic image generation |
US11600084B2 (en) | 2017-05-05 | 2023-03-07 | Symbol Technologies, Llc | Method and apparatus for detecting and interpreting price label text |
US11662739B2 (en) | 2019-06-03 | 2023-05-30 | Zebra Technologies Corporation | Method, system and apparatus for adaptive ceiling-based localization |
US11822333B2 (en) | 2020-03-30 | 2023-11-21 | Zebra Technologies Corporation | Method, system and apparatus for data capture illumination control |
US11954882B2 (en) | 2021-06-17 | 2024-04-09 | Zebra Technologies Corporation | Feature-based georegistration for mobile computing devices |
US11960286B2 (en) | 2019-06-03 | 2024-04-16 | Zebra Technologies Corporation | Method, system and apparatus for dynamic task sequencing |
US11978011B2 (en) | 2017-05-01 | 2024-05-07 | Symbol Technologies, Llc | Method and apparatus for object status detection |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11847832B2 (en) | 2020-11-11 | 2023-12-19 | Zebra Technologies Corporation | Object classification for autonomous navigation systems |
Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP2256690A1 (en) * | 2009-05-29 | 2010-12-01 | Honda Research Institute Europe GmbH | Object motion detection system based on combining 3D warping techniques and a proper object motion detection |
-
2012
- 2012-12-13 MY MYPI2012005406A patent/MY172143A/en unknown
-
2013
- 2013-12-13 WO PCT/MY2013/000266 patent/WO2014092552A2/en active Application Filing
Patent Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP2256690A1 (en) * | 2009-05-29 | 2010-12-01 | Honda Research Institute Europe GmbH | Object motion detection system based on combining 3D warping techniques and a proper object motion detection |
Non-Patent Citations (2)
Title |
---|
Cheng Yuan Tang ET AL: "A 3D Feature-Based Tracker for Multiple Object Tracking", Proc. Natl. Sci. Counc. ROC(A), Vol. 21, No. 1, 1 January 1999 (1999-01-01), pages 151-168, XP055126436, Retrieved from the Internet: URL:http://pdf.aminer.org/000/361/024/a_d_predictive_visual_tracker_for_tracking_multiple_moving_objects.pdf [retrieved on 2014-07-02] * |
WEI DU ET AL: "Tracking by cluster analysis of feature points using a mixture particle filter", PROCEEDINGS. IEEE CONFERENCE ON ADVANCED VIDEO AND SIGNAL BASED SURVEILLANCE, 2005. COMO, ITALY SEPT. 15-16, 2005, PISCATAWAY, NJ, USA,IEEE, PISCATAWAY, NJ, USA, 15 September 2005 (2005-09-15), pages 165-170, XP010881168, DOI: 10.1109/AVSS.2005.1577261 ISBN: 978-0-7803-9385-1 * |
Cited By (54)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10140725B2 (en) | 2014-12-05 | 2018-11-27 | Symbol Technologies, Llc | Apparatus for and method of estimating dimensions of an object associated with a code in automatic response to reading the code |
US10375288B2 (en) | 2015-06-09 | 2019-08-06 | Oxford Metrics Plc | Motion capture system |
GB2539387B (en) * | 2015-06-09 | 2021-04-14 | Oxford Metrics Plc | Motion capture system |
GB2539387A (en) * | 2015-06-09 | 2016-12-21 | Omg Plc | Motion capture system |
US10352689B2 (en) | 2016-01-28 | 2019-07-16 | Symbol Technologies, Llc | Methods and systems for high precision locationing with depth values |
US10145955B2 (en) | 2016-02-04 | 2018-12-04 | Symbol Technologies, Llc | Methods and systems for processing point-cloud data with a line scanner |
US10721451B2 (en) | 2016-03-23 | 2020-07-21 | Symbol Technologies, Llc | Arrangement for, and method of, loading freight into a shipping container |
US10776661B2 (en) | 2016-08-19 | 2020-09-15 | Symbol Technologies, Llc | Methods, systems and apparatus for segmenting and dimensioning objects |
US11042161B2 (en) | 2016-11-16 | 2021-06-22 | Symbol Technologies, Llc | Navigation control method and apparatus in a mobile automation system |
WO2018097915A1 (en) * | 2016-11-22 | 2018-05-31 | Symbol Technologies, Llc | Dimensioning system for, and method of, dimensioning freight in motion along an unconstrained path in a venue |
US10451405B2 (en) | 2016-11-22 | 2019-10-22 | Symbol Technologies, Llc | Dimensioning system for, and method of, dimensioning freight in motion along an unconstrained path in a venue |
US10354411B2 (en) | 2016-12-20 | 2019-07-16 | Symbol Technologies, Llc | Methods, systems and apparatus for segmenting objects |
US10591918B2 (en) | 2017-05-01 | 2020-03-17 | Symbol Technologies, Llc | Fixed segmented lattice planning for a mobile automation apparatus |
US10726273B2 (en) | 2017-05-01 | 2020-07-28 | Symbol Technologies, Llc | Method and apparatus for shelf feature and object placement detection from shelf images |
US11978011B2 (en) | 2017-05-01 | 2024-05-07 | Symbol Technologies, Llc | Method and apparatus for object status detection |
US11093896B2 (en) | 2017-05-01 | 2021-08-17 | Symbol Technologies, Llc | Product status detection system |
US10663590B2 (en) | 2017-05-01 | 2020-05-26 | Symbol Technologies, Llc | Device and method for merging lidar data |
US10949798B2 (en) | 2017-05-01 | 2021-03-16 | Symbol Technologies, Llc | Multimodal localization and mapping for a mobile automation apparatus |
US11449059B2 (en) | 2017-05-01 | 2022-09-20 | Symbol Technologies, Llc | Obstacle detection for a mobile automation apparatus |
US11367092B2 (en) | 2017-05-01 | 2022-06-21 | Symbol Technologies, Llc | Method and apparatus for extracting and processing price text from an image set |
US11600084B2 (en) | 2017-05-05 | 2023-03-07 | Symbol Technologies, Llc | Method and apparatus for detecting and interpreting price label text |
US10572763B2 (en) | 2017-09-07 | 2020-02-25 | Symbol Technologies, Llc | Method and apparatus for support surface edge detection |
US10521914B2 (en) | 2017-09-07 | 2019-12-31 | Symbol Technologies, Llc | Multi-sensor object recognition system and method |
US10832436B2 (en) | 2018-04-05 | 2020-11-10 | Symbol Technologies, Llc | Method, system and apparatus for recovering label positions |
US11327504B2 (en) | 2018-04-05 | 2022-05-10 | Symbol Technologies, Llc | Method, system and apparatus for mobile automation apparatus localization |
US10823572B2 (en) | 2018-04-05 | 2020-11-03 | Symbol Technologies, Llc | Method, system and apparatus for generating navigational data |
US10809078B2 (en) | 2018-04-05 | 2020-10-20 | Symbol Technologies, Llc | Method, system and apparatus for dynamic path generation |
US10740911B2 (en) | 2018-04-05 | 2020-08-11 | Symbol Technologies, Llc | Method, system and apparatus for correcting translucency artifacts in data representing a support structure |
US11010920B2 (en) | 2018-10-05 | 2021-05-18 | Zebra Technologies Corporation | Method, system and apparatus for object detection in point clouds |
US11506483B2 (en) | 2018-10-05 | 2022-11-22 | Zebra Technologies Corporation | Method, system and apparatus for support structure depth determination |
US11003188B2 (en) | 2018-11-13 | 2021-05-11 | Zebra Technologies Corporation | Method, system and apparatus for obstacle handling in navigational path generation |
US11090811B2 (en) | 2018-11-13 | 2021-08-17 | Zebra Technologies Corporation | Method and apparatus for labeling of support structures |
US11079240B2 (en) | 2018-12-07 | 2021-08-03 | Zebra Technologies Corporation | Method, system and apparatus for adaptive particle filter localization |
US11416000B2 (en) | 2018-12-07 | 2022-08-16 | Zebra Technologies Corporation | Method and apparatus for navigational ray tracing |
US11100303B2 (en) | 2018-12-10 | 2021-08-24 | Zebra Technologies Corporation | Method, system and apparatus for auxiliary label detection and association |
US11015938B2 (en) | 2018-12-12 | 2021-05-25 | Zebra Technologies Corporation | Method, system and apparatus for navigational assistance |
US10731970B2 (en) | 2018-12-13 | 2020-08-04 | Zebra Technologies Corporation | Method, system and apparatus for support structure detection |
US11592826B2 (en) | 2018-12-28 | 2023-02-28 | Zebra Technologies Corporation | Method, system and apparatus for dynamic loop closure in mapping trajectories |
US11151743B2 (en) | 2019-06-03 | 2021-10-19 | Zebra Technologies Corporation | Method, system and apparatus for end of aisle detection |
US11341663B2 (en) | 2019-06-03 | 2022-05-24 | Zebra Technologies Corporation | Method, system and apparatus for detecting support structure obstructions |
US11960286B2 (en) | 2019-06-03 | 2024-04-16 | Zebra Technologies Corporation | Method, system and apparatus for dynamic task sequencing |
US11662739B2 (en) | 2019-06-03 | 2023-05-30 | Zebra Technologies Corporation | Method, system and apparatus for adaptive ceiling-based localization |
US11402846B2 (en) | 2019-06-03 | 2022-08-02 | Zebra Technologies Corporation | Method, system and apparatus for mitigating data capture light leakage |
US11200677B2 (en) | 2019-06-03 | 2021-12-14 | Zebra Technologies Corporation | Method, system and apparatus for shelf edge detection |
US11080566B2 (en) | 2019-06-03 | 2021-08-03 | Zebra Technologies Corporation | Method, system and apparatus for gap detection in support structures with peg regions |
US11507103B2 (en) | 2019-12-04 | 2022-11-22 | Zebra Technologies Corporation | Method, system and apparatus for localization-based historical obstacle handling |
US11107238B2 (en) | 2019-12-13 | 2021-08-31 | Zebra Technologies Corporation | Method, system and apparatus for detecting item facings |
US11822333B2 (en) | 2020-03-30 | 2023-11-21 | Zebra Technologies Corporation | Method, system and apparatus for data capture illumination control |
US11450024B2 (en) | 2020-07-17 | 2022-09-20 | Zebra Technologies Corporation | Mixed depth object detection |
US11593915B2 (en) | 2020-10-21 | 2023-02-28 | Zebra Technologies Corporation | Parallax-tolerant panoramic image generation |
US11392891B2 (en) | 2020-11-03 | 2022-07-19 | Zebra Technologies Corporation | Item placement detection and optimization in material handling systems |
US11954882B2 (en) | 2021-06-17 | 2024-04-09 | Zebra Technologies Corporation | Feature-based georegistration for mobile computing devices |
WO2023236514A1 (en) * | 2022-06-06 | 2023-12-14 | 苏州元脑智能科技有限公司 | Cross-camera multi-object tracking method and apparatus, device, and medium |
CN114708304A (en) * | 2022-06-06 | 2022-07-05 | 苏州浪潮智能科技有限公司 | Cross-camera multi-target tracking method, device, equipment and medium |
Also Published As
Publication number | Publication date |
---|---|
MY172143A (en) | 2019-11-14 |
WO2014092552A3 (en) | 2014-09-12 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2014092552A2 (en) | Method for non-static foreground feature extraction and classification | |
US11195038B2 (en) | Device and a method for extracting dynamic information on a scene using a convolutional neural network | |
CN104751491B (en) | A kind of crowd's tracking and people flow rate statistical method and device | |
CN103824070B (en) | A kind of rapid pedestrian detection method based on computer vision | |
US8744122B2 (en) | System and method for object detection from a moving platform | |
CN117095349A (en) | Appearance search system, method, and non-transitory computer readable medium | |
CN102867386B (en) | Intelligent video analysis-based forest smoke and fire detection method and special system thereof | |
CN107657244B (en) | Human body falling behavior detection system based on multiple cameras and detection method thereof | |
CN104966062B (en) | Video monitoring method and device | |
CN112514373B (en) | Image processing apparatus and method for feature extraction | |
CN108596157B (en) | Crowd disturbance scene detection method and system based on motion detection | |
Denman et al. | Multi-spectral fusion for surveillance systems | |
CN111145223A (en) | Multi-camera personnel behavior track identification analysis method | |
CN110544271B (en) | Parabolic motion detection method and related device | |
CN103336947A (en) | Method for identifying infrared movement small target based on significance and structure | |
Malhi et al. | Vision based intelligent traffic management system | |
CN111833380B (en) | Multi-view image fusion space target tracking system and method | |
CN112270253A (en) | High-altitude parabolic detection method and device | |
KR20140132140A (en) | Method and apparatus for video surveillance based on detecting abnormal behavior using extraction of trajectories from crowd in images | |
Zhang et al. | Real-time moving object classification with automatic scene division | |
KR101214858B1 (en) | Moving object detecting apparatus and method using clustering | |
Sharma et al. | Automatic vehicle detection using spatial time frame and object based classification | |
Piérard et al. | A probabilistic pixel-based approach to detect humans in video streams | |
Xu et al. | A novel method for people and vehicle classification based on Hough line feature | |
Benedek et al. | Lidar-based gait analysis in people tracking and 4D visualization |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
122 | Ep: pct application non-entry in european phase |
Ref document number: 13831952 Country of ref document: EP Kind code of ref document: A2 |