WO2014092552A2 - Method for non-static foreground feature extraction and classification - Google Patents

Method for non-static foreground feature extraction and classification Download PDF

Info

Publication number
WO2014092552A2
WO2014092552A2 PCT/MY2013/000266 MY2013000266W WO2014092552A2 WO 2014092552 A2 WO2014092552 A2 WO 2014092552A2 MY 2013000266 W MY2013000266 W MY 2013000266W WO 2014092552 A2 WO2014092552 A2 WO 2014092552A2
Authority
WO
WIPO (PCT)
Prior art keywords
clouds
cloud
feature points
disparity
frame
Prior art date
Application number
PCT/MY2013/000266
Other languages
French (fr)
Other versions
WO2014092552A3 (en
Inventor
Binti Kadim ZULAIKHA
Hock Woon Hon
Binti Samudin NORSHUHADA
Original Assignee
Mimos Berhad
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Mimos Berhad filed Critical Mimos Berhad
Publication of WO2014092552A2 publication Critical patent/WO2014092552A2/en
Publication of WO2014092552A3 publication Critical patent/WO2014092552A3/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/292Multi-camera tracking
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/194Segmentation; Edge detection involving foreground-background segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/215Motion-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/246Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/285Analysis of motion using a sequence of stereo image pairs
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence
    • G06T2207/10021Stereoscopic video; Stereoscopic image sequence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30232Surveillance

Definitions

  • the present invention relates to image processing. More particularly, the present invention relates to a method for extracting and classifying non-static foreground features.
  • US patent no. 7,729,512 provides a stereo image processing system and method for detecting moving objects.
  • disparity threshold is used for classifying the feature points. This system require extensive searching and computing of all the individual the feature points and their disparity infomiation, which may lead to high processing power needed for real-time tracking.
  • US Patent publication no. 2011/0134221 too suggests a system and method of processing stereo images for extracting moving object. Similarly, feature vectors are being extracted and matched to detect moving object. [0005] It is recognized that feature points/vectors are widely used for detecting moving object from a video stream. However, it is a challenge to provide a reliable system and method to process these feature points in an effective and efficient way to extract the moving object. It is also desired that the system and method are adapted with adaptive capability to refine the detection results in real time.
  • a surveillance system having a pair of stereo camera capturing videos having a plurality of frames, each frames includes at least a left and a right images, wherein the stereo camera having an accelerometer attached thereto.
  • the surveillance system comprises a video acquisition module for acquiring the videos captured by the pair of stereo cameras and acceleration information of the stereo camera; a feature point disparity computational unit adapted for extracting feature points in right and left images of a current frame and matching the feature points between the right and left images, wherein disparity value of each feature points is computed; a feature point tracking unit adapted for clustering the feature points on a current frame into one or more cloud based on their appearance and distance, and determining a correspondence between the clouds of the current frame and a previous frame, a feature point classification unit for classifying the clouds through estimating the statistical disparity distribution of each cloud to identify if the respective clouds are new clouds, existing clouds or unknown clouds; and an object extraction unit for merging the clouds that are determined to be a same object based on the tracking information and changes of disparity of the clouds over time.
  • the surveillance system may further comprise an event rules unit; a display module for displaying overplayed information on live video; and a post detection module to trigger post event actions defined by the event rules unit.
  • an event rules unit for processing videos captured through a stereo camera, the videos having a plurality of frames, each frames includes at least a left and a right images, wherein the stereo camera having an accelerometer attached thereto.
  • the method comprises acquiring videos captured by the pair of stereo cameras and acceleration information of the stereo camera detected through the accelerometer; extracting feature points in the right and left images of a current frame; matching the feature points between the right and left images, wherein disparity value of each feature point is computed; clustering the feature points on a current frame into one or more clouds based on their appearance and distance; determining a correspondence between the clouds of the current frame and previous frame; classifying the clouds through estimating statistical disparity distribution of each cloud to identify if the respective clouds are new clouds, existing clouds or unknown clouds; and merging the clouds that are determined to be a same object based on the tracking information and changes of disparity of the clouds over time.
  • the method further comprises filtering noises while matching the feature points between the left and the right images; computing distances between the matched features points; and quantizing the distance based on a current disparity value range. It is also possible that matching the feature points between the left and right image comprising searching area for fining a best matched feature point based on matching feature points detennined in the previous frame.
  • searching area for finding the best matched feature point may comprise matching feature points within an entire image of a first frame; and matching feature points for all subsequent frames based on areas corresponding to feature points matched in a previous frame.
  • clustering the feature points into the one or more clouds follows by the steps of tracking each cloud by determining the correspondences between clouds detected in current frame and in previous frames; determining a tracking status of each cloud detected in current frame, wherein the tracking status include new object, existing moving object, existing background and existing undetermined object; performing feature points matching between one of the left and right images of the current frame and previous frame; and redistribute the feature point clouds into different or same cloud based on the matching information.
  • classifying feature point may further comprise estimating a statistical disparity distribution of each feature point cloud through mean, mpd or standard deviation; Storing the information when the cloud is determined as a new cloud; computing the change of disparity of each clouds that are tracked between a current frame and a previous frame; determining the acceleration information of the stereo camera for the instance frame.
  • the cloud is determined to be a moving object when the rate of change is not substantially zero, and the cloud is determined to be a static object or background when the rate of change is substantially zero; and when the acceleration is not substantially zero, the rate of disparity change between a current frame and a previous frame is computed, and the cloud is determined to be a static object or background if the rate of change is increasing/decreasing, and the cloud is determined to be a moving object when the rate of disparity change is substantially zero.
  • FIG. 1 illustrates a schematic diagram a surveillance system in accordance with one embodiment of the present invention
  • FIG. 2 illustrates the process flow operationally carries out by a feature point disparity computational unit in accordance with one embodiment of the present invention
  • FIG. 3 illustrates the process flow operationally carries out by a feature point tracking unit in accordance with one embodiment of the present invention
  • FIG. 4 illustrates the process flow of a feature point classification unit in accordance with one embodiment of the present invention.
  • FIG. 5 illustrates the process flow carries out by the moving object extraction unit in accordance with one embodiment of the present invention.
  • the objective of this invention is to extract foreground feature points captured by a pair of non-static cameras and classify them into a moving pixel or static / background pixel.
  • the resultant output can be used to delineate moving objects from background for further processing or event analysis.
  • the feature point extraction and classification for non-static cameras is difficult as the computer vision is unable to differentiate the movement on the feature point either it is caused by the actual moving object or by the motion of the cameras.
  • proposed invention is to solve the problem on how to differentiate between moving object and background when everything in the scene is non-static between the frames.
  • FIG. 1 illustrates a block diagram of a surveillance system 100 in accordance with one embodiment of the present invention.
  • the surveillance system 100 is adapted for processing surveillance videos captured through a pair of surveillance cameras 101.
  • the pair of cameras is adapted for capturing stereo images/videos operationally.
  • the pair of cameras shall herewith below refer to as stereo cameras.
  • stereo camera can be a commercially available stereo camera or a camera assembly that comprise two suitable videos that are installed on a platform, or frames.
  • the surveillance cameras 101 equipped or attached with an accelerometer 102 for detecting the movements of the respective surveillance cameras 101b.
  • the surveillance system 100 comprises a video acquisition module 103, database 112, an image-processing module 104, a display module 110, and a post detection module 111.
  • the video acquisition module 103 is connected to the pair of cameras 101 to acquire images or videos captured through the cameras.
  • the acceleration information concerning the plurality of cameras 101 recorded through the accelerometers 102 is also transmitted to the video acquisition module 103.
  • the captured videos/images and the acceleration information are stored on the database module 112, which are retrievable by the imaging-processing module 104 for processing anytime.
  • the image-processing module 104 responsible for processing the input videos and acceleration information transmitted from the video acquisition module 103, and analyzing the same to extract and detects moving feature points from the images/videos. Once the feature points are being detected and extracted, they are output to the display module 110 as visual outputs, and to the post detection module 111 for triggering further processing or control as necessary .
  • the image-processing module 104 further comprises a feature point disparity computational unit 105, a feature point tracking unit 106, a feature point classification unit 107, a moving object extraction unit 108, and an event rule unit 109.
  • the feature point disparity computational unit 105 performs feature point matching between right and left images of the stereo videos at any instance of time, and subsequently estimating the disparity values for each of the matched points.
  • the disparity values are used to track and classify the matched points through the feature point-tracking unit 106.
  • the feature point tracking unit 106 tracks the matched feature points by processing subsequent stereo image frames to extract the movement information of each feature point.
  • the feature points are tracked in a group of similar properties of the features.
  • some feature points may also be represented by the same or substantially the same feature vector.
  • the feature vector may be appearance such as e.g. color, edges etc. or the like. It can be challenging to find the correct correspondences between the multiple feature points having same or substantially the same feature vector, as opposed to finding the correspondence between the groups of feature points between two video frames.
  • the group of feature points that shares similar properties shall herein referred to as "the cloud”. As the cloud of feature points are being tracked, instead of the individual feature point, the computational processing needed by the image processing module 104 and the information stored over time can be reduced.
  • the internal intrinsic parameters of the stereo camera or the camera assembly are known. These internal intrinsic parameters provide distance between the two cameras (or imaging sensors) and the rotation of one camera with respect to the other.
  • the two cameras are also horizontally aligned with a 3D point projected in a relatively same horizontal line on both images.
  • the matching can be done by assuming that any point in one line in one image is correspond to one point in same corresponding line in the other image. Then with these point corresponding in hand, together with the intrinsic parameter known earlier, disparity value of this 3D point from the camera can be computed.
  • each feature point is being classified as either moving paints or background points through the feature point classification unit 107.
  • the classification is made up based on the information from the feature point tracking unit 106.
  • the information includes tracking and disparity information. It is desired that the classification be done on each cloud.
  • the moving object extraction unit 108 further group one or more clouds into a moving entity based on the feature point's classification result, tracking information and the disparity information.
  • the event rule unit 109 applies an appropriate rule based on the extraction results.
  • the event rule unit 109 applies an intrusion event rule when the moving object is detected to be an intruder.
  • These rules are predefined and they may be stored in the database 112 for retrieval.
  • the image-processing module 104 then output the video processing results to the display module 110 for visual output and when appropriate, the post detection module 111 is triggered to activate a corresponding alarm based on the processed results.
  • Feature points that have almost the same or similar appearance to each other will be group as one cloud. For example, all the feature points that are extracted from the subject's shirt are presumably having the same or similar appearances and will therefore be group as one cloud. The appearance that includes color would have certain level of similarity.
  • all feature points extracted from the pant part shall be group as the same cloud, which is a separate cloud from the cloud for the shirt.
  • the motion of both clouds may form specific corresponding patterns and therefore can further be grouped into one cloud defining the subject.
  • FIG. 2 illustrates the process flow operationally carries out by the feature point disparity computational unit 105 of FIG. 1 in accordance with one embodiment of the present invention.
  • the feature point disparity unit 105 derives the disparity information of feature points between the two stereo (i.e. left and right images) of the current frame.
  • the process flow includes extracting feature points in cun-ently left and right images in step 201; matching feature points in cunent left and right images in step 202; filtering noises in step 203; computing distance between the respective matched feature points in step 204; and quantizing the distance based on cuirent disparity value in step 205.
  • the process starts with extracting feature points light and left images.
  • salient points are extracted as the feature points in the images.
  • These salient points can be corner and edges points that can be detected through any known techniques. For example, Harris corner point detector, scale-invariant feature transform (SIFT), speeded up robust feature (SURF) etc. can be utilized for extracting the desired salient points.
  • Harris corner point detector scale-invariant feature transform (SIFT), speeded up robust feature (SURF) etc.
  • SIFT scale-invariant feature transform
  • SURF speeded up robust feature
  • the feature points i.e. the salient points
  • they are matched between right and left images of the same frame in the step 202.
  • searching and matching can be simplified to within a single or one dimensional (ID) space because the horizontal line in one image can be presumed to be in parallel to the same horizontal line in the other image.
  • searching for matching or corresponding feature points may requires searching in two-dimensional (2D) space, as two imaging means may not be perfectly aligned. It is more so when the imaging means are two independent cameras.
  • One possible approach for example is to perform image rectification between right and left cameras before matching. Such image rectification for processing stereo images is well known in the art.
  • Another desired approach that can be taken is to adaptively refine the 2D searching area for matching, of which no rectification is required. In this adaptive refinery process, previous matching information is used to determine the search area in the current frame.
  • the each point on the whole horizontal lines or rows of the right image of the first frame will be considered as the searching area.
  • the results will be analyzed for range of possible matching horizontal lines on the right image for searching through, given a point on certain horizontal line in left image is determined. Accordingly, when the next frame is to be searched, the searching area will be constraint within these ranges. That could effectively refine the number of horizontal lines to be search over time.
  • the distance or disparity value of these matched feature points is computed.
  • the distance or disparity value provides a relative distance of these points in the real world with respect to focal point of the surveillance camera.
  • the distance or disparity values are quantized into discrete levels. Accordingly, the feature points extracted from the right/left image of the current frame can be outputted by the feature point disparity computational unit 105 along with the respective disparity values.
  • FIG. 3 illustrates the process flow operationally carries out by the feature point tracking unit 106 of FIG. 1 in accordance with one embodiment of the present invention.
  • the feature point tracking unit 106 receives input from the feature point disparity computational unit 105.
  • the input includes either one of the stereo image (i.e. left or right image) of the current frame and previous frames, the extracted feature points from current and previous frames, and previous tracking information of feature points in previous frames.
  • the output from the feature point tracking unit 106 provides the correspondences between feature points in current frame and previous frames.
  • the process includes clustering the feature points in the current frame into clouds in step 301; tracking each cloud in step 302; determining tracking status of the clouds in step 303; matching feature point between the right/left image of the current frame and previous frame in step 304; and redistributing the feature points into the cloud in step 305.
  • the feature point tracking unit 106 primarily forms and group the feature points into clouds and track the same accordingly.
  • the process starts with clustering the feature points in current frame into clouds in the step 301.
  • the clustering is done based on the similarity in appearance and proximity within the image of the current frame.
  • the feature point tracking unit 106 tracks the clouds by determining correspondences of all the resultant clouds in current frame with respect to previous frame in the step 302.
  • the correspondences are determined based on similarity of the clouds with reference to the appearance and shape of the feature point distribution.
  • Various correspondences determination methods are widely available, and they can be desired for the present invention. Therefore, no f irther detail is provided herewith.
  • a tracking status is being identified for each cloud.
  • the tracking status includes new cloud, existing moving cloud, existing background cloud and existing unknown cloud.
  • Current and previous clouds i.e. from current and previous frames, are referring to clouds in current and previous frame respectively.
  • the feature point distributions i.e. into clouds
  • the matching results that defines which points in the current frames is matched with the points in the previous frame can be used to redistribute feature point initially assigned to one cloud to another cloud in the step 305.
  • FIG. 4 illustrates the process flow of the feature point classification unit
  • the process is adapted for classifying each cloud detected by the feature point tracking unit 106. It classifies the clouds into moving objects, background/static objects or unknown objects.
  • the feature point classification unit 107 receives inputs such as the clouds, tracking information and disparity information from the feature point tracking unit 106 for classification.
  • the clouds are classified based on the disparity tracking information of each cloud. Assuming that the camera is moving in constant speed, then the rate of change of the cloud disparity information over subsequent frames would be substantially constant as well. For the cloud that is new, it will not be classified in current frame.
  • the global disparity information for this cloud will be stored.
  • the statistical information such as the mean or mod or the standard deviation of the disparity distribution with a cloud can be used to represent the global disparity information of the cloud.
  • the process flow starts with estimating a statistical disparity distribution of each cloud (i.e. mean, mpd, standard deviation etc.) in current frame in step 401. Then each of the cloud will be examined whether it is new in step 402. If the cloud is new, the disparity information will be stored in step 403 and the process ends accordingly. Returning to the step 402, if the cloud is determined not being a new cloud, the cloud is further determined whether it is an existing cloud or an unknown cloud in step 404. If it is an existing cloud at the step 404, the feature point classification unit 107 computes the change in disparity of each cloud between the current and previous frame in step 405.
  • each cloud i.e. mean, mpd, standard deviation etc.
  • the feature point classification unit 107 further retrieves the acceleration information of the surveillance camera recorded by the accelerometer in step 406. If the acceleration is zero or substantially zero in the step 407, it provides that the surveillance camera is moving at a constant speed.
  • the rate of change of disparity information of each cloud between the current frame and subsequent frames is computed in step 408. At least three frames would be required to compute the rate of change. This computation takes into consideration the changes of disparity information tracked over subsequent frames. If the rate of change is determined to be zero or substantially zero in the step 409, it can conclude that the cloud belongs to background or static object in step 413. If the rate of change is determined to be non-zero in the step 409, it an then conclude that the cloud is a moving object in step 410.
  • the surveillance camera is determined to be accelerating, i.e. not zero acceleration, it can either be accelerating or decelerating.
  • the feature point classification unit 107 further computes the rate of disparity change over subsequent frames in the step 411.
  • the rate of disparity change of the cloud over the subsequent frames is in conformance with the movement of the surveillance camera in step 412, the cloud shall be classified as background or static object in the step 413. Otherwise, it is classified as moving object in the step 410.
  • FIG. 5 illustrates the process flow carries out by the moving object extraction unit 108 of FIG. 1 in accordance with one embodiment of the present invention.
  • the moving objection extraction unit 108 receives the clouds and the corresponding classifications from the feature point classification unit 107.
  • step 501 it merges all the clouds determined belonging to the same object based on the tracking information and changes of disparity collected over certain time period.
  • the processing units are able to effectively determine moving objects from a stereo surveillance camera, and track the same accordingly.
  • one object may be defined by two or more independent clouds when the video is processed initially.
  • a human who is weaiing red shirt and black pant may be defined by two different clouds since the object comprises two groups of different appearances.
  • the two clouds After tracking these clouds over subsequent frame, it is expected that the two clouds that have similar tracking information, e.g. moving path, relative position between the two clouds over subsequent names and tracking disparity information.
  • tracking information e.g. moving path
  • the output provides the group of feature point clouds, which belong to the same moving object.

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Multimedia (AREA)
  • Image Analysis (AREA)

Abstract

The present invention provides a method for processing videos captured through a stereo camera, the videos having a plurality of frames, each frames includes at least a left and a right images. The method includes acquiring videos captured by the pair of stereo cameras (101) and acceleration information of the stereo camera (101) detected through the accelerometer (102); extracting feature points in the right and left images of a current frame; matching the feature points between the right and left images, wherein disparity value of each feature point is computed; clustering the feature points on a current frame into one or more clouds based on their appearance and distance; determining a correspondence between the clouds of the current frame and previous frame; classifying the clouds through estimating statistical disparity distribution of each cloud to identify if the respective clouds are new clouds, existing clouds or unknown clouds; and merging the clouds that are determined to be a same object based on the tracking information and changes of disparity of the clouds over time. A system carrying out the above method is also provided herein.

Description

Method For Non-Static Foreground Feature Extraction and
Classification
Field of the Invention
[0001] The present invention relates to image processing. More particularly, the present invention relates to a method for extracting and classifying non-static foreground features.
Background
[0002] The feature point extraction and classification for non-static cameras is difficult as the computer vision is unable to differentiate the movement on the feature point either it is caused by the actual moving object or by the motion of the cameras.
[0003] Many studies and research had been earned out extensively. US patent no. 7,729,512 provides a stereo image processing system and method for detecting moving objects. In this patent, disparity threshold is used for classifying the feature points. This system require extensive searching and computing of all the individual the feature points and their disparity infomiation, which may lead to high processing power needed for real-time tracking.
[0004] US Patent publication no. 2011/0134221 too suggests a system and method of processing stereo images for extracting moving object. Similarly, feature vectors are being extracted and matched to detect moving object. [0005] It is recognized that feature points/vectors are widely used for detecting moving object from a video stream. However, it is a challenge to provide a reliable system and method to process these feature points in an effective and efficient way to extract the moving object. It is also desired that the system and method are adapted with adaptive capability to refine the detection results in real time.
Summary [0006] In one aspect of the present invention, there is provided a surveillance system having a pair of stereo camera capturing videos having a plurality of frames, each frames includes at least a left and a right images, wherein the stereo camera having an accelerometer attached thereto. The surveillance system comprises a video acquisition module for acquiring the videos captured by the pair of stereo cameras and acceleration information of the stereo camera; a feature point disparity computational unit adapted for extracting feature points in right and left images of a current frame and matching the feature points between the right and left images, wherein disparity value of each feature points is computed; a feature point tracking unit adapted for clustering the feature points on a current frame into one or more cloud based on their appearance and distance, and determining a correspondence between the clouds of the current frame and a previous frame, a feature point classification unit for classifying the clouds through estimating the statistical disparity distribution of each cloud to identify if the respective clouds are new clouds, existing clouds or unknown clouds; and an object extraction unit for merging the clouds that are determined to be a same object based on the tracking information and changes of disparity of the clouds over time.
[0007] In one embodiment, the surveillance system may further comprise an event rules unit; a display module for displaying overplayed information on live video; and a post detection module to trigger post event actions defined by the event rules unit. [0008] In another aspect of the present invention, there is further provided a method for processing videos captured through a stereo camera, the videos having a plurality of frames, each frames includes at least a left and a right images, wherein the stereo camera having an accelerometer attached thereto. The method comprises acquiring videos captured by the pair of stereo cameras and acceleration information of the stereo camera detected through the accelerometer; extracting feature points in the right and left images of a current frame; matching the feature points between the right and left images, wherein disparity value of each feature point is computed; clustering the feature points on a current frame into one or more clouds based on their appearance and distance; determining a correspondence between the clouds of the current frame and previous frame; classifying the clouds through estimating statistical disparity distribution of each cloud to identify if the respective clouds are new clouds, existing clouds or unknown clouds; and merging the clouds that are determined to be a same object based on the tracking information and changes of disparity of the clouds over time.
[0009] In one embodiment, the method further comprises filtering noises while matching the feature points between the left and the right images; computing distances between the matched features points; and quantizing the distance based on a current disparity value range. It is also possible that matching the feature points between the left and right image comprising searching area for fining a best matched feature point based on matching feature points detennined in the previous frame.
[0010] In yet another embodiment, searching area for finding the best matched feature point may comprise matching feature points within an entire image of a first frame; and matching feature points for all subsequent frames based on areas corresponding to feature points matched in a previous frame.
[0011] In yet a further embodiment, clustering the feature points into the one or more clouds follows by the steps of tracking each cloud by determining the correspondences between clouds detected in current frame and in previous frames; determining a tracking status of each cloud detected in current frame, wherein the tracking status include new object, existing moving object, existing background and existing undetermined object; performing feature points matching between one of the left and right images of the current frame and previous frame; and redistribute the feature point clouds into different or same cloud based on the matching information.
[0012] In a further embodiment, classifying feature point may further comprise estimating a statistical disparity distribution of each feature point cloud through mean, mpd or standard deviation; Storing the information when the cloud is determined as a new cloud; computing the change of disparity of each clouds that are tracked between a current frame and a previous frame; determining the acceleration information of the stereo camera for the instance frame. When the acceleration is substantially zero, the rate of disparity changes of each tracked cloud in the current frame over a subsequent frame is computed, the cloud is determined to be a moving object when the rate of change is not substantially zero, and the cloud is determined to be a static object or background when the rate of change is substantially zero; and when the acceleration is not substantially zero, the rate of disparity change between a current frame and a previous frame is computed, and the cloud is determined to be a static object or background if the rate of change is increasing/decreasing, and the cloud is determined to be a moving object when the rate of disparity change is substantially zero. Brief Description of the Drawings
[0013] Preferred embodiments according to the present invention will now be described with reference to the figures accompanied herein, in which like reference numerals denote like elements; [0014] FIG. 1 illustrates a schematic diagram a surveillance system in accordance with one embodiment of the present invention;
[0015] FIG. 2 illustrates the process flow operationally carries out by a feature point disparity computational unit in accordance with one embodiment of the present invention; [0016] FIG. 3 illustrates the process flow operationally carries out by a feature point tracking unit in accordance with one embodiment of the present invention;
[0017] FIG. 4 illustrates the process flow of a feature point classification unit in accordance with one embodiment of the present invention; and
[0018] FIG. 5 illustrates the process flow carries out by the moving object extraction unit in accordance with one embodiment of the present invention.
Detailed Description
[0019] Embodiments of the present invention shall now be described in detail, with reference to the attached drawings. It is to be understood that no limitation of the scope of the invention is thereby intended, such alterations and further modifications in the illustrated device, and such further applications of the principles of the invention as illustrated therein being contemplated as would normally occur to one skilled in the art to which the invention relates.
[0020] The objective of this invention is to extract foreground feature points captured by a pair of non-static cameras and classify them into a moving pixel or static / background pixel. The resultant output (sparse information) can be used to delineate moving objects from background for further processing or event analysis. The feature point extraction and classification for non-static cameras is difficult as the computer vision is unable to differentiate the movement on the feature point either it is caused by the actual moving object or by the motion of the cameras. Thus, proposed invention is to solve the problem on how to differentiate between moving object and background when everything in the scene is non-static between the frames.
[0021] FIG. 1 illustrates a block diagram of a surveillance system 100 in accordance with one embodiment of the present invention. The surveillance system 100 is adapted for processing surveillance videos captured through a pair of surveillance cameras 101. The pair of cameras is adapted for capturing stereo images/videos operationally. For the purpose of description below, the pair of cameras shall herewith below refer to as stereo cameras. Such stereo camera can be a commercially available stereo camera or a camera assembly that comprise two suitable videos that are installed on a platform, or frames. Preferably, the surveillance cameras 101 equipped or attached with an accelerometer 102 for detecting the movements of the respective surveillance cameras 101b. The surveillance system 100 comprises a video acquisition module 103, database 112, an image-processing module 104, a display module 110, and a post detection module 111. The video acquisition module 103 is connected to the pair of cameras 101 to acquire images or videos captured through the cameras. The acceleration information concerning the plurality of cameras 101 recorded through the accelerometers 102 is also transmitted to the video acquisition module 103. The captured videos/images and the acceleration information are stored on the database module 112, which are retrievable by the imaging-processing module 104 for processing anytime. The image-processing module 104 responsible for processing the input videos and acceleration information transmitted from the video acquisition module 103, and analyzing the same to extract and detects moving feature points from the images/videos. Once the feature points are being detected and extracted, they are output to the display module 110 as visual outputs, and to the post detection module 111 for triggering further processing or control as necessary .
[0022] Still referring to FIG. 1, the image-processing module 104 further comprises a feature point disparity computational unit 105, a feature point tracking unit 106, a feature point classification unit 107, a moving object extraction unit 108, and an event rule unit 109. The feature point disparity computational unit 105 performs feature point matching between right and left images of the stereo videos at any instance of time, and subsequently estimating the disparity values for each of the matched points. The disparity values are used to track and classify the matched points through the feature point-tracking unit 106. The feature point tracking unit 106 tracks the matched feature points by processing subsequent stereo image frames to extract the movement information of each feature point. In accordance with the present embodiment, the feature points are tracked in a group of similar properties of the features. It is recognized that some feature points may also be represented by the same or substantially the same feature vector. The feature vector may be appearance such as e.g. color, edges etc. or the like. It can be challenging to find the correct correspondences between the multiple feature points having same or substantially the same feature vector, as opposed to finding the correspondence between the groups of feature points between two video frames. For simplicity, the group of feature points that shares similar properties shall herein referred to as "the cloud". As the cloud of feature points are being tracked, instead of the individual feature point, the computational processing needed by the image processing module 104 and the information stored over time can be reduced.
[0023] When a stereo camera or a camera assembly that comprising two cameras disposed on a frame with a fix distance apart, the internal intrinsic parameters of the stereo camera or the camera assembly are known. These internal intrinsic parameters provide distance between the two cameras (or imaging sensors) and the rotation of one camera with respect to the other. Typically, the two cameras are also horizontally aligned with a 3D point projected in a relatively same horizontal line on both images. Thus, for any instant of two images taken by these two cameras, the matching can be done by assuming that any point in one line in one image is correspond to one point in same corresponding line in the other image. Then with these point corresponding in hand, together with the intrinsic parameter known earlier, disparity value of this 3D point from the camera can be computed.
[0024] Through the present invention however, it is cater for stereo camera or the camera assembly that is may not be properly calibrated or misaligned in the cause of operating the same. Accordingly, one line in an image may not always correspond to the same line in the other image. The proposed invention has therefore proposed the optimized matching process in light of the above. [0025] Each feature point is being classified as either moving paints or background points through the feature point classification unit 107. The classification is made up based on the information from the feature point tracking unit 106. The information includes tracking and disparity information. It is desired that the classification be done on each cloud. The moving object extraction unit 108 further group one or more clouds into a moving entity based on the feature point's classification result, tracking information and the disparity information. Accordingly, the event rule unit 109 applies an appropriate rule based on the extraction results. For example, the event rule unit 109 applies an intrusion event rule when the moving object is detected to be an intruder. These rules are predefined and they may be stored in the database 112 for retrieval. The image-processing module 104 then output the video processing results to the display module 110 for visual output and when appropriate, the post detection module 111 is triggered to activate a corresponding alarm based on the processed results. [0026] Feature points that have almost the same or similar appearance to each other will be group as one cloud. For example, all the feature points that are extracted from the subject's shirt are presumably having the same or similar appearances and will therefore be group as one cloud. The appearance that includes color would have certain level of similarity. In the case where the same subject containing more than one colors, for example, wearing a pant with a color different from the shirt, then all feature points extracted from the pant part shall be group as the same cloud, which is a separate cloud from the cloud for the shirt. After monitoring this subject, i.e. formed by the two detected cloud, over subsequent frames, the motion of both clouds may form specific corresponding patterns and therefore can further be grouped into one cloud defining the subject.
[0027] FIG. 2 illustrates the process flow operationally carries out by the feature point disparity computational unit 105 of FIG. 1 in accordance with one embodiment of the present invention. The feature point disparity unit 105 derives the disparity information of feature points between the two stereo (i.e. left and right images) of the current frame. Briefly, the process flow includes extracting feature points in cun-ently left and right images in step 201; matching feature points in cunent left and right images in step 202; filtering noises in step 203; computing distance between the respective matched feature points in step 204; and quantizing the distance based on cuirent disparity value in step 205.
[0028] In the step 201, the process starts with extracting feature points light and left images. In one embodiment of the present invention, salient points are extracted as the feature points in the images. These salient points can be corner and edges points that can be detected through any known techniques. For example, Harris corner point detector, scale-invariant feature transform (SIFT), speeded up robust feature (SURF) etc. can be utilized for extracting the desired salient points.
[0029] Once the feature points (i.e. the salient points) are extracted, they are matched between right and left images of the same frame in the step 202. On a stereo camera where two imaging means are aligned to be coplanar, searching and matching can be simplified to within a single or one dimensional (ID) space because the horizontal line in one image can be presumed to be in parallel to the same horizontal line in the other image. In some camera configurations, searching for matching or corresponding feature points may requires searching in two-dimensional (2D) space, as two imaging means may not be perfectly aligned. It is more so when the imaging means are two independent cameras. One possible approach for example is to perform image rectification between right and left cameras before matching. Such image rectification for processing stereo images is well known in the art. Another desired approach that can be taken is to adaptively refine the 2D searching area for matching, of which no rectification is required. In this adaptive refinery process, previous matching information is used to determine the search area in the current frame.
[0030] In one illustrative example for matching feature points on the right image that correspond to one feature point on the left image, the each point on the whole horizontal lines or rows of the right image of the first frame will be considered as the searching area. Once the matching is done for all feature points, the results will be analyzed for range of possible matching horizontal lines on the right image for searching through, given a point on certain horizontal line in left image is determined. Accordingly, when the next frame is to be searched, the searching area will be constraint within these ranges. That could effectively refine the number of horizontal lines to be search over time.
[0031] Once the feature points matching between the right and left images of the current frame is done, matching noises are removed in the step 203. Methods such as RANSAC can be used in the step. In the step 204, the distance or disparity value of these matched feature points is computed. The distance or disparity value provides a relative distance of these points in the real world with respect to focal point of the surveillance camera. Further, in the step 205, the distance or disparity values are quantized into discrete levels. Accordingly, the feature points extracted from the right/left image of the current frame can be outputted by the feature point disparity computational unit 105 along with the respective disparity values.
[0032] FIG. 3 illustrates the process flow operationally carries out by the feature point tracking unit 106 of FIG. 1 in accordance with one embodiment of the present invention. The feature point tracking unit 106 receives input from the feature point disparity computational unit 105. The input includes either one of the stereo image (i.e. left or right image) of the current frame and previous frames, the extracted feature points from current and previous frames, and previous tracking information of feature points in previous frames. The output from the feature point tracking unit 106 provides the correspondences between feature points in current frame and previous frames. Briefly, the process includes clustering the feature points in the current frame into clouds in step 301; tracking each cloud in step 302; determining tracking status of the clouds in step 303; matching feature point between the right/left image of the current frame and previous frame in step 304; and redistributing the feature points into the cloud in step 305.
[0033] The feature point tracking unit 106 primarily forms and group the feature points into clouds and track the same accordingly. The process starts with clustering the feature points in current frame into clouds in the step 301. The clustering is done based on the similarity in appearance and proximity within the image of the current frame. Once the feature points are clustered into clouds, the feature point tracking unit 106 tracks the clouds by determining correspondences of all the resultant clouds in current frame with respect to previous frame in the step 302. The correspondences are determined based on similarity of the clouds with reference to the appearance and shape of the feature point distribution. Various correspondences determination methods are widely available, and they can be desired for the present invention. Therefore, no f irther detail is provided herewith. In the step 303, a tracking status is being identified for each cloud. The tracking status includes new cloud, existing moving cloud, existing background cloud and existing unknown cloud. A new cloud assigned to one that has no correspondence found between current cloud and any of previous clouds. Current and previous clouds, i.e. from current and previous frames, are referring to clouds in current and previous frame respectively. In the step 304, the feature point distributions (i.e. into clouds) are further refined by feature point matching within certain clouds. The matching results that defines which points in the current frames is matched with the points in the previous frame can be used to redistribute feature point initially assigned to one cloud to another cloud in the step 305.
[0034] FIG. 4 illustrates the process flow of the feature point classification unit
107 of FIG. 1 in accordance with one embodiment of the present invention. The process is adapted for classifying each cloud detected by the feature point tracking unit 106. It classifies the clouds into moving objects, background/static objects or unknown objects. The feature point classification unit 107 receives inputs such as the clouds, tracking information and disparity information from the feature point tracking unit 106 for classification. The clouds are classified based on the disparity tracking information of each cloud. Assuming that the camera is moving in constant speed, then the rate of change of the cloud disparity information over subsequent frames would be substantially constant as well. For the cloud that is new, it will not be classified in current frame. The global disparity information for this cloud will be stored. The statistical information such as the mean or mod or the standard deviation of the disparity distribution with a cloud can be used to represent the global disparity information of the cloud.
[0035] With the above in mind, the process flow starts with estimating a statistical disparity distribution of each cloud (i.e. mean, mpd, standard deviation etc.) in current frame in step 401. Then each of the cloud will be examined whether it is new in step 402. If the cloud is new, the disparity information will be stored in step 403 and the process ends accordingly. Returning to the step 402, if the cloud is determined not being a new cloud, the cloud is further determined whether it is an existing cloud or an unknown cloud in step 404. If it is an existing cloud at the step 404, the feature point classification unit 107 computes the change in disparity of each cloud between the current and previous frame in step 405. Following that, the feature point classification unit 107 further retrieves the acceleration information of the surveillance camera recorded by the accelerometer in step 406. If the acceleration is zero or substantially zero in the step 407, it provides that the surveillance camera is moving at a constant speed. The rate of change of disparity information of each cloud between the current frame and subsequent frames is computed in step 408. At least three frames would be required to compute the rate of change. This computation takes into consideration the changes of disparity information tracked over subsequent frames. If the rate of change is determined to be zero or substantially zero in the step 409, it can conclude that the cloud belongs to background or static object in step 413. If the rate of change is determined to be non-zero in the step 409, it an then conclude that the cloud is a moving object in step 410. Returning to the step 407, the surveillance camera is determined to be accelerating, i.e. not zero acceleration, it can either be accelerating or decelerating. The feature point classification unit 107 further computes the rate of disparity change over subsequent frames in the step 411. The rate of disparity change of the cloud over the subsequent frames is in conformance with the movement of the surveillance camera in step 412, the cloud shall be classified as background or static object in the step 413. Otherwise, it is classified as moving object in the step 410. [0036] FIG. 5 illustrates the process flow carries out by the moving object extraction unit 108 of FIG. 1 in accordance with one embodiment of the present invention. The moving objection extraction unit 108 receives the clouds and the corresponding classifications from the feature point classification unit 107. In step 501, it merges all the clouds determined belonging to the same object based on the tracking information and changes of disparity collected over certain time period. Through the series of tracking and grouping, the processing units are able to effectively determine moving objects from a stereo surveillance camera, and track the same accordingly.
[0037] In an illustrative example, one object may be defined by two or more independent clouds when the video is processed initially. For example, a human who is weaiing red shirt and black pant may be defined by two different clouds since the object comprises two groups of different appearances. After tracking these clouds over subsequent frame, it is expected that the two clouds that have similar tracking information, e.g. moving path, relative position between the two clouds over subsequent names and tracking disparity information. Thus these similaiities can be exploited to assign all these clouds to belong to a same object entity in the step 501. As the entire process illustrated above is performed on the clouds that are classified as moving object, the output provides the group of feature point clouds, which belong to the same moving object. [0038] While specific embodiments have been described and illustrated, it is understood that many changes, modifications, variations, and combinations thereof could be made to the present invention without depaiting from the scope of the invention.

Claims

Claims
1. A surveillance system having a pair of stereo camera capturing videos having a plurality of frames, each frames includes at least a left and a right images, wherein the stereo camera (101) having an accelerometer (102) attached thereto, the surveillance system comprising:
a video acquisition module (103) for acquiring the videos captured by the pair of stereo cameras (101) and acceleration information of the stereo camera;
a feature point disparity computational unit (105) adapted for extracting feature points in right and left images of a current frame and matching the feature points between the right and left images, wherein disparity value of each feature points is computed;
a feature point tracking unit (106) adapted for clustering the feature points on a current frame into one or more cloud based on their appearance and distance, and determining a correspondence between the clouds of the current frame and a previous frame,
a feature point classification unit (107) for classifying the clouds through estimating the statistical disparity distribution of each cloud to identify if the respective clouds are new clouds, existing clouds or unknown clouds; and
an object extraction unit (108) for merging the clouds that are determined to be a same object based on the tracking information and changes of disparity of the clouds over time.
2. The surveillance system according to claim 1, further comprising:
an event rules unit (109);
a display module for displaying overplayed information on live video; and a post detection module to trigger post event actions (11 1) defined by the event rales unit (109).
3. A method for processing videos captured through a stereo camera, the videos having a plurality of frames, each frames includes at least a left and a right images, wherein the stereo camera (101 ) having an accelerometer attached thereto, the method comprising: acquiring videos captured by the pair of stereo cameras (101) and acceleration information of the stereo camera (101) detected through the accelerometer (102);
extracting feature points in the right and left images of a current frame;
matching the feature points between the right and left images, wherein disparity value of each feature point is computed;
clustering the feature points on a current frame into one or more clouds based on their appearance and distance;
determining a correspondence between the clouds of the current frame and previous frame;
classifying the clouds through estimating statistical disparity distribution of each cloud to identify if the respective clouds are new clouds, existing clouds or unknown clouds; and
merging the clouds that are determined to be a same object based on the tracking information and changes of disparity of the clouds over time.
4. The method according to claim 3, further comprising:
filtering noises while matching the feature points between the left and the right images;
computing distances between the matched features points; and
quantizing the distance based on a current disparity value range.
5. The method according to claim 4, wherein matching the feature points between the left and right image comprising searching area for fining a best matched feature point based on matching feature points determined in the previous frame.
6. The method according to claim 5, wherein searching area for finding the best matched feature point comprising:
matching feature points within an entire image of a first frame; and
matching feature points for all subsequent frames based on areas corresponding to feature points matched in a previous frame.
7. The method according to claim 3, wherein clustering the feature points into the one or more clouds follows by the steps comprising: tracking (302) each cloud by determining the correspondences between clouds detected in cuirent frame and in previous frames;
deteimining (303) a tracking status of each cloud detected in current frame, wherein the tracking status include new object, existing moving object, existing background and existing undetermined object;
perforrning feature points matching (304) between one of the left and right images of the current frame and previous frame; and
redistribute (305) the feature point clouds into different or same cloud based on the matching information.
8. The method according to claim 3, wherein classifying feature point further comprising:
estimating a statistical disparity distribution of each feature point cloud through mean, mpd or standard deviation;
storing the information when the cloud is determined as a new cloud;
computing the change of disparity of each clouds that are tracked between a current frame and a previous frame;
deteimining the acceleration information of the stereo camera for the instance frame,
wherein when the acceleration is substantially zero, the rate of disparity changes of each tracked cloud in the cun'ent frame over a subsequent frame is computed, the cloud is determined to be a moving object when the rate of change is not substantially zero, and the cloud is determined to be a static object or background when the rate of change is substantially zero; and when the acceleration is not substantially zero, the rate of dispaiity change between a cuirent frame and a previous frame is computed, and the cloud is determined to be a static object or background if the rate of change is increasing/decreasing, and the cloud is determined to be a moving object when the rate of disparity change is substantially zero.
PCT/MY2013/000266 2012-12-13 2013-12-13 Method for non-static foreground feature extraction and classification WO2014092552A2 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
MYPI2012005406A MY172143A (en) 2012-12-13 2012-12-13 Method for non-static foreground feature extraction and classification
MYPI2012005406 2012-12-13

Publications (2)

Publication Number Publication Date
WO2014092552A2 true WO2014092552A2 (en) 2014-06-19
WO2014092552A3 WO2014092552A3 (en) 2014-09-12

Family

ID=50179900

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/MY2013/000266 WO2014092552A2 (en) 2012-12-13 2013-12-13 Method for non-static foreground feature extraction and classification

Country Status (2)

Country Link
MY (1) MY172143A (en)
WO (1) WO2014092552A2 (en)

Cited By (50)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB2539387A (en) * 2015-06-09 2016-12-21 Omg Plc Motion capture system
WO2018097915A1 (en) * 2016-11-22 2018-05-31 Symbol Technologies, Llc Dimensioning system for, and method of, dimensioning freight in motion along an unconstrained path in a venue
US10140725B2 (en) 2014-12-05 2018-11-27 Symbol Technologies, Llc Apparatus for and method of estimating dimensions of an object associated with a code in automatic response to reading the code
US10145955B2 (en) 2016-02-04 2018-12-04 Symbol Technologies, Llc Methods and systems for processing point-cloud data with a line scanner
US10352689B2 (en) 2016-01-28 2019-07-16 Symbol Technologies, Llc Methods and systems for high precision locationing with depth values
US10354411B2 (en) 2016-12-20 2019-07-16 Symbol Technologies, Llc Methods, systems and apparatus for segmenting objects
US10521914B2 (en) 2017-09-07 2019-12-31 Symbol Technologies, Llc Multi-sensor object recognition system and method
US10572763B2 (en) 2017-09-07 2020-02-25 Symbol Technologies, Llc Method and apparatus for support surface edge detection
US10591918B2 (en) 2017-05-01 2020-03-17 Symbol Technologies, Llc Fixed segmented lattice planning for a mobile automation apparatus
US10663590B2 (en) 2017-05-01 2020-05-26 Symbol Technologies, Llc Device and method for merging lidar data
US10721451B2 (en) 2016-03-23 2020-07-21 Symbol Technologies, Llc Arrangement for, and method of, loading freight into a shipping container
US10726273B2 (en) 2017-05-01 2020-07-28 Symbol Technologies, Llc Method and apparatus for shelf feature and object placement detection from shelf images
US10731970B2 (en) 2018-12-13 2020-08-04 Zebra Technologies Corporation Method, system and apparatus for support structure detection
US10740911B2 (en) 2018-04-05 2020-08-11 Symbol Technologies, Llc Method, system and apparatus for correcting translucency artifacts in data representing a support structure
US10776661B2 (en) 2016-08-19 2020-09-15 Symbol Technologies, Llc Methods, systems and apparatus for segmenting and dimensioning objects
US10809078B2 (en) 2018-04-05 2020-10-20 Symbol Technologies, Llc Method, system and apparatus for dynamic path generation
US10823572B2 (en) 2018-04-05 2020-11-03 Symbol Technologies, Llc Method, system and apparatus for generating navigational data
US10832436B2 (en) 2018-04-05 2020-11-10 Symbol Technologies, Llc Method, system and apparatus for recovering label positions
US10949798B2 (en) 2017-05-01 2021-03-16 Symbol Technologies, Llc Multimodal localization and mapping for a mobile automation apparatus
US11003188B2 (en) 2018-11-13 2021-05-11 Zebra Technologies Corporation Method, system and apparatus for obstacle handling in navigational path generation
US11010920B2 (en) 2018-10-05 2021-05-18 Zebra Technologies Corporation Method, system and apparatus for object detection in point clouds
US11015938B2 (en) 2018-12-12 2021-05-25 Zebra Technologies Corporation Method, system and apparatus for navigational assistance
US11042161B2 (en) 2016-11-16 2021-06-22 Symbol Technologies, Llc Navigation control method and apparatus in a mobile automation system
US11080566B2 (en) 2019-06-03 2021-08-03 Zebra Technologies Corporation Method, system and apparatus for gap detection in support structures with peg regions
US11079240B2 (en) 2018-12-07 2021-08-03 Zebra Technologies Corporation Method, system and apparatus for adaptive particle filter localization
US11093896B2 (en) 2017-05-01 2021-08-17 Symbol Technologies, Llc Product status detection system
US11090811B2 (en) 2018-11-13 2021-08-17 Zebra Technologies Corporation Method and apparatus for labeling of support structures
US11100303B2 (en) 2018-12-10 2021-08-24 Zebra Technologies Corporation Method, system and apparatus for auxiliary label detection and association
US11107238B2 (en) 2019-12-13 2021-08-31 Zebra Technologies Corporation Method, system and apparatus for detecting item facings
US11151743B2 (en) 2019-06-03 2021-10-19 Zebra Technologies Corporation Method, system and apparatus for end of aisle detection
US11200677B2 (en) 2019-06-03 2021-12-14 Zebra Technologies Corporation Method, system and apparatus for shelf edge detection
US11327504B2 (en) 2018-04-05 2022-05-10 Symbol Technologies, Llc Method, system and apparatus for mobile automation apparatus localization
US11341663B2 (en) 2019-06-03 2022-05-24 Zebra Technologies Corporation Method, system and apparatus for detecting support structure obstructions
US11367092B2 (en) 2017-05-01 2022-06-21 Symbol Technologies, Llc Method and apparatus for extracting and processing price text from an image set
CN114708304A (en) * 2022-06-06 2022-07-05 苏州浪潮智能科技有限公司 Cross-camera multi-target tracking method, device, equipment and medium
US11392891B2 (en) 2020-11-03 2022-07-19 Zebra Technologies Corporation Item placement detection and optimization in material handling systems
US11402846B2 (en) 2019-06-03 2022-08-02 Zebra Technologies Corporation Method, system and apparatus for mitigating data capture light leakage
US11416000B2 (en) 2018-12-07 2022-08-16 Zebra Technologies Corporation Method and apparatus for navigational ray tracing
US11449059B2 (en) 2017-05-01 2022-09-20 Symbol Technologies, Llc Obstacle detection for a mobile automation apparatus
US11450024B2 (en) 2020-07-17 2022-09-20 Zebra Technologies Corporation Mixed depth object detection
US11507103B2 (en) 2019-12-04 2022-11-22 Zebra Technologies Corporation Method, system and apparatus for localization-based historical obstacle handling
US11506483B2 (en) 2018-10-05 2022-11-22 Zebra Technologies Corporation Method, system and apparatus for support structure depth determination
US11592826B2 (en) 2018-12-28 2023-02-28 Zebra Technologies Corporation Method, system and apparatus for dynamic loop closure in mapping trajectories
US11593915B2 (en) 2020-10-21 2023-02-28 Zebra Technologies Corporation Parallax-tolerant panoramic image generation
US11600084B2 (en) 2017-05-05 2023-03-07 Symbol Technologies, Llc Method and apparatus for detecting and interpreting price label text
US11662739B2 (en) 2019-06-03 2023-05-30 Zebra Technologies Corporation Method, system and apparatus for adaptive ceiling-based localization
US11822333B2 (en) 2020-03-30 2023-11-21 Zebra Technologies Corporation Method, system and apparatus for data capture illumination control
US11954882B2 (en) 2021-06-17 2024-04-09 Zebra Technologies Corporation Feature-based georegistration for mobile computing devices
US11960286B2 (en) 2019-06-03 2024-04-16 Zebra Technologies Corporation Method, system and apparatus for dynamic task sequencing
US11978011B2 (en) 2017-05-01 2024-05-07 Symbol Technologies, Llc Method and apparatus for object status detection

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11847832B2 (en) 2020-11-11 2023-12-19 Zebra Technologies Corporation Object classification for autonomous navigation systems

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP2256690A1 (en) * 2009-05-29 2010-12-01 Honda Research Institute Europe GmbH Object motion detection system based on combining 3D warping techniques and a proper object motion detection

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP2256690A1 (en) * 2009-05-29 2010-12-01 Honda Research Institute Europe GmbH Object motion detection system based on combining 3D warping techniques and a proper object motion detection

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
Cheng Yuan Tang ET AL: "A 3D Feature-Based Tracker for Multiple Object Tracking", Proc. Natl. Sci. Counc. ROC(A), Vol. 21, No. 1, 1 January 1999 (1999-01-01), pages 151-168, XP055126436, Retrieved from the Internet: URL:http://pdf.aminer.org/000/361/024/a_d_predictive_visual_tracker_for_tracking_multiple_moving_objects.pdf [retrieved on 2014-07-02] *
WEI DU ET AL: "Tracking by cluster analysis of feature points using a mixture particle filter", PROCEEDINGS. IEEE CONFERENCE ON ADVANCED VIDEO AND SIGNAL BASED SURVEILLANCE, 2005. COMO, ITALY SEPT. 15-16, 2005, PISCATAWAY, NJ, USA,IEEE, PISCATAWAY, NJ, USA, 15 September 2005 (2005-09-15), pages 165-170, XP010881168, DOI: 10.1109/AVSS.2005.1577261 ISBN: 978-0-7803-9385-1 *

Cited By (54)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10140725B2 (en) 2014-12-05 2018-11-27 Symbol Technologies, Llc Apparatus for and method of estimating dimensions of an object associated with a code in automatic response to reading the code
US10375288B2 (en) 2015-06-09 2019-08-06 Oxford Metrics Plc Motion capture system
GB2539387B (en) * 2015-06-09 2021-04-14 Oxford Metrics Plc Motion capture system
GB2539387A (en) * 2015-06-09 2016-12-21 Omg Plc Motion capture system
US10352689B2 (en) 2016-01-28 2019-07-16 Symbol Technologies, Llc Methods and systems for high precision locationing with depth values
US10145955B2 (en) 2016-02-04 2018-12-04 Symbol Technologies, Llc Methods and systems for processing point-cloud data with a line scanner
US10721451B2 (en) 2016-03-23 2020-07-21 Symbol Technologies, Llc Arrangement for, and method of, loading freight into a shipping container
US10776661B2 (en) 2016-08-19 2020-09-15 Symbol Technologies, Llc Methods, systems and apparatus for segmenting and dimensioning objects
US11042161B2 (en) 2016-11-16 2021-06-22 Symbol Technologies, Llc Navigation control method and apparatus in a mobile automation system
WO2018097915A1 (en) * 2016-11-22 2018-05-31 Symbol Technologies, Llc Dimensioning system for, and method of, dimensioning freight in motion along an unconstrained path in a venue
US10451405B2 (en) 2016-11-22 2019-10-22 Symbol Technologies, Llc Dimensioning system for, and method of, dimensioning freight in motion along an unconstrained path in a venue
US10354411B2 (en) 2016-12-20 2019-07-16 Symbol Technologies, Llc Methods, systems and apparatus for segmenting objects
US10591918B2 (en) 2017-05-01 2020-03-17 Symbol Technologies, Llc Fixed segmented lattice planning for a mobile automation apparatus
US10726273B2 (en) 2017-05-01 2020-07-28 Symbol Technologies, Llc Method and apparatus for shelf feature and object placement detection from shelf images
US11978011B2 (en) 2017-05-01 2024-05-07 Symbol Technologies, Llc Method and apparatus for object status detection
US11093896B2 (en) 2017-05-01 2021-08-17 Symbol Technologies, Llc Product status detection system
US10663590B2 (en) 2017-05-01 2020-05-26 Symbol Technologies, Llc Device and method for merging lidar data
US10949798B2 (en) 2017-05-01 2021-03-16 Symbol Technologies, Llc Multimodal localization and mapping for a mobile automation apparatus
US11449059B2 (en) 2017-05-01 2022-09-20 Symbol Technologies, Llc Obstacle detection for a mobile automation apparatus
US11367092B2 (en) 2017-05-01 2022-06-21 Symbol Technologies, Llc Method and apparatus for extracting and processing price text from an image set
US11600084B2 (en) 2017-05-05 2023-03-07 Symbol Technologies, Llc Method and apparatus for detecting and interpreting price label text
US10572763B2 (en) 2017-09-07 2020-02-25 Symbol Technologies, Llc Method and apparatus for support surface edge detection
US10521914B2 (en) 2017-09-07 2019-12-31 Symbol Technologies, Llc Multi-sensor object recognition system and method
US10832436B2 (en) 2018-04-05 2020-11-10 Symbol Technologies, Llc Method, system and apparatus for recovering label positions
US11327504B2 (en) 2018-04-05 2022-05-10 Symbol Technologies, Llc Method, system and apparatus for mobile automation apparatus localization
US10823572B2 (en) 2018-04-05 2020-11-03 Symbol Technologies, Llc Method, system and apparatus for generating navigational data
US10809078B2 (en) 2018-04-05 2020-10-20 Symbol Technologies, Llc Method, system and apparatus for dynamic path generation
US10740911B2 (en) 2018-04-05 2020-08-11 Symbol Technologies, Llc Method, system and apparatus for correcting translucency artifacts in data representing a support structure
US11010920B2 (en) 2018-10-05 2021-05-18 Zebra Technologies Corporation Method, system and apparatus for object detection in point clouds
US11506483B2 (en) 2018-10-05 2022-11-22 Zebra Technologies Corporation Method, system and apparatus for support structure depth determination
US11003188B2 (en) 2018-11-13 2021-05-11 Zebra Technologies Corporation Method, system and apparatus for obstacle handling in navigational path generation
US11090811B2 (en) 2018-11-13 2021-08-17 Zebra Technologies Corporation Method and apparatus for labeling of support structures
US11079240B2 (en) 2018-12-07 2021-08-03 Zebra Technologies Corporation Method, system and apparatus for adaptive particle filter localization
US11416000B2 (en) 2018-12-07 2022-08-16 Zebra Technologies Corporation Method and apparatus for navigational ray tracing
US11100303B2 (en) 2018-12-10 2021-08-24 Zebra Technologies Corporation Method, system and apparatus for auxiliary label detection and association
US11015938B2 (en) 2018-12-12 2021-05-25 Zebra Technologies Corporation Method, system and apparatus for navigational assistance
US10731970B2 (en) 2018-12-13 2020-08-04 Zebra Technologies Corporation Method, system and apparatus for support structure detection
US11592826B2 (en) 2018-12-28 2023-02-28 Zebra Technologies Corporation Method, system and apparatus for dynamic loop closure in mapping trajectories
US11151743B2 (en) 2019-06-03 2021-10-19 Zebra Technologies Corporation Method, system and apparatus for end of aisle detection
US11341663B2 (en) 2019-06-03 2022-05-24 Zebra Technologies Corporation Method, system and apparatus for detecting support structure obstructions
US11960286B2 (en) 2019-06-03 2024-04-16 Zebra Technologies Corporation Method, system and apparatus for dynamic task sequencing
US11662739B2 (en) 2019-06-03 2023-05-30 Zebra Technologies Corporation Method, system and apparatus for adaptive ceiling-based localization
US11402846B2 (en) 2019-06-03 2022-08-02 Zebra Technologies Corporation Method, system and apparatus for mitigating data capture light leakage
US11200677B2 (en) 2019-06-03 2021-12-14 Zebra Technologies Corporation Method, system and apparatus for shelf edge detection
US11080566B2 (en) 2019-06-03 2021-08-03 Zebra Technologies Corporation Method, system and apparatus for gap detection in support structures with peg regions
US11507103B2 (en) 2019-12-04 2022-11-22 Zebra Technologies Corporation Method, system and apparatus for localization-based historical obstacle handling
US11107238B2 (en) 2019-12-13 2021-08-31 Zebra Technologies Corporation Method, system and apparatus for detecting item facings
US11822333B2 (en) 2020-03-30 2023-11-21 Zebra Technologies Corporation Method, system and apparatus for data capture illumination control
US11450024B2 (en) 2020-07-17 2022-09-20 Zebra Technologies Corporation Mixed depth object detection
US11593915B2 (en) 2020-10-21 2023-02-28 Zebra Technologies Corporation Parallax-tolerant panoramic image generation
US11392891B2 (en) 2020-11-03 2022-07-19 Zebra Technologies Corporation Item placement detection and optimization in material handling systems
US11954882B2 (en) 2021-06-17 2024-04-09 Zebra Technologies Corporation Feature-based georegistration for mobile computing devices
WO2023236514A1 (en) * 2022-06-06 2023-12-14 苏州元脑智能科技有限公司 Cross-camera multi-object tracking method and apparatus, device, and medium
CN114708304A (en) * 2022-06-06 2022-07-05 苏州浪潮智能科技有限公司 Cross-camera multi-target tracking method, device, equipment and medium

Also Published As

Publication number Publication date
MY172143A (en) 2019-11-14
WO2014092552A3 (en) 2014-09-12

Similar Documents

Publication Publication Date Title
WO2014092552A2 (en) Method for non-static foreground feature extraction and classification
US11195038B2 (en) Device and a method for extracting dynamic information on a scene using a convolutional neural network
CN104751491B (en) A kind of crowd's tracking and people flow rate statistical method and device
CN103824070B (en) A kind of rapid pedestrian detection method based on computer vision
US8744122B2 (en) System and method for object detection from a moving platform
CN117095349A (en) Appearance search system, method, and non-transitory computer readable medium
CN102867386B (en) Intelligent video analysis-based forest smoke and fire detection method and special system thereof
CN107657244B (en) Human body falling behavior detection system based on multiple cameras and detection method thereof
CN104966062B (en) Video monitoring method and device
CN112514373B (en) Image processing apparatus and method for feature extraction
CN108596157B (en) Crowd disturbance scene detection method and system based on motion detection
Denman et al. Multi-spectral fusion for surveillance systems
CN111145223A (en) Multi-camera personnel behavior track identification analysis method
CN110544271B (en) Parabolic motion detection method and related device
CN103336947A (en) Method for identifying infrared movement small target based on significance and structure
Malhi et al. Vision based intelligent traffic management system
CN111833380B (en) Multi-view image fusion space target tracking system and method
CN112270253A (en) High-altitude parabolic detection method and device
KR20140132140A (en) Method and apparatus for video surveillance based on detecting abnormal behavior using extraction of trajectories from crowd in images
Zhang et al. Real-time moving object classification with automatic scene division
KR101214858B1 (en) Moving object detecting apparatus and method using clustering
Sharma et al. Automatic vehicle detection using spatial time frame and object based classification
Piérard et al. A probabilistic pixel-based approach to detect humans in video streams
Xu et al. A novel method for people and vehicle classification based on Hough line feature
Benedek et al. Lidar-based gait analysis in people tracking and 4D visualization

Legal Events

Date Code Title Description
122 Ep: pct application non-entry in european phase

Ref document number: 13831952

Country of ref document: EP

Kind code of ref document: A2