CN101026759A - Visual tracking method and system based on particle filtering - Google Patents

Visual tracking method and system based on particle filtering Download PDF

Info

Publication number
CN101026759A
CN101026759A CN 200710090883 CN200710090883A CN101026759A CN 101026759 A CN101026759 A CN 101026759A CN 200710090883 CN200710090883 CN 200710090883 CN 200710090883 A CN200710090883 A CN 200710090883A CN 101026759 A CN101026759 A CN 101026759A
Authority
CN
China
Prior art keywords
many visual
visual signatures
synthetic weights
signatures
current
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN 200710090883
Other languages
Chinese (zh)
Other versions
CN100571392C (en
Inventor
杨杰
程建
凌建国
张翼
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Huawei Technologies Co Ltd
Original Assignee
Huawei Technologies Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Huawei Technologies Co Ltd filed Critical Huawei Technologies Co Ltd
Priority to CN 200710090883 priority Critical patent/CN100571392C/en
Publication of CN101026759A publication Critical patent/CN101026759A/en
Application granted granted Critical
Publication of CN100571392C publication Critical patent/CN100571392C/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Landscapes

  • Image Analysis (AREA)
  • Image Processing (AREA)

Abstract

The method includes steps: based on calculation of synthesized weight value set of current multiple vision characters to obtain filtering estimation of interfusion state of multiple vision characters in current sub frame; the said filtering estimation includes information of target position; and based on information of target position to determine target position of current sub frame of vision track. The invention also discloses vision track system based on particle filtering. An instance in the invention embeds democracy synthesis mechanism to state estimation procedure of particle filtering in order to implement vision track. The invention raises reliability and robustness of vision track.

Description

Visual tracking method and system based on particle filter
Technical field
The present invention relates to the vision tracking technique, be specifically related to visual tracking method and system based on particle filter.
Background technology
The essence that vision is followed the tracks of is that recursion estimates interested position with target of specific image feature in video sequence, and characteristics of image comprises color, shape, texture and motion etc.The vision tracking problem can be considered the state estimation problem of recursion.That is to say to obtain the sensation target state estimation, just can determine the position of tracking target.In the prior art, there are following several visual tracking methods:
In the prior art, merge angle from many visual signatures self adaptation, self-organizing and proposed democracy synthetic method (Democratic Integration), and apply to face tracking.The democracy synthetic method refers to a kind of many visual signatures self-organizing blending algorithm, can organize together correlated characteristic well, brings into play bigger effect.The visual tracking method that the democracy synthetic method is applied to people's face is specially: at each pixel in the image, carry out many Feature Extraction, obtain the similarity degree of each pixel representative target and real goal by the democracy synthetic method, find the center of the highest pixel of similarity degree then, to realize tracing process as target.This is actually an exhaustive process this method, lacks effective theoretical frame of realizing that vision is followed the tracks of, causes vision track algorithm efficient lower.And, find also that by experiment this method is to the discontinuous fluid sensitivity of sensation target; To the interference sensitivity of false target, in case generation because of the tracking error that false target produces, will be difficult to recover.Like this, the visual tracking method of prior art has caused the reliability and the robustness of vision tracking lower.
This shows that there is the shortcoming of the low and robustness difference of reliability in the visual tracking method of prior art.
Summary of the invention
The embodiment of the invention provides a kind of visual tracking method based on particle filter, and this method can improve reliability and the robustness that vision is followed the tracks of.
The embodiment of the invention provides a kind of vision tracking system based on particle filter, and this system can improve reliability and the robustness that vision is followed the tracks of.
A kind of visual tracking method based on particle filter, this method comprises:
Many visual signatures that synthetic weights set calculates current subframe according to current many visual signatures merge state filtering to be estimated, described many visual signatures merge state filtering estimates to comprise target position information; And
Determine the target location of current sub frame of vision track according to described target position information.
A kind of vision tracking system based on particle filter, this system comprise that synthetic weights acquisition module, many visual signatures merge state filtering estimation module and target location determination module;
Synthetic weights acquisition module, the set of obtaining the synthetic weights of current many visual signatures sends many visual signatures to and merges the state filtering estimation module;
Many visual signatures merge the state filtering estimation module, the set that the current many visual signatures that transmit according to synthetic weights acquisition module synthesize weights, and the many visual signatures that calculate current subframe merge state filterings to be estimated;
The target location determination module merges state filtering according to described many visual signatures and estimates, determines the target location of current subframe.
From such scheme as can be seen, the embodiment of the invention obtains the set of the synthetic weights of many visual signatures, set according to the synthetic weights of many visual signatures calculates the estimation of many visual signatures fusion state filterings, merges the position that state filterings estimate to determine current subframe target by many visual signatures.Like this, follow the tracks of to realize vision, improved reliability and robustness that vision is followed the tracks of by the state estimation procedure that the democracy synthesis mechanism is embedded into particle filter.
Description of drawings
Fig. 1 is the structural representation of the embodiment of the invention based on the vision tracking system of particle filter;
Fig. 2 is the principle flow chart of the embodiment of the invention based on the visual tracking method of particle filter;
Fig. 3 carries out the method flow diagram of a specific embodiment of vision tracking for utilization principle shown in Figure 2;
Fig. 4 is the tracing process schematic diagram of the embodiment of the invention based on the visual tracking method of particle filter.
Embodiment
Please referring to Fig. 1, be the structural representation of the embodiment of the invention based on the vision tracking system of particle filter, this system comprises that synthetic weights acquisition module 110, many visual signatures merge state filtering estimation module 120 and target location determination module 130.
Synthetic weights acquisition module 110, the set of obtaining the synthetic weights of current many visual signatures sends many visual signatures to and merges state filtering estimation module 120.
Many visual signatures merge state filtering estimation module 120, the set that the current many visual signatures that transmit according to synthetic weights acquisition module 110 synthesize weights, and the many visual signatures that calculate current subframe merge state filterings to be estimated.
Target location determination module 130 merges many visual signatures fusion state filterings estimation that state filtering estimation module 120 calculates according to many visual signatures, determines the target location of current subframe.
Synthetic weights acquisition module 110 can comprise single vision significant condition filtering estimator module 111 and state estimation deviation calculation submodule 112, quality metrics calculating sub module 113 and synthetic weights calculating sub module 114.
Single vision significant condition filtering estimator module 111 is obtained the single vision significant condition filtering of current subframe and is estimated.
The single vision significant condition filtering of obtaining current subframe is estimated as: according to the weights of sampling particle, calculate the filtering of single vision significant condition and estimate that its concrete computational process is referring to the description at step 201 place.
State estimation deviation calculation submodule 112, estimate according to many visual signatures fusion state filterings that the filtering of single vision significant condition is estimated and many visual signatures fusion state filtering estimation module 120 transmits that single vision significant condition filtering estimator module 111 transmits, calculate the state estimation deviation.Its concrete computational process is referring to the description at step 204 place.
Quality metrics calculating sub module 113, the state estimation deviation according to state estimation deviation calculation submodule meter 112 obtains calculates the visual signature quality metrics.Its concrete computational process is referring to the description at step 204 place.
Synthetic weights calculating sub module 114 according to the vision quality tolerance that quality metrics calculating sub module 113 obtains, calculates the set of the synthetic weights of current many visual signatures.Its concrete computational process is referring to the description at step 205 place.
The vision tracking system of the embodiment of the invention can also comprise that characteristic disappear judging module 140 and feature produce judging module 150.
Characteristic disappear judging module 140, whether the interior synthetic weights of each many visual signatures of set of judging the synthetic weights of many visual signatures that synthetic weights acquisition module 110 transmits are less than or equal to judgment value, if be less than or equal to judgment value, then disappear, the synthetic weights of these many visual signatures are deleted from the set of the synthetic weights of many visual signatures with the synthetic weights characteristic of correspondence of these many visual signatures.This judgment value is a constant that sets in advance, and can be 0, also can be other numerical value.The concrete judging process of characteristic disappear is referring to the description at step 3031 place.
Feature generation module 150, the set characteristic of correspondence number of judging the synthetic weights of many visual signatures that synthetic weights acquisition module 110 transmits less than the visual signature threshold value after, produce feature indication, for described feature indication is provided with the synthetic weights of corresponding many visual signatures, the synthetic weights of many visual signatures of this setting are added the set of the synthetic weights of current many visual signatures.The synthetic weights of specifically generation feature indication, and many visual signatures that will be provided with add the process of the set of the synthetic weights of current many visual signatures, referring to the description at step 3032 and 3033 places.
Please referring to Fig. 2, be the principle flow chart of the embodiment of the invention based on the visual tracking method of particle filter, this method may further comprise the steps:
Step 201 obtains the filtering of single vision significant condition and estimates.Single vision significant condition filtering estimated statement is shown
Figure A20071009088300091
Be the k sensation target state filtering estimation of m visual signature constantly, be used for step 204 and calculate the state estimation deviation.Its procurement process is specially:
At first, for m visual signature, calculate sampling particle X k (i)Weight w k (i), m, calculation expression is:
w k ( i ) , m ∝p ( Z k ( i ) , m | X k ( i ) ) - - - ( 1 )
Wherein, m=1,2......M, the sequence number of m representation feature, i represents to adopt the sequence number of particle, and k represents constantly, if k=0, the expression initial time, then current subframe is initial subframe; w k (i), mThe weights of representing m feature, Z k (i), mThe observation situation of representing m visual signature; P (Z k (i), m| X k (i)) expression m visual signature observation probability, ∝ represents to be proportional to.
Then, according to the sampling particle weight w k (i), m, the sensation target state filtering that obtains m visual signature estimates that single vision significant condition filtering is just estimated
Figure A20071009088300101
Calculation expression is:
X ^ k m = Σ i = 1 N w k ( i ) , m X k ( i ) - - - ( 2 )
Step 202, the set of synthesizing weights by many visual signatures calculates the estimation of many visual signatures fusion state filterings
Figure A20071009088300103
To obtain the set of the synthetic weights of many visual signatures after the vision of last subframe is followed the tracks of and finished, it is retrieved as the process of step 204 and 205.If current subframe is first subframe, then the set of the synthetic weights of pre-configured many visual signatures is synthesized the set of weights as described many visual signatures.The process that obtains the estimation of many visual signatures fusion state filterings is specially:
At first, definition sampling particle X k (i)Observation probability p (Z k (i)| X k (i)) calculation expression, for:
p ( Z k ( i ) | X k ( i ) ) = Σ m = 1 , . . . , M π k m p ( Z k ( i ) , m | X k ( i ) ) - - - ( 3 )
Wherein, m=1,2......M, π k mThe synthetic weights of many visual signatures of representing m visual signature of last subframe are called the set that many visions are synthesized weights with the synthetic weights of all many visual signatures of M visual signature correspondence of last subframe, and Σ m = 1 , . . . , M π k m = 1 ; P (Z k (i)| X k (i)) expression sampling particle X k (i)Observation probability.Here, (3) formula is called the weighting synthetic model.
Then, with the synthetic weights π of many visual signatures of last subframe k m, substitution (3) formula, particle X obtains sampling k (i)Observation probability p (Z k (i)| X k (i)), with the observation probability p (Z that calculates k (i)| X k (i)) substitution (4) formula, by the proportional relation of (4), obtain many visual signatures and merge down-sampling particle X k (i)Weight w k (i)
w k ( i ) ∝ p ( Z k ( i ) | X k ( i ) ) - - - ( 4 )
At last, according to the sampling particle X that calculates by (4) formula k (i)Weight w k (i), calculate many visual signatures and merge sensation target state filtering estimation down
Figure A20071009088300107
Calculation expression is:
X ^ k = Σ i = 1 N w k ( i ) X k ( i ) - - - ( 5 )
Many visual signatures merge sensation target state filtering estimation down
Figure A20071009088300112
Be used for step 203 and determine the target location of current subframe, and be used for step 204 calculating vision quality tolerance.
Step 203 merges state estimation according to the many visual signatures that obtain
Figure A20071009088300113
Determine the target location of current subframe.
Many visual signatures of the current subframe that obtains merge state estimation The information that has comprised the target location of current subframe, according to Just can determine the target location of current subframe.
For many visual signatures, the state filtering of sensation target estimates to exist two levels, i.e. single vision significant condition filtering is estimated and many visual signatures merge the state filtering estimation.Step 201 is for calculating the process that the filtering of single vision significant condition is estimated, step 202 merges the process that state filterings are estimated for calculating many visual signatures, and step 201 and step 202 can be carried out simultaneously, also can first execution in step 202 and 203, and execution in step 201 again.
Step 204 obtains visual signature quality metrics q k m, its concrete procurement process comprises:
At first, the single vision significant condition filtering that calculates according to step 201 is estimated
Figure A20071009088300116
The many visual signatures that calculate with step 202 merge the state filtering estimation
Figure A20071009088300117
Obtain the state estimation deviation of two levels, the state estimation deviation chart is shown d k m, calculation expression is:
d k m = | | X ^ k - X ^ k m | | 2 - - - ( 6 )
Wherein, ‖ ‖ 2Be the Euclidean norm.
Then, according to the state estimation deviation d that tries to achieve k m, obtain visual signature quality metrics q k m, calculation expression is:
q k m = A ( d k m ) - - - ( 7 )
Wherein, A () arrives the interior mapping of (0,1) scope for set of real numbers, 0 ≤ q k m ≤ 1 . q k mReflected the contribution rate of visual signature m to the vision tracking results.With (d k m) to multiply by A be in order to make the visual signature quality metrics that obtains Paint Gloss.A is an empirical value.Mapping A () may be defined as negative exponential function, or other functions such as 0-1 function.Here, suppose to shine upon A () and be defined as the 0-1 function, be:
q k m = A ( d k m ) = 1 if d k m ≤ δ 0 if d k m ≥ δ - - - ( 8 )
Wherein, δ is the state estimation deviation threshold, and typical state estimation deviation threshold can be taken as 4 pixels, i.e. δ=4 pixels. q k m = 1 Expression k moment visual signature m follows the tracks of vision contribution preferably, and q k m = 0 Expression k visual signature m constantly follows the tracks of the nothing contribution to vision.Under this 0-1 mapping, visual signature quality metrics q k mInitialization value, q 0 m, determine according to visual signature characteristic priori in the scene.
Step 205 is according to visual signature quality metrics q k mObtain the set of the synthetic weights of many visual signatures.The set of the synthetic weights of the many visual signatures that calculate is used to calculate many visual signatures fusion state filterings estimation of next subframe.The process that obtains the set of the synthetic weights of many visual signatures specifically comprises:
At first, under the particle filter theoretical frame, will be based on the synthetic weights π of many visual signatures of visual signature quality metrics k mThe self-organizing correction chart be shown
Figure A20071009088300124
It is defined as:
τ π · k m = q k m - π k m - - - ( 9 )
In actual vision tracing process, adopt the difference form of (9) formula to represent through the synthetic weights π of the revised many visual signatures of self-organizing k m, expression formula is:
π k m = π k - 1 m + β ( q k - 1 m - π k - 1 m ) - - - ( 10 )
Wherein, q k - 1 m - π k - 1 m = τ · π k - 1 m , β is synthetic weights adjusted rate, and typically synthetic weights adjusted rate value is 0.05, i.e. β=0.05.
Then, to the synthetic weights π of the many visual signatures that obtain k mCarry out the normalization computing.
Because (8) formula is calculated q k mProcess in, carry out [0,1] mapping computing make q k mNo longer have the normalization feature, thereby make the π that calculates by (10) formula k mAlso no longer have the normalization feature, so, need be to the synthetic weights π of many visual signatures k mCarry out normalization, method for normalizing is: to the m value is 1, the M of 2......M the synthetic weights π of many visual signatures k mSummation, the synthetic weights of many visual signatures that use m feature correspondence are divided by this summing value, just obtain the synthetic weights of many visual signatures after the normalization, the m value is 1, and the synthetic weights of the many visual signatures after the M of the 2......M normalization are formed the set of the synthetic weights of many visual signatures after the normalization.Normalized computing expression is: π k m / Σ m = 1 M π k m .
Normalization computing described here is the normalization of broad sense, can be the computing that is classified as " 1 ", can be the computing that is classified as " 2 " also, or is classified as the computing of other numerical value.
Step 204 and 205 is the π after the normalization that calculates k mThe set vision that will apply to next subframe follow the tracks of in the computing.Just, merging state filtering at the many visual signatures that calculate next subframe estimates The time, with the π after the normalization k mIn substitution (3) formula, continue execution in step 201~205.
Such scheme is the basic principle of visual tracking method of the present invention.Among Fig. 2, step 201~203 estimate for single vision significant condition filtering of the present invention and many visual signatures merge state filtering estimation procedure, just particle filter process; Step 204 and 205 is for adopting the democracy composition algorithm to obtain the process of the synthetic weights of many visual signatures.Below in conjunction with specific embodiment shown in Figure 3, the described method of Fig. 2 is elaborated.
Please referring to Fig. 3, carry out the method flow diagram of a specific embodiment of vision tracking for using principle shown in Figure 2, this method may further comprise the steps:
Step 301 is carried out particle state and is shifted.
Carrying out the particle state transfer is specially: the particle that is obtained current time by the sampling calculating particles of specific state transition model and last subframe.Here, the sampling particle with last subframe is expressed as X K-1 (i), the sampling particle that will work as last subframe is expressed as The sampling particle set representations of last subframe is { X ~ k ( i ) , w k ( i ) } i = 1 N . The method that is obtained the particle of current time by the sampling calculating particles of specific state transition model and last subframe is a process well known to those skilled in the art, repeats no more here.Described state transition model can be the state equation of motion.
If current subframe is the initial subframe that vision is followed the tracks of, the process of then carrying out the particle state transfer is the initialization procedure that vision is followed the tracks of, and this process specifically comprises:
Initial subframe is promptly during k=0.According to prior distribution p (X 0) determine initial sampling particle collection, be expressed as { X 0 ( i ) , 1 N } i = 1 N ; For sampling particle collection, the many visual signatures of initialization synthesize weights, and hypothesis is shown the synthetic weight table of initialized many visual signatures here π 0 m = 1 M , Wherein, m=1 ..., M, is called the summation of M the synthetic weights of many visual signatures the set of the synthetic weights of many visual signatures here; According to the visual signature characteristic of tracked sensation target in the scene, initialization visual signature quality metrics is expressed as q 0 mDetermine state estimation deviation threshold value and and synthetic weights adjusted rate, hypothesis is respectively 4 and 0.05 here, i.e. δ=4 and β=0.05.The process of synthetic weights of the many visual signatures of initialization and initialization visual signature quality metrics, the process of synthetic weights of just pre-configured many visions and visual signature quality metrics.
Step 302 is calculated the filtering of single vision significant condition and is estimated
Figure A20071009088300142
The value of m is: m=1 ..., M.Calculate the filtering of single vision significant condition according to step 201 and estimate that detailed process comprises:
At first, to m feature, calculate the sampling particle
Figure A20071009088300143
Weight w k (i), m, the calculating expression is:
w k ( i ) , m ∝ p ( Z ~ k ( i ) , m | X k ( i ) ) - - - ( 11 )
Wherein, m=1,2......M, the sequence number of m representation feature, i represents to adopt the sequence number of particle, and k represents constantly, if k=0, the expression initial time, then current subframe is initial subframe; w k (i), mThe weights of representing m feature, Z k (i), mThe observation situation of representing m visual signature; P (Z k (i), m| X k (i)) expression m visual signature observation probability, ∝ represents to be proportional to.
Then, the weights that calculate are carried out normalization, method for normalizing is, is 1 to the i value, the N of a 2......N weight w k (i), mSummation, the weights that use i particle correspondence are divided by this summing value, just obtain the weights after the normalization, and the computing expression is:
w k ( i ) , m / Σ i = 1 N w k ( i ) , m - - - ( 12 )
Normalization computing described here is the normalization of broad sense, can be the computing that is classified as " 1 ", can be the computing that is classified as " 2 " also, or is classified as the computing of other numerical value.
At last, the weight w that will calculate by (12) formula k (i), mIn substitution (13) formula, the sensation target state filtering that obtains m visual signature estimates that single vision significant condition filtering is just estimated
Figure A20071009088300146
Calculation expression is:
X ^ k m = Σ i = 1 N w k ( i ) , m X ~ k ( i ) - - - ( 13 )
Step 303 calculates many visual signatures and merges the weight w of particle down k (i), concrete computational process is:
Step 3031 is carried out visual signature disappearance judgement.
Supposing to can be used for the visual signature number that vision follows the tracks of is M, and each visual signature has a feature indication, is Ω with the feature indication set representations of M feature indication formation.Suppose that it is M that k is used for the characteristic that vision follows the tracks of constantly k, with M kThe feature indication set representations that individual feature indication constitutes is Ω kTo be used for visual signature that vision the follows the tracks of threshold value table that rolls off the production line and be shown M 0Here, establish M 0Be representative value 2.
Visual signature disappears to adjudicate and is specially: whether the interior synthetic weights of each many visual signatures of set of judging the synthetic weights of current many visual signatures are less than or equal to judgment value, if be less than or equal to judgment value, then disappear, the synthetic weights of these many visual signatures are deleted from the set of the synthetic weights of described many visual signatures with the synthetic weights characteristic of correspondence of these many visual signatures.This judgment value is a constant that sets in advance, and can be 0, also can be other numerical value.For example, to the synthetic weights π of many visual signatures of a certain feature m k mJudge, if judge π k mBe less than or equal to 0, promptly π k m ≤ 0 , Representation feature m disappears, and then the synthetic weights of many visual signatures of this feature correspondence are deleted from the set of the synthetic weights of many visual signatures, with this feature indication from Ω kMiddle deletion, and order M k = Ω = k ; Otherwise this feature m exists.Wherein, Ω kRepresentation feature indicates collection Ω kGesture.
Above-mentioned many visual signatures synthesize weights π k m, for last subframe is carried out the value that vision calculates in following the tracks of, π k mMay be one and approach very much zero value, also may be negative value.
Step 3032 is carried out visual signature and is produced judgement, and its detailed process comprises:
(1) to characteristic M kJudge, if characteristic M kLess than the visual signature lower threshold M that is provided with 0, i.e. M k<M 0, then carry out (2); Otherwise, execution in step 3033.
(2) produce random number, be expressed as u.This random number can be that (0,1) goes up equally distributed random number, also can be other interval arbitrarily random number that distributes that goes up.
(3) u be multiply by the long-pending of M and round, rounding the result is feature indication, is expressed as label.The process of rounding can be expressed as: label=round (uM).
The corresponding feature of feature indication label.
(4) whether judging characteristic sign label belongs to feature indication collection Ω k, if do not belong to, i.e. label  Ω k, then this feature indication label characteristic of correspondence is effective, and feature indication label is added feature indication collection Ω kIn, and order M k = Ω = k , Execution in step (5); Otherwise this feature indication label characteristic of correspondence is invalid, execution in step (2).
(5) according to the characteristic of label characteristic of correspondence in the scene, add feature indication collection Ω in the initialization step (4) kFeature vision quality tolerance and synthesize weights, the synthetic weights of initialized many visual signatures are added in the set of the synthetic weights of many visual signatures execution in step (1).In initial method and the step 301 to initial subframe initialization identical, repeat no more here.
Certain feature is chosen for concentrating from feature indication at random in step (2)~(5), and to the vision quality tolerance of this feature with synthesize weights and carry out initialized process.
Generate judgement by visual signature disappearance judgement and visual signature, reduced the computation complexity that many visual signatures merge tracking, improved the real-time of track algorithm.Step 303 can not comprise step 3031 and step 3032.
Step 3033, the synthetic weights π of each the many visual signatures in the set of the synthetic weights of many visual signatures that step (5) is obtained k mSubstitution weighting synthetic model, promptly formula (3) calculates sampling particle X k (i)Observation probability p (Z k (i)| X k (i)).With the observation probability p (Z that calculates k (i)| X k (i)) substitution (4) formula, obtain many visual signatures and merge down-sampling particle X k (i)Weight w k (i)To weight w k (i)Carry out normalization, method for normalizing is, is 1 to the i value, the N of a 2......N weight w k (i), m summation, the weights that use i particle correspondence are divided by this summing value, just obtain the weights after the normalization, and the computing expression is:
w k ( i ) , m / Σ i = 1 N w k ( i ) , m .
Normalization computing described here is the normalization of broad sense, can be the computing that is classified as " 1 ", can be the computing that is classified as " 2 " also, or is classified as the computing of other numerical value.
Step 302 and step 303 can be carried out simultaneously, also can first execution in step 303 execution in step 302 again.
Step 304 merges the many visual signatures after the normalization that calculates in the step 303 with the weight w of particle down k (i)Substitution formula (5), the many visual signatures that obtain current subframe merge state filtering to be estimated
Figure A20071009088300171
Calculation expression is: X ^ k m Σ i = 1 N w k ( i ) , m X k ( i ) .
Step 305 is estimated according to many visual signatures fusion state filterings of the current subframe that calculates in the step 304
Figure A20071009088300173
Determine the target location of current subframe.
Many visual signatures of the current subframe that obtains merge state estimation
Figure A20071009088300174
The information that has comprised the target location of current subframe, according to
Figure A20071009088300175
Just can determine the target location of current subframe.
Step 306 calculates the visual signature quality metrics q of current subframe k m, concrete computational process may further comprise the steps:
At first, the single vision significant condition filtering that calculates according to step 302 is estimated
Figure A20071009088300176
The normalized many visual signatures that calculate with step 304 merge the state filtering estimation
Figure A20071009088300177
Calculate the state estimation deviation d of two levels k m, calculation expression is:
d k m = | | X ^ k - X ^ k m | | 2
Wherein, ‖ ‖ 2Be the Euclidean norm.
Then, according to the state estimation deviation d that tries to achieve k m, obtain visual signature quality metrics q k m, calculation expression is:
q k m = A ( d k m )
Wherein, A () arrives the interior mapping of (0,1) scope for set of real numbers, 0 ≤ q k m ≤ 1 . q k mReflected the contribution rate of visual signature m to the vision tracking results.With (d k m) to multiply by A be in order to make the visual signature quality metrics that obtains Paint Gloss.A is an empirical value.Mapping A () may be defined as negative exponential function, or other functions such as 0-1 function.Here, suppose to shine upon A () and be defined as the 0-1 function, be:
q k m = A ( d k m ) = 1 if d k m ≤ δ 0 if d k m ≥ δ
Wherein, δ is the state estimation deviation threshold, and typical state estimation deviation threshold can be taken as 4 pixels, i.e. δ=4 pixels. q k m = 1 Expression k moment visual signature m follows the tracks of vision contribution preferably, and q k m = 0 Expression k visual signature m constantly follows the tracks of the nothing contribution to vision.Under this 0-1 mapping, visual signature quality metrics q k mInitial value, q 0 m, determine according to visual signature characteristic priori in the scene.
Step 307 is according to visual signature quality metrics q k mObtain the synthetic weights π of many visual signatures of current subframe k mThe many visual signatures that calculate synthesize weights π k mBe used to calculate many visual signatures fusion state filterings estimation of next subframe.Obtain the synthetic weights π of many visual signatures of current subframe k mProcess specifically comprise:
At first, under the particle filter theoretical frame, will be based on the synthetic weights π of many visual signatures of visual signature quality metrics k mThe self-organizing correction be defined as:
τ π · k m = q k m - π k m
In actual vision tracing process, adopt the difference form of (9) formula to represent through the synthetic weights π of the revised many visual signatures of self-organizing k m, expression formula is:
π k m = π k - 1 m + β ( q k - 1 m - π k - 1 m )
Wherein, q k - 1 m - π k - 1 m = τ · π k - 1 m , β is synthetic weights adjusted rate, and typically synthetic weights adjusted rate value is 0.05, i.e. β=0.05.
Then, to the synthetic weights π of the many visual signatures that obtain k mCarry out the normalization computing.
Because (8) formula is calculated q k mProcess in, carry out [0,1] mapping computing make q k mNo longer have the normalization feature, thereby make the π that calculates by (10) formula k mAlso no longer have the normalization feature, so, need be to π k mCarry out the normalization computing, the m value is 1, and the synthetic weights of the many visual signatures after the M of the 2......M normalization are formed the set of the synthetic weights of many visual signatures after the normalization.Operation expression is:
π k m / Σ m = 1 M π k m .
Normalization computing described here is the normalization of broad sense, can be the computing that is classified as " 1 ", can be the computing that is classified as " 2 " also, or is classified as the computing of other numerical value.
Step 306 and 307 is in order to calculate the set of the synthetic weights of many visual signatures after the normalization.The set of the synthetic weights of the many visual signatures after the normalization will be used with the vision of next subframe and be followed the tracks of in the computing.
Step 308 is carried out particle and is resampled.
Merge the weight w of particle down according to the many visual signatures after the normalization that calculates in the step 3033 k (i), from sampling particle collection { X ~ k ( i ) , w k ( i ) } i = 1 N In extract N particle again, form new sampling particle collection, be expressed as { X k ( i ) , 1 N } i = 1 N .
Behind k=k+1, execution in step 301 is carried out the vision of next subframe and is followed the tracks of computing.
Please, be the tracing process schematic diagram of the embodiment of the invention based on the visual tracking method of particle filter referring to Fig. 4.The target that Fig. 4 follows the tracks of is the head portrait of Ms in the image, and as can be seen, embodiment of the invention tracing process is locked as Ms's head portrait always.As seen, the visual tracking method of the utilization embodiment of the invention has higher reliability and robustness than prior art.
The technical scheme of the embodiment of the invention can be applicable to video monitoring, video compression coding, robot navigation and location, intelligent human-machine interaction, virtual reality, and field such as Imaging Guidance.
The technical scheme of the embodiment of the invention obtains the set of the synthetic weights of many visual signatures by the democracy composition algorithm; Calculate after many visual signatures merge state filterings and estimate according to the set of the synthetic weights of many visual signatures again, determine the target location that the vision of current subframe is followed the tracks of.Just, the embodiment of the invention is embedded into the democracy synthesis mechanism in the particle filter process, has realized the vision tracking.Like this, vision is followed the tracks of have higher reliability and robustness, realized the vision tracking under the complex scene, improved the efficient of vision track algorithm, satisfied the real-time requirement that actual vision is followed the tracks of application system.
Above-described specific embodiment; purpose of the present invention, technical scheme and beneficial effect are further described; institute is understood that; the above only is specific embodiments of the invention; and be not intended to limit the scope of the invention; within the spirit and principles in the present invention all, any modification of being made, be equal to replacement, improvement etc., all should be included within protection scope of the present invention.

Claims (12)

1, a kind of visual tracking method based on particle filter is characterized in that the method comprising the steps of:
Many visual signatures that synthetic weights set calculates current subframe according to current many visual signatures merge state filtering to be estimated, described many visual signatures merge state filtering estimates to comprise target position information; And
Determine the target location of current sub frame of vision track according to described target position information.
2, the method for claim 1 is characterized in that, if described current subframe is first subframe, the synthetic weights set of pre-configured many visual signatures is gathered as the synthetic weights of described many visual signatures.
3, the method for claim 1 is characterized in that, described many visual signatures that synthetic weights set calculates current subframe according to current many visual signatures merge state filtering to be estimated, comprising:
For the synthetic weights of each many visual signature in the synthetic weights set of many visual signatures, judge whether the synthetic weights of these many visual signatures are less than or equal to judgment value, if be less than or equal to described judgment value, then disappear with the synthetic weights characteristic of correspondence of these many visual signatures, should the deletion from the synthetic weights set of described many visual signatures of the synthetic weights of many visual signatures.
As claim 1 or 3 described methods, it is characterized in that 4, described many visual signatures that synthetic weights set calculates current subframe according to current many visual signatures merge state filtering to be estimated, further comprises step:
Judge that whether the synthetic weights characteristic of correspondence number of all many visual signatures is less than the visual signature lower threshold that sets in advance in the synthetic weights set of current many visual signatures, if, carry out the step that produces random number, otherwise, continue to carry out the described step of determining the target location of current sub frame of vision track according to described target position information;
Produce random number;
Described random number be multiply by described characteristic, the product that obtains is rounded, obtain feature indication;
Judge whether resulting feature indication belongs to the feature indication collection,, resulting feature indication is added the feature indication collection, be implemented as the step that the feature indication that adds the feature indication collection is provided with the synthetic weights of corresponding many visual signatures if do not belong to; Otherwise this feature indication characteristic of correspondence is invalid, carries out the step of described generation random number, and wherein, the feature indication collection is the set that the feature indication of the synthetic weights characteristic of correspondence of all many visual signatures in the synthetic weights set of current many visual signatures is formed;
For the feature indication that adds the feature indication collection is provided with the synthetic weights of corresponding many visual signatures, the synthetic weights of many visual signatures of this setting are added in the synthetic weights set of many visual signatures, carry out and describedly judge that the synthetic weights characteristic of correspondence number of all many visual signatures in the synthetic weights set of current many visual signatures is whether less than the step of the visual signature lower threshold that sets in advance.
5, the method for claim 1 is characterized in that, the step that the described many visual signatures that calculate current subframe merge the state filtering estimation specifically comprises:
Determine current sampling particle collection;
According to synthetic weights set of current many visual signatures and weighting synthetic model, obtain the observation probability of current sampling particle set sampling particle;
The many visual signatures that obtain after the normalization according to described observation probability merge the weights of described sampling particles down;
Merge the weights of described sampling particle down according to the many visual signatures after the normalization, the many visual signatures that obtain current subframe merge state filtering to be estimated.
6, method as claimed in claim 5, it is characterized in that, describedly determine that the step of current sampling particle collection comprises: merge the weights of down-sampling particles according to many visual signatures in the last subframe, extract particle again, form current sampling particle collection from the employing particle set of previous frame.
7, the method for claim 1 is characterized in that, obtaining of the synthetic weights set of described current many visual signatures comprises:
The single vision significant condition filtering that obtains last subframe is estimated;
Merge the state filterings estimation according to the single vision significant condition filtering estimation of last subframe and many visual signatures of last subframe, obtain the state estimation deviation;
Obtain the visual signature quality metrics according to described state estimation deviation;
Obtain the synthetic weights set of many visual signatures according to described vision quality tolerance.
8, method as claimed in claim 7 is characterized in that, the single vision significant condition filtering of described last subframe is estimated to obtain according to the weights of sampling particle in the last subframe.
9, a kind of vision tracking system based on particle filter is characterized in that, this system comprises that synthetic weights acquisition module, many visual signatures merge state filtering estimation module and target location determination module;
Synthetic weights acquisition module obtains the synthetic weights set of current many visual signatures;
Many visual signatures merge the state filtering estimation module, and according to the synthetic weights set of current many visual signatures, the many visual signatures that calculate current subframe merge state filtering to be estimated;
The target location determination module merges state filtering according to described many visual signatures and estimates, determines the target location of current subframe.
10, system as claimed in claim 9 is characterized in that, described synthetic weights acquisition module comprises single vision significant condition filtering estimator module, state estimation deviation calculation submodule, quality metrics calculating sub module and synthetic weights calculating sub module;
Single vision significant condition filtering estimator module, the single vision significant condition filtering that obtains current subframe is estimated;
State estimation deviation calculation submodule merges state filterings according to many visual signatures that described single vision significant condition filtering is estimated and many visual signatures fusion state filtering estimation module transmits and estimates, calculates the state estimation deviation;
The quality metrics calculating sub module obtains the visual signature quality metrics according to described state estimation deviation calculation;
Synthetic weights calculating sub module according to the vision quality tolerance that the quality metrics computing module obtains, calculates the synthetic weights set of current many visual signatures.
11, as claim 9 or 10 described systems, it is characterized in that this system further comprises the characteristic disappear judging module;
The characteristic disappear judging module, judge whether the synthetic weights of each many visual signatures in the synthetic weights set of many visual signatures are less than or equal to judgment value, if be less than or equal to described judgment value, then disappear with the synthetic weights characteristic of correspondence of these many visual signatures, should the deletion from the synthetic weights set of many visual signatures of the synthetic weights of many visual signatures.
12, as claim 9 or 10 described systems, it is characterized in that, this system further comprises the feature generation module, judge the synthetic weights set of many visual signatures characteristic of correspondence number less than the visual signature threshold value after, produce feature indication, for described feature indication is provided with the synthetic weights of corresponding many visual signatures, the synthetic weights of many visual signatures of this setting are added the synthetic weights set of current many visual signatures.
CN 200710090883 2007-04-09 2007-04-09 Visual tracking method and system based on particle filter Expired - Fee Related CN100571392C (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN 200710090883 CN100571392C (en) 2007-04-09 2007-04-09 Visual tracking method and system based on particle filter

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN 200710090883 CN100571392C (en) 2007-04-09 2007-04-09 Visual tracking method and system based on particle filter

Publications (2)

Publication Number Publication Date
CN101026759A true CN101026759A (en) 2007-08-29
CN100571392C CN100571392C (en) 2009-12-16

Family

ID=38744594

Family Applications (1)

Application Number Title Priority Date Filing Date
CN 200710090883 Expired - Fee Related CN100571392C (en) 2007-04-09 2007-04-09 Visual tracking method and system based on particle filter

Country Status (1)

Country Link
CN (1) CN100571392C (en)

Cited By (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101339655B (en) * 2008-08-11 2010-06-09 浙江大学 Visual sense tracking method based on target characteristic and bayesian filtering
CN101394546B (en) * 2007-09-17 2010-08-25 华为技术有限公司 Video target profile tracing method and device
CN102186078A (en) * 2011-05-11 2011-09-14 浙江大学 Particle filter based multi-frame reference motion estimation method
CN102194234A (en) * 2010-03-03 2011-09-21 中国科学院自动化研究所 Image tracking method based on sequential particle swarm optimization
CN101923716B (en) * 2009-06-10 2012-07-18 新奥特(北京)视频技术有限公司 Method for improving particle filter tracking effect
CN102722702A (en) * 2012-05-28 2012-10-10 河海大学 Multiple feature fusion based particle filter video object tracking method
CN102982555A (en) * 2012-11-01 2013-03-20 江苏科技大学 Guidance infrared small target tracking method based on self-adaption manifold particle filters
CN103024344A (en) * 2011-09-20 2013-04-03 佳都新太科技股份有限公司 Automatic PTZ (Pan/Tilt/Zoom) target tracking method based on particle filter
CN103150546A (en) * 2012-12-26 2013-06-12 冉阳 Video face identification method and device
CN103557792A (en) * 2013-11-12 2014-02-05 中国科学院自动化研究所 Method for visual tracking and position measurement of drogue object
CN104200495A (en) * 2014-09-25 2014-12-10 重庆信科设计有限公司 Multi-target tracking method in video surveillance
CN106127808A (en) * 2016-06-20 2016-11-16 浙江工业大学 A kind of block particle filter method for tracking target based on color and the anti-of local binary patterns Feature Fusion
CN107403222A (en) * 2017-07-19 2017-11-28 燕山大学 A kind of motion tracking method based on auxiliary more new model and validity check

Cited By (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101394546B (en) * 2007-09-17 2010-08-25 华为技术有限公司 Video target profile tracing method and device
CN101339655B (en) * 2008-08-11 2010-06-09 浙江大学 Visual sense tracking method based on target characteristic and bayesian filtering
CN101923716B (en) * 2009-06-10 2012-07-18 新奥特(北京)视频技术有限公司 Method for improving particle filter tracking effect
CN102194234A (en) * 2010-03-03 2011-09-21 中国科学院自动化研究所 Image tracking method based on sequential particle swarm optimization
CN102186078A (en) * 2011-05-11 2011-09-14 浙江大学 Particle filter based multi-frame reference motion estimation method
CN103024344A (en) * 2011-09-20 2013-04-03 佳都新太科技股份有限公司 Automatic PTZ (Pan/Tilt/Zoom) target tracking method based on particle filter
CN102722702B (en) * 2012-05-28 2015-01-28 河海大学 Multiple feature fusion based particle filter video object tracking method
CN102722702A (en) * 2012-05-28 2012-10-10 河海大学 Multiple feature fusion based particle filter video object tracking method
CN102982555A (en) * 2012-11-01 2013-03-20 江苏科技大学 Guidance infrared small target tracking method based on self-adaption manifold particle filters
CN102982555B (en) * 2012-11-01 2016-12-21 江苏科技大学 Guidance Tracking Method of IR Small Target based on self adaptation manifold particle filter
CN103150546A (en) * 2012-12-26 2013-06-12 冉阳 Video face identification method and device
CN103150546B (en) * 2012-12-26 2016-03-16 冉阳 video face identification method and device
CN103557792B (en) * 2013-11-12 2015-10-28 中国科学院自动化研究所 A kind of vision of drogue target is followed the tracks of and location measurement method
CN103557792A (en) * 2013-11-12 2014-02-05 中国科学院自动化研究所 Method for visual tracking and position measurement of drogue object
CN104200495A (en) * 2014-09-25 2014-12-10 重庆信科设计有限公司 Multi-target tracking method in video surveillance
CN104200495B (en) * 2014-09-25 2017-03-29 重庆信科设计有限公司 A kind of multi-object tracking method in video monitoring
CN106127808A (en) * 2016-06-20 2016-11-16 浙江工业大学 A kind of block particle filter method for tracking target based on color and the anti-of local binary patterns Feature Fusion
CN106127808B (en) * 2016-06-20 2018-09-07 浙江工业大学 It is a kind of that particle filter method for tracking target is blocked based on color and the anti-of local binary patterns Fusion Features
CN107403222A (en) * 2017-07-19 2017-11-28 燕山大学 A kind of motion tracking method based on auxiliary more new model and validity check

Also Published As

Publication number Publication date
CN100571392C (en) 2009-12-16

Similar Documents

Publication Publication Date Title
CN100571392C (en) Visual tracking method and system based on particle filter
CN103106667B (en) A kind of towards blocking the Moving Objects method for tracing with scene change
CN109669049B (en) Particle image velocity measurement method based on convolutional neural network
CN104215249B (en) Smoothening method of driving track
EP2071515A1 (en) Visually tracking an object in real world using 2D appearance and multicue depth estimations
EP2418622A2 (en) Image processing method and image processing apparatus
CN107563286A (en) A kind of dynamic gesture identification method based on Kinect depth information
CN102103689B (en) Frontal face image synthesis-based face recognition method
Jia et al. Sensor fusion-based visual target tracking for autonomous vehicles with the out-of-sequence measurements solution
CN102156537A (en) Equipment and method for detecting head posture
CN103150728A (en) Vision positioning method in dynamic environment
CN106127125A (en) Distributed DTW human body behavior intension recognizing method based on human body behavior characteristics
CN104881029A (en) Mobile robot navigation method based on one point RANSAC and FAST algorithm
CN103003846A (en) Articulation region display device, articulation region detection device, articulation region relatedness computation device, articulation shape region relatedness computation device, and articulation region display method
CN104794737A (en) Depth-information-aided particle filter tracking method
CN110827320B (en) Target tracking method and device based on time sequence prediction
Argyros et al. Binocular hand tracking and reconstruction based on 2D shape matching
CN114693720A (en) Design method of monocular vision odometer based on unsupervised deep learning
CN102968615B (en) A kind of three-dimensional somatic data identification method of the anti-interference function had in intensive people flow
CN101799927B (en) Cartoon role contour tracing method based on key frame
Jia et al. Vision based data fusion for autonomous vehicles target tracking using interacting multiple dynamic models
CN102214301A (en) Multi-target tracking method for associated cooperation of adaptive motion
CN114283265A (en) Unsupervised face correcting method based on 3D rotation modeling
Zhao et al. An approach based on mean shift and kalman filter for target tracking under occlusion
Huang et al. Fast initialization method for monocular slam based on indoor model

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20091216

Termination date: 20170409

CF01 Termination of patent right due to non-payment of annual fee