CN106952288A - Based on convolution feature and global search detect it is long when block robust tracking method - Google Patents

Based on convolution feature and global search detect it is long when block robust tracking method Download PDF

Info

Publication number
CN106952288A
CN106952288A CN201710204379.1A CN201710204379A CN106952288A CN 106952288 A CN106952288 A CN 106952288A CN 201710204379 A CN201710204379 A CN 201710204379A CN 106952288 A CN106952288 A CN 106952288A
Authority
CN
China
Prior art keywords
target
scale
init
convolution
feature
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201710204379.1A
Other languages
Chinese (zh)
Other versions
CN106952288B (en
Inventor
李映
林彬
杭涛
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Northwestern Polytechnical University
Original Assignee
Northwestern Polytechnical University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Northwestern Polytechnical University filed Critical Northwestern Polytechnical University
Priority to CN201710204379.1A priority Critical patent/CN106952288B/en
Publication of CN106952288A publication Critical patent/CN106952288A/en
Application granted granted Critical
Publication of CN106952288B publication Critical patent/CN106952288B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20048Transform domain processing
    • G06T2207/20056Discrete and fast Fourier transform, [DFT, FFT]

Landscapes

  • Image Analysis (AREA)

Abstract

The present invention relates to it is a kind of based on convolution feature and global search detect it is long when block robust tracking method, by using convolution feature and multiple dimensioned correlation filtering method in tracking module, enhance the feature representation ability of tracking target appearance model so that tracking result has very strong robustness for factors such as illumination variation, target scale change, target rotations;Further through the global search testing mechanism of introducing, so that when target by it is long when block cause tracking failure when, detection module can detect target again, tracking module is recovered from mistake, accordingly even when in the case where target appearance changes, can also be tracked long lasting for ground.

Description

Based on convolution feature and global search detect it is long when block robust tracking method
Technical field
The invention belongs to computer vision field, it is related to a kind of method for tracking target, and in particular to one kind is based on convolution feature With global search detect it is long when block robust tracking method.
Background technology
The main task of target following is to obtain the position of specific objective and movable information in video sequence, in video prison The fields such as control, man-machine interaction have a wide range of applications.During tracking, illumination variation, background are complicated, target is rotated or contracted The complexity of Target Tracking Problem can all be increased by the factor such as putting, especially when target by it is long when block when, then be easier to cause with Track fails.
Document " Tracking-learning-detection, IEEE Transactions on Pattern Analysis and Machine Intelligence,2012,34(7):The tracking that 1409-1422 " is proposed is (referred to as TLD) traditional track algorithm and detection algorithm are combined first, tracking result is improved using testing result, improves and is The reliability and robustness of system.Its track algorithm is based on optical flow method, and detection algorithm produces substantial amounts of detection window, for each inspection Survey window, it is necessary to received that last testing result could be turned into by three detectors.For occlusion issue, TLD provides one Individual effective resolving ideas, can carry out long time-tracking (Long-term Tracking) to target.But, TLD is used Be shallow-layer manual features, the sign to target is limited in one's ability, and the design of detection algorithm is also complex, there is certain change Enter space.
The content of the invention
The technical problem to be solved
In order to avoid in place of the deficiencies in the prior art, the present invention proposes a kind of to detect based on convolution feature and global search Block robust tracking method when long, solve video frequency motion target during tracking due to it is long when block or target translates out the visual field Outside etc. factor cause display model to drift about so that be easily caused tracking failure the problem of.
Technical scheme
It is a kind of based on convolution feature and global search detect it is long when block robust tracking method, it is characterised in that step is such as Under:
Step 1 reads the first frame image data and the initial position message [x, y, w, h] where target in video, wherein X, y represent the abscissa and ordinate of target's center, and w, h represents the wide and height of target.(x, y) corresponding coordinate points are designated as P, Centered on P, size is designated as R for w × h target prime areainit, then the yardstick of target is designated as scale, it is initialized as 1.
Step 2 determines a region R comprising target and background information centered on Pbkg, RbkgSize be M × N, M =2w, N=2h.Using VGGNet-19 as CNN models, convolution characteristic pattern is extracted to R' in the 5th layer of convolutional layer (conv5-4) ztarget_init.Then according to ztarget_initBuild the object module of tracking moduleT ∈ { 1,2 ..., T }, T are CNN moulds Type port number, computational methods are as follows:
Wherein, the variable of capitalization is expression of the corresponding non-capitalized variables on frequency domain, gaussian filtering templateM, n be Gaussian function independent variable, m ∈ { 1,2 ..., M }, n ∈ { 1,2 ..., N }, σtargetFor the bandwidth of Gaussian kernel,⊙ representative element multiplication operations, upper line represents complex conjugate, λ1For adjustment Parameter (in order to avoid denominator introduces for 0), is set as 0.0001.
Step 3 extracts the image subblock of S different scale centered on P, and S is set as 33.The size of each sub-block is w × h × s, variable s are the scale factor of image subblock, s ∈ [0.7,1.4].Then the HOG for extracting each image subblock respectively is special Levy, turn into the HOG characteristic vectors of a S dimension after merging, scale feature vector is named as here, z is designated asscale_init.Again According to zscale_initBuild the Scale Model W of tracking modulescale, computational methods in step 2 with calculatingIt is similar (yardstick is special Levy vector and replace convolution characteristic pattern), it is specific as follows:
Wherein,S' is Gaussian function independent variable, s' ∈ { 1,2 ..., S }, σscaleFor Gaussian kernel Bandwidth,λ2For adjusting parameter, it is set as 0.0001.
Step 4 is to target prime area RinitGray feature is extracted, obtained gray feature represents it is a two-dimensional matrix, Here the matrix is named as target appearance representing matrix, is designated as Ak, subscript k represents current frame number, k=1 when initial.Then will The Filtering Model D of detection module is initialized as A1, i.e. D=A1, the history that reinitializes object representation set of matrices Ahis。AhisWork With being current and each frame before the target appearance representing matrix of storage, i.e. Ahis={ A1,A2,...,Ak, A when initialhis= {A1}。
Step 5 reads next two field picture, still centered on P, and extraction size is Rbkg× scale process scaling Target search region afterwards.Then by the convolution feature in the CNN network extraction target searches region in step 2, and with bilateral The mode of interpolation samples RbkgSize obtain the convolution characteristic pattern z of present frametarget_cur, recycle object moduleMeter Calculate target confidence map ftarget, computational methods are as follows:
Wherein,For inverse Fourier transform.Final updating P coordinate, f is modified to by (x, y)targetIn peak response The corresponding coordinate of value:
Step 6 is extracted the image subblock of S different scale, each image subblock is then extracted respectively centered on P The scale feature vector z of present frame is obtained after HOG features, mergingscale_cur(with z in step 3scale_initComputational methods).Again Utilize Scale Model WscaleCalculate yardstick confidence map:
The yardstick scale of final updating target, computational methods are as follows:
So far, output of the tracking module in present frame (kth frame) can be obtained:Centered on using coordinate as the P of (x, y), greatly Small is Rinit× scale image subblock TPatchk.In addition, the f that completion will be computedtargetIn maximum response brief note For TPeakk, i.e. TPeakk=ftarget(x,y)。
The entire image of Filtering Model D and present frame is carried out convolution, meter by step 7 detection module in the way of global search Calculate Filtering Model D and the similarity degree of each position of present frame.J value (j is set as 10) before responsiveness highest is taken, and respectively Centered on the corresponding location point of this j value, extraction size is Rinit× scale j image subblock.By this j image Block generates an image subblock set DPatches as elementk, i.e. output of the detection module in kth frame.
The set DPatches that step 8 is exported to detection modulekIn each image subblock, it is calculated respectively defeated with tracking module The TPatch gone outkBetween pixel Duplication, j value can be obtained, wherein highest value is designated asIfLess than threshold value(It is set as 0.05), being determined as that target is blocked, it is necessary to suppress learning rate β of the tracking module in model modification completely, And go to step 9;Otherwise initial learning rate β is pressedinitinitIt is set as 0.02) being updated, and goes to step 10.β computing formula It is as follows:
Step 9 is according to DPatcheskIn each image subblock center, respectively extract size be Rbkg× scale j mesh Region of search is marked, according to the method in step 5 is to each target search extracted region convolution characteristic pattern and calculates target confidence Figure, can obtain the maximum response on j target search region.It is compared again in this j response, by maximum value It is designated as DPeakk.If DPeakkMore than TPeakk, then P coordinate is updated again, and (x, y) is modified to DPeakkCorresponding Coordinate.And recalculate target scale characteristic vector and target scale scale (with the calculation in step 6).
Step 10 target is defined as P in the optimal place-centric of present frame, and optimal scale is defined as scale.In the picture Indicate new target area Rnew, i.e., centered on P, wide and high respectively w × scale, h × scale rectangle frame.In addition, The convolution characteristic pattern for being computed completing and can obtain optimal objective place-centric P is abbreviated as ztarget;Equally, by energy The scale feature vector for accessing optimal objective yardstick scale is abbreviated as zscale
Step 11 utilizes ztarget、zscale, and the object module in the tracking module of previous frame foundationAnd yardstick Model Wscale, model modification is carried out in the way of weighted sum respectively, computational methods are as follows:
Wherein, β is the learning rate after step 8 is calculated.
The new target area R of step 12 pairnewExtract the target appearance representing matrix A that present frame is obtained after gray featurek, By AkIt is added to history object representation set of matrices Ahis.If set AhisMiddle element number is more thanc(cBe set as 20), then from AhisMiddle random selectioncIndividual one three-dimensional matrice C of Element generationk, Ck(:, i) corresponding is AhisIn any one element it is (i.e. two-dimentional Matrix Ak);Otherwise A is usedhisMiddle all elements generator matrix Ck.Then to CkAverage and obtain two-dimensional matrix, by this two Matrix is tieed up as the new Filtering Model D of detection module, computational methods are as follows:
Step 13 judges whether to have handled picture frames all in video, and algorithm terminates if having handled, and otherwise goes to step 5 Continue executing with.
Beneficial effect
It is proposed by the present invention it is a kind of based on convolution feature and global search detect it is long when block robust tracking method, respectively Tracking module and detection module are devised, during tracking, two module cooperative work:Tracking module mainly utilizes convolutional Neural The convolution feature that network (Convolutional Neural Network, CNN) extracts target is used for the target mould for building robust Type, and by histograms of oriented gradients (Histogram of Oriented Gradient, HOG) feature construction Scale Model, Determine the place-centric and yardstick of target respectively with reference to correlation filtering method;Detection module extracts gray feature and builds target Filtering Model, is used for quickly detecting and judges the generation blocked, one by the way of global search in entire image to target Denier target is blocked (or other factorses cause target appearance acute variation) completely, and detection module is tracked using testing result amendment The position of target, and suppress the model modification of tracking module, prevent from introducing unnecessary noise so as to cause model drift about and with Track fails.
Superiority:By using convolution feature and multiple dimensioned correlation filtering method in tracking module, tracking mesh is enhanced Mark the feature representation ability of display model so that tracking result is for factors such as illumination variation, target scale change, target rotations With very strong robustness;Further through the global search testing mechanism of introducing so that when target by it is long when block cause tracking lose When losing, detection module can detect target again, tracking module is recovered from mistake, accordingly even when in target appearance In the case of change, it can also be tracked long lasting for ground.
Brief description of the drawings
Fig. 1:Based on convolution feature and global search detect it is long when block robust tracking method flow chart
Embodiment
In conjunction with embodiment, accompanying drawing, the invention will be further described:
Step 1 reads the first frame image data and the initial position message [x, y, w, h] where target in video, wherein X, y represent the abscissa and ordinate of target's center, and w, h represents the wide and height of target.(x, y) corresponding coordinate points are designated as P, Centered on P, size is designated as R for w × h target prime areainit, then the yardstick of target is designated as scale, it is initialized as 1.
Step 2 determines a region R comprising target and background information centered on Pbkg, RbkgSize be M × N, M =2w, N=2h.Using VGGNet-19 as CNN models, convolution characteristic pattern is extracted to R' in the 5th layer of convolutional layer (conv5-4) ztarget_init.Then according to ztarget_initBuild the object module of tracking moduleT ∈ { 1,2 ..., T }, T are CNN moulds Type port number, computational methods are as follows:
Wherein, the variable of capitalization is expression of the corresponding non-capitalized variables on frequency domain, gaussian filtering templateM, n be Gaussian function independent variable, m ∈ { 1,2 ..., M }, n ∈ { 1,2 ..., N }, σtargetFor the bandwidth of Gaussian kernel,⊙ representative element multiplication operations, upper line represents complex conjugate, λ1To adjust Whole parameter (in order to avoid denominator introduces for 0), is set as 0.0001.
Step 3 extracts the image subblock of S different scale centered on P, and S is set as 33.The size of each sub-block is w × h × s, variable s are the scale factor of image subblock, s ∈ [0.7,1.4].Then the HOG for extracting each image subblock respectively is special Levy, turn into the HOG characteristic vectors of a S dimension after merging, scale feature vector is named as here, z is designated asscale_init.Again According to zscale_initBuild the Scale Model W of tracking modulescale, computational methods in step 2 with calculatingIt is similar (yardstick is special Levy vector and replace convolution characteristic pattern), it is specific as follows:
Wherein,S' is Gaussian function independent variable, s' ∈ { 1,2 ..., S }, σscaleFor Gaussian kernel Bandwidth,λ2For adjusting parameter, it is set as 0.0001.
Step 4 is to target prime area RinitGray feature is extracted, obtained gray feature represents it is a two-dimensional matrix, Here the matrix is named as target appearance representing matrix, is designated as Ak, subscript k represents current frame number, k=1 when initial.Then will The Filtering Model D of detection module is initialized as A1, i.e. D=A1, the history that reinitializes object representation set of matrices Ahis。AhisWork With being current and each frame before the target appearance representing matrix of storage, i.e. Ahis={ A1,A2,...,Ak, A when initialhis= {A1}。
Step 5 reads next two field picture, still centered on P, and extraction size is Rbkg× scale process scaling Target search region afterwards.Then by the convolution feature in the CNN network extraction target searches region in step 2, and with bilateral The mode of interpolation samples RbkgSize obtain the convolution characteristic pattern z of present frametarget_cur, recycle object moduleMeter Calculate target confidence map ftarget, computational methods are as follows:
Wherein,For inverse Fourier transform.Final updating P coordinate, f is modified to by (x, y)targetIn peak response The corresponding coordinate of value:
Step 6 is extracted the image subblock of S different scale, each image subblock is then extracted respectively centered on P The scale feature vector z of present frame is obtained after HOG features, mergingscale_cur(with z in step 3scale_initComputational methods).Again Utilize Scale Model WscaleCalculate yardstick confidence map:
The yardstick scale of final updating target, computational methods are as follows:
So far, output of the tracking module in present frame (kth frame) can be obtained:Centered on using coordinate as the P of (x, y), greatly Small is Rinit× scale image subblock TPatchk.In addition, the f that completion will be computedtargetIn maximum response brief note For TPeakk, i.e. TPeakk=ftarget(x,y)。
The entire image of Filtering Model D and present frame is carried out convolution, meter by step 7 detection module in the way of global search Calculate Filtering Model D and the similarity degree of each position of present frame.J value (j is set as 10) before responsiveness highest is taken, and respectively Centered on the corresponding location point of this j value, extraction size is Rinit× scale j image subblock.By this j image Block generates an image subblock set DPatches as elementk, i.e. output of the detection module in kth frame.
The set DPatches that step 8 is exported to detection modulekIn each image subblock, it is calculated respectively defeated with tracking module The TPatch gone outkBetween pixel Duplication, j value can be obtained, wherein highest value is designated asIfLess than threshold value(It is set as 0.05), being determined as that target is blocked, it is necessary to suppress learning rate β of the tracking module in model modification completely, And go to step 9;Otherwise initial learning rate β is pressedinitinitIt is set as 0.02) being updated, and goes to step 10.β computing formula It is as follows:
Step 9 is according to DPatcheskIn each image subblock center, respectively extract size be Rbkg× scale j mesh Region of search is marked, according to the method in step 5 is to each target search extracted region convolution characteristic pattern and calculates target confidence Figure, can obtain the maximum response on j target search region.It is compared again in this j response, by maximum value It is designated as DPeakk.If DPeakkMore than TPeakk, then P coordinate is updated again, and (x, y) is modified to DPeakkCorresponding Coordinate.And recalculate target scale characteristic vector and target scale scale (with the calculation in step 6).
Step 10 target is defined as P in the optimal place-centric of present frame, and optimal scale is defined as scale.In the picture Indicate new target area Rnew, i.e., centered on P, wide and high respectively w × scale, h × scale rectangle frame.In addition, The convolution characteristic pattern for being computed completing and can obtain optimal objective place-centric P is abbreviated as ztarget;Equally, by energy The scale feature vector for accessing optimal objective yardstick scale is abbreviated as zscale
Step 11 utilizes ztarget、zscale, and the object module in the tracking module of previous frame foundationAnd yardstick Model Wscale, model modification is carried out in the way of weighted sum respectively, computational methods are as follows:
Wherein, β is the learning rate after step 8 is calculated.
The new target area R of step 12 pairnewExtract the target appearance representing matrix A that present frame is obtained after gray featurek, By AkIt is added to history object representation set of matrices Ahis.If set AhisMiddle element number is more thanc(cBe set as 20), then from AhisMiddle random selectioncIndividual one three-dimensional matrice C of Element generationk, Ck(:, i) corresponding is AhisIn any one element it is (i.e. two-dimentional Matrix Ak);Otherwise A is usedhisMiddle all elements generator matrix Ck.Then to CkAverage and obtain two-dimensional matrix, by this two Matrix is tieed up as the new Filtering Model D of detection module, computational methods are as follows:
Step 13 judges whether to have handled picture frames all in video, and algorithm terminates if having handled, and otherwise goes to step 5 Continue executing with.

Claims (6)

1. it is a kind of based on convolution feature and global search detect it is long when block robust tracking method, it is characterised in that step is such as Under:
Step 1:Read the first frame image data and the initial position message [x, y, w, h] where target, wherein x, y in video The abscissa and ordinate of target's center are represented, w, h represents the wide and height of target;(x, y) corresponding coordinate points are designated as P, with P Centered on, size is designated as R for w × h target prime areainit, then the yardstick of target is designated as scale, it is initialized as 1;
Step 2:Centered on P, a region R comprising target and background information is determinedbkg, RbkgSize be M × N, M= 2w, N=2h;Using VGGNet-19 as CNN models, convolution characteristic pattern is extracted to R' in the 5th layer of convolutional layer (conv5-4) ztarget_init;Then according to ztarget_initBuild the object module of tracking moduleT ∈ { 1,2 ..., T }, T are CNN moulds Type port number, computational methods are as follows:
Wherein:The variable of capitalization is expression of the corresponding non-capitalized variables on frequency domain, gaussian filtering templateM, n be Gaussian function independent variable, m ∈ { 1,2 ..., M }, n ∈ { 1,2 ..., N }, σtargetFor the bandwidth of Gaussian kernel,⊙ representative element multiplication operations, upper line represents complex conjugate, λ1To adjust Whole parameter;
Step 3:Centered on P, the image subblock of S different scale is extracted, S is set as 33;The size of each sub-block is w × h × s, variable s are the scale factor of image subblock, s ∈ [0.7,1.4];Then the HOG features of each image subblock are extracted respectively, Turn into the HOG characteristic vectors of a S dimension after merging, and be named as scale feature vector, be designated as zscale_init;Further according to zscale_initBuild the Scale Model W of tracking modulescale, computational methods are as follows:
Wherein,S' is Gaussian function independent variable, s' ∈ { 1,2 ..., S }, σscaleFor the band of Gaussian kernel Width,λ2For adjusting parameter;
Step 4:To target prime area RinitGray feature is extracted, the two-dimensional matrix that obtained gray feature is represented is named as mesh Outward appearance representing matrix is marked, A is designated ask, subscript k represents current frame number, k=1 when initial;Then by the Filtering Model D of detection module It is initialized as A1, i.e. D=A1, the history that reinitializes object representation set of matrices Ahis;AhisTo store current and each frame before Target appearance representing matrix, i.e. Ahis={ A1,A2,...,Ak, A when initialhis={ A1};
Step 5:Next two field picture is read, still centered on P, extraction size is Rbkg× scale after scaling Target search region;Then by the convolution feature in the CNN network extraction target searches region in step 2, and with bilateral interpolation Mode sample RbkgSize obtain the convolution characteristic pattern z of present frametarget_cur, recycle object moduleCalculate mesh Mark confidence map ftarget, computational methods are as follows:
Wherein,For inverse Fourier transform.Final updating P coordinate, f is modified to by (x, y)targetIn maximum response institute Corresponding coordinate:
( x , y ) = arg m a x x ′ , y ′ ( f t arg e t ( x ′ , y ′ ) ) ;
Step 6:Centered on P, the image subblock of S different scale is extracted, the HOG that each image subblock is then extracted respectively is special Levy, the scale feature vector z of present frame is obtained after mergingscale_cur, with z in step 3scale_initComputational methods;Recycle chi Spend model WscaleCalculate yardstick confidence map:
The yardstick scale of final updating target, computational methods are as follows:
s c a l e = arg m a x s ′ ′ ( f s c a l e ( s ′ ′ ) )
Obtain output of the tracking module in present frame (kth frame):Centered on using coordinate as the P of (x, y), size is Rinit×scale Image subblock TPatchk;In addition, the f that completion will be computedtargetIn maximum response be abbreviated as TPeakk, i.e., TPeakk=ftarget(x,y);
Step 7:The entire image of Filtering Model D and present frame is carried out convolution by detection module in the way of global search, is calculated Filtering Model D and the similarity degree of each position of present frame;J value before responsiveness highest is taken, and it is corresponding with j value respectively Centered on location point, extraction size is Rinit× scale j image subblock;It regard j image subblock as element, generation one Individual image subblock set DPatchesk, i.e. output of the detection module in kth frame;
Step 8:The set DPatches of detection module output is calculated respectivelykIn each image subblock and tracking module export TPatchkBetween pixel Duplication, obtain j value, wherein highest value be designated asIfLess than threshold valueJudge Blocked completely for target, it is necessary to suppress learning rate β of the tracking module in model modification, and go to step 9;Otherwise by initial Habit rate βinitIt is updated, and goes to step 10;
The computing formula of the β is as follows:
Step 9:According to DPatcheskIn each image subblock center, respectively extract size be Rbkg× scale j target is searched Rope region, according to the method in step 5 is to each target search extracted region convolution characteristic pattern and calculates target confidence map, is obtained Maximum response onto j target search region;Maximum value in j response is designated as DPeakk;If DPeakkGreatly In TPeakk, then P coordinate is updated again, and (x, y) is modified to DPeakkCorresponding coordinate;And recalculate target scale Characteristic vector and target scale scale, using the calculation in step 6;
Step 10:Target is defined as P in the optimal place-centric of present frame, and optimal scale is defined as scale;Indicate in the picture Go out new target area Rnew, centered on P, wide and high respectively w × scale, h × scale rectangle frame;In addition, by Calculate and complete and can obtain optimal objective place-centric P convolution characteristic pattern and be abbreviated as ztarget;Equally, it is possible to obtain Optimal objective yardstick scale scale feature vector is abbreviated as zscale
Step 11:Utilize ztarget、zscale, and the object module in the tracking module of previous frame foundationAnd Scale Model Wscale, model modification is carried out in the way of weighted sum respectively, computational methods are as follows:
W t arg e t t = W t arg e t _ n e w t
Wscale=Wscale_new
Step 12:To new target area RnewExtract the target appearance representing matrix A that present frame is obtained after gray featurek, by Ak It is added to history object representation set of matrices Ahis;If set AhisMiddle element number is more than c, then from AhisMiddle random selection c One three-dimensional matrice C of Element generationk, Ck(:, i) corresponding is AhisIn any one element, i.e. two-dimensional matrix Ak;Otherwise A is usedhis Middle all elements generator matrix Ck;Then to CkAverage and obtain two-dimensional matrix, regard this two-dimensional matrix as detection module New Filtering Model D, computational methods are as follows:
D = 1 c Σ i = 1 , ... , c C ( : , i ) , i f | A h i s | > c 1 | A h i s | Σ i = 1 , ... , | A h i s | C ( : , i ) , o t h e r w i s e
Step 13:Algorithm terminates if picture frames all in video have been handled, and otherwise goes to step 5 and continues executing with.
2. according to claim 1 based on convolution feature and global search detect it is long when block robust tracking method, it is special Levy and be:The adjusting parameter λ1And λ2It is set as 0.0001.
3. according to claim 1 based on convolution feature and global search detect it is long when block robust tracking method, it is special Levy and be:The j values are set as 10.
4. according to claim 1 based on convolution feature and global search detect it is long when block robust tracking method, it is special Levy and be:The threshold valueIt is set as 0.05.
5. according to claim 1 based on convolution feature and global search detect it is long when block robust tracking method, it is special Levy and be:The initial learning rate βinitIt is set as 0.02.
6. according to claim 1 based on convolution feature and global search detect it is long when block robust tracking method, it is special Levy and be:The c is set as 20.
CN201710204379.1A 2017-03-31 2017-03-31 Based on convolution feature and global search detect it is long when block robust tracking method Expired - Fee Related CN106952288B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201710204379.1A CN106952288B (en) 2017-03-31 2017-03-31 Based on convolution feature and global search detect it is long when block robust tracking method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710204379.1A CN106952288B (en) 2017-03-31 2017-03-31 Based on convolution feature and global search detect it is long when block robust tracking method

Publications (2)

Publication Number Publication Date
CN106952288A true CN106952288A (en) 2017-07-14
CN106952288B CN106952288B (en) 2019-09-24

Family

ID=59475259

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710204379.1A Expired - Fee Related CN106952288B (en) 2017-03-31 2017-03-31 Based on convolution feature and global search detect it is long when block robust tracking method

Country Status (1)

Country Link
CN (1) CN106952288B (en)

Cited By (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107452022A (en) * 2017-07-20 2017-12-08 西安电子科技大学 A kind of video target tracking method
CN107491742A (en) * 2017-07-28 2017-12-19 西安因诺航空科技有限公司 Stable unmanned plane target tracking when a kind of long
CN107644430A (en) * 2017-07-27 2018-01-30 孙战里 Target following based on self-adaptive features fusion
CN108734151A (en) * 2018-06-14 2018-11-02 厦门大学 Robust long-range method for tracking target based on correlation filtering and the twin network of depth
CN109271865A (en) * 2018-08-17 2019-01-25 西安电子科技大学 Motion target tracking method based on scattering transformation multilayer correlation filtering
CN109410249A (en) * 2018-11-13 2019-03-01 深圳龙岗智能视听研究院 A kind of method for tracing of combination depth characteristic and Freehandhand-drawing feature adaptive targets
CN109596649A (en) * 2018-11-29 2019-04-09 昆明理工大学 A kind of method and device that host element concentration is influenced based on convolutional network coupling microalloy element
CN109740448A (en) * 2018-12-17 2019-05-10 西北工业大学 Video object robust tracking method of taking photo by plane based on correlation filtering and image segmentation
CN109754424A (en) * 2018-12-17 2019-05-14 西北工业大学 Correlation filtering track algorithm based on fusion feature and adaptive updates strategy
CN110276782A (en) * 2018-07-09 2019-09-24 西北工业大学 A kind of EO-1 hyperion method for tracking target of combination sky spectrum signature and correlation filtering
WO2020056903A1 (en) * 2018-09-21 2020-03-26 北京字节跳动网络技术有限公司 Information generating method and device
CN111260687A (en) * 2020-01-10 2020-06-09 西北工业大学 Aerial video target tracking method based on semantic perception network and related filtering
CN111652910A (en) * 2020-05-22 2020-09-11 重庆理工大学 Target tracking algorithm based on object space relationship
CN112762841A (en) * 2020-12-30 2021-05-07 天津大学 Bridge dynamic displacement monitoring system and method based on multi-resolution depth features

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105631895A (en) * 2015-12-18 2016-06-01 重庆大学 Temporal-spatial context video target tracking method combining particle filtering
CN105741316A (en) * 2016-01-20 2016-07-06 西北工业大学 Robust target tracking method based on deep learning and multi-scale correlation filtering
CN106326924A (en) * 2016-08-23 2017-01-11 武汉大学 Object tracking method and object tracking system based on local classification

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105631895A (en) * 2015-12-18 2016-06-01 重庆大学 Temporal-spatial context video target tracking method combining particle filtering
CN105741316A (en) * 2016-01-20 2016-07-06 西北工业大学 Robust target tracking method based on deep learning and multi-scale correlation filtering
CN106326924A (en) * 2016-08-23 2017-01-11 武汉大学 Object tracking method and object tracking system based on local classification

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
CHAO MA等: "Hierarchical convolutional features for visual tracking", 《2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER》 *
YAN CHEN等: "CNNTracker: Online discriminative object tracking via deep convolutional neural network", 《APPLIED SOFT COMPUTING》 *
YOUNGBIN PARK等: "Tracking Human-like Natural Motion Using Deep Recurrent Neural Networks", 《ARXIV: COMPUTER VISION AND PATTERN RECOGNITION》 *

Cited By (20)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107452022A (en) * 2017-07-20 2017-12-08 西安电子科技大学 A kind of video target tracking method
CN107644430A (en) * 2017-07-27 2018-01-30 孙战里 Target following based on self-adaptive features fusion
CN107491742A (en) * 2017-07-28 2017-12-19 西安因诺航空科技有限公司 Stable unmanned plane target tracking when a kind of long
CN108734151A (en) * 2018-06-14 2018-11-02 厦门大学 Robust long-range method for tracking target based on correlation filtering and the twin network of depth
CN110276782A (en) * 2018-07-09 2019-09-24 西北工业大学 A kind of EO-1 hyperion method for tracking target of combination sky spectrum signature and correlation filtering
CN110276782B (en) * 2018-07-09 2022-03-11 西北工业大学 Hyperspectral target tracking method combining spatial spectral features and related filtering
CN109271865A (en) * 2018-08-17 2019-01-25 西安电子科技大学 Motion target tracking method based on scattering transformation multilayer correlation filtering
CN109271865B (en) * 2018-08-17 2021-11-09 西安电子科技大学 Moving target tracking method based on scattering transformation multilayer correlation filtering
WO2020056903A1 (en) * 2018-09-21 2020-03-26 北京字节跳动网络技术有限公司 Information generating method and device
CN109410249B (en) * 2018-11-13 2021-09-28 深圳龙岗智能视听研究院 Self-adaptive target tracking method combining depth characteristic and hand-drawn characteristic
CN109410249A (en) * 2018-11-13 2019-03-01 深圳龙岗智能视听研究院 A kind of method for tracing of combination depth characteristic and Freehandhand-drawing feature adaptive targets
CN109596649A (en) * 2018-11-29 2019-04-09 昆明理工大学 A kind of method and device that host element concentration is influenced based on convolutional network coupling microalloy element
CN109754424A (en) * 2018-12-17 2019-05-14 西北工业大学 Correlation filtering track algorithm based on fusion feature and adaptive updates strategy
CN109740448A (en) * 2018-12-17 2019-05-10 西北工业大学 Video object robust tracking method of taking photo by plane based on correlation filtering and image segmentation
CN109740448B (en) * 2018-12-17 2022-05-10 西北工业大学 Aerial video target robust tracking method based on relevant filtering and image segmentation
CN109754424B (en) * 2018-12-17 2022-11-04 西北工业大学 Correlation filtering tracking algorithm based on fusion characteristics and self-adaptive updating strategy
CN111260687A (en) * 2020-01-10 2020-06-09 西北工业大学 Aerial video target tracking method based on semantic perception network and related filtering
CN111260687B (en) * 2020-01-10 2022-09-27 西北工业大学 Aerial video target tracking method based on semantic perception network and related filtering
CN111652910A (en) * 2020-05-22 2020-09-11 重庆理工大学 Target tracking algorithm based on object space relationship
CN112762841A (en) * 2020-12-30 2021-05-07 天津大学 Bridge dynamic displacement monitoring system and method based on multi-resolution depth features

Also Published As

Publication number Publication date
CN106952288B (en) 2019-09-24

Similar Documents

Publication Publication Date Title
CN106952288B (en) Based on convolution feature and global search detect it is long when block robust tracking method
CN105741316B (en) Robust method for tracking target based on deep learning and multiple dimensioned correlation filtering
Liu et al. Detection of multiclass objects in optical remote sensing images
CN108665481A (en) Multilayer depth characteristic fusion it is adaptive resist block infrared object tracking method
CN106570893A (en) Rapid stable visual tracking method based on correlation filtering
CN111062973A (en) Vehicle tracking method based on target feature sensitivity and deep learning
CN110120064B (en) Depth-related target tracking algorithm based on mutual reinforcement and multi-attention mechanism learning
CN107316316A (en) The method for tracking target that filtering technique is closed with nuclear phase is adaptively merged based on multiple features
CN107330357A (en) Vision SLAM closed loop detection methods based on deep neural network
CN107644430A (en) Target following based on self-adaptive features fusion
CN103886325B (en) Cyclic matrix video tracking method with partition
CN107424171A (en) A kind of anti-shelter target tracking based on piecemeal
CN108447078A (en) The interference of view-based access control model conspicuousness perceives track algorithm
CN112232134B (en) Human body posture estimation method based on hourglass network and attention mechanism
CN110175649A (en) It is a kind of about the quick multiscale estimatiL method for tracking target detected again
CN107067410B (en) Manifold regularization related filtering target tracking method based on augmented samples
CN107452022A (en) A kind of video target tracking method
Yang et al. Visual tracking with long-short term based correlation filter
CN105844665A (en) Method and device for tracking video object
CN111882546A (en) Weak supervised learning-based three-branch convolutional network fabric defect detection method
CN111027586A (en) Target tracking method based on novel response map fusion
CN111640138A (en) Target tracking method, device, equipment and storage medium
CN110111369A (en) A kind of dimension self-adaption sea-surface target tracking based on edge detection
CN110135435B (en) Saliency detection method and device based on breadth learning system
CN110309729A (en) Tracking and re-detection method based on anomaly peak detection and twin network

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20190924

CF01 Termination of patent right due to non-payment of annual fee