CN106960176A - A kind of pedestrian's gender identification method based on transfinite learning machine and color characteristic fusion - Google Patents

A kind of pedestrian's gender identification method based on transfinite learning machine and color characteristic fusion Download PDF

Info

Publication number
CN106960176A
CN106960176A CN201710096262.6A CN201710096262A CN106960176A CN 106960176 A CN106960176 A CN 106960176A CN 201710096262 A CN201710096262 A CN 201710096262A CN 106960176 A CN106960176 A CN 106960176A
Authority
CN
China
Prior art keywords
feature
pedestrian
learning machine
image
fusion
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201710096262.6A
Other languages
Chinese (zh)
Other versions
CN106960176B (en
Inventor
曾焕强
蔡磊
朱建清
曹九稳
蔡灿辉
马凯光
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Huaqiao University
Original Assignee
Huaqiao University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Huaqiao University filed Critical Huaqiao University
Priority to CN201710096262.6A priority Critical patent/CN106960176B/en
Publication of CN106960176A publication Critical patent/CN106960176A/en
Application granted granted Critical
Publication of CN106960176B publication Critical patent/CN106960176B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/52Surveillance or monitoring of activities, e.g. for recognising suspicious objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2411Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on the proximity to a decision surface, e.g. support vector machines
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • G06F18/253Fusion techniques of extracted features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/56Extraction of image or video features relating to colour

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • Artificial Intelligence (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Evolutionary Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Software Systems (AREA)
  • Multimedia (AREA)
  • Medical Informatics (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Image Analysis (AREA)
  • Image Processing (AREA)

Abstract

A kind of pedestrian's gender identification method based on transfinite learning machine and color characteristic fusion of the present invention, including:Extract the learning machine feature that transfinites of the training image of unmarked gender attribute;The input training image hsv color feature of unmarked gender attribute is extracted, the learning machine feature that will transfinite is combined with color characteristic, obtain fusion feature, pedestrian's gender sorter is trained using linear SVM SVM according to fusion feature and training image label;Utilize training gained model extraction test image feature, extract its hsv color feature simultaneously, then two category features are merged, obtains the fusion feature of test image, fusion feature is classified with linear SVM SVM pedestrian's gender sorter obtained by training process.The present invention is extracted to input picture to transfinite learning characteristic and color characteristic and carries out effective integration, realizes the complementation of two kinds of features, pedestrian's gender attribute is more effectively caught, so as to improve pedestrian's sex discrimination.

Description

A kind of pedestrian's gender identification method based on transfinite learning machine and color characteristic fusion
Technical field
It is more particularly to a kind of based on transfiniting learning machine and color characteristic melts the present invention relates to computer vision and pattern-recognition Pedestrian's gender identification method of conjunction.
Background technology
With actively promoting for " smart city ", thousands of video monitoring cameras gradually cover various public arenas, Basic guarantee facility is provided for municipal public safety management.For the consideration of many factors such as anti-terrorism, national public safety, depending on Frequency monitoring system is in the urgent need to the quick identity recognizing technology under a kind of remote, non-mated condition of target, so as to remote When can quickly confirm pedestrian's identity, realize intelligent early-warning.And it is used as the important supplementary means of remote identity recognizing technology, pedestrian Techniques of Gender Recognition is an indispensable part in intelligent video monitoring system.
Pedestrian's Techniques of Gender Recognition has extensive market application foreground, such as Intelligent Human-Machine Interface field, and pedestrian's sex is known Other quantity statistics system and automatic passenger number statistical system etc., its foundation that can be determined as business and other important decisions.
It is expert in people's gender identification method, relatively common method is based on traditional manual feature (such as HOG, SIFT Deng) extract, then carry out feature learning (as with SVM SVMs) realization identification.Either based on popular at present The method of deep learning carries out depth characteristic extraction, study and identification.But due to single traditional-handwork feature extracting method Accuracy of identification is not universal high, although and deep learning can improve sex accuracy of identification, but its precision is dependent on a large amount of instructions Practice data, adjust ginseng skill and high-capability computing device.
The content of the invention
It is an object of the invention to overcome the shortcomings of existing method, the present invention proposes a kind of based on the learning machine that transfinites Pedestrian's gender identification method of (Extreme Learning Machine) and color characteristic fusion, solves traditional-handwork extraction Characterization method accuracy of identification is not high and deep learning relies on big data, adjusts the problem of joining skill and high-performance calculation hardware.
The present invention includes training stage and test phase, in the training stage, to one point of input training pedestrian's picture training Layer transfinites learning machine, obtains model and extracts the feature of training image, while extract input picture hsv color feature, by this two Category feature carries out effective integration and obtains a kind of fusion feature with high distinction, then utilizes linear branch to gained fusion feature Hold vector machine training grader.In test phase, the study spy of the learning machine model extraction test image that transfinites trained is utilized Levy, and merged with the hsv color feature of test image, fusion feature is entered with training gained linear SVM SVM Row classification, exports recognition result.
A kind of pedestrian's gender identification method based on transfinite learning machine and color characteristic fusion includes training process and identification Process, it is specific as follows:
Training process S1:The training image of unmarked gender attribute is inputted, is then fed into the learning machine that transfinites of layering and instructs Practice, obtain model M (M is to solve the output weight beta transfinited in hierarchical network obtained by the object function of learning machine autocoder), And the feature a of training image is extracted with training gained model M;
Training process S2:The input training image hsv color feature b of unmarked gender attribute is extracted, by training process S1 Obtain learning machine feature a and the color characteristic b that transfinites be combined, obtain fusion feature V=[a, b], according to fusion feature V with (training and test image select gained to training image label from PETA pedestrian images storehouse, and the image library provides every image Pedestrian's sex label) utilize linear SVM SVM training pedestrian's gender sorter;
Identification process S3:Gained model M is trained to extract test image feature a' with training process S1, test image feature is carried Take process identical with training image characteristic extraction procedure.While extraction hsv color feature b ', then two category features are melted Close, obtain the fusion feature V' of test image, gained linear SVM SVM is trained to fusion feature V' with training process S2 Classified.
It is preferred that, training process S1 also includes, and carries out the super of pretreatment retraining layering to the training image of input first Learning machine is limited, and is transfinited learning machine feature a with model extraction obtained by training;Identification process S3 also includes, first the survey to input Attempt extract again test image feature a' as to be pre-processed.
It is preferred that, described pretreatment, including:
3.1) gray processing processing is carried out to input color pedestrian image, the coloured image of input is converted into gray level image, Only retain monochrome information;
3.2) specific quantity m sub-block { x is extracted to pedestrian's gray level image of acquisition(1),...,x(m), sub-block x(i)Name For local receptor field (Local receptive field), m=8;
3.3) local contrast normalized is performed to each local receptor field respectively;
3.4) whitening processing is performed to each local receptor field respectively.
It is preferred that, it is described to each local receptor field execution local contrast normalized, it is specific as follows:
Wherein:For the pixel value of i-th of sub-block (local receptor field) (j, k) position,To return by local contrast The pixel value of i-th of sub-block (local receptor field) (j, k) position, k after one change1For the length of local receptor field, k2For local experiences Wild width, C is integer constant, and value is 10.
It is preferred that, it is described to each local receptor field execution whitening processing, it is specific as follows:
[D, U]=Eig (cov (Y)),
z(i)=U (D+diag (ε))-1/2UTy(i), i=1 ..., m
Wherein:z(i)For the sub-block (local receptor field) after whitening processing, size is k1×k2, Y is by m (m=8) height Block y(i)Constitute, cov () is covariance function, Eig () is characterized value analytic function, and D is characterized value, and U is characterized vector, diag () represents diagonal matrix, and ε is 0.1.
It is preferred that, the learning machine feature a that transfinites for extracting training image, including:
6.1) by k1×k2Local receptor field z(i)Column vector is converted to, dimension is k1k2, it is expressed as Z(i);Then by m (m =8) individual local receptor field be combined into the form of matrix:
The transfinite object function of learning machine autocoder of first layer is:
Wherein, W is random orthogonal weight,It is L for size1×L1Unit matrix, b to standardize random column vector,
Dimension is L1.I is the row vector that all elements are 1, and dimension is m (m=8), and σ () is sigmoid functions;α1, α2
To change the factor of influence that concealed nodes are distributed with output node;K is regularization factors.
Make H=σ (α1(WZ1+ bi)), then the output weight of first layer is:
6.2) the Feature Mapping figure of first layer is calculated according to 6.1) gained output weight betaIts In:I is input picture,Represent convolution operation;
6.3) in the layer network of learning machine second that transfinites, to the output characteristic mapping graph of first layerExtract specific quantity m (m =local receptor field 8), and local contrast normalized and whitening processing are performed to the local receptor field of acquisition;
6.4) to 6.3) obtaining local receptor field repeat step 6.1) and 6.2) operation, the learning machine that the transfinites acquisition second layer Feature Mapping figureWherein:L2For the number of second layer output characteristic mapping graph;
6.5) L for obtaining first layer1The L that individual Feature Mapping figure is obtained with the second layer2Individual Feature Mapping figure is cascaded, Obtain L1×(L2+ 1) individual Feature Mapping figure;
6.6) in the learning machine third layer network that transfinites, to the L for learning machine acquisition of transfiniting1×(L2+ 1) individual Feature Mapping figure enters Row binary quantization processing so that Feature Mapping figure only has 0 and 1 two kind of value;
6.7) to every L1Individual binary features mapping graph is quantified, will be per L1Individual binary features mapping graph is compressed into one Individual Feature Mapping figure, obtains L2Feature Mapping figure after+1 compressionFeature Mapping after compression The pixel value span of figure is 0 to 255;
6.8) block division is carried out to the characteristic pattern after compression, can obtain block is:
Wherein, BiFor to characteristic pattern obtain after block division i-th piece, x × y represents the Feature Mapping figure after compression Size, w1×w2Represent to divide block size, s1×s2Represent step-length;
6.9) using histc to each block BiBuild its histogram;
6.10) histogram of all pieces of cascade, obtains the feature of input pedestrian image:
Wherein:Wherein, NB=(L2+1)×[(x-w1)/s1+1]×[(y-w2)/s2+1];hist(Bi) represent to block BiStructure The histogram built;Represent that dimension is N in real number fieldBVector;Represent the number of bits in histogram;X is compression The length of Feature Mapping figure afterwards, y is the width of Feature Mapping figure after compression, w1For the length of divided block, w2For the width of divided block, s1For water Square to division step-length, s2For the division step-length of vertical direction.
It is preferred that, the input training image hsv color feature b for extracting unmarked gender attribute, including:
7.1) input pedestrian image is converted into HSV color spaces by rgb color space;
7.2) series is quantified to different colours (H, S, V) path setting difference for inputting pedestrian image:
HQL=8, SQL=2, VQL=2;
7.3) the respective max pixel value H in H in input pedestrian image, S, V passages is found outmax, Smax, Vmax
7.4) H, S, the quantized value of V passages of each pixel in input pedestrian image are calculated:
HQV=HQL×h(x,y)/Hmax,
SQV=SQL×s(x,y)/Smax,
VQV=VQL×v(x,y)/Vmax
Wherein, h (x, y) represents the pixel value of input pedestrian image H passages (x, y) point, and s (x, y) represents input pedestrian's figure As the pixel value of channel S (x, y) point, v (x, y) represents the pixel value of input pedestrian image V passages (x, y) point;
7.5) the hsv color histogram matrix A that size is 8 × 2 × 2 is created, 0 is initialized with;According to step 7.4) The quantized value of three passages of each pixel obtained by calculating counts each matrix coordinate [H as matrix coordinateQV,SQV,VQV] The number of times of appearance, is recorded as Q [HQV,SQV,VQV];
7.6) 8 × 2 × 2HSV color histogram matrix As are converted to 1 × 32 characteristic vector b.
It is preferred that, the obtaining step of the fusion feature V includes:
8.1) 2 kinds of different classes of features of n pedestrian images are obtained:a1, b1, a2, b2..., an, bnCarry out Fusion Features, Obtain fusion feature V:
Wherein, often row represents the fusion feature of single sample to matrix V;
8.2) normalized matrix V* is obtained by row standardization to matrix V:
Wherein, vminRepresent the minimum value of each row, vmaxThe maximum of each row is represented, v is the element of original matrix, and v* is Element obtained by after normalizing operation in matrix.
The present invention constructs a kind of pedestrian's sex identification model based on transfinite learning machine and color characteristic fusion, the model Transfinited learning machine using pedestrian image training layering, and optimize relevant parameter.Transfinited learning machine mould using the obtained layering of training Type extracts the feature of training image, then merges the hsv color feature of manual extraction, is instructed using Linear SVM SVMs Practice grader.In identification process, obtained layering is trained to transfinite the feature simultaneously combined test of learning machine model extraction test image The hsv color feature of image is then carried out with obtaining fusion feature with training gained linear SVM SVM to fusion feature Classification, exports recognition result.The present invention can be widely used in intelligent video monitoring system, Intelligent Human-Machine Interface field, OK People's sex is recognized among quantity statistics and automatic passenger number statistical system, is used as business and the foundation of other important decisions decision.
The present invention is described in further detail below in conjunction with drawings and Examples, but one kind of the present invention is based on transfinite Habit machine and pedestrian's gender identification method of color characteristic fusion are not limited to embodiment.
Brief description of the drawings
Fig. 1 is pedestrian sex recognizer flow chart of the present invention based on transfinite learning machine and color characteristic fusion;
Fig. 2 carries out block to the characteristic pattern after compression for the present invention and divides figure.
Embodiment
The present invention is described further below in conjunction with the accompanying drawings.
Pedestrian sex identification model of the present embodiment based on transfinite learning machine and color characteristic fusion, the model includes feature Extraction module and perceptron functional module, the embodiment specifically includes training process and identification process, shown in Figure 1, this hair A kind of bright pedestrian's sex in transfinite learning machine and color characteristic fusion knows method, and step is as follows:
1) training image of unmarked gender attribute is inputted, is then fed into the learning machine that transfinites of layering and trains, mould is obtained Type M (M for solve the output weight beta that transfinites obtained by the object function of learning machine autocoder in hierarchical network), and with training Gained model M extracts the feature a of training image;
2) the input training image hsv color feature b of unmarked gender attribute is extracted, by transfiniting that training process S1 is obtained Learning machine feature a is combined with color characteristic b, fusion feature V=[a, b] is obtained, according to fusion feature V and training image mark (training and test image select gained to label from PETA pedestrian images storehouse, and the image library provides pedestrian's sex of every image Label) pedestrian's gender sorter is trained (it should be noted that in machine learning field, supporting using linear SVM SVM Vector machine SVM, which is one, the learning model of supervision, is generally used for two classification problems, the present invention relates to support vector machines Place utilize prior art);
3) use training process 1) training gained model M extract test image feature a', test image characteristic extraction procedure with Training image characteristic extraction procedure is identical.While extraction hsv color feature b ', then two category features are merged, surveyed The fusion feature V' of picture is attempted, with training process 2) train gained linear SVM SVM to classify fusion feature V'.
Further, step 1) and step 3) in, input training to the coloured image of input with after test pedestrian image, entering The processing of row gray processing, gray level image is converted into by the coloured image of input, only retains monochrome information.
Further, the learning machine feature a that transfinites of training image is extracted, process is as follows:
In the first layer of learning machine that transfinites, specific quantity m sub-block { x is extracted to pedestrian's gray level image of acquisition(1),..., x(m), sub-block x(i)It is named as local receptor field (Local receptive field), m=8.
Then local contrast normalized is performed to each local receptor field of acquisition, it is specific as follows:
Wherein:For the pixel value of i-th of sub-block (local receptor field) (j, k) position,To pass through local contrast The pixel value of i-th of sub-block (local receptor field) (j, k) position, k after normalization1For the length of local receptor field, k2For local sense By wild width, C is integer constant, and value is 10.
Whitening processing is performed to each local receptor field of acquisition, it is specific as follows:
[D, U]=Eig (cov (Y)),
z(i)=U (D+diag (ε))-1/2UTy(i), i=1 ..., m
Wherein:z(i)For the sub-block (local receptor field) after whitening processing, size is k1×k2, Y is by m (m=8) height Block y(i)Constitute, cov () is covariance function, Eig () is characterized value analytic function, and D is characterized value, and U is characterized vector, diag () represents diagonal matrix, and ε is 0.1.
It is each after local contrast normalization and whitening processing followed by learning machine autocoder study of transfiniting Local receptor field, comprise the following steps that:
By k1×k2Local receptor field z(i)Column vector is converted to, dimension is k1k2, it is expressed as Z(i).Then by m (m=8) Individual local receptor field is combined into the form of matrix:
The transfinite object function of learning machine autocoder of first layer is:
Wherein:W is random orthogonal weight.It is L for size1×L1Unit matrix.B ties up to standardize random column vector Spend for L1.I is the row vector that all elements are 1, and dimension is m (m=8).σ () is sigmoid functions.α1, α2Hidden to change The factor of influence that node is distributed with output node.K is regularization factors.
Make H=σ (α1(WZ1+ bi)), then the output weight of first layer is:
The Feature Mapping figure of first layer is calculated according to the output weight beta of acquisitionWherein:I is Input picture,Represent convolution operation.
In the layer network of learning machine second that transfinites, to the output characteristic mapping graph of first layerExtract specific quantity m (m=8) Local receptor field, and local contrast normalized and whitening processing are performed to each local receptor field of acquisition.
Then each local receptor field obtained also with the learning machine autocoder (ELM-AE) that transfinites to the second layer Learnt, obtain the Feature Mapping figure of the second layerWherein:L2For of second layer output characteristic mapping graph Number.Then L first layer obtained1The L that individual Feature Mapping figure is obtained with the second layer2Individual Feature Mapping figure is cascaded, and obtains L1 ×(L2+ 1) individual Feature Mapping figure.
In the learning machine third layer network that transfinites, to the L for learning machine acquisition of transfiniting1×(L2+ 1) individual Feature Mapping figure carries out two System quantification treatment so that Feature Mapping figure only has two kinds of values, 0 and 1.
Then to every L1Individual binary features mapping graph is quantified, will be per L1Individual binary features mapping graph is compressed into one Individual Feature Mapping figure, obtains L2Feature Mapping figure after+1 compressionFeature Mapping figure after compression Pixel value span be 0 to 255.
Described in reference picture 2, block division is carried out to the characteristic pattern after compression, can obtain block is:
Wherein, BiFor to characteristic pattern obtain after block division i-th piece, x × y represents the Feature Mapping figure after compression Size, w1×w2Represent to divide block size, s1×s2Step-length is represented, x is the length of Feature Mapping figure after compression, and y is feature after compression The width of mapping graph.w1For the length of divided block, w2For the width of divided block.s1For the division step-length of horizontal direction, s2For vertical direction Divide step-length.
Using histc to each block BiBuild its histogram.
All pieces of histogram is finally cascaded, input picture feature is obtained:
Wherein, NB=(L2+1)×[(x-w1)/s1+1]×[(y-w2)/s2+1];hist(Bi) represent to block BiWhat is built is straight Fang Tu;Represent that dimension is N in real number fieldBVector;Represent the number of bits in histogram.
It should be noted that test image feature a' extraction process is identical with training image characteristic extraction procedure, this reality Apply in example and be not repeated.
Further, step 2) and step 3) in, hsv color feature extraction is extracted, extraction process is following (not to be marked to extract The input training image hsv color feature b of note gender attribute is illustrated, the hsv color feature b of test image ' extraction Journey is the same, and the present embodiment is not repeated):
Input pedestrian image is converted into HSV color spaces by rgb color space.
Path setting differences different to image quantify series:
HQL=8, SQL=2, VQL=2.
Find out the respective max pixel value in H in input pedestrian image, S, V passages:Hmax, Smax, Vmax
Calculate the H, S, the quantized value of V passages of each pixel in input pedestrian image:
HQV=HQL×h(x,y)/Hmax,
SQV=SQL×s(x,y)/Smax,
VQV=VQL×v(x,y)/Vmax
Wherein h (x, y) represents the pixel value of input pedestrian image H passages (x, y) point, and s (x, y) represents input pedestrian image The pixel value of channel S (x, y) point, v (x, y) represents the pixel value of input pedestrian image V passages (x, y) point.
The hsv color histogram matrix A that size is 8 × 2 × 2 is created, 0 is initialized with.Obtained by calculating The quantized value of three passages of each pixel counts each matrix coordinate [H as matrix coordinateQV,SQV,VQV] occur number of times, note Record as Q [HQV,SQV,VQV].If for example, [HQV,SQV,VQV]=[7,2,1] number of times that occurs is 3, then Q [HQV,SQV,VQV]= 3。
8 × 2 × 2HSV color histogram matrix As are converted to 1 × 32 characteristic vector b.
Further, the learning machine feature a that will transfinite is merged with color characteristic b, and fusion steps are as follows:
Obtain 2 kinds of different classes of features of n pedestrian images:a1, b1, a2, b2..., an, bnFusion Features are carried out, are obtained Fusion feature V:
Wherein:Often row represents the fusion feature of single sample to matrix V.
Normalized matrix V is obtained by row standardization to matrix V*
Wherein:vminRepresent the minimum value of each row, vmaxThe maximum of each row is represented, v is the element of original matrix, v*For Element obtained by after normalizing operation in matrix.
Above-described embodiment is intended merely to the explanation present invention, and is not used as limitation of the invention.As long as according to this hair Bright technical spirit, is changed, modification etc. will all fall in the range of the claim of the present invention to above-described embodiment.

Claims (8)

1. a kind of pedestrian's gender identification method based on transfinite learning machine and color characteristic fusion, it is characterised in that methods described Including training process and identification process, step is as follows:
Training process S1:The training image of unmarked gender attribute is inputted, being then fed into layering transfinites training in learning machine, obtains Model M, and with training gained model M extract training image the learning machine feature a that transfinites;
Training process S2:The input training image hsv color feature b of unmarked gender attribute is extracted, training process S1 is obtained The learning machine feature a and color characteristic b that transfinites be combined, obtain fusion feature V=[a, b], according to fusion feature V and training Image tag trains pedestrian's gender sorter using linear SVM SVM;
Identification process S3:Gained model M is trained to extract test image feature a' with training process S1, while extracting its hsv color Feature b ', then two category features are merged, the fusion feature V' of test image is obtained, with linear branch obtained by training process S2 Vector machine SVM pedestrian gender sorter is held to classify to fusion feature V'.
2. pedestrian's gender identification method according to claim 1 based on transfinite learning machine and color characteristic fusion, it is special Levy and be, training process S1 also includes, the training image of input is pre-processed transfinite again learning machine feature a first;Identification Process S3 also includes, and the test image of input is pre-processed extract again test image feature a' first.
3. pedestrian's gender identification method according to claim 2 based on transfinite learning machine and color characteristic fusion, it is special Levy and be, the pretreatment, including:
3.1) gray processing processing is carried out to the colorful pedestrian image of input, the coloured image of input is converted into gray level image, only Retain monochrome information;
3.2) specific quantity m sub-block { x is extracted to pedestrian's gray level image of acquisition(1),...,x(m)};Wherein, sub-block x(i)Name For local receptor field, m=8;
3.3) local contrast normalized is performed to each local receptor field respectively;
3.4) whitening processing is performed to each local receptor field respectively.
4. pedestrian's gender identification method according to claim 3 based on transfinite learning machine and color characteristic fusion, it is special Levy and be, it is described to each local receptor field execution local contrast normalized, it is specific as follows:
y j , k ( i ) = x j , k ( i ) - 1 k 1 k 2 Σ j = 1 k 1 Σ k = 1 k 2 x j , k ( i ) ( 1 k 1 k 2 Σ j = 1 k 1 Σ k = 1 k 2 ( x j , k ( i ) - 1 k 1 k 2 Σ j = 1 k 1 Σ k = 1 k 2 x j , k ( i ) ) 2 + C ) ,
Wherein:For the pixel value of i-th of sub-block (j, k) position,For i-th of sub-block after local contrast is normalized The pixel value of (j, k) position, k1For the length of local receptor field, k2For the width of local receptor field, C is integer constant, and value is 10。
5. pedestrian's gender identification method according to claim 3 based on transfinite learning machine and color characteristic fusion, it is special Levy and be, it is described to each local receptor field execution whitening processing, it is specific as follows:
[D, U]=Eig (cov (Y)),
z(i)=U (D+diag (ε))-1/2UTy(i), i=1 ..., m
Wherein:z(i)For the sub-block after whitening processing, size is k1×k2, Y is by m sub-block y(i)Constitute, cov () is association side Difference function, Eig () is characterized value analytic function, and D is characterized value, and U is characterized vector, and diag () represents diagonal matrix, and ε is 0.1。
6. pedestrian's gender identification method according to claim 2 based on transfinite learning machine and color characteristic fusion, it is special Levy and be, the learning machine feature a that transfinites for extracting training image, including:
6.1) by k1×k2Local receptor field z(i)Column vector is converted to, dimension is k1k2, it is expressed as Z(i);By m local experiences Open country is combined into the form of matrix:
Z 1 = [ Z ( 1 ) , Z ( 2 ) , ... , Z ( m ) ] ∈ R k 1 k 2 × m
The transfinite object function of learning machine autocoder of first layer is:
m i n K W , b | | Z 1 - α 2 β σ ( α 1 ( WZ 1 + b i ) ) | | L 2 2 + | | β | | L 2 2
s . t WW T = I L 1 , b T b = 1
Wherein:W is random orthogonal weight,It is L for size1×L1Unit matrix, b is standardizes random column vector, and dimension is L1, i is the row vector that all elements are 1, and dimension is m, and σ () is sigmoid functions;α1、α2To change concealed nodes and output The factor of influence of Node distribution;K is regularization factors;
Make H=σ (α1(WZ1+ bi)), then the output weight of first layer is:
β = ( β 1 , ... , β L 1 ) = 1 α 2 Z 1 H T ( I L 1 K + HH T ) - 1
6.2) the Feature Mapping figure of first layer is calculated according to output weight betaWherein, I schemes for input Picture,Represent convolution operation;
6.3) in the layer network of learning machine second that transfinites, to the output characteristic mapping graph of first layerExtract specific quantity m part Receptive field, and local contrast normalized and whitening processing are performed to the local receptor field of acquisition;
6.4) to 6.3) obtaining local receptor field repeat step 6.1) and 6.2) operation, the spy of the learning machine that the transfinites acquisition second layer Levy mapping graphWherein, L2For the number of second layer output characteristic mapping graph;
6.5) L for obtaining first layer1The L that individual Feature Mapping figure is obtained with the second layer2Individual Feature Mapping figure is cascaded, and is obtained L1×(L2+ 1) individual Feature Mapping figure;
6.6) in the learning machine third layer network that transfinites, to the L for learning machine acquisition of transfiniting1×(L2+ 1) individual Feature Mapping figure carries out two System quantification treatment so that Feature Mapping figure only has 0 and 1 two kind of value;
6.7) to every L1Individual binary features mapping graph is quantified, will be per L1Individual binary features mapping graph is compressed into a spy Mapping graph is levied, L is obtained2Feature Mapping figure after+1 compressionFeature Mapping figure after compression Pixel value span is 0 to 255;
6.8) block division is carried out to the characteristic pattern after compression, obtaining block is:
{ B 1 B 2 , ... , B ( L 2 + 1 ) × [ ( x - w 1 ) / s 1 + 1 ] × [ ( y - w 2 ) / s 2 + 1 ] }
Wherein, BiFor to characteristic pattern obtain after block division i-th piece, x × y represents the Feature Mapping figure size after compression, w1×w2Represent to divide block size, s1×s2Represent step-length;
6.9) using histc to each block BiBuild its histogram;
6.10) histogram of all pieces of cascade, obtains the learning machine feature that transfinites of input pedestrian image:
a = f ( Im a g e ) = [ h i s t ( B 1 ) T h i s t ( B 2 ) T , ... , h i s t ( B N B ) T ] T ∈ R N B ( 2 L 1 )
Wherein, NB=(L2+1)×[(x-w1)/s1+1]×[(y-w2)/s2+1];hist(Bi) represent to block BiThe Nogata of structure Figure;Represent that dimension is N in real number fieldBVector;Represent the number of bits in histogram;X reflects for feature after compression The length of figure is penetrated, y is the width of Feature Mapping figure after compression, w1For the length of divided block, w2For the width of divided block, s1For horizontal direction Divide step-length, s2For the division step-length of vertical direction.
7. pedestrian's gender identification method according to claim 1 based on transfinite learning machine and color characteristic fusion, it is special Levy and be, the input training image hsv color feature b for extracting unmarked gender attribute, including:
7.1) input pedestrian image is converted into HSV color spaces by rgb color space;
7.2) series is quantified to different colours (H, S, V) path setting difference for inputting pedestrian image:
HQL=8, SQL=2, VQL=2;
7.3) the respective max pixel value H in H in input pedestrian image, S, V passages is found outmax, Smax, Vmax
7.4) H, S, the quantized value of V passages of each pixel in input pedestrian image are calculated:
HQV=HQL×h(x,y)/Hmax,
SQV=SQL×s(x,y)/Smax,
VQV=VQL×v(x,y)/Vmax
Wherein, h (x, y) represents the pixel value of input pedestrian image H passages (x, y) point, and s (x, y) represents that input pedestrian image S leads to The pixel value of road (x, y) point, v (x, y) represents the pixel value of input pedestrian image V passages (x, y) point;
7.5) the hsv color histogram matrix A that size is 8 × 2 × 2 is created, 0 is initialized with;According to step 7.4) calculate The quantized value of resulting three passages of each pixel counts each matrix coordinate [H as matrix coordinateQV,SQV,VQV] occur Number of times, be recorded as Q [HQV,SQV,VQV];
7.6) 8 × 2 × 2HSV color histogram matrix As are converted to 1 × 32 characteristic vector b.
8. pedestrian's gender identification method according to claim 1 based on transfinite learning machine and color characteristic fusion, it is special Levy and be, the obtaining step of the fusion feature V includes:
8.1) 2 kinds of different classes of features of n pedestrian images are obtained:a1, b1, a2, b2..., an, bnFusion Features are carried out, are obtained Fusion feature V:
V = a 1 b 1 a 2 b 2 . . . . . . a n b n
Wherein, often row represents the fusion feature of single sample to matrix V;
8.2) normalized matrix V is obtained by row standardization to matrix V*
v * = v - v m i n v m a x - v m i n ,
Wherein, vminRepresent the minimum value of each row, vmaxThe maximum of each row is represented, v is the element of original matrix, v*For standard Element obtained by changing after operation in matrix.
CN201710096262.6A 2017-02-22 2017-02-22 Pedestrian gender identification method based on transfinite learning machine and color feature fusion Active CN106960176B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201710096262.6A CN106960176B (en) 2017-02-22 2017-02-22 Pedestrian gender identification method based on transfinite learning machine and color feature fusion

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710096262.6A CN106960176B (en) 2017-02-22 2017-02-22 Pedestrian gender identification method based on transfinite learning machine and color feature fusion

Publications (2)

Publication Number Publication Date
CN106960176A true CN106960176A (en) 2017-07-18
CN106960176B CN106960176B (en) 2020-03-10

Family

ID=59480972

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710096262.6A Active CN106960176B (en) 2017-02-22 2017-02-22 Pedestrian gender identification method based on transfinite learning machine and color feature fusion

Country Status (1)

Country Link
CN (1) CN106960176B (en)

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109063535A (en) * 2018-05-30 2018-12-21 华侨大学 It is a kind of based on combined depth study pedestrian recognize again and pedestrian's gender classification method
CN109389092A (en) * 2018-10-22 2019-02-26 北京工业大学 A kind of local enhancement multitask depth migration transfinites the facial video fatigue detection method of learning machine and individual robust
CN109584172A (en) * 2018-11-12 2019-04-05 北京农业信息技术研究中心 Backlight compensation method and device based on the fuzzy learning machine that transfinites of iteration
CN110838150A (en) * 2019-11-18 2020-02-25 重庆邮电大学 Color recognition method for supervised learning
CN111191526A (en) * 2019-12-16 2020-05-22 汇纳科技股份有限公司 Pedestrian attribute recognition network training method, system, medium and terminal
CN111461147A (en) * 2020-04-30 2020-07-28 柳州智视科技有限公司 Binary coding organization algorithm based on image characteristics
CN113077388A (en) * 2021-04-25 2021-07-06 中国人民解放军国防科技大学 Data-augmented deep semi-supervised over-limit learning image classification method and system
CN113450347A (en) * 2021-07-20 2021-09-28 中国石油大学(华东) Tower-assembling progress identification method for power transmission iron tower

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101706964A (en) * 2009-08-27 2010-05-12 北京交通大学 Color constancy calculating method and system based on derivative structure of image
US20110257893A1 (en) * 2008-10-10 2011-10-20 Ian Taylor Methods for classifying samples based on network modularity
US20120105613A1 (en) * 2010-11-01 2012-05-03 Robert Bosch Gmbh Robust video-based handwriting and gesture recognition for in-car applications
WO2012170393A2 (en) * 2011-06-06 2012-12-13 Aaron Marshall System and method for providing thermal gender recognition
CN104463252A (en) * 2014-12-18 2015-03-25 南京信息工程大学 Foundation cloud classification method based on self-adaptive extreme learning machine
CN105069400A (en) * 2015-07-16 2015-11-18 北京工业大学 Face image gender recognition system based on stack type sparse self-coding
CN105183748A (en) * 2015-07-13 2015-12-23 电子科技大学 Combined forecasting method based on content and score
CN106203528A (en) * 2016-07-19 2016-12-07 华侨大学 A kind of feature based merges and the 3D of KNN draws intelligent classification algorithm

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110257893A1 (en) * 2008-10-10 2011-10-20 Ian Taylor Methods for classifying samples based on network modularity
CN101706964A (en) * 2009-08-27 2010-05-12 北京交通大学 Color constancy calculating method and system based on derivative structure of image
US20120105613A1 (en) * 2010-11-01 2012-05-03 Robert Bosch Gmbh Robust video-based handwriting and gesture recognition for in-car applications
WO2012170393A2 (en) * 2011-06-06 2012-12-13 Aaron Marshall System and method for providing thermal gender recognition
CN104463252A (en) * 2014-12-18 2015-03-25 南京信息工程大学 Foundation cloud classification method based on self-adaptive extreme learning machine
CN105183748A (en) * 2015-07-13 2015-12-23 电子科技大学 Combined forecasting method based on content and score
CN105069400A (en) * 2015-07-16 2015-11-18 北京工业大学 Face image gender recognition system based on stack type sparse self-coding
CN106203528A (en) * 2016-07-19 2016-12-07 华侨大学 A kind of feature based merges and the 3D of KNN draws intelligent classification algorithm

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
LIJUAN DUAN等: "Multimodal Fusion Using Kernel-Based", 《PROCEEDINGS IN ADAPTATION, LEARNING AND OPTIMIZATION 6》 *
童冰等: "基于局部域超限学习机的石材识别算法", 《闽南师范大学学报(自然科学版)》 *

Cited By (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109063535B (en) * 2018-05-30 2021-09-28 华侨大学 Pedestrian re-identification and pedestrian gender classification method based on joint deep learning
CN109063535A (en) * 2018-05-30 2018-12-21 华侨大学 It is a kind of based on combined depth study pedestrian recognize again and pedestrian's gender classification method
CN109389092A (en) * 2018-10-22 2019-02-26 北京工业大学 A kind of local enhancement multitask depth migration transfinites the facial video fatigue detection method of learning machine and individual robust
CN109389092B (en) * 2018-10-22 2023-05-02 北京工业大学 Locally-enhanced multitasking depth migration overrun learning machine and individual robust facial video fatigue detection method
CN109584172A (en) * 2018-11-12 2019-04-05 北京农业信息技术研究中心 Backlight compensation method and device based on the fuzzy learning machine that transfinites of iteration
CN110838150B (en) * 2019-11-18 2022-07-15 重庆邮电大学 Color recognition method for supervised learning
CN110838150A (en) * 2019-11-18 2020-02-25 重庆邮电大学 Color recognition method for supervised learning
CN111191526A (en) * 2019-12-16 2020-05-22 汇纳科技股份有限公司 Pedestrian attribute recognition network training method, system, medium and terminal
CN111191526B (en) * 2019-12-16 2023-10-10 汇纳科技股份有限公司 Pedestrian attribute recognition network training method, system, medium and terminal
CN111461147A (en) * 2020-04-30 2020-07-28 柳州智视科技有限公司 Binary coding organization algorithm based on image characteristics
CN111461147B (en) * 2020-04-30 2023-05-23 柳州智视科技有限公司 Binary coding organization algorithm based on image features
CN113077388A (en) * 2021-04-25 2021-07-06 中国人民解放军国防科技大学 Data-augmented deep semi-supervised over-limit learning image classification method and system
CN113077388B (en) * 2021-04-25 2022-08-09 中国人民解放军国防科技大学 Data-augmented deep semi-supervised over-limit learning image classification method and system
CN113450347A (en) * 2021-07-20 2021-09-28 中国石油大学(华东) Tower-assembling progress identification method for power transmission iron tower

Also Published As

Publication number Publication date
CN106960176B (en) 2020-03-10

Similar Documents

Publication Publication Date Title
CN106650806B (en) A kind of cooperating type depth net model methodology for pedestrian detection
CN106960176A (en) A kind of pedestrian's gender identification method based on transfinite learning machine and color characteristic fusion
CN103761531B (en) The sparse coding license plate character recognition method of Shape-based interpolation contour feature
CN106815604B (en) Method for viewing points detecting based on fusion of multi-layer information
CN110399821B (en) Customer satisfaction acquisition method based on facial expression recognition
CN105574550B (en) A kind of vehicle identification method and device
CN104077613B (en) Crowd density estimation method based on cascaded multilevel convolution neural network
CN109359559B (en) Pedestrian re-identification method based on dynamic shielding sample
CN110889672B (en) Student card punching and class taking state detection system based on deep learning
CN108549926A (en) A kind of deep neural network and training method for refining identification vehicle attribute
CN108491797A (en) A kind of vehicle image precise search method based on big data
CN101667245B (en) Human face detection method by cascading novel detection classifiers based on support vectors
CN108171184A (en) Method for distinguishing is known based on Siamese networks again for pedestrian
CN109255284B (en) Motion trajectory-based behavior identification method of 3D convolutional neural network
CN107945153A (en) A kind of road surface crack detection method based on deep learning
CN106126585B (en) The unmanned plane image search method combined based on quality grading with perceived hash characteristics
CN105590099B (en) A kind of more people's Activity recognition methods based on improvement convolutional neural networks
CN103679191B (en) An automatic fake-licensed vehicle detection method based on static state pictures
CN106909902A (en) A kind of remote sensing target detection method based on the notable model of improved stratification
CN103514456A (en) Image classification method and device based on compressed sensing multi-core learning
CN105427309A (en) Multiscale hierarchical processing method for extracting object-oriented high-spatial resolution remote sensing information
CN106408030A (en) SAR image classification method based on middle lamella semantic attribute and convolution neural network
CN103942564A (en) High-resolution remote sensing image scene classifying method based on unsupervised feature learning
CN104978567A (en) Vehicle detection method based on scenario classification
CN109919106A (en) Gradual target finely identifies and description method

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
EE01 Entry into force of recordation of patent licensing contract
EE01 Entry into force of recordation of patent licensing contract

Application publication date: 20170718

Assignee: YEALINK (XIAMEN) NETWORK TECHNOLOGY Co.,Ltd.

Assignor: HUAQIAO University

Contract record no.: X2021990000281

Denomination of invention: A pedestrian gender recognition method based on transfinite learning machine and color feature fusion

Granted publication date: 20200310

License type: Common License

Record date: 20210517