CN104573744A - Fine granularity classification recognition method and object part location and feature extraction method thereof - Google Patents

Fine granularity classification recognition method and object part location and feature extraction method thereof Download PDF

Info

Publication number
CN104573744A
CN104573744A CN201510026025.3A CN201510026025A CN104573744A CN 104573744 A CN104573744 A CN 104573744A CN 201510026025 A CN201510026025 A CN 201510026025A CN 104573744 A CN104573744 A CN 104573744A
Authority
CN
China
Prior art keywords
feature
expression
detector
detecting device
lambda
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201510026025.3A
Other languages
Chinese (zh)
Other versions
CN104573744B (en
Inventor
熊红凯
张晓鹏
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shanghai Jiaotong University
Original Assignee
Shanghai Jiaotong University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shanghai Jiaotong University filed Critical Shanghai Jiaotong University
Priority to CN201510026025.3A priority Critical patent/CN104573744B/en
Publication of CN104573744A publication Critical patent/CN104573744A/en
Application granted granted Critical
Publication of CN104573744B publication Critical patent/CN104573744B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/23Clustering techniques
    • G06F18/232Non-hierarchical techniques
    • G06F18/2321Non-hierarchical techniques using statistics or function optimisation, e.g. modelling of probability density functions
    • G06F18/23213Non-hierarchical techniques using statistics or function optimisation, e.g. modelling of probability density functions with fixed number of clusters, e.g. K-means clustering
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/103Static body considered as a whole, e.g. static pedestrian or occupant recognition

Landscapes

  • Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • General Engineering & Computer Science (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Probability & Statistics with Applications (AREA)
  • Human Computer Interaction (AREA)
  • Multimedia (AREA)
  • Image Analysis (AREA)

Abstract

The invention provides a fine granularity classification recognition method and an object part location and feature extraction method thereof. The fine granularity classification recognition method and the object part location and feature extraction method thereof well achieve object part location and feature expression in fine granularity classification recognition. For object part location, a series of part detectors trained by supervised learning are utilized, the methods just detect the part with small deformation in consideration of the posture change and deformation influence of targets to be located, different detectors are trained for the same object part by adopting the posture clustering method, and therefore the posture change of objects is taken into account. For feature expression of the objects or parts, features are extracted at multiple dimensions and multiple positions according to the methods and then fused to be used for final object expression, and therefore the features have certain dimension and translation invariance. According to the methods, object part location and feature expression have certain complementarity at the same time, and therefore the accuracy of fine granularity classification recognition can be effectively improved.

Description

The part location of the identification of fine granulation classification and object and feature extracting method
Technical field
The present invention relates to a kind of method of technical field of image processing, specifically, what relate to is the recognition methods of a kind of fine granulation classification, and the part of the object related in this identification problem is located and feature extracting method.
Background technology
The target of fine granulation classification problem is the hundreds of multiple subclass under the same large class of differentiation, such as distinguishes different classes of flower, bird, dog etc.For layman, identify that these subclasses are very difficult, the proposition of fine granulation classification problem solves the problem that layman identifies these similar subclasses.User only needs given target object, by meticulous classification recognition methods, just can return the classification of target object, and then can obtain this subclass series of characteristics.Be different from general category identification problem (such as distinguishing car and people), because the comparison in difference between subclass is little and high localized, it is very difficult for distinguishing these subclasses.The spatial pyramid model being widely used in general category identification problem, owing to can not catch this high localized sub-class inherited, thus can not reach satisfied recognition result.
Through finding the literature search of prior art, the difficult point of fine granulation classification problem is mainly in two, and namely part is located and iamge description.That part location extensively adopts is the ``A discriminativelytrained that P.Felzenszwalb in 2010, " IEEETransactions on Pattern Analysis and Machine Intelligence " delivers, multiscale, deformable part model ", i.e. deformable segment model and its mutation.This model finds target object or partial target object by training template detector, and take into account the geometric relativity between department pattern.But only smaller to the deformation part of this model has good Detection results, the part larger to deformation ratio, the wing of such as bird, the poor performance of part detection model.For iamge description, most employing D.G.Lowe was published in the ``Distinctive imagefeatures from scale-invariant keypoints on " International Journal of Computer Vision " in 2004 ", i.e. scale invariant feature.But this feature is only the combination of some gradient informations, independent of concrete data set, do not possess good separating capacity.Other feature such as Krizhevsky was published in the ``Imagenet classificationwith deep convolutional neural networks on " Neural Information Processing Systems " in 2010 "; i.e. volume and neural network characteristics; although this feature is for the abundant feature of the semanteme of design data, lack enough yardsticks and translation invariance.When the fractional object detected and actual position have relatively large deviation, this feature well can not overcome this translation change.
Summary of the invention
For defect of the prior art, the object of this invention is to provide part location and the feature extracting method of the identification of a kind of fine granulation classification and object thereof, improve scale invariability and the translation invariance of precision and the feature representation of partly locating, thus improve the accuracy of identification of meticulous category classification problem.
The present invention is achieved by the following technical solutions:
According to a first aspect of the invention, a kind of part localization method of object is provided, that one divides sector of breakdown localization method for fine granulation, the method utilizes object detector and partial detector to detect target object and its deformation smaller portions, described detecting device utilizes the measure of supervision that has of attitude cluster to learn to obtain, and take into account the attitudes vibration of object or part; Object detector and partial detector independently carry out, and return surveyed area that in each detecting device, score is high alternatively, and final testing result is by correcting object and partial detection obtains.
Preferably, described detecting device utilizes the measure of supervision that has of attitude cluster to learn to obtain, and is specially: for object and each part, assembles positive example sample to some mixture models according to attitude;
Suppose each part p iall use a bounding box definition, whole object is with bounding box p 0express, wherein (l, t, r, b) represents the left side of bounding box, top, right side and bottom coordinate position; By following vector, these parts of demarcating are used for the attitude θ of parametrization sample I i:
θ I=(p′ 1,p′ 2,...,p′ n)
p i ′ = ( p i l + p i r 2 w , p i t + p i b 2 h ) , i = 1,2 , . . . , n
Wherein, w and h represents object p 0width and height, n represents the quantity .p ' of object parts ip inormalization express, this normalized expression makes it possible to only to consider the relative position of part, and ignores the different scale between different objects part; All positive samples, according to attitude characteristic, utilize k-means clustering method to be clustered into C composition.
Further, likely inconsistent with the position of object for solving object parts in the testing result that returns, described object detector and partial detector return surveyed area that in each detecting device, score is high alternatively, are specially:
Make X={x 0, x 1..., x nrepresent the testing result that the score of object and corresponding n part thereof is high, φ (X)={ φ (x 0), φ (x 1) ..., φ (x n) represent corresponding convolution feature, a series of detecting device { w that given training obtains 0, w 1..., w n, upgrade testing result by optimizing following expression:
arg max X Ψ ( w 0 T φ ( x 0 ) ) + Σ i = 1 n [ λ i ] ϵ Ψ ( w i T φ ( x i ) )
Wherein
&Psi; ( z ) = e z - e - z e z + e - z [ &lambda; i ] &epsiv; = &lambda; i , if &lambda; i &GreaterEqual; &epsiv; 0 , if &lambda; i < &epsiv;
Ψ [] is a nonlinear function, and detection score is mapped to scope [-1,1], [] εit is a loss function; Parameter lambda ithe degree of overlapping of tolerance part and object, scope is [0,1]; Weighted term [λ i] εbe used for punishing the part situation inconsistent with object detected.
Part location of the present invention is only for the part that object deformation is less, and the training of detecting device take into account the change of aspect, is upgraded the relation of object and the part thereof detected, obtain the positioning precision of reliable object and part thereof by geometry.
According to a second aspect of the invention, a kind of feature extracting method is provided, the method extracts constant convolution feature on each object parts positioning result, namely convolution feature is extracted at multiple yardstick and multiple visual angle, these convolution features are carried out merging and are obtained final feature representation, and this expression is used for final classification.
Further, described feature extracting method, comprises the steps:
Step one: to given scalogram picture, extracts the 5th convolutional layer characteristic pattern f w × h × C, wherein w × h represents convolved image size, the port number of C representative feature figure; Input picture is 16 relative to the down-sampling ratio of the 5th convolutional layer, means that the 5th convolutional layer characteristic pattern is 16 relative to the step-length of input picture;
Step 2: carry out zero padding operation to the border of each channel of characteristic pattern, every side increases by two pixels, obtains the characteristic pattern f ' that zero padding is later thus w ' × h ' × C; The characteristic pattern f ' later to zero padding w ' × h ' × C, each passage uses slip window sampling select any subgraph f with step-length 1 w × h × C, therefore always having 5 × 5 relative to the upper left corner biased (Δ x, Δ y) is { subgraph of 0,1,2,3,4}; Then carry out pondization operation to each subgraph, obtaining target, to export size be the later subgraph of the pondization of n × n;
Step 3: use the pond beggar figure calculated for subsequent full UNICOM layer characteristic pattern obtained in step 2.
Preferably, aforesaid operations carries out on 5 yardsticks of input picture and flip horizontal image thereof, finally altogether obtain 25 × 5 × 2 proper vectors, these proper vectors obtain the single features expression on each yardstick carry out pondization operation respectively on each yardstick after, the feature on these multiple yardsticks of last cascade is used for the final expression to image.This feature makes it have certain yardstick and translation invariance.
According to a third aspect of the invention we, provide a kind of image fine granulation that improves to know method for distinguishing, comprise the steps:
The first step: for test pattern, utilizes the part that object detector and partial detector detection target object and deformation thereof are little, and this detecting device utilizes the measure of supervision that has of attitude cluster to learn to obtain, and take into account the attitudes vibration of object or part.Because object detector and partial detector independently carry out, do not consider the geometric relationship between them.As improvement, the method returns some high surveyed areas of score in each detecting device alternatively, and final testing result is obtained by correction object and partial detection.
Second step, extracts convolution feature to object of detecting each in the first step or part at multiple yardstick and multiple visual angle, and these convolution features are carried out merging and obtained final feature representation, and this expression is used for final classification.The present invention can improve the identification of image fine granulation.
To sum up, the inventive method solves part orientation problem and the feature representation problem of object in fine granulation classification identification problem preferably, the method increases part detection perform, and makes this feature have certain yardstick and translation invariance.Between object parts location of the present invention and feature representation, there is certain complementarity simultaneously, thus effectively can improve the precision of meticulous classification identification problem.
Compared with prior art, the present invention has following beneficial effect:
Technique scheme of the present invention solves part orientation problem and the feature representation problem of object in fine granulation classification identification problem preferably.Part location of the present invention and feature representation all make use of the current convolutional neural networks with better ability to express.Present invention employs the strong supervised learning method training objective detecting device based on attitude cluster, and geometry renewal is carried out to final testing result, part positioning precision comparatively accurately can be obtained.Meanwhile, constant feature representation technology can overcome the inaccuracy of location to a certain extent, makes it have certain yardstick and translation invariance.The combination of two kinds of methods makes the present invention can obtain good recognition performance in fine granulation classification problem.
Accompanying drawing explanation
By reading the detailed description done non-limiting example with reference to the following drawings, other features, objects and advantages of the present invention will become more obvious:
Fig. 1 is the principle framework figure of one embodiment of the invention;
Fig. 2 is the invariant feature extraction process flow diagram of one embodiment of the invention.
Embodiment
Below in conjunction with specific embodiment, the present invention is described in detail.Following examples will contribute to those skilled in the art and understand the present invention further, but not limit the present invention in any form.It should be pointed out that to those skilled in the art, without departing from the inventive concept of the premise, some distortion and improvement can also be made.These all belong to protection scope of the present invention.
As shown in Figure 1, be the principle framework of one embodiment of the invention, this framework comprises two parts, i.e. the part localization part of object and yardstick and translation invariant feature representation part.A given width test pattern, first utilizes object detector and partial detector to detect target object and its deformation smaller portions, and this detecting device utilizes the measure of supervision that has of attitude cluster to learn to obtain, and take into account the attitudes vibration of object or part.Because object detector and partial detector independently carry out, do not consider the geometric relationship between them.As improvement, return surveyed area that in each detecting device, score is higher alternatively, final testing result is by correcting object and partial detection obtains.Then, characteristic extracting module extracts convolution feature to the object of each detection or part at multiple yardstick and multiple visual angle, and these convolution features are carried out merging and obtained final feature representation, and this expression is used for final classification.
As a preferred implementation, described object and part detect specific implementation process and comprise the steps:
Step one:
Attitude cluster: the method uses the learning method training detecting device of strong supervision, and for training sample, the bounding box of whole object and some object parts is all known.For object and each part, the method assembles positive example sample to some mixture models according to attitude.Suppose each part p iall use a bounding box (whole object is with bounding box p in definition 0express), wherein (l, t, r, b) represents the left side of bounding box, top, right side and bottom coordinate position.By following vector, these parts of demarcating are used for the attitude θ of parametrization sample I i:
θ I=(p′ 1,p′ 2,...,p′ n)
p i &prime; = ( p i l + p i r 2 w , p i t + p i b 2 h ) , i = 1,2 , . . . , n
Wherein, w and h represents object p 0width and height, n represents the quantity .p ' of object parts ip inormalization express, this normalized expression makes it possible to only to consider the relative position of part, and ignores the different scale between different objects part.All positive samples, according to attitude characteristic, utilize k-means clustering method to be clustered into C composition.This cluster take into account the attitudes vibration of object, and this training for detecting device is very important.
Step 2:
Convolutional network training and the study of detection: the feature for detecting device training is extracted from convolutional network.In order to make convolutional network adapt to concrete meticulous categorical data collection, first convolutional neural networks should be finely tuned.Because training sample is limited, first use and select searching method to produce a series of subregion image, wherein the overlapping subsample being greater than 0.5 of all and original positive example sample is all regarded as positive example, and every other subsample is regarded as negative example, obtains thus finely tuning later convolutional neural networks.In detecting device training process, only only have original sample feature to be regarded as positive example, the subsample that those and original sample degree of overlapping are less than 0.3 is regarded as negative example.To each part of object and object, a series of detecting device { w can be obtained by stand-alone training 0, w 1..., w n.
When test, a width test pattern is used equally and selects the method for search to produce a series of candidate's subregion.The feature of each candidate's subregion x represents with φ (x), accordingly for detecting device w iscore be then expressed as the region (such as 100, this region quantity can set as required) that wherein score is higher is selected as couple candidate detection result.
Step 3:
Object and part detect renewal: owing to being independently carry out the detection of object and object parts, in the testing result returned, object parts is likely inconsistent with the position of object.A kind of geometry update method is used to address this problem.Make X={x 0, x 1..., x nrepresent the testing result in object and higher 100 regions (this region quantity can set as required) of corresponding n score partly, φ (X)={ φ (x 0), φ (x 1) ..., φ (x n) represent corresponding convolution feature.A series of detecting device { w that given training obtains 0, w 1..., w n, upgrade testing result by optimizing following expression:
arg max X &Psi; ( w 0 T &phi; ( x 0 ) ) + &Sigma; i = 1 n [ &lambda; i ] &epsiv; &Psi; ( w i T &phi; ( x i ) )
Wherein
&Psi; ( z ) = e z - e - z e z + e - z [ &lambda; i ] &epsiv; = &lambda; i , if &lambda; i &GreaterEqual; &epsiv; 0 , if &lambda; i < &epsiv;
Ψ [z] is a nonlinear function, and detection score is mapped to scope [-1,1], [λ i] εit is a loss function.Parameter lambda ithe degree of overlapping of tolerance part and object, scope is [0,1].Weighted term [λ i] ε(ε=0.6) is used for punishing the part situation inconsistent with object detected.
As shown in Figure 2, test pattern to different yardsticks by resize ratio, for each yardstick, is extracted feature and mainly comprises the steps: characteristic extraction part
Step one: to given scalogram picture, extracts the 5th convolutional layer characteristic pattern f w × h × C, wherein w × h represents convolved image size, the port number of C representative feature figure.Input picture is 16 relative to the down-sampling ratio of the 5th convolutional layer, means that the 5th convolutional layer characteristic pattern is 16 relative to the step-length of input picture.
Step 2: carry out zero padding operation to the border of each channel of characteristic pattern, every side increases by two pixels, obtains the characteristic pattern f ' that zero padding is later thus w ' × h ' × C.The characteristic pattern f ' later to zero padding w ' × h ' × C, each passage uses slip window sampling select any subgraph f with step-length 1 w × h × C, therefore always having 5 × 5 relative to the upper left corner biased (Δ x, Δ y) is { subgraph of 0,1,2,3,4}.Then carry out pondization operation to each subgraph, obtaining target, to export size be the later subgraph of the pondization of n × n.
Step 3: use the pond beggar figure calculated for subsequent full UNICOM layer characteristic pattern obtained in step 2.
Aforesaid operations carries out on 5 yardsticks of input picture and flip horizontal image thereof, therefore finally altogether obtains 25 × 5 × 2 proper vectors.These proper vectors obtain the single features expression on each yardstick carry out pondization operation respectively on each yardstick after, the feature on these multiple yardsticks of last cascade is used for the final expression to image.
Implementation result:
Test and carry out on the fine granulation data set CUB-200-2011 extensively adopted.This data set comprises 200 different types of birds, altogether 11788 width images.Identify that these subclasses are all very difficult concerning people.Choosing in deformation smaller portions, only chooses head and health as part detected object.Choosing in attitude cluster, each detects target and is clustered into 3 mixture models, and during feature extraction, 5 yardsticks are chosen for { 227,280,340,400,454}.Final experimental standard is weighed with nicety of grading.
Object/part positioning precision result:
Positioning precision is weighed with the ratio of correctly locating, and the principle of correct location is that the target that detects and realistic objective degree of overlapping are greater than 0.5.To object, head and body part, the method can obtain the position precision of 96.36%, 75.22%70.14% respectively.
Classification results:
Have benefited from higher object/part positioning precision, the classifying identification method based on part finally can obtain the discrimination of 77.51%, under same experiment condition, far above existing accuracy of identification.The validity of the method comes from part positioning precision and constant feature representation form accurately, these two parts have complementary relationship again simultaneously, namely the unchangeability of feature expresses the inaccuracy that compensate for location to a certain extent, further increases final image recognition precision.
Above specific embodiments of the invention are described.It is to be appreciated that the present invention is not limited to above-mentioned particular implementation, those skilled in the art can make various distortion or amendment within the scope of the claims, and this does not affect flesh and blood of the present invention.

Claims (7)

1. the part localization method of object in fine granulation classification identification, it is characterized in that, the method utilizes object detector and partial detector to detect target object and its deformation smaller portions, described detecting device utilizes the measure of supervision that has of attitude cluster to learn to obtain, and take into account the attitudes vibration of object or part; Object detector and partial detector independently carry out, and return surveyed area that in each detecting device, score is high alternatively, and final testing result is by correcting object and partial detection obtains.
2. the part localization method of object according to claim 1, it is characterized in that, described detecting device utilizes the measure of supervision that has of attitude cluster to learn to obtain, and is specially: for object and each part, assembles positive example sample to some mixture models according to attitude;
Suppose each part p iall use a bounding box definition, whole object is with bounding box p 0express, wherein (l, t, r, b) represents the left side of bounding box, top, right side and bottom coordinate position; By following vector, these parts of demarcating are used for the attitude θ of parametrization sample I i:
θ I=(p′ 1,p′ 2,...,p′ n)
p i &prime; = ( p i l + p i r 2 w , p i t + p i b 2 h ) , i = 1,2 , . . . , n
Wherein, w and h represents object p 0width and height, n represents the quantity .p ' of object parts ip inormalization express, this normalized expression makes it possible to only to consider the relative position of part, and ignores the different scale between different objects part; All positive samples, according to attitude characteristic, utilize k-means clustering method to be clustered into C composition.
3. the part localization method of object according to claim 2, it is characterized in that, for object parts in the testing result that solution returns is likely inconsistent with the position of object, described object detector and partial detector return surveyed area that in each detecting device, score is high alternatively, are specially:
Make X={x 0, x 1..., x nrepresent the testing result that the score of object and corresponding n part thereof is high, φ (X)={ φ (x 0), φ (x 1) ..., φ (x n) represent corresponding convolution feature, a series of detecting device { w that given training obtains 0, w 1..., w n, upgrade testing result by optimizing following expression:
arg max X &Psi; ( w 0 T &phi; ( x 0 ) ) + &Sigma; i = 1 n [ &lambda; i ] &Element; &Psi; ( w i T &phi; ( x i ) )
Wherein
&Psi; ( z ) = e z - e - z e z + e - z [ &lambda; i ] &Element; = &lambda; i , if &lambda; i &GreaterEqual; &Element; 0 , if &lambda; i < &Element;
Ψ [] is a nonlinear function, and detection score is mapped to scope [-1,1], [] it is a loss function; Parameter lambda ithe degree of overlapping of tolerance part and object, scope is [0,1]; Weighted term [λ i] be used for punishing the part situation inconsistent with object detected.
4. feature extracting method in fine granulation classification identification, it is characterized in that, each object parts positioning result extracts constant convolution feature, namely convolution feature is extracted at multiple yardstick and multiple visual angle, these convolution features are carried out merging and are obtained final feature representation, and this expression is used for final classification.
5. feature extracting method according to claim 4, is characterized in that comprising the steps:
Step one: to given scalogram picture, extracts the 5th convolutional layer characteristic pattern f w × h × C, wherein w × h represents convolved image size, the port number of C representative feature figure; Input picture is 16 relative to the down-sampling ratio of the 5th convolutional layer, means that the 5th convolutional layer characteristic pattern is 16 relative to the step-length of input picture;
Step 2: carry out zero padding operation to the border of each channel of characteristic pattern, every side increases by two pixels, obtains the characteristic pattern f ' that zero padding is later thus w ' × h ' × C; The characteristic pattern f ' later to zero padding w ' × h ' × C, each passage uses slip window sampling select any subgraph f with step-length 1 w × h × C, therefore always having 5 × 5 relative to the upper left corner biased (Δ x, Δ y) is { subgraph of 0,1,2,3,4}; Then carry out pondization operation to each subgraph, obtaining target, to export size be the later subgraph of the pondization of n × n;
Step 3: use the pond beggar figure calculated for subsequent full UNICOM layer characteristic pattern obtained in step 2.
6. feature extracting method according to claim 5, it is characterized in that aforesaid operations carries out on 5 yardsticks of input picture and flip horizontal image thereof, finally altogether obtain 25 × 5 × 2 proper vectors, these proper vectors obtain the single features expression on each yardstick carry out pondization operation respectively on each yardstick after, the feature on these multiple yardsticks of last cascade is used for the final expression to image.
7. adopt a fine granulation classification recognition methods for method described in above-mentioned any one claim, it is characterized in that comprising two steps:
The first step: for test pattern, utilizes object detector and partial detector to detect target object and its deformation smaller portions, and described detecting device utilizes the measure of supervision that has of attitude cluster to learn to obtain, and take into account the attitudes vibration of object or part; Object detector and partial detector independently carry out, and return surveyed area that in each detecting device, score is high alternatively, and final testing result is by correcting object and partial detection obtains;
Second step, extracts convolution feature to object of detecting each in the first step or part at multiple yardstick and multiple visual angle, and these convolution features are carried out merging and obtained final feature representation, and this expression is used for final classification.
CN201510026025.3A 2015-01-19 2015-01-19 Fine granulation classification identifies and the part of object positions and feature extracting method Active CN104573744B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201510026025.3A CN104573744B (en) 2015-01-19 2015-01-19 Fine granulation classification identifies and the part of object positions and feature extracting method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201510026025.3A CN104573744B (en) 2015-01-19 2015-01-19 Fine granulation classification identifies and the part of object positions and feature extracting method

Publications (2)

Publication Number Publication Date
CN104573744A true CN104573744A (en) 2015-04-29
CN104573744B CN104573744B (en) 2018-07-20

Family

ID=53089763

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201510026025.3A Active CN104573744B (en) 2015-01-19 2015-01-19 Fine granulation classification identifies and the part of object positions and feature extracting method

Country Status (1)

Country Link
CN (1) CN104573744B (en)

Cited By (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2016183766A1 (en) * 2015-05-18 2016-11-24 Xiaogang Wang Method and apparatus for generating predictive models
CN106780479A (en) * 2016-12-31 2017-05-31 天津大学 A kind of high precision image fuzzy detection method based on deep learning
WO2017215622A1 (en) * 2016-06-15 2017-12-21 北京市商汤科技开发有限公司 Object segmentation method and apparatus and computing device
CN107766890A (en) * 2017-10-31 2018-03-06 天津大学 The improved method that identification segment learns in a kind of fine granularity identification
CN108509939A (en) * 2018-04-18 2018-09-07 北京大学深圳研究生院 A kind of birds recognition methods based on deep learning
CN108764247A (en) * 2018-04-13 2018-11-06 中国科学院自动化研究所 Deep learning object detecting method and device based on dense connection
CN109086792A (en) * 2018-06-26 2018-12-25 上海理工大学 Based on the fine granularity image classification method for detecting and identifying the network architecture
CN111091150A (en) * 2019-12-12 2020-05-01 哈尔滨市科佳通用机电股份有限公司 Railway wagon cross rod cover plate fracture detection method
CN111177305A (en) * 2019-12-30 2020-05-19 广州骏伯网络科技有限公司 Method, system, device and storage medium for time segment selection
CN112329603A (en) * 2020-11-03 2021-02-05 西南科技大学 Dam face crack defect positioning method based on image cascade
CN112613371A (en) * 2020-12-16 2021-04-06 上海大学 Hyperspectral image road extraction method based on dense connection convolution neural network
CN113255568A (en) * 2021-06-15 2021-08-13 湖南星汉数智科技有限公司 Bill image classification method and device, computer equipment and storage medium

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
A KRIZHEVSKY ET AL.: ""ImageNet classification with deep convolutional neural networks"", 《INTERNATIONAL CONFERENCE ON NEURAL INFORMATION PROCESSING SYSTEMS》 *
STEVE BRANSON ET AL.: ""Bird Species Categorization Using Pose Normalized Deep Convolutional Nets"", 《EPRINT ARXIV》 *
张颖: ""车型识别技术研究与应用"", 《中国优秀硕士学位论文全文数据库·信息科技辑》 *
王文: ""细粒度表情分类研究与设计"", 《中国优秀硕士学位论文全文数据库•信息科技辑》 *

Cited By (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107624189A (en) * 2015-05-18 2018-01-23 北京市商汤科技开发有限公司 Method and apparatus for generating forecast model
CN107624189B (en) * 2015-05-18 2020-11-20 北京市商汤科技开发有限公司 Method and apparatus for generating a predictive model
WO2016183766A1 (en) * 2015-05-18 2016-11-24 Xiaogang Wang Method and apparatus for generating predictive models
US10489913B2 (en) 2016-06-15 2019-11-26 Beijing Sensetime Technology Development Co., Ltd. Methods and apparatuses, and computing devices for segmenting object
WO2017215622A1 (en) * 2016-06-15 2017-12-21 北京市商汤科技开发有限公司 Object segmentation method and apparatus and computing device
CN106780479A (en) * 2016-12-31 2017-05-31 天津大学 A kind of high precision image fuzzy detection method based on deep learning
CN107766890B (en) * 2017-10-31 2021-09-14 天津大学 Improved method for discriminant graph block learning in fine-grained identification
CN107766890A (en) * 2017-10-31 2018-03-06 天津大学 The improved method that identification segment learns in a kind of fine granularity identification
CN108764247A (en) * 2018-04-13 2018-11-06 中国科学院自动化研究所 Deep learning object detecting method and device based on dense connection
CN108764247B (en) * 2018-04-13 2020-11-10 中国科学院自动化研究所 Dense connection-based deep learning object detection method and device
CN108509939A (en) * 2018-04-18 2018-09-07 北京大学深圳研究生院 A kind of birds recognition methods based on deep learning
CN109086792A (en) * 2018-06-26 2018-12-25 上海理工大学 Based on the fine granularity image classification method for detecting and identifying the network architecture
CN111091150A (en) * 2019-12-12 2020-05-01 哈尔滨市科佳通用机电股份有限公司 Railway wagon cross rod cover plate fracture detection method
CN111177305A (en) * 2019-12-30 2020-05-19 广州骏伯网络科技有限公司 Method, system, device and storage medium for time segment selection
CN111177305B (en) * 2019-12-30 2021-01-05 广州骏伯网络科技有限公司 Method, system, device and storage medium for time segment selection
CN112329603A (en) * 2020-11-03 2021-02-05 西南科技大学 Dam face crack defect positioning method based on image cascade
CN112613371A (en) * 2020-12-16 2021-04-06 上海大学 Hyperspectral image road extraction method based on dense connection convolution neural network
CN113255568A (en) * 2021-06-15 2021-08-13 湖南星汉数智科技有限公司 Bill image classification method and device, computer equipment and storage medium

Also Published As

Publication number Publication date
CN104573744B (en) 2018-07-20

Similar Documents

Publication Publication Date Title
CN104573744A (en) Fine granularity classification recognition method and object part location and feature extraction method thereof
Zeng et al. Wsod2: Learning bottom-up and top-down objectness distillation for weakly-supervised object detection
Grinias et al. MRF-based segmentation and unsupervised classification for building and road detection in peri-urban areas of high-resolution satellite images
Tao et al. Airport detection from large IKONOS images using clustered SIFT keypoints and region information
Sirmacek et al. A probabilistic framework to detect buildings in aerial and satellite images
Tao et al. Unsupervised detection of built-up areas from multiple high-resolution remote sensing images
WO2018005413A1 (en) Method and system for cell annotation with adaptive incremental learning
Kovács et al. Improved harris feature point set for orientation-sensitive urban-area detection in aerial images
CN102208038A (en) Image classification method based on visual dictionary
Zhang et al. Road recognition from remote sensing imagery using incremental learning
CN105069811A (en) Multi-temporal remote sensing image change detection method
CN103310195A (en) LLC-feature-based weak-supervision recognition method for vehicle high-resolution remote sensing images
Yuan et al. Learning to count buildings in diverse aerial scenes
CN113705570B (en) Deep learning-based few-sample target detection method
Lian et al. Weakly supervised road segmentation in high-resolution remote sensing images using point annotations
Löw et al. Per-field crop classification in irrigated agricultural regions in middle Asia using random forest and support vector machine ensemble
CN104751475A (en) Feature point optimization matching method for static image object recognition
Jing et al. Unsupervised oil tank detection by shape-guide saliency model
CN114998748A (en) Remote sensing image target fine identification method, electronic equipment and storage medium
CN116740758A (en) Bird image recognition method and system for preventing misjudgment
Chandra et al. Building detection methods from remotely sensed images
CN115588178A (en) Method for automatically extracting high-precision map elements
CN114927236A (en) Detection method and system for multiple target images
Bai et al. Semantic segmentation of sparse irregular point clouds for leaf/wood discrimination
Tsai et al. Detection of roadway sign condition changes using multi-scale sign image matching (M-SIM)

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant