CN107301376A - A kind of pedestrian detection method stimulated based on deep learning multilayer - Google Patents

A kind of pedestrian detection method stimulated based on deep learning multilayer Download PDF

Info

Publication number
CN107301376A
CN107301376A CN201710385952.3A CN201710385952A CN107301376A CN 107301376 A CN107301376 A CN 107301376A CN 201710385952 A CN201710385952 A CN 201710385952A CN 107301376 A CN107301376 A CN 107301376A
Authority
CN
China
Prior art keywords
pedestrian
msub
mrow
multilayer
frame
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201710385952.3A
Other languages
Chinese (zh)
Other versions
CN107301376B (en
Inventor
李玺
李健
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhejiang University ZJU
Original Assignee
Zhejiang University ZJU
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhejiang University ZJU filed Critical Zhejiang University ZJU
Priority to CN201710385952.3A priority Critical patent/CN107301376B/en
Publication of CN107301376A publication Critical patent/CN107301376A/en
Application granted granted Critical
Publication of CN107301376B publication Critical patent/CN107301376B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/25Determination of region of interest [ROI] or a volume of interest [VOI]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/20Movements or behaviour, e.g. gesture recognition

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Data Mining & Analysis (AREA)
  • Human Computer Interaction (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Evolutionary Biology (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Artificial Intelligence (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Psychiatry (AREA)
  • Social Psychology (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Image Analysis (AREA)
  • Traffic Control Systems (AREA)

Abstract

The invention discloses a kind of pedestrian detection method stimulated based on deep learning multilayer, for after given video monitoring and the target that need to be detected, marking the position that target occurs in video.Specifically include following steps:Pedestrian's data set for training objective detection model is obtained, and defines algorithm target;Position deviation and apparent semanteme to pedestrian target are modeled;Modeling result in step S2, which sets up pedestrian's multilayer, stimulates network model;The pedestrian position in monitoring image is detected using the detection model.Pedestrian detection of the present invention suitable for real video monitoring image, has preferably effect and robustness in face of all kinds of complex situations.

Description

A kind of pedestrian detection method stimulated based on deep learning multilayer
Technical field
The invention belongs to computer vision field, a kind of particularly pedestrian detection stimulated based on deep learning multilayer Method.
Background technology
Since 20 end of the centurys, with the development of computer vision, intelligent video treatment technology is widely paid close attention to and ground Study carefully.Pedestrian detection is the important and challenging task of one of which, and its target is accurately to detect in video monitoring image The position of pedestrian.The problem has very high application value in the field such as video monitoring and intelligent robot, is largely senior regard The basis of feel task.But same, the problem has larger challenge, and one is how to express target area information;How two be The extraction of candidate region and target classification unified Modeling are optimized, these challenges are proposed to the performance and robustness of respective algorithms Higher requirement.
General pedestrian detection algorithm is divided into three parts:1st, the candidate region that target is included in input picture is found out.2nd, base In candidate region manual extraction target signature.3rd, Detection task is realized using sorting algorithm to feature.This kind of method is primarily present Following problem:1) it is based on traditional visual signature, and these visual signatures can only express the visual information of lower level, but row People's Detection task needs model to possess the semantic understanding ability of higher level of abstraction;2) extraction of candidate region and the classification of feature do not have End-to-end study optimization;3) combination is not stimulated by multilayer based on the feature that deep learning is extracted, target signature is not abstract enough It is abundant.
The content of the invention
To solve the above problems, it is an object of the invention to provide a kind of pedestrian detection stimulated based on deep learning multilayer Method, for detecting the pedestrian position in given monitoring image.This method is based on deep neural network, the depth stimulated using multilayer Spend visual signature and characterize target area information, pedestrian detection is modeled using Faster R-CNN frameworks, can better adapt to true Complex situations in real video monitoring scene.
To achieve the above object, the technical scheme is that:
A kind of pedestrian detection method stimulated based on deep learning multilayer, is comprised the following steps:
S1, pedestrian's data set for training objective detection model is obtained, and define algorithm target;
S2, the position deviation to pedestrian target and apparent semanteme are modeled;
S3, the modeling result in step S2, which set up pedestrian's multilayer, stimulates network model;
S4, use the detection model detection monitoring image in pedestrian position.
Further, in step S1, described pedestrian's data set for training objective detection model, including pedestrian image Xtrain, the pedestrian position B manually marked;
Defining algorithm target is:Detect the pedestrian position P in a width monitoring image X.
Further, in step S2, position deviation and apparent semanteme to pedestrian target are modeled and specifically included:
S21, according to pedestrian's data set XtrainPosition deviation is modeled with pedestrian position P:
Wherein, x, y are the middle point coordinates of pedestrian's box label, and w, h is the width and length of pedestrian's box label, xa,yaIt is pedestrian The coordinate of candidate frame, wa,haIt is the width and length of pedestrian candidate frame;txFor pedestrian's frame x coordinate relative to callout box x coordinate Deviation correspondence mark width of frame ratio, tyFor deviation correspondence callout box of the y-coordinate relative to callout box y-coordinate of pedestrian's frame The ratio of length, twFor ratio of the width relative to mark width of frame of pedestrian's frame, thFor pedestrian's frame length relative to callout box The ratio of length;
S22, according to pedestrian's data set XtrainApparent semanteme is modeled with pedestrian position P:
S=<w,d>
Wherein s represents projection values of the feature d on projection vector w, and w is pedestrian's weight projection vector, and d is that pedestrian's feature is retouched State son,<.,.>It is that inner product operation is accorded with, and p (C=k | d) it is softmax functions, represent to belong to the probable value of kth class;sjIt is characterized d Projection value on j-th of projection vector w;C is the discrete random variable that value number is k;J is whole projection vector w jth Individual w index.
Further, in step S3, the modeling result in step S2, which sets up pedestrian's multilayer, stimulates network model specific Including:
S31, set up multilayer and stimulate convolutional neural networks, the input of neutral net is that a width monitoring image X and pedestrian mark Frame B, is output as the probable value p of correspondence pedestrian candidate frame, and the pedestrian position deviation O in X;The representation of neutral net is Map X → (p, O);
S32, sub- mapping X → p use soft maximum Softmax loss functions, are expressed as
Lcls(X,Y;θ)=- ∑jYjLogp (C | d) formula (3)
Wherein Y is binary set, if belonging to kth class, respective value is 1, and remaining is 0;Lcls(X,Y;θ) represent whole instruction Practice the softmax loss functions of data set;
S33, sub- mapping X → O use Euclid's loss function, are expressed as
Lloc(t, v)=∑ismooth(ti,vi)
Wherein tiIt is pedestrian position deviation label, viIt is pedestrian position deflection forecast value;I represents i-th of training sample;
S34, the loss function of whole multilayer stimulation neutral net are
L=Lcls+LlocFormula (5)
Whole neutral net is trained under loss function L using stochastic gradient descent and back-propagation algorithm.
Further, in step S4, the pedestrian position in detection monitoring image includes:Monitoring image X to be detected is defeated Enter the neutral net trained, the candidate frame probable value according to its output determines whether pedestrian, finally according to the position of prediction Deviation O corrections obtain pedestrian position P.
The present invention is applied to the pedestrian detection method of video monitoring scene, compared to existing pedestrian detection method, has Following beneficial effect:
First, pedestrian detection method of the invention sets up model based on depth convolutional neural networks.The present invention is by candidate regions The generation in domain and the classification of feature, which are unified in same network frame, learns optimization, improves the final effect of method.
Secondly, multilayer proposed by the present invention stimulate algorithm can more feature-rich abstracting power, while the Algorithm Learning The feature gone out is so that grader learns the classifying rules of more robust.
The present invention is applied to the pedestrian detection method of video monitoring scene, has well in intelligent video analysis system Application value, can effectively improve efficiency and the degree of accuracy of pedestrian detection.For example, in traffic video monitoring, row of the invention People's detection method can quickly and correctly detect all pedestrian positions, and pedestrian's search mission for after provides data, greatly Release human resources.
Brief description of the drawings
Fig. 1 is the schematic flow sheet of the pedestrian detection method applied to video monitoring scene of the present invention;
Fig. 2 stimulates the loss function schematic diagram of neutral net for the whole multilayer of the present invention.
Embodiment
In order to make the purpose , technical scheme and advantage of the present invention be clearer, it is right below in conjunction with drawings and Examples The present invention is further elaborated.It should be appreciated that the specific embodiments described herein are merely illustrative of the present invention, and It is not used in the restriction present invention.
On the contrary, the present invention covers any replacement done in the spirit and scope of the present invention being defined by the claims, repaiied Change, equivalent method and scheme.Further, in order that the public has a better understanding to the present invention, below to the thin of the present invention It is detailed to describe some specific detail sections in section description.Part without these details for a person skilled in the art Description can also understand the present invention completely.
With reference to Fig. 1, in the preferred embodiment, a kind of pedestrian detection side stimulated based on deep learning multilayer Method, comprises the following steps:
First, pedestrian's data set for training objective detection model, including pedestrian image X are obtainedtrain, manually mark Pedestrian position B;
Defining algorithm target is:Detect the pedestrian position P in a width monitoring image X.
Secondly, position deviation and apparent semanteme to pedestrian target are modeled and specifically included:
The first step, according to pedestrian's data set XtrainPosition deviation is modeled with pedestrian position P:
Wherein, x, y are the middle point coordinates of pedestrian's box label, and w, h is the width and length of pedestrian's box label, xa,yaIt is pedestrian The coordinate of candidate frame, wa,haIt is the width and length of pedestrian candidate frame;txFor pedestrian's frame x coordinate relative to callout box x coordinate Deviation correspondence mark width of frame ratio, tyFor deviation correspondence callout box of the y-coordinate relative to callout box y-coordinate of pedestrian's frame The ratio of length, twFor ratio of the width relative to mark width of frame of pedestrian's frame, thFor pedestrian's frame length relative to callout box The ratio of length;
Second step, according to pedestrian's data set XtrainApparent semanteme is modeled with pedestrian position P:
S=<w,d>
Wherein s represents projection values of the feature d on projection vector w, and w is pedestrian's weight projection vector, and d is that pedestrian's feature is retouched State son,<.,.>It is that inner product operation is accorded with, and p (C=k | d) it is softmax functions, represent to belong to the probable value of kth class;sjIt is characterized d Projection value on j-th of projection vector w;C is the discrete random variable that value number is k;J is whole projection vector w jth Individual w index.
Afterwards, according to the detection model of appeal modeling result pre-training billboard target.Specifically include:
The first step, setting up multilayer stimulates convolutional neural networks, and the input of neutral net is that a width monitoring image X and pedestrian mark Frame B is noted, the probable value p of correspondence pedestrian candidate frame, and the pedestrian position deviation O in X is output as;So as to the knot of neutral net Structure can be expressed as mapping X → (p, O);
Second step, sub- mapping X → p is expressed as using soft maximum (Softmax) loss function
Lcls(X,Y;θ)=- ∑jYjLogp (C | d) formula (3)
Wherein Y is binary set, if belonging to kth class, respective value is 1, and remaining is 0;Lcls(X,Y;θ) represent whole instruction Practice the softmax loss functions of data set;
3rd step, sub- mapping X → O uses Euclid's loss function, is expressed as
Lloc(t, v)=∑ismooth(ti,vi)
Wherein tiIt is pedestrian position deviation label, viIt is pedestrian position deflection forecast value, i represents i-th of training sample.
4th step, with reference to Fig. 2, the loss function of whole multilayer stimulation neutral net is
L=Lcls+LlocFormula (5)
Whole neutral net is trained under loss function L using stochastic gradient descent and back-propagation algorithm.
Finally, the pedestrian in monitoring image is detected using the detection model trained.Specifically include:Will pretreatment Good image, which is put into multilayer, stimulates calculating in detection framework.Multilayer stimulates detection framework to extract candidate frame with 3 RPN networks, The characteristic information that each RPN networks are utilized is different, and candidate frame size and yardstick are also different obtained from.First obtain each The candidate frame of RPN network extractions, 300 candidate regions are filtrated to get according to respective confidence level size.Then by 3 RPN networks In candidate region merge, obtain 900 candidate regions.Arrange, be filtrated to get most from big to small then according to classification confidence 300 whole object candidate areas.Whether the candidate frame class probability value according to its output, which is more than given threshold value, is filtered candidate frame, Overlapping detection block is eliminated using non-maxima suppression algorithm simultaneously, is corrected finally according to the position deviation O of prediction To pedestrian position P.
In above-described embodiment, the position deviation and apparent semanteme of pedestrian detection method of the invention first to pedestrian target are entered Row modeling.On this basis, former problem is converted into multi-task learning problem, and pedestrian detection is set up based on deep neural network Model.Finally, the pedestrian position in monitoring image is detected using the detection model trained.
By above technical scheme, the embodiment of the present invention has been developed a kind of many based on deep learning based on depth learning technology The pedestrian detection algorithm that layer is stimulated.The present invention can effectively model the position deviation and apparent semantic information of target simultaneously, so that Detect accurate pedestrian position.
The foregoing is merely illustrative of the preferred embodiments of the present invention, is not intended to limit the invention, all essences in the present invention Any modifications, equivalent substitutions and improvements made within refreshing and principle etc., should be included in the scope of the protection.

Claims (5)

1. a kind of pedestrian detection method stimulated based on deep learning multilayer, it is characterised in that comprise the following steps:
S1, pedestrian's data set for training objective detection model is obtained, and define algorithm target;
S2, the position deviation to pedestrian target and apparent semanteme are modeled;
S3, the modeling result in step S2, which set up pedestrian's multilayer, stimulates network model;
S4, use the detection model detection monitoring image in pedestrian position.
2. the pedestrian detection method as claimed in claim 1 stimulated based on deep learning multilayer, it is characterised in that step S1 In, described pedestrian's data set for training objective detection model, including pedestrian image Xtrain, the pedestrian position manually marked B;
Defining algorithm target is:Detect the pedestrian position P in a width monitoring image X.
3. the pedestrian detection method as claimed in claim 2 stimulated based on deep learning multilayer, it is characterised in that step S2 In, position deviation and apparent semanteme to pedestrian target are modeled and specifically included:
S21, according to pedestrian's data set XtrainPosition deviation is modeled with pedestrian position P:
<mfenced open = "" close = ""> <mtable> <mtr> <mtd> <mrow> <msub> <mi>t</mi> <mi>x</mi> </msub> <mo>=</mo> <mfrac> <mrow> <mi>x</mi> <mo>-</mo> <msub> <mi>x</mi> <mi>a</mi> </msub> </mrow> <msub> <mi>w</mi> <mi>a</mi> </msub> </mfrac> <mo>,</mo> </mrow> </mtd> <mtd> <mrow> <msub> <mi>t</mi> <mi>y</mi> </msub> <mo>=</mo> <mfrac> <mrow> <mi>y</mi> <mo>-</mo> <msub> <mi>y</mi> <mi>a</mi> </msub> </mrow> <msub> <mi>h</mi> <mi>a</mi> </msub> </mfrac> </mrow> </mtd> </mtr> </mtable> </mfenced>
Wherein, x, y are the middle point coordinates of pedestrian's box label, and w, h is the width and length of pedestrian's box label, xa,yaIt is pedestrian candidate The coordinate of frame, wa,haIt is the width and length of pedestrian candidate frame;txIt is inclined relative to callout box x coordinate for the x coordinate of pedestrian's frame The ratio of difference correspondence mark width of frame, tyFor deviation correspondence callout box length of the y-coordinate relative to callout box y-coordinate of pedestrian's frame Ratio, twFor ratio of the width relative to mark width of frame of pedestrian's frame, thFor pedestrian's frame length relative to callout box length Ratio;
S22, according to pedestrian's data set XtrainApparent semanteme is modeled with pedestrian position P:
S=<w,d>
Wherein s represents projection values of the feature d on projection vector w, and w is pedestrian's weight projection vector, and d is the description of pedestrian's feature Son,<.,.>It is that inner product operation is accorded with, and p (C=k | d) it is softmax functions, represent to belong to the probable value of kth class;sjD is characterized to exist Projection value on j-th of projection vector w;C is the discrete random variable that value number is k;J is j-th of whole projection vector w W index.
4. the pedestrian detection method as claimed in claim 3 stimulated based on deep learning multilayer, it is characterised in that step S3 In, the modeling result in step S2, which sets up pedestrian's multilayer, stimulates network model to specifically include:
S31, set up multilayer and stimulate convolutional neural networks, the input of neutral net is a width monitoring image X and pedestrian callout box B, It is output as the probable value p of correspondence pedestrian candidate frame, and the pedestrian position deviation O in X;The representation of neutral net is mapping X→(p,O);
S32, sub- mapping X → p use soft maximum Softmax loss functions, are expressed as
<mrow> <mi>p</mi> <mrow> <mo>(</mo> <mi>C</mi> <mo>=</mo> <mi>k</mi> <mo>|</mo> <mi>d</mi> <mo>)</mo> </mrow> <mo>=</mo> <msup> <mi>e</mi> <msub> <mi>s</mi> <mi>k</mi> </msub> </msup> <mo>/</mo> <msub> <mi>&amp;Sigma;</mi> <mi>j</mi> </msub> <msup> <mi>e</mi> <msub> <mi>s</mi> <mi>j</mi> </msub> </msup> </mrow>
Lcls(X,Y;θ)=- ∑jYjLogp (C | d) formula (3)
Wherein Y is binary set, if belonging to kth class, respective value is 1, and remaining is 0;Lcls(X,Y;θ) represent whole training number According to the softmax loss functions of collection;
S33, sub- mapping X → O use Euclid's loss function, are expressed as
Lloc(t, v)=∑ismooth(ti,vi)
Wherein tiIt is pedestrian position deviation label, viIt is pedestrian position deflection forecast value;I represents i-th of training sample;
S34, the loss function of whole multilayer stimulation neutral net are
L=Lcls+LlocFormula (5)
Whole neutral net is trained under loss function L using stochastic gradient descent and back-propagation algorithm.
5. the pedestrian detection method as claimed in claim 4 stimulated based on deep learning multilayer, it is characterised in that step S4 In, the pedestrian position in detection monitoring image includes:The neutral net that monitoring image X inputs to be detected are trained, foundation Its candidate frame probable value exported determines whether pedestrian, and the position deviation O corrections finally according to prediction obtain pedestrian position P.
CN201710385952.3A 2017-05-26 2017-05-26 Pedestrian detection method based on deep learning multi-layer stimulation Active CN107301376B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201710385952.3A CN107301376B (en) 2017-05-26 2017-05-26 Pedestrian detection method based on deep learning multi-layer stimulation

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710385952.3A CN107301376B (en) 2017-05-26 2017-05-26 Pedestrian detection method based on deep learning multi-layer stimulation

Publications (2)

Publication Number Publication Date
CN107301376A true CN107301376A (en) 2017-10-27
CN107301376B CN107301376B (en) 2021-04-13

Family

ID=60138099

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710385952.3A Active CN107301376B (en) 2017-05-26 2017-05-26 Pedestrian detection method based on deep learning multi-layer stimulation

Country Status (1)

Country Link
CN (1) CN107301376B (en)

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108446662A (en) * 2018-04-02 2018-08-24 电子科技大学 A kind of pedestrian detection method based on semantic segmentation information
CN108537117A (en) * 2018-03-06 2018-09-14 哈尔滨思派科技有限公司 A kind of occupant detection method and system based on deep learning
CN110163224A (en) * 2018-01-23 2019-08-23 天津大学 It is a kind of can on-line study auxiliary data mask method
CN110969657A (en) * 2018-09-29 2020-04-07 杭州海康威视数字技术股份有限公司 Gun and ball coordinate association method and device, electronic equipment and storage medium
CN111178267A (en) * 2019-12-30 2020-05-19 成都数之联科技有限公司 Video behavior identification method for monitoring illegal fishing
CN111476089A (en) * 2020-03-04 2020-07-31 上海交通大学 Pedestrian detection method, system and terminal based on multi-mode information fusion in image
CN111523478A (en) * 2020-04-24 2020-08-11 中山大学 Pedestrian image detection method acting on target detection system

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2016149881A1 (en) * 2015-03-20 2016-09-29 Intel Corporation Object recogntion based on boosting binary convolutional neural network features
CN106022237A (en) * 2016-05-13 2016-10-12 电子科技大学 Pedestrian detection method based on end-to-end convolutional neural network
CN106250812A (en) * 2016-07-15 2016-12-21 汤平 A kind of model recognizing method based on quick R CNN deep neural network
WO2017062610A1 (en) * 2015-10-06 2017-04-13 Evolv Technologies, Inc. Augmented machine decision making

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2016149881A1 (en) * 2015-03-20 2016-09-29 Intel Corporation Object recogntion based on boosting binary convolutional neural network features
WO2017062610A1 (en) * 2015-10-06 2017-04-13 Evolv Technologies, Inc. Augmented machine decision making
CN106022237A (en) * 2016-05-13 2016-10-12 电子科技大学 Pedestrian detection method based on end-to-end convolutional neural network
CN106250812A (en) * 2016-07-15 2016-12-21 汤平 A kind of model recognizing method based on quick R CNN deep neural network

Non-Patent Citations (5)

* Cited by examiner, † Cited by third party
Title
JIE LIU ET AL.: "Deep Convolutional Neural Networks for Pedestrian Detection with Skip Pooling", 《 2017 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS》 *
JIFENG DAI ET AL.: "R-FCN: Object Detection via Region-based Fully Convolutional Networks", 《ARXIV:1605.06409V2》 *
ROSS GIRSHICK: "Fast R-CNN", 《ARXIV:1504.08083V2》 *
ZHAOWEI CAI ET AL.: "A Unified Multi-scale Deep Convolutional Neural Network for Fast Object Detection", 《EUROPEAN CONFERENCE ON COMPUTER VISION》 *
任少卿: "基于特征共享的高效物体检测", 《中国博士学位论文全文数据库 信息科技辑》 *

Cited By (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110163224A (en) * 2018-01-23 2019-08-23 天津大学 It is a kind of can on-line study auxiliary data mask method
CN110163224B (en) * 2018-01-23 2023-06-20 天津大学 Auxiliary data labeling method capable of online learning
CN108537117A (en) * 2018-03-06 2018-09-14 哈尔滨思派科技有限公司 A kind of occupant detection method and system based on deep learning
CN108446662A (en) * 2018-04-02 2018-08-24 电子科技大学 A kind of pedestrian detection method based on semantic segmentation information
CN110969657A (en) * 2018-09-29 2020-04-07 杭州海康威视数字技术股份有限公司 Gun and ball coordinate association method and device, electronic equipment and storage medium
CN110969657B (en) * 2018-09-29 2023-11-03 杭州海康威视数字技术股份有限公司 Gun ball coordinate association method and device, electronic equipment and storage medium
CN111178267A (en) * 2019-12-30 2020-05-19 成都数之联科技有限公司 Video behavior identification method for monitoring illegal fishing
CN111476089A (en) * 2020-03-04 2020-07-31 上海交通大学 Pedestrian detection method, system and terminal based on multi-mode information fusion in image
CN111476089B (en) * 2020-03-04 2023-06-23 上海交通大学 Pedestrian detection method, system and terminal for multi-mode information fusion in image
CN111523478A (en) * 2020-04-24 2020-08-11 中山大学 Pedestrian image detection method acting on target detection system
CN111523478B (en) * 2020-04-24 2023-04-28 中山大学 Pedestrian image detection method acting on target detection system

Also Published As

Publication number Publication date
CN107301376B (en) 2021-04-13

Similar Documents

Publication Publication Date Title
CN107301376A (en) A kind of pedestrian detection method stimulated based on deep learning multilayer
CN109635694B (en) Pedestrian detection method, device and equipment and computer readable storage medium
CN106447658B (en) Conspicuousness object detection method based on global and local convolutional network
CN107862261A (en) Image people counting method based on multiple dimensioned convolutional neural networks
CN104182772B (en) A kind of gesture identification method based on deep learning
CN104143079B (en) The method and system of face character identification
CN107506722A (en) One kind is based on depth sparse convolution neutral net face emotion identification method
CN107123123A (en) Image segmentation quality evaluating method based on convolutional neural networks
CN106682697A (en) End-to-end object detection method based on convolutional neural network
CN107704877A (en) A kind of image privacy cognitive method based on deep learning
CN106778835A (en) The airport target by using remote sensing image recognition methods of fusion scene information and depth characteristic
CN110111340A (en) The Weakly supervised example dividing method cut based on multichannel
CN107341452A (en) Human bodys&#39; response method based on quaternary number space-time convolutional neural networks
CN105005774A (en) Face relative relation recognition method based on convolutional neural network and device thereof
CN107633511A (en) A kind of blower fan vision detection system based on own coding neutral net
CN106529448A (en) Method for performing multi-visual-angle face detection by means of integral channel features
CN107134144A (en) A kind of vehicle checking method for traffic monitoring
CN107945153A (en) A kind of road surface crack detection method based on deep learning
CN106295506A (en) A kind of age recognition methods based on integrated convolutional neural networks
CN107506786A (en) A kind of attributive classification recognition methods based on deep learning
CN105678278A (en) Scene recognition method based on single-hidden-layer neural network
CN108280397A (en) Human body image hair detection method based on depth convolutional neural networks
CN107766890A (en) The improved method that identification segment learns in a kind of fine granularity identification
CN104240256A (en) Image salient detecting method based on layering sparse modeling
CN108256462A (en) A kind of demographic method in market monitor video

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant