CN110008834A - A kind of the steering wheel intervention detection and statistical method of view-based access control model - Google Patents

A kind of the steering wheel intervention detection and statistical method of view-based access control model Download PDF

Info

Publication number
CN110008834A
CN110008834A CN201910150734.0A CN201910150734A CN110008834A CN 110008834 A CN110008834 A CN 110008834A CN 201910150734 A CN201910150734 A CN 201910150734A CN 110008834 A CN110008834 A CN 110008834A
Authority
CN
China
Prior art keywords
intervention
steering wheel
frame
network structure
detection
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201910150734.0A
Other languages
Chinese (zh)
Other versions
CN110008834B (en
Inventor
程球
张雪莲
毛泉涌
文凌艳
周明政
赵云
胡芳芳
王军
谢兰青
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
CETHIK Group Ltd
Original Assignee
CETHIK Group Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by CETHIK Group Ltd filed Critical CETHIK Group Ltd
Priority to CN201910150734.0A priority Critical patent/CN110008834B/en
Publication of CN110008834A publication Critical patent/CN110008834A/en
Application granted granted Critical
Publication of CN110008834B publication Critical patent/CN110008834B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • G06V20/41Higher-level, semantic clustering, classification or understanding of video scenes, e.g. detection, labelling or Markovian modelling of sport events or news items
    • G06V20/42Higher-level, semantic clustering, classification or understanding of video scenes, e.g. detection, labelling or Markovian modelling of sport events or news items of sport video content
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/59Context or environment of the image inside of a vehicle, e.g. relating to seat occupancy, driver state or inner lighting conditions
    • G06V20/597Recognising the driver's state or behaviour, e.g. attention or drowsiness
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/20Movements or behaviour, e.g. gesture recognition
    • G06V40/28Recognition of hand or arm movements, e.g. recognition of deaf sign language

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Software Systems (AREA)
  • Computational Linguistics (AREA)
  • Computing Systems (AREA)
  • Biomedical Technology (AREA)
  • Molecular Biology (AREA)
  • Data Mining & Analysis (AREA)
  • General Engineering & Computer Science (AREA)
  • Biophysics (AREA)
  • Mathematical Physics (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Psychiatry (AREA)
  • Social Psychology (AREA)
  • Human Computer Interaction (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Image Analysis (AREA)

Abstract

Steering wheel intervention detection and statistical method the invention discloses a kind of view-based access control model, comprising: building detects comprising steering wheel and intervene the Design Integrative Network Structure of Attribute Recognition, and utilizes the sample image training Design Integrative Network Structure;Using the Design Integrative Network Structure, using single-frame images to be detected as the input of Design Integrative Network Structure, using steering wheel as detection target, the intervention attribute information exported according to Design Integrative Network Structure, judge whether steering wheel is intervened on current single-frame images, obtains intervening judging result;Using the intervention judging result in Statistics of Density method processing preset time, the beginning and ending time point and intervene duration that steering wheel is intervened are obtained.The hardware device that this method is relied on is few, is easy to implement, and can eliminate the erroneous detection in detection process and missing inspection, and the accuracy of testing result is high, and can be counted to testing result in order to more intuitively embody.

Description

A kind of the steering wheel intervention detection and statistical method of view-based access control model
Technical field
The invention belongs to steering wheels to intervene detection field, and in particular to a kind of steering wheel intervention detection of view-based access control model and system Meter method.
Background technique
Hand detection and gesture identification are important and grinding with good prospect of human-computer interaction and robot application Study carefully problem, has important application, in recent years, nothing in sports field, security field, traffic safety driving field People drive in match the detection method intervened of steering wheel more and more attention has been paid to.
In unmanned match, driver is to intervention number, each time duration of intervening of vehicle steering wheel Evaluate two important indicators of Unmanned Systems' superiority and inferiority degree.The direct embodiment of driver's intervention steering wheel is hand and steering wheel It touches, therefore steering wheel intervention detection can be converted into hand and the touching of steering wheel detects.Most of steering wheel touching inspections Examining system is all detected (such as pressure sensor) using sensor, such as Authorization Notice No. is the special of 105143015 B of CN Sharp document, then the detection method used is sentences in sensor arrangement to steering wheel according to the unlike signal that sensor generates Whether the hand of disconnected driver touches steering wheel, and it is auxiliary that this steering wheel touching detection technique is widely used in various Senior Officers In auxiliary system (ADAS).In fact, touching signal is outer except through special sensor acquisition, can also be received by video image Collection, with the logic of similar human eye to determine whether touching.Using the method for vision, we only need a camera, without in side The sensor that various complexity are arranged on disk greatlies simplify steering wheel and cab space design, and low in cost.
Summary of the invention
Steering wheel intervention detection and statistical method the purpose of the present invention is to provide a kind of view-based access control model, this method institute according to Bad hardware device is few, is easy to implement, and can eliminate the erroneous detection in detection process and missing inspection, and the accuracy of testing result is high, and Testing result can be counted in order to more intuitively embody.
To achieve the above object, the technical solution used in the present invention are as follows:
A kind of the steering wheel intervention detection and statistical method of view-based access control model, the steering wheel intervention detection of the view-based access control model With statistical method, comprising:
Building detects comprising steering wheel and intervenes the Design Integrative Network Structure of Attribute Recognition, and utilizes sample image training institute State Design Integrative Network Structure;
Using the Design Integrative Network Structure, using single-frame images to be detected as the input of Design Integrative Network Structure, with Steering wheel judges on current single-frame images as detection target according to the intervention attribute information that Design Integrative Network Structure exports Whether steering wheel is intervened, and obtains intervening judging result;
Using the intervention judging result in Statistics of Density method processing preset time, that steering wheel is intervened is obtained Only time point and intervention duration.
Preferably, the building detects comprising steering wheel and intervenes the Design Integrative Network Structure of Attribute Recognition, comprising:
Infrastructure network is constructed, which includes 9 convolutional layers and 5 maximum pond layers;
Several candidate windows are arranged in each candidate region on the last layer characteristic pattern of the infrastructure network, Include the coordinate information of the circumscribed rectangle of steering wheel in each candidate window, whether have object judgement information, target category probability letter Breath and the intervention attribute information form Design Integrative Network Structure.
Preferably, described utilize the sample image training Design Integrative Network Structure, comprising:
Driving video is obtained, 1 frame is extracted at interval of N frame from the driving video and is saved;
Be labeled to obtain sample image to the image saved, marked content include: the circumscribed rectangle of steering wheel coordinate, Whether target category and steering wheel are intervened;
Sample image random division is taken to obtain test set and training set;
Using the sample image training Design Integrative Network Structure in training set, until the sample image in test set is surveyed It tries the Design Integrative Network Structure and reaches preset condition.
Preferably, the infrastructure network since input layer I, successively passes through convolutional layer C1, maximum pond layer M1, convolutional layer C2, maximum pond layer M2, convolutional layer C3, maximum pond layer M3, convolutional layer C4, maximum pond layer M4, convolutional layer C5, maximum pond layer M5, convolutional layer C6, convolutional layer C7, convolutional layer C8, convolutional layer C9.
Preferably, the loss function for intervening attribute information calculates, comprising: li=(yp_i-pi)2;Wherein, liIt indicates Loss function;yp_iIndicate the output valve for intervening attribute information;piIndicate true value;
The gradient for intervening attribute information calculates, comprising:Wherein, δ indicates ladder Degree;yp_iIndicate the output valve for intervening attribute information, piIndicate true value.
Preferably, the intervention judging result using in Statistics of Density method processing preset time, comprising:
Setting fusion width is N frame, and a frame f is specified in the intervention judging resultt, count frame ftPreceding N frame, rear N frame And including frame ftTotalframes inside is 2N+1, and counting the intervention frame number in totalframes in intervention states is nt, then calculate Obtaining the intervention density within the scope of 2N+1 frame is dt, and
If dt>=0.5, then current 2N+1 frame range is defined as intervention states;If dt< 0.5, then current 2N+ 1 frame range is defined as non-intervention states.
Preferably, described obtain the beginning and ending time point and intervene duration that steering wheel is intervened, comprising: in the period dt-1To dt+TIt is interior:
If dt-1< 0.5, and dt>=0.5, then it represents that dtCorresponding time point is the start time point of this intervention states;
If dt+T-1>=0.5, and dt+T< 0.5, and dtTo dt+T-1Between the intervention density that judges be more than or equal to 0.5, then Indicate dt+TCorresponding time point is the termination time point of this intervention states;
And obtain a length of T when the intervention of this intervention states;Wherein, T indicates the period;dtIndicate current 2N+1 frame range Interior intervention density;dt-1It indicates relative to dtPrevious 2N+1 frame within the scope of intervention density;dt+TIt indicates relative to dtBy Intervention density within the scope of the later 2N+1 frame of T duration;dt+T-1It indicates relative to dt+TPrevious 2N+1 frame within the scope of intervention Density.
Preferably, the value of the N is 5~10.
The steering wheel intervention detection of view-based access control model provided by the invention and statistical method devise a target detection and category The property integrated end-to-end deep learning network of identification mission, the detection of hand and steering wheel is integrated into a network, is avoided The real-time detection of hand touching steering wheel is carried out dependent on multiple networks or tactful mode, simplification detection method improves detection efficiency; And the video sequence based on timing is intervened into Statistics of Density algorithm and is detected applied to steering wheel intervention, eliminate the mistake in detection process It examines and rising, stopping time point and intervene time duration for generation is intervened in missing inspection, estimation.
Detailed description of the invention
Fig. 1 is that the present invention is based on a kind of embodiment frame diagrams of the steering wheel intervention detection of vision and statistical method;
Fig. 2 is that the present invention is based on a kind of embodiment flow charts of the steering wheel intervention detection of vision and statistical method;
Fig. 3 is that steering wheel intervention of the present invention detects application scenarios schematic diagram;
Fig. 4 is the schematic diagram after image labeling of the present invention;
Fig. 5 is a kind of embodiment schematic diagram of Design Integrative Network Structure of the present invention;
Fig. 6 is that steering wheel intervenes detection effect figure in single-frame images of the present invention;
Fig. 7 is the intervention timing diagram that the present invention ideally intervenes detection output;
Fig. 8 is the intervention timing diagram for intervening detection output under virtual condition of the present invention;
Fig. 9 is that the present invention intervenes density calculating schematic diagram;
Figure 10 is the intervention density map that the intervention timing diagram in Fig. 8 of the present invention obtains after intervening density and calculating;
Figure 11 is another intervention density map of the invention.
Specific embodiment
Following will be combined with the drawings in the embodiments of the present invention, and technical solution in the embodiment of the present invention carries out clear, complete Site preparation description, it is clear that the described embodiment is only a part of the embodiment of the present invention, instead of all the embodiments.Based on this Embodiment in invention, every other reality obtained by those of ordinary skill in the art without making creative efforts Example is applied, shall fall within the protection scope of the present invention.
Unless otherwise defined, all technical and scientific terms used herein and belong to technical field of the invention The normally understood meaning of technical staff is identical.Term as used herein in the specification of the present invention is intended merely to description tool The purpose of the embodiment of body is not to be to limit the present invention.
As shown in Figure 1, present embodiments providing the steering wheel intervention detection and statistical method of a kind of view-based access control model, this method Mainly include two stages:
First stage is the steering wheel intervention detection based on deep learning, is input, using steering wheel as inspection with single-frame images Target is surveyed, the target frame comprising steering wheel is detected using algorithm of target detection, it is right on End features figure to be multiplexed the target frame The feature vector answered judges whether steering wheel is intervened on the frame image;
Second stage is the intervention Statistics of Density based on image sequence, is eliminated in the first stage accidentally using Statistics of Density algorithm Inspection and the interference of missing inspection bring, thus the beginning and ending time point intervened every time and intervention duration.
As shown in Fig. 2, the steering wheel intervention detection of the view-based access control model of the present embodiment and statistical method are broadly divided into following 4 Step.
S1, monitoring system is built in driver's cabin, collect application scenarios video image.
In obtaining driving video, application scenarios video image mainly consists of two parts: first is that unmanned ratio is better than True driver's monitor video in journey;Second is that monitoring view using the driver that more private cars imitate unmanned heat Frequently.Why sample data set includes two parts, is the monitoring view because true unmanned match monitor video resource is limited Angle, environment inside car are excessively single, and a large amount of period steering wheels lead to serious imbalanced training sets all in no intervention states.It is private The monitor video of family's vehicle simulation mainly makes up the deficiency of real games scene, increases sample diversity.
Firstly, adjusting multiple monitoring visual angles, more vehicles being used to increase angles and environmental diversity;Secondly, private car drives When the steering wheel most of the time be in by intervention states (steering wheel is not intervened when fraction of time such as dead ship condition), with this The problem of reducing imbalanced training sets;Again, driver specially imitates various vehicle driving postures and driving habit in simulation process, such as double Disk, singlehanded large face machine, palm scratching plate, finger hook disk etc. are held, increases the diversity of intervention states sample with this;Finally, mould Intend various extraneous natural environments, including when passing through tunnel dim illumination, fine day when heavy exposure etc..Application Scenarios-Example is such as Shown in Fig. 3.
Comprehensive driving video is obtained using scene abundant, in order to the test and training in later period.
S2, image is manually marked, every image labeling content includes: that (such as rectangle is left for the coordinate of the circumscribed rectangle of steering wheel Coordinate upper, at bottom righthand side angle), whether steering wheel by hand touching (such as 0 indicate to be intervened, 1 indicates not intervened)
Video is disassembled frame by frame, and extracts 1 frame every N frame and saves as sample, N=25 is taken in the present embodiment, and further It removes wherein because fuzzy, dark, transition such as exposes at the invalid picture caused by reasons, remaining picture is labeled.
It under normal circumstances, at most only include a steering wheel in every picture, the content of mark is that steering wheel target is circumscribed Whether coordinate (xmin, ymin, xmax, ymax), target category c and steering wheel at the upper left of rectangle, bottom righthand side angle are intervened Indicate i.Due to only having steering wheel one kind target, therefore c is always 0 (classification id is since 0);Intervene in flag bit, is indicated using 0 Intervened, 1 indicates not intervened.
When training, target frame coordinate is converted into center point coordinate and the high form (x, y, w, h) of width, and respectively divided by original For the wide height of figure to normalize, finally formed label form is [c, x, y, w, h, i], picture finally mark after the completion of schematic diagram such as Shown in Fig. 4, under normal circumstances, label is directly to be labeled on picture with colour, shows label in the present embodiment for clarity, therefore Label is separated with picture, makes label for labelling above picture.
It includes 45798 pictures that valid data collection is ultimately formed in the present embodiment, wherein selecting 5798 at random as survey Examination collection, remaining 40000 are used as training set.
S3, building detect comprising steering wheel and intervene the Design Integrative Network Structure of Attribute Recognition, and are instructed using sample image Practice the Design Integrative Network Structure, obtains the network knot that whether can be touched by hand compared with steering wheel in the every frame image of accurate judgement Structure.
The main purpose in this stage is to judge whether steering wheel is intervened in single-frame images.It is real from the point of view of technological layer Matter is two classification problems, has obtained very high accuracy rate currently based on the image classification problem of deep learning, but net of classifying Network requires to input most of pixel that subject goal in picture occupies entire image, because the principle of sorter network is by several The feature that entire image is extracted after secondary convolution, pondization operation is classified, if subject goal only occupies entire image seldom one Partial pixel, feature are easy to lose or flooded by background during the extraction process, it is difficult to obtain ideal classification accuracy.
Unlike usual classification task, the class object of the present embodiment is not jobbie but two kinds of object (hands And steering wheel) three-dimensional space position relationship, this spatial relation on 2d appearance form multiplicity, this is described The useful pixel accounting of kind positional relationship is few, and subtlety, which is easy to generate, to be obscured.
In view of the particularity of application scenarios, the present embodiment does not use the sorter network of standard, but combining target detection Thought detects intervention relationship as an attribute of each target, and with target frame position coordinates, objective degrees of confidence, mesh The information such as mark classification export simultaneously.Design in this way there are two benefit, one is object detection task be more concerned with it is important in image Clarification of objective extract, intervene relationship judgement also only in accordance with target frame instruction local feature rather than entire image, so as to To reduce background interference, Detection accuracy is improved;The second is using the relationship of intervention as an attribute of each target, rather than will Cascade sort network judges again after local feature is taken off, and sufficiently realizes feature multiplexing in this way, avoid the design of complex network with Training, while detection process being made to have more real-time.
Common-denominator target has 2 classes, respectively hand and steering wheel in the detection.Detect whether that the intuitive thinking intervened is to examine respectively The position coordinates of hand and steering wheel are measured, then judge whether there is intervention according to position coordinates, at this moment detection-phase can be deposited Essence is reduced to the conventional object detection task of hand and steering wheel, however infers that three-dimensional space position relationship is extremely by two-dimensional coordinate Difficult, therefore scheme is difficult to carry out.Another thinking is, directly detect hand whether intervene steering wheel or steering wheel whether by Hand intervention goes reasoning to intervene relationship without using artificial rule here, and is handed over to deep neural network and learns simultaneously reasoning automatically. The present embodiment is further simplified as a detection direction disk using second of thinking, is used as one for whether it is intervened Attribute exports together with object detection information.
S31, as shown in figure 5, building comprising steering wheel detect and intervene Attribute Recognition Design Integrative Network Structure, comprising:
Infrastructure network is constructed, which includes 9 convolutional layers and 5 maximum pond layers.Specifically, base Plinth network structure successively passes through convolutional layer C1, maximum pond layer M1, convolutional layer C2, maximum pond layer M2 since input layer I, Convolutional layer C3, maximum pond layer M3, convolutional layer C4, maximum pond layer M4, convolutional layer C5, maximum pond layer M5, convolutional layer C6, volume Lamination C7, convolutional layer C8, convolutional layer C9.
The network parameter of convolutional layer uses the form of [k_size, k_size, channels, stride] to indicate, wherein k_ Size is convolution kernel size, and channels is output feature port number, and stride is step-length;The network parameter of pond layer uses [k_size, k_size, stride] is indicated, wherein k_size is Chi Huahe size, and stride is step-length;The input parameter of each layer Using [resolution, resolution, channel], wherein resolution is the resolution ratio of image, and channel is logical Road number.Specifically, infrastructure network is as shown in table 1.
1 infrastructure network of table
Network layer type Network parameter Input parameter
Convolutional layer C1 [3,3,16,1] [416,416,3]
Maximum pond layer M1 [2,2,2] [416,416,3]
Convolutional layer C2 [3,3,32,1] [208,208,16]
Maximum pond layer M2 [2,2,2] [208,208,16]
Convolutional layer C3 [3,3,64,1] [104,104,32]
Maximum pond layer M3 [2,2,2] [104,104,32]
Convolutional layer C4 [3,3,128,1] [52,52,64]
Maximum pond layer M4 [2,2,2] [52,52,64]
Convolutional layer C5 [3,3,256,1] [26,26,128]
Maximum pond layer M5 [2,2,2] [26,26,128]
Convolutional layer C6 [3,3,512,1] [13,13,256]
Convolutional layer C7 [3,3,1024,1] [13,13,256]
Convolutional layer C8 [3,3,1024,1] [13,13,256]
Convolutional layer C9 [1,1,35,1] [13,13,256]
In the tiny-yolov2 algorithm of target detection of standard, each anchor is set on the last layer characteristic pattern of convolution Whether set 5 anchor box, each anchor box has target to sentence comprising 4 target frame coordinate x, y, w, h information, 1 Determining the other possibility probabilistic information P_c of information P_o, every type, (due to only having steering wheel one kind target here, therefore only 1 is believed Breath), therefore the feature port number of the target detection the last layer characteristic pattern of standard is 5* (4+1+1)=30.
However, the steering wheel intervention of the present embodiment has detected mostly one and has done unlike the algorithm of target detection of standard Pre- attribute information P_i.Each candidate region of the present embodiment on the last layer characteristic pattern of the infrastructure network (anchor) 5 candidate windows (anchor box) are set, and the coordinate comprising the circumscribed rectangle of steering wheel is believed in each candidate window Whether breath (x, y, w, h information) has object judgement information (P_o), target category probabilistic information (P_c) and the intervention attribute Information (P_i) forms Design Integrative Network Structure.
Further, will intervene after attribute information P_i is placed in target category probabilistic information P_c, for describing mesh in the frame Target attribute, therefore the parameter of the last layer convolutional layer C9 of the Design Integrative Network Structure of the present embodiment output is [13,13,35], I.e. the feature port number of the network the last layer characteristic pattern is 5* (4+1+1+1)=35.
In this way, just form steering wheel intervention detection and intervene Attribute Recognition Design Integrative Network Structure, using one Change network structure, using single-frame images to be detected as the input of Design Integrative Network Structure, using steering wheel as detection target, root According to the intervention attribute information that Design Integrative Network Structure exports, judge whether steering wheel is intervened on current single-frame images It obtains intervening judging result.
Since the attribute of the present embodiment only accounts for whether steering wheel is intervened this, the network structure and standard The only more intervention attribute positions of target detection network structure, in fact, attribute position can extend, it might even be possible to according to need It asks and extends to tens of, Shuo Baiwei.Common Attribute Recognition system generally uses target detection network and the cascade side of sorter network Formula detects the coordinate of targets of interest in image using algorithm of target detection first, is then taken off according to the coordinate and is wrapped in original image The regional area is finally input in a sorter network by the regional area containing the target, exports objective attribute target attribute.
And target detection and Attribute Recognition Design Integrative Network Structure that the present embodiment proposes, it will test task and Attribute Recognition Task is integrated into a network, has sufficiently been multiplexed detection network feature information, avoids time-consuming serious, the consumption memory of cascade network The problems such as serious.The network structure that the present embodiment proposes still keeps the characteristic of end-to-end training, reduces cascade network substep Trained trouble.The target detection and Attribute Recognition integrated network of the present embodiment design are detected as with being applied to steering wheel intervention Example is described in detail, but this thought can extend in the application of other intelligent visions, such as target following, target positioning etc. Deng.
S32, loss function design.
Standard target detection output (x, y, w, h, P_o, P_c) loss function design and gradient calculation with it is existing It is consistent in technology, such as described in yolov2 paper, no longer repeated herein.It is further illustrated below by specific formula That adds in the Design Integrative Network Structure of the present embodiment intervenes the loss function design and calculating of attribute information output P_i.
It can see from sample labeling form, the output of P_i is in fact carrying out simple two classification, using linear It returns to classify, its loss is described using mean square error, loss function calculates, comprising: loss function li=(yp_i-pi)2; Wherein, liIndicate loss function;yp_iIndicate the output valve for intervening attribute information, piTrue value is indicated, to intervene attribute information True value.
When gradient calculates, in order to mitigate intervention states and non-intervention states imbalanced training sets problem, less sample is produced Raw gradient value is multiplied by a coefficient, to increase it to the modified influence power of network weight.Since intervention states sample is less, such as The markup information of the pre- flag bit of dried fruit is yp_i=0, i.e., steering wheel is in by intervention states in the picture, we are by calculating Multiplied by 1.1, then gradient calculates to design includes: gradientWherein, δ indicates gradient;yp_iIt indicates Intervene the output valve of attribute information, piTrue value is indicated, for the true value for intervening attribute information.
The training and test of S33, Design Integrative Network Structure.
In training process, all pictures are input in network by reisize to 416x416 resolution ratio, total the number of iterations Max_batches=45000, each batch size are 64, and weight decay factor λ=0.0005, backpropagation uses momentum Method, factor of momentum v=0.9.Learning rate damped manner is piecewise constant decaying, and initial learning rate is learning_rate= 0.0001, it is promoted after iteration 100 times to 0.001, hereafter every iteration 10000 times, learning rate is multiplied by a decay factor 0.1.
Middle training set size is 40000 in the present embodiment, and test set size is 5798.After 40000 training, surveying It is 99% that steering wheel target detection accuracy of the mean (Average Precision, AP) is measured on examination collection, intervenes Detection accuracy 90%, detecting speed on GTX1070 is 6ms/f.Detection effect schematic diagram as shown in fig. 6, wherein normal expression do not intervene, I.e. steering wheel is not touched by hand, and automobile is in normal automatic Pilot state;Unnormal indicates to intervene, i.e., steering wheel is touched by hand It touches, automobile is in pilot steering state.
Calculate whether judgement is current is in intervention states according to the loss function of intervention attribute information: when intervention attribute information Output valve yp_i> 0.5 indicates not intervene, and output valve indicates that not intervene confidence level higher closer to 1;When intervention attribute The output valve y of informationp_i≤ 0.5 indicates to intervene, and output valve indicates that the confidence level intervened is higher closer to 0.
S4, the intervention judging result in preset time is handled using Statistics of Density method, obtains steering wheel and is intervened Beginning and ending time point and intervene duration.
It is successfully identifying after each frame image obtained in driving video, in order to further increase the accuracy of identification, Simultaneously in order to intuitively be presented in video whether have intervention states, the present embodiment further finds steering wheel in time series The start time point intervened terminates time point, and the duration intervened every time is calculated with this and counts the intervention of whole section of video Number.
Ideal recognition result that whether steering wheel is intervened as shown in fig. 7, intervening Detection accuracy is 100%, at this moment when Sequence analysis is just very simple, need to only judge whether intervention states occur to count on easily and intervene starting point and intervene duration.So And intervention detection in practice is often as shown in figure 8, being constantly present mistake since steering wheel detects and intervenes detection in present condition Poor (missing inspection, erroneous detection), even if accuracy rate reaches 99%, average 100 frames can also generate an erroneous detection, calculate to table in time series Now an erroneous detection will be generated for average every 4s.Obviously directly rely on intervene detection network output as a result, it is difficult to correct parsing Intervene beginning and ending time point and intervenes duration.
To solve the above-mentioned problems, sentenced in the present embodiment using the intervention in Statistics of Density method processing preset time Break as a result, as shown in figure 9, specifically including:
Setting fusion width is N frame, and a frame f is specified in intervening judging resultt, count frame ftPreceding N frame, rear N frame and Including frame ftTotalframes inside is 2N+1, and counting the intervention frame number in totalframes in intervention states is nt, then it is calculated Intervention density within the scope of 2N+1 frame is dt, and
If dt>=0.5, then current 2N+1 frame range is defined as intervention states;If dt< 0.5, then current 2N+ 1 frame range is defined as non-intervention states.
As shown in Figure 10, the intervention density map obtain after intervention density calculates for Fig. 8, it can be seen that detection Although there are a large amount of missing inspections and erroneous detections for the intervention timing of network output, after intervening density and calculating, remains to accurate statistics and do Pre- number.The intervention states presented in figure and the boundary of non-intervention states are obvious, can be obtained by the intervention density map more clear The case where whether clear accurate steering wheel is intervened.
At the judgement in the above described manner within the scope of each 2N+1 frame whether intervention states after, steering wheel can be obtained and done Pre- beginning and ending time point and intervention duration specifically includes as shown in figure 11: in period dt-1To dt+TIt is interior:
If dt-1< 0.5, and dt>=0.5, then it represents that dtCorresponding time point is the start time point of this intervention states;
If dt+T-1>=0.5, and dt+T< 0.5, and dtTo dt+T-1Between the intervention density that judges be more than or equal to 0.5, then Indicate dt+TCorresponding time point is the termination time point of this intervention states;
And obtain a length of T when the intervention of this intervention states;Wherein, T indicates period and T > 0;dtIndicate current 2N+1 Intervention density within the scope of frame;dt-1It indicates relative to dtPrevious 2N+1 frame within the scope of intervention density;dt+TIndicate opposite In dtThe intervention density within the scope of 2N+1 frame after T duration;dt+T-1It indicates relative to dt+TPrevious 2N+1 frame range Interior intervention density.
It is noted that taking N frame before and after sequence due to intervening Statistics of Density intervenes testing result, when the start-stop intervened Between put all can than really put delay N frame.Between 5~10, that is, between 0.2s to 0.4s, which exists usual N value It is acceptable in the statistics of the present embodiment.
Each technical characteristic of embodiment described above can be combined arbitrarily, for simplicity of description, not to above-mentioned reality It applies all possible combination of each technical characteristic in example to be all described, as long as however, the combination of these technical characteristics is not present Contradiction all should be considered as described in this specification.
The embodiments described above only express several embodiments of the present invention, and the description thereof is more specific and detailed, but simultaneously It cannot therefore be construed as limiting the scope of the patent.It should be pointed out that coming for those of ordinary skill in the art It says, without departing from the inventive concept of the premise, various modifications and improvements can be made, these belong to protection of the invention Range.Therefore, the scope of protection of the patent of the invention shall be subject to the appended claims.

Claims (8)

1. a kind of steering wheel intervention detection of view-based access control model and statistical method, which is characterized in that the direction of the view-based access control model Disk intervention detection and statistical method, comprising:
Building detects comprising steering wheel and intervenes the Design Integrative Network Structure of Attribute Recognition, and utilizes sample image training described one Body network structure;
Using the Design Integrative Network Structure, using single-frame images to be detected as the input of Design Integrative Network Structure, with direction Disk judges direction on current single-frame images according to the intervention attribute information that Design Integrative Network Structure exports as detection target Whether disk is intervened, and obtains intervening judging result;
Using the intervention judging result in Statistics of Density method processing preset time, when obtaining the start-stop that steering wheel is intervened Between put and intervene duration.
2. the steering wheel intervention detection of view-based access control model as described in claim 1 and statistical method, which is characterized in that the building The Design Integrative Network Structure of Attribute Recognition is detected and intervened comprising steering wheel, comprising:
Infrastructure network is constructed, which includes 9 convolutional layers and 5 maximum pond layers;
Several candidate windows are arranged in each candidate region on the last layer characteristic pattern of the infrastructure network, each In candidate window comprising the circumscribed rectangle of steering wheel coordinate information, whether have object judgement information, target category probabilistic information with And the intervention attribute information, form Design Integrative Network Structure.
3. the steering wheel intervention detection of view-based access control model as claimed in claim 2 and statistical method, which is characterized in that the utilization The sample image training Design Integrative Network Structure, comprising:
Driving video is obtained, 1 frame is extracted at interval of N frame from the driving video and is saved;
The image saved is labeled to obtain sample image, marked content includes: the coordinate of the circumscribed rectangle of steering wheel, target Whether classification and steering wheel are intervened;
Sample image random division is taken to obtain test set and training set;
Using the sample image training Design Integrative Network Structure in training set, until the sample image test in test set should Design Integrative Network Structure reaches preset condition.
4. the steering wheel intervention detection of view-based access control model as claimed in claim 2 and statistical method, which is characterized in that the base Plinth network structure successively passes through convolutional layer C1, maximum pond layer M1, convolutional layer C2, maximum pond layer M2 since input layer I, Convolutional layer C3, maximum pond layer M3, convolutional layer C4, maximum pond layer M4, convolutional layer C5, maximum pond layer M5, convolutional layer C6, volume Lamination C7, convolutional layer C8, convolutional layer C9.
5. the steering wheel intervention detection of view-based access control model as described in claim 1 and statistical method, which is characterized in that the intervention The loss function of attribute information calculates, comprising: li=(yp_i-pi)2;Wherein, liIndicate loss function;yp_iIt indicates to intervene attribute The output valve of information;piIndicate true value;
The gradient for intervening attribute information calculates, comprising:Wherein, δ indicates gradient;yp_i Indicate the output valve for intervening attribute information, piIndicate true value.
6. the steering wheel intervention detection of view-based access control model as described in claim 1 and statistical method, which is characterized in that the use Statistics of Density method handles the intervention judging result in preset time, comprising:
Setting fusion width is N frame, and a frame f is specified in the intervention judging resultt, count frame ftPreceding N frame, rear N frame and Including frame ftTotalframes inside is 2N+1, and counting the intervention frame number in totalframes in intervention states is nt, then it is calculated Intervention density within the scope of 2N+1 frame is dt, and
If dt>=0.5, then current 2N+1 frame range is defined as intervention states;If dt< 0.5, then current 2N+1 frame Range is defined as non-intervention states.
7. the steering wheel intervention detection of view-based access control model as claimed in claim 6 and statistical method, which is characterized in that described to obtain The beginning and ending time point and intervene duration that steering wheel is intervened, comprising: in period dt-1To dt+TIt is interior:
If dt-1< 0.5, and dt>=0.5, then it represents that dtCorresponding time point is the start time point of this intervention states;
If dt+T-1>=0.5, and dt+T< 0.5, and dtTo dt+T-1Between the intervention density that judges be more than or equal to 0.5, then it represents that dt+TCorresponding time point is the termination time point of this intervention states;
And obtain a length of T when the intervention of this intervention states;Wherein, T indicates the period;dtIt indicates within the scope of current 2N+1 frame Intervene density;dt-1It indicates relative to dtPrevious 2N+1 frame within the scope of intervention density;dt+TIt indicates relative to dtWhen by T Intervention density within the scope of long later 2N+1 frame;dt+T-1It indicates relative to dt+TPrevious 2N+1 frame within the scope of intervention it is close Degree.
8. the steering wheel intervention detection of view-based access control model as claimed in claim 7 and statistical method, which is characterized in that the N's Value is 5~10.
CN201910150734.0A 2019-02-28 2019-02-28 Steering wheel intervention detection and statistics method based on vision Active CN110008834B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910150734.0A CN110008834B (en) 2019-02-28 2019-02-28 Steering wheel intervention detection and statistics method based on vision

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910150734.0A CN110008834B (en) 2019-02-28 2019-02-28 Steering wheel intervention detection and statistics method based on vision

Publications (2)

Publication Number Publication Date
CN110008834A true CN110008834A (en) 2019-07-12
CN110008834B CN110008834B (en) 2021-04-06

Family

ID=67166379

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910150734.0A Active CN110008834B (en) 2019-02-28 2019-02-28 Steering wheel intervention detection and statistics method based on vision

Country Status (1)

Country Link
CN (1) CN110008834B (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111310841A (en) * 2020-02-24 2020-06-19 中南大学湘雅医院 Medical image classification method, apparatus, device, computer device and storage medium
CN114360321A (en) * 2021-11-09 2022-04-15 易显智能科技有限责任公司 Hand action sensing system, training system and training method for motor vehicle driver
CN118107605A (en) * 2024-04-30 2024-05-31 润芯微科技(江苏)有限公司 Vehicle control method and system based on steering wheel gesture interaction

Citations (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102263937A (en) * 2011-07-26 2011-11-30 华南理工大学 Driver's driving behavior monitoring device and monitoring method based on video detection
CN102289660A (en) * 2011-07-26 2011-12-21 华南理工大学 Method for detecting illegal driving behavior based on hand gesture tracking
CN102324016A (en) * 2011-05-27 2012-01-18 郝红卫 Statistical method for high-density crowd flow
CN102547139A (en) * 2010-12-30 2012-07-04 北京新岸线网络技术有限公司 Method for splitting news video program, and method and system for cataloging news videos
CN104078039A (en) * 2013-03-27 2014-10-01 广东工业大学 Voice recognition system of domestic service robot on basis of hidden Markov model
US20140292692A1 (en) * 2013-03-27 2014-10-02 Honda Motor Co., Ltd. Input apparatus, input method, and input program
CN104092988A (en) * 2014-07-10 2014-10-08 深圳市中控生物识别技术有限公司 Method, device and system for managing passenger flow in public place
CN104207791A (en) * 2014-08-26 2014-12-17 江南大学 Fatigue driving detection method
CN104228845A (en) * 2013-06-13 2014-12-24 福特全球技术公司 Hand/steering wheel contact detection using observer
CN105488957A (en) * 2015-12-15 2016-04-13 小米科技有限责任公司 Fatigue driving detection method and apparatus
CN105513354A (en) * 2015-12-22 2016-04-20 电子科技大学 Video-based urban road traffic jam detecting system
CN106372584A (en) * 2016-08-26 2017-02-01 浙江银江研究院有限公司 Video image mosaic detection method
CN106845344A (en) * 2016-12-15 2017-06-13 重庆凯泽科技股份有限公司 Demographics' method and device
CN107274678A (en) * 2017-08-14 2017-10-20 河北工业大学 A kind of night vehicle flowrate and model recognizing method based on Kinect
CN107479044A (en) * 2017-08-23 2017-12-15 西安电子工程研究所 Based on an adaptive track initiation method of mark density real-time statistics
CN107944341A (en) * 2017-10-27 2018-04-20 荆门程远电子科技有限公司 Driver based on traffic monitoring image does not fasten the safety belt automatic checkout system
CN108399388A (en) * 2018-02-28 2018-08-14 福州大学 A kind of middle-high density crowd quantity statistics method
CN108647617A (en) * 2018-05-02 2018-10-12 深圳市唯特视科技有限公司 A kind of positioning of driver's hand and grasping analysis method based on convolutional neural networks
CN109151501A (en) * 2018-10-09 2019-01-04 北京周同科技有限公司 A kind of video key frame extracting method, device, terminal device and storage medium

Patent Citations (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102547139A (en) * 2010-12-30 2012-07-04 北京新岸线网络技术有限公司 Method for splitting news video program, and method and system for cataloging news videos
CN102324016A (en) * 2011-05-27 2012-01-18 郝红卫 Statistical method for high-density crowd flow
CN102263937A (en) * 2011-07-26 2011-11-30 华南理工大学 Driver's driving behavior monitoring device and monitoring method based on video detection
CN102289660A (en) * 2011-07-26 2011-12-21 华南理工大学 Method for detecting illegal driving behavior based on hand gesture tracking
CN104078039A (en) * 2013-03-27 2014-10-01 广东工业大学 Voice recognition system of domestic service robot on basis of hidden Markov model
US20140292692A1 (en) * 2013-03-27 2014-10-02 Honda Motor Co., Ltd. Input apparatus, input method, and input program
CN104228845A (en) * 2013-06-13 2014-12-24 福特全球技术公司 Hand/steering wheel contact detection using observer
CN104092988A (en) * 2014-07-10 2014-10-08 深圳市中控生物识别技术有限公司 Method, device and system for managing passenger flow in public place
CN104207791A (en) * 2014-08-26 2014-12-17 江南大学 Fatigue driving detection method
CN105488957A (en) * 2015-12-15 2016-04-13 小米科技有限责任公司 Fatigue driving detection method and apparatus
CN105513354A (en) * 2015-12-22 2016-04-20 电子科技大学 Video-based urban road traffic jam detecting system
CN106372584A (en) * 2016-08-26 2017-02-01 浙江银江研究院有限公司 Video image mosaic detection method
CN106845344A (en) * 2016-12-15 2017-06-13 重庆凯泽科技股份有限公司 Demographics' method and device
CN107274678A (en) * 2017-08-14 2017-10-20 河北工业大学 A kind of night vehicle flowrate and model recognizing method based on Kinect
CN107479044A (en) * 2017-08-23 2017-12-15 西安电子工程研究所 Based on an adaptive track initiation method of mark density real-time statistics
CN107944341A (en) * 2017-10-27 2018-04-20 荆门程远电子科技有限公司 Driver based on traffic monitoring image does not fasten the safety belt automatic checkout system
CN108399388A (en) * 2018-02-28 2018-08-14 福州大学 A kind of middle-high density crowd quantity statistics method
CN108647617A (en) * 2018-05-02 2018-10-12 深圳市唯特视科技有限公司 A kind of positioning of driver's hand and grasping analysis method based on convolutional neural networks
CN109151501A (en) * 2018-10-09 2019-01-04 北京周同科技有限公司 A kind of video key frame extracting method, device, terminal device and storage medium

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111310841A (en) * 2020-02-24 2020-06-19 中南大学湘雅医院 Medical image classification method, apparatus, device, computer device and storage medium
CN111310841B (en) * 2020-02-24 2023-06-20 中南大学湘雅医院 Medical image classification method, medical image classification device, medical image classification apparatus, medical image classification computer device, and medical image classification storage medium
CN114360321A (en) * 2021-11-09 2022-04-15 易显智能科技有限责任公司 Hand action sensing system, training system and training method for motor vehicle driver
CN118107605A (en) * 2024-04-30 2024-05-31 润芯微科技(江苏)有限公司 Vehicle control method and system based on steering wheel gesture interaction

Also Published As

Publication number Publication date
CN110008834B (en) 2021-04-06

Similar Documents

Publication Publication Date Title
CN105069472B (en) A kind of vehicle checking method adaptive based on convolutional neural networks
CN105512720B (en) A kind of public transit vehicle passenger flow statistics method and system
CN109447033A (en) Vehicle front obstacle detection method based on YOLO
CN109508715A (en) A kind of License Plate and recognition methods based on deep learning
CN103886308B (en) A kind of pedestrian detection method of use converging channels feature and soft cascade grader
CN110097109A (en) A kind of road environment obstacle detection system and method based on deep learning
CN105335966B (en) Multiscale morphology image division method based on local homogeney index
CN110018524A (en) A kind of X-ray safety check contraband recognition methods of view-based access control model-attribute
CN109800629A (en) A kind of Remote Sensing Target detection method based on convolutional neural networks
CN104123549B (en) Eye positioning method for real-time monitoring of fatigue driving
CN109033950A (en) Vehicle based on multiple features fusion cascade deep model, which is disobeyed, stops detection method
CN106682586A (en) Method for real-time lane line detection based on vision under complex lighting conditions
CN103049751A (en) Improved weighting region matching high-altitude video pedestrian recognizing method
CN106203274A (en) Pedestrian&#39;s real-time detecting system and method in a kind of video monitoring
CN110008834A (en) A kind of the steering wheel intervention detection and statistical method of view-based access control model
CN106355602A (en) Multi-target locating and tracking video monitoring method
CN109766936A (en) Image change detection method based on information transmitting and attention mechanism
CN111553201A (en) Traffic light detection method based on YOLOv3 optimization algorithm
CN109190444A (en) A kind of implementation method of the lane in which the drivers should pay fees vehicle feature recognition system based on video
CN109543632A (en) A kind of deep layer network pedestrian detection method based on the guidance of shallow-layer Fusion Features
CN110009648A (en) Trackside image Method of Vehicle Segmentation based on depth Fusion Features convolutional neural networks
CN106128121A (en) Vehicle queue length fast algorithm of detecting based on Local Features Analysis
CN106127812A (en) A kind of passenger flow statistical method of non-gate area, passenger station based on video monitoring
CN109948690A (en) A kind of high-speed rail scene perception method based on deep learning and structural information
CN106373146A (en) Target tracking method based on fuzzy learning

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant