CN109101108A - Method and system based on three decision optimization intelligence cockpit human-computer interaction interfaces - Google Patents
Method and system based on three decision optimization intelligence cockpit human-computer interaction interfaces Download PDFInfo
- Publication number
- CN109101108A CN109101108A CN201810823980.3A CN201810823980A CN109101108A CN 109101108 A CN109101108 A CN 109101108A CN 201810823980 A CN201810823980 A CN 201810823980A CN 109101108 A CN109101108 A CN 109101108A
- Authority
- CN
- China
- Prior art keywords
- gesture
- granularity
- area image
- computer interaction
- gesture area
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/017—Gesture based interaction, e.g. based on a set of recognized hand gestures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/107—Static hand or arm
- G06V40/113—Recognition of static hand signs
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/20—Movements or behaviour, e.g. gesture recognition
- G06V40/28—Recognition of hand or arm movements, e.g. recognition of deaf sign language
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Psychiatry (AREA)
- Social Psychology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Image Analysis (AREA)
Abstract
The invention belongs to intelligent driving fields, are a kind of method and system based on three decision optimization intelligence cockpit human-computer interaction interfaces, and method includes the gesture video acquired in cockpit, are pre-processed, obtain images of gestures;To the gesture and background segment of images of gestures, gesture area image is obtained;More granularity expression are carried out for it, more grain size characteristics of gesture area image are extracted using convolutional neural networks;Each granularity gesture area image classification is calculated from coarseness to fine granularity to conditional probability of all categories, utilizes the completion gesture identification that three decisions are sequential;Gesture after identification is subjected to semantic conversion, human-computer interaction interface is operated according to the result after semantic conversion;Optimal granularity is obtained by the way of weighted sum, using the optimal granularity as most fine granularity.The present invention can not only more accurately identify gesture in cockpit, execute gesture command, additionally it is possible to reduce the interaction time of cockpit human-computer interaction interface, provide more comfortable interactive experience for user.
Description
Technical field
The invention belongs to intelligent driving fields, more particularly to one kind to be based on three decision optimization intelligence cockpit human-computer interactions circle
The method and system in face.
Background technique
With the development of artificial intelligence and depth learning technology, intelligent driving has attracted the concern of many people.Gesture
Identification is as one of man-machine interaction mode typical in intelligent driving, to the optimization design at the interface human-computer interaction in cockpit (HMI)
It is very important.Precisely quickly gesture identification can not only provide more comfortable interactive experience, can also improve driver's
Safety.
Current gesture identification method mainly has based on sensor device and two ways based on computer vision.Though the former
So there is preferable discrimination, its cost is larger, and interactive experience is not able to satisfy current demand, although the latter acquires gesture figure
As being easier, but it is existing include: based on template matching, based on Extraction of Geometrical Features, based on hidden Markov model and
Gesture identification method neural network based still has the problems such as model accuracy of identification is low or recognition speed is slow, can not be very
The current precisely real-time gesture identification demand of good adaptation.The main reason for model accuracy of identification is low is to fail to extract well
The feature of gesture, and recognition speed is slow is primarily due to caused by model too complex, existing method tends not to solve simultaneously
Certainly both of these problems.
Summary of the invention
Based on problem above, the ability of feature is extracted using deep neural network, in conjunction with more granular information expression ways and
Three decision thoughts, selecting suitable granularity that can solve simultaneously, gesture identification precision is low and the slow optimization problem of recognition speed.
The present invention provides a kind of methods based on three decision optimization intelligence cockpit human-computer interaction interfaces, including following step
It is rapid:
Gesture video in S1, acquisition cockpit, pre-processes it, obtains static images of gestures;
S2, in images of gestures gesture and background be split processing, obtain gesture area image;
S3, more granularity expression are carried out by coarseness to fine granularity for gesture area image;It is extracted using convolutional neural networks
More grain size characteristics of gesture area image;
S4, each granularity gesture area image classification is calculated from coarseness to fine granularity to conditional probability of all categories, benefit
With the completion gesture identification that three decisions are sequential;
S5, the gesture after identification is subjected to semantic conversion, to human-computer interaction interface according to the gesture identification after semantic conversion
As a result corresponding operation is carried out;
S6, optimal granularity is obtained by the way of weighted sum, using the optimal granularity as most fine granularity, repeat step
Rapid S3~S5.
Further, described that more granularity expression are carried out by coarseness to fine granularity for gesture area image, for same proficiency
Gesture area image, more granular information representations are specific as follows:
Wherein, AiIndicate gesture area image in varigrained information, A1Indicate gesture area image in the letter of coarseness
Breath, AnGesture area image is indicated in fine-grained information, i.e. fine granularity includes coarseness;I=1,2 ..., n, n indicate granularity
Number.
Further, more grain size characteristics that gesture area image is extracted using convolutional neural networks, including the use of volume
Different convolution kernels, extracts more granularity image features of images of gestures in product neural network.
Further, the step S4 includes carrying out three decisions from gesture area image zooming-out coarseness feature, if energy
It determines the class categories of gesture, does not then continue fine-grained feature extraction and further three decisions, otherwise extract more particulate
The feature of degree carries out three decisions, the class categories until determining gesture area image.
Further, the step S6 includes human-computer interaction circle that final each granularity is obtained by the way of weighted sum
Face optimum results, so that it is determined that gesture is to the optimal granularity of human-computer interaction interface effect of optimization;
Result=w × Acc+ (1-w) × Time
Time=T1+T2
Wherein, Result is the optimal granularity of gesture area image, and Acc indicates gesture identification precision, and Time indicates gesture
The time spent in identification process, w indicate weight, T1Indicate the time of more grain size characteristics of extraction gesture area image;T2It indicates
Identify the time of gesture.
The present invention provides a kind of systems based on three decision optimization intelligence cockpit human-computer interaction interfaces, including electrically connect
The camera that connects, cockpit gesture obtain module, images of gestures segmentation module, more grain size characteristic extraction modules, three decision gestures
Identification module, gesture semantic conversion module and optimal granularity obtain module;
The cockpit gesture obtains module by the gesture video in camera acquisition cockpit, and video frame is changed into a system
The static gesture image of column;
The images of gestures segmentation module is used to for the gesture of images of gestures and background to be split processing, obtains gesture area
Area image;
The more grain size characteristic extraction modules of gesture are for extracting gesture area image from coarseness to fine-grained more
Grain size characteristic;
Three decision gesture recognition modules are used for according to extracted more grain size characteristics in each granularity to gesture area
Area image carries out three decisions, thus by gesture classification;
The gesture semantic conversion module is used to sorted gesture carrying out semantic conversion;
The optimal granularity obtains module for obtaining optimal granularity, and sends more grain size characteristics for the optimal granularity
Extraction module.
Further, the more grain size characteristic extraction modules of the gesture include convolutional neural networks unit, utilize convolutional Neural
Different convolution kernels in network unit extract more granularity image features of gesture area image;More granular information representations
SpeciallyWherein, AiIndicate gesture area image in varigrained information, A1Indicate gesture area
Information of the image in coarseness, AnGesture area image is indicated in fine-grained information, i.e. fine granularity includes coarseness;I=1,
2 ..., n, n indicate granularity number.
Further, three decision gesture recognition modules include carrying out three to the coarseness feature of gesture area image
Branch decision does not continue fine-grained feature extraction and further three decisions, otherwise if can determine that the class categories of gesture
It extracts more fine-grained feature and carries out three decisions, the class categories until determining gesture area image.
Further, it includes that final each granularity is obtained by the way of weighted sum that the optimal granularity, which obtains module,
Human-computer interaction interface optimum results, so that it is determined that the optimal granularity of gesture area image;
Result=w × Acc+ (1-w) × Time
Time=T1+T2
Wherein, Result is the optimal granularity of gesture area image, and Acc indicates gesture identification precision, and Time indicates gesture
The time spent in identification process, w indicate weight, T1Indicate the time of more grain size characteristics of extraction gesture area image;T2It indicates
Identify the time of gesture.
Beneficial effects of the present invention:
The present invention constructs more granular information expression ways using the thought of " step by step calculation " in Granule Computing for images of gestures,
The images of gestures feature that more granularities are extracted using convolutional neural networks applies three in each granularity from coarseness to fine granularity
The method of decision carries out gesture identification, then carries out corresponding semantic conversion to the gesture of identification, and gesture identification result is answered
In HMI interface optimization in cockpit.
The present invention can more accurately be identified using the varigrained feature of acquired gesture in conjunction with three decision thoughts
Gesture faster executes corresponding semantic operation, this can not only reduce the interaction time at the interface cockpit HMI, additionally it is possible to be
User provides more comfortable interactive experience.
Detailed description of the invention
Fig. 1 is the flow diagram that the present invention uses;
Fig. 2 is that more grain size characteristics that the present invention uses extract schematic diagram;
Fig. 3 is three decision gesture identification flow charts that the present invention uses.
Fig. 4 is the HMI interface optimization design method that the present invention uses.
Specific embodiment
In order to make the objectives, technical solutions, and advantages of the present invention clearer, below in conjunction with attached drawing to of the invention real
The technical solution applied in example is clearly and completely described, it is clear that described embodiment is only that present invention a part is implemented
Example, instead of all the embodiments.
In order to better illustrate the specific implementation step of this method, it is described as follows in conjunction with Fig. 1 and example way:
Embodiment 1
The present invention the following steps are included:
Gesture video in S1, acquisition cockpit, pre-processes it, obtains static images of gestures;
S2, in images of gestures gesture and background be split processing, obtain gesture area image;
S3, more granularity expression are carried out by coarseness to fine granularity for gesture area image;It is extracted using convolutional neural networks
More grain size characteristics of gesture area image;
S4, each granularity gesture area image classification is calculated from coarseness to fine granularity to conditional probability of all categories, benefit
With the completion gesture identification that three decisions are sequential;
S5, by after identification gesture area image carry out semantic conversion, to human-computer interaction interface according to semantic conversion after
Gesture identification result is operated;
It is described that more granularity expression are carried out by coarseness to fine granularity for gesture area image, for same gesture area figure
Picture, more granular information representations are specific as follows:
Wherein, AiIndicate gesture area image in varigrained information, A1Indicate gesture area image in the letter of coarseness
Breath, AnGesture area image is indicated in fine-grained information, i.e. fine granularity includes coarseness;I=1,2 ..., n, n indicate granularity
Number.
More grain size characteristics that gesture area image is extracted using convolutional neural networks, including the use of convolutional neural networks
In different convolution kernels, more granularity image features of images of gestures are extracted, as shown in Fig. 2, mentioning using convolutional neural networks CNN
Take the feature (being followed successively by coarseness to fine-grained feature) of n granularity of gesture area image.
Further, the step S4 includes carrying out three decisions from the coarseness feature to gesture area image, if energy
It determines the class categories of gesture, does not then continue fine-grained feature extraction and further three decisions, otherwise extract more particulate
The feature of degree carries out three decisions, the class categories until determining gesture area image.
Wherein, the flow chart of three decisions is as shown in figure 3, by the data set of input, more for extracting gesture area image
Feature is spent, design conditions probability simultaneously carries out three decisions.
Softmax function design conditions probability is selected, gesture x is classified as to the conditional probability of classification j are as follows:
Wherein, l=1,2 ..., k, k indicate the classification sum of gesture area image;θ is parameter vector.
Three decision models utilize one group of decision-making value α, and gesture object is divided into positive domain (POS), Boundary Region by beta, gamma
(BND) and in negative domain (NEG), for positive domain and negative domain using receiving and refusing rule, gesture identification is directly obtained as a result, and side
Boundary domain uses Delayed Decision, when continuing when more fine granularity gets more information using three decisions.
The expression formula in positive domain, Boundary Region and negative domain is as follows:
POS(α,β)=x ∈ U | and p (X | [x]) >=α }
BND(α,β)=x ∈ U | and β < p (X | [x]) < α }
NEG(α,β)=x ∈ U | and p (X | [x])≤β }
Wherein, p (X | [x]) is the conditional probability of classification, and [x] is the equivalence class comprising x.
The threshold alpha of three decisionsi, βi, γiCalculation it is as follows:
The loss function of different action is respectively taken,Point
Receiving, delay and the loss function for refusing decision Biao Shi not be taken respectively when the i-th granularity gesture x belongs to classification X,It is illustrated respectively in and takes receiving, delay and refusal decision when the i-th granularity gesture x is not belonging to classification X respectively
Loss function, the loss function of each granularity are rule of thumb provided by expert respectively.
The setting principle of more three decision-making values of granularity is as follows, i.e., only in the case where must or be beneficial to decision
Carry out more fine-grained decision.This provides foundation, i.e. coarseness selection more for the setting of lower three decision-making values of different grain size
Big acceptance threshold and smaller refusal threshold value, i=1,2 ..., n-1 indicate by coarseness to fine-grained sequence, then difference grains
The threshold value of degree is described in detail below:
0≤βi<αi≤1,1≤i<n,
β1≤β2≤…≤βi<αi≤…≤α2≤α1
When i=n granularity, three decisions become two decisions, decision-making value calculation are as follows:
Three decisions are a kind of decision modes for meeting human thinking, and one is not held more than two traditional decisions
Promise selection, i.e., when information is not enough to receive or refuse using the third Delayed Decision.Two decision processes are quickly succinct, but
It is that three decisions are more suitable when obtaining insufficient information or the acquisition information certain cost of needs.Three decisions are selected to carry out hand
The purpose of gesture identification is exactly to be different because obtaining the time spent required for varigrained gesture feature, for real-time
For exigent HMI interface operation, consider that time cost is necessary.In three decision gesture identifications, close
The step of key be extract more grain size characteristics, calculate the threshold values of three decisions to and conditional probability.
Embodiment 2
On the basis of step S1~S5, the present embodiment is also added step S6, is obtained most by the way of weighted sum
Good granularity repeats step S3~S5 using the optimal granularity as most fine granularity.
HMI interface optimization design method by the way of weighted sum as shown in figure 4, obtain the man-machine of final each granularity
Interactive interface optimum results, so that it is determined that the optimal granularity of gesture area image is utilized using the optimal granularity as most fine granularity
Convolutional neural networks extract more grain size characteristics, and three decisions of sequential carry out to new gesture;
Result=w × Acc+ (1-w) × Time
Time=T1+T2
Wherein, Result is the optimal granularity of gesture area image, and Acc indicates gesture identification precision, and Time indicates gesture
The time spent in identification process, w indicate weight, T1Indicate the time of more grain size characteristics of extraction gesture area image;T2It indicates
Identify the time of gesture.
The present embodiment can save more time resources compared with embodiment 1, there is smaller computation complexity, for example,
In the case where embodiment 1, optimal granularity is not used, the feature of 5 granularities is extracted in setting, and the time of cost is 100, and if known
The effect of 3 granularities in road than 5 granularities recognition effect slightly almost, but the time is 40, then comprehensively considering, 5 to 3
Its 5 to 1 granularity of fullsized of granularity is more suitble to practical application.
Wherein, after the expression of human-computer interaction interface optimization design finds out optimal granularity, as subsequent gesture image procossing
Most fine granularity.Since the information content of the extracted feature of different grain size is different, different recognition results, and fine granularity feature can be obtained
Extraction is more compared to the time that coarseness is spent, and considers gesture accuracy of identification and recognition time weighting, can be special for gesture
Sign extracts one most suitable granularity of selection, to meet the HMI interface optimization design object in cockpit based on gesture.
The present invention provides a kind of systems based on three decision optimization intelligence cockpit human-computer interaction interfaces, including electrically connect
The camera that connects, cockpit gesture obtain module, images of gestures segmentation module, more grain size characteristic extraction modules, three decision gestures
Identification module, gesture semantic conversion module and optimal granularity obtain module;
The cockpit gesture obtains module by the gesture video in camera acquisition cockpit, and video frame is changed into a system
The static gesture image of column;
The images of gestures segmentation module is used to for the gesture of images of gestures and background to be split processing, obtains gesture area
Area image;
The more grain size characteristic extraction modules of gesture are for extracting gesture area image from coarseness to fine-grained more
Grain size characteristic;
Three decision gesture recognition modules are used for according to extracted more grain size characteristics in each granularity to gesture area
Area image carries out three decisions, thus by gesture classification;
The gesture semantic conversion module is used to sorted gesture carrying out semantic conversion;
The optimal granularity obtains module for obtaining optimal granularity, and sends more grain size characteristics for the optimal granularity
Extraction module.
Further, the more grain size characteristic extraction modules of the gesture include convolutional neural networks unit, utilize convolutional Neural
Different convolution kernels in network unit extract more granularity image features of gesture area image;More granular information representations
SpeciallyWherein, AiIndicate gesture area image in varigrained information, A1Indicate gesture area
Information of the image in coarseness, AnGesture area image is indicated in fine-grained information, i.e. fine granularity includes coarseness;I=1,
2 ..., n, n indicate granularity number.
Further, three decision gesture recognition modules include carrying out three to the coarseness feature of gesture area image
Branch decision does not continue fine-grained feature extraction and further three decisions, otherwise if can determine that the class categories of gesture
It extracts more fine-grained feature and carries out three decisions, the class categories until determining gesture area image.
Further, it includes that final each granularity is obtained by the way of weighted sum that the optimal granularity, which obtains module,
Human-computer interaction interface optimum results, so that it is determined that the optimal granularity of gesture area image;
Result=w × Acc+ (1-w) × Time
Time=T1+T2
Wherein, Result is the optimal granularity of gesture area image, and Acc indicates gesture identification precision, and Time indicates gesture
The time spent in identification process, w indicate weight, T1Indicate the time of more grain size characteristics of extraction gesture area image;T2It indicates
Identify the time of gesture.
Those of ordinary skill in the art will appreciate that all or part of the steps in the various methods of above-described embodiment is can
With what is be done through the relevant hardware of the program instructions, which be can store in computer-readable storage medium, storage
Medium may include: ROM, RAM, disk or CD etc..
Embodiment provided above has carried out further detailed description, institute to the object, technical solutions and advantages of the present invention
It should be understood that embodiment provided above is only the preferred embodiment of the present invention, be not intended to limit the invention, it is all
Any modification, equivalent substitution, improvement and etc. made for the present invention, should be included in the present invention within the spirit and principles in the present invention
Protection scope within.
Claims (9)
1. a kind of method based on three decision optimization intelligence cockpit human-computer interaction interfaces, which is characterized in that the described method includes:
Gesture video in S1, acquisition cockpit, pre-processes it, obtains a series of images of gestures of static state;
S2, in images of gestures gesture and background be split processing, obtain gesture area image;
S3, more granularity expression are carried out by coarseness to fine granularity for gesture area image;Gesture is extracted using convolutional neural networks
More grain size characteristics of area image;
S4, each granularity gesture area image classification is calculated from coarseness to fine granularity to conditional probability of all categories, utilize three
The sequential completion gesture identification of branch decision;
S5, the gesture after identification is subjected to semantic conversion, to human-computer interaction interface according to the gesture identification result after semantic conversion
Carry out corresponding operation;
S6, optimal granularity is obtained by the way of weighted sum, using the optimal granularity as most fine granularity, repeat step S3
~S5.
2. a kind of method based on three decision optimization intelligence cockpit human-computer interaction interfaces according to claim 1, special
Sign is, described is gesture area image by coarseness to the more granularity expression of fine granularity progress, for same gesture area image,
Its more granular information representation is specific as follows:
Wherein, AiIndicate gesture area image in varigrained information, A1Indicate gesture area image coarseness information,
AnGesture area image is indicated in fine-grained information, i.e. fine granularity includes coarseness;I=1,2 ..., n, n indicate granularity number.
3. a kind of method based on three decision optimization intelligence cockpit human-computer interaction interfaces according to claim 2, special
Sign is that more grain size characteristics that gesture area image is extracted using convolutional neural networks include utilizing convolutional neural networks
Middle different convolution kernel extracts more granularity image features of gesture area image.
4. a kind of method based on three decision optimization intelligence cockpit human-computer interaction interfaces according to claim 1, special
Sign is that the step S4 includes carrying out three decisions from the coarseness feature to gesture area image, if can determine that gesture
Class categories do not continue fine-grained feature extraction and further three decisions then, otherwise extract more fine-grained feature into
Three decisions of row, the class categories until determining gesture area image.
5. a kind of method based on three decision optimization intelligence cockpit human-computer interaction interfaces according to claim 1, special
Sign is that the step S6 includes the human-computer interaction interface optimization knot that final each granularity is obtained by the way of weighted sum
Fruit, so that it is determined that the optimal granularity of gesture area image;
Result=w × Acc+ (1-w) × Time
Time=T1+T2
Wherein, Result is the optimal granularity of gesture area image, and Acc indicates gesture identification precision, and Time indicates gesture identification
The time spent in the process, w indicate weight, T1Indicate the time of more grain size characteristics of extraction gesture area image;T2Indicate identification
The time of gesture.
6. a kind of system based on three decision optimization intelligence cockpit human-computer interaction interfaces, which is characterized in that the system comprises
The camera of electric connection, cockpit gesture obtain module, images of gestures segmentation module, more grain size characteristic extraction modules, three certainly
Plan gesture recognition module, gesture semantic conversion module and optimal granularity obtain module;
The cockpit gesture obtains module by the gesture video in camera acquisition cockpit, video frame is changed into a series of
Static gesture image;
The images of gestures segmentation module is used to for the gesture of images of gestures and background to be split processing, obtains gesture area figure
Picture;
The more grain size characteristic extraction modules of gesture are for extracting gesture area image from coarseness to fine-grained more granularities
Feature;
Three decision gesture recognition modules are used for according to extracted more grain size characteristics in each granularity to gesture administrative division map
As carrying out three decisions, thus by gesture classification;
The gesture semantic conversion module is used to sorted gesture carrying out semantic conversion;
The optimal granularity obtains module for obtaining optimal granularity, and sends more grain size characteristics for the optimal granularity and extract
Module.
7. a kind of system based on three decision optimization intelligence cockpit human-computer interaction interfaces according to claim 6, special
Sign is that the more grain size characteristic extraction modules of gesture include convolutional neural networks unit, using in convolutional neural networks unit
Different convolution kernels, extract more granularity image features of gesture area image;More granular information representations are speciallyWherein, AiIndicate gesture area image in varigrained information, A1Indicate that gesture area image exists
The information of coarseness, AnGesture area image is indicated in fine-grained information, i.e. fine granularity includes coarseness;I=1,2 ...,
N, n indicate granularity number.
8. a kind of system based on three decision optimization intelligence cockpit human-computer interaction interfaces according to claim 6, special
Sign is that three decision gesture recognition modules include coarseness feature three decisions of progress to gesture area image, if
It can determine that the class categories of gesture, then do not continue fine-grained feature extraction and further three decisions, otherwise extract thinner
The feature of granularity carries out three decisions, the class categories until determining gesture area image.
9. a kind of system based on three decision optimization intelligence cockpit human-computer interaction interfaces according to claim 6, special
Sign is that it includes that human-computer interaction circle of final each granularity is obtained by the way of weighted sum that the optimal granularity, which obtains module,
Face optimum results, so that it is determined that the optimal granularity of gesture area image;
Result=w × Acc+ (1-w) × Time
Time=T1+T2
Wherein, Result is the optimal granularity of gesture area image, and Acc indicates gesture identification precision, and Time indicates gesture identification
The time spent in the process, w indicate weight, T1Indicate the time of more grain size characteristics of extraction gesture area image;T2Indicate identification
The time of gesture.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810823980.3A CN109101108B (en) | 2018-07-25 | 2018-07-25 | Method and system for optimizing human-computer interaction interface of intelligent cabin based on three decisions |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810823980.3A CN109101108B (en) | 2018-07-25 | 2018-07-25 | Method and system for optimizing human-computer interaction interface of intelligent cabin based on three decisions |
Publications (2)
Publication Number | Publication Date |
---|---|
CN109101108A true CN109101108A (en) | 2018-12-28 |
CN109101108B CN109101108B (en) | 2021-06-18 |
Family
ID=64847467
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201810823980.3A Active CN109101108B (en) | 2018-07-25 | 2018-07-25 | Method and system for optimizing human-computer interaction interface of intelligent cabin based on three decisions |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN109101108B (en) |
Cited By (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109816022A (en) * | 2019-01-29 | 2019-05-28 | 重庆市地理信息中心 | A kind of image-recognizing method based on three decisions and CNN |
CN109820479A (en) * | 2019-01-08 | 2019-05-31 | 西北大学 | A kind of fluorescent molecular tomography feasible zone optimization method |
CN110298263A (en) * | 2019-06-10 | 2019-10-01 | 中南大学 | Real-time accurate and contactless gesture identification method and system based on RFID system |
CN110458233A (en) * | 2019-08-13 | 2019-11-15 | 腾讯云计算(北京)有限责任公司 | Combination grain object identification model training and recognition methods, device and storage medium |
CN111046732A (en) * | 2019-11-11 | 2020-04-21 | 华中师范大学 | Pedestrian re-identification method based on multi-granularity semantic analysis and storage medium |
CN111104339A (en) * | 2019-12-31 | 2020-05-05 | 上海艺赛旗软件股份有限公司 | Software interface element detection method and system based on multi-granularity learning, computer equipment and storage medium |
CN111814737A (en) * | 2020-07-27 | 2020-10-23 | 西北工业大学 | Target intention identification method based on three sequential decisions |
CN112580785A (en) * | 2020-12-18 | 2021-03-30 | 河北工业大学 | Neural network topological structure optimization method based on three-branch decision |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR101639351B1 (en) * | 2015-01-15 | 2016-07-13 | 주식회사 엔씨소프트 | Wearable input system and method for recognizing motion |
CN107578023A (en) * | 2017-09-13 | 2018-01-12 | 华中师范大学 | Man-machine interaction gesture identification method, apparatus and system |
CN107958255A (en) * | 2017-11-21 | 2018-04-24 | 中国科学院微电子研究所 | A kind of object detection method and device based on image |
-
2018
- 2018-07-25 CN CN201810823980.3A patent/CN109101108B/en active Active
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR101639351B1 (en) * | 2015-01-15 | 2016-07-13 | 주식회사 엔씨소프트 | Wearable input system and method for recognizing motion |
CN107578023A (en) * | 2017-09-13 | 2018-01-12 | 华中师范大学 | Man-machine interaction gesture identification method, apparatus and system |
CN107958255A (en) * | 2017-11-21 | 2018-04-24 | 中国科学院微电子研究所 | A kind of object detection method and device based on image |
Non-Patent Citations (1)
Title |
---|
时煜斌: ""基于三支决策的机器人触感手势识别研究"", 《中国优秀硕士学位论文全文数据库信息科技辑》 * |
Cited By (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109820479A (en) * | 2019-01-08 | 2019-05-31 | 西北大学 | A kind of fluorescent molecular tomography feasible zone optimization method |
CN109820479B (en) * | 2019-01-08 | 2021-08-27 | 西北大学 | Fluorescence molecular tomography feasible region optimization method |
CN109816022A (en) * | 2019-01-29 | 2019-05-28 | 重庆市地理信息中心 | A kind of image-recognizing method based on three decisions and CNN |
CN110298263A (en) * | 2019-06-10 | 2019-10-01 | 中南大学 | Real-time accurate and contactless gesture identification method and system based on RFID system |
CN110458233A (en) * | 2019-08-13 | 2019-11-15 | 腾讯云计算(北京)有限责任公司 | Combination grain object identification model training and recognition methods, device and storage medium |
CN110458233B (en) * | 2019-08-13 | 2024-02-13 | 腾讯云计算(北京)有限责任公司 | Mixed granularity object recognition model training and recognition method, device and storage medium |
CN111046732A (en) * | 2019-11-11 | 2020-04-21 | 华中师范大学 | Pedestrian re-identification method based on multi-granularity semantic analysis and storage medium |
CN111046732B (en) * | 2019-11-11 | 2023-11-28 | 华中师范大学 | Pedestrian re-recognition method based on multi-granularity semantic analysis and storage medium |
CN111104339B (en) * | 2019-12-31 | 2023-06-16 | 上海艺赛旗软件股份有限公司 | Software interface element detection method, system, computer equipment and storage medium based on multi-granularity learning |
CN111104339A (en) * | 2019-12-31 | 2020-05-05 | 上海艺赛旗软件股份有限公司 | Software interface element detection method and system based on multi-granularity learning, computer equipment and storage medium |
CN111814737A (en) * | 2020-07-27 | 2020-10-23 | 西北工业大学 | Target intention identification method based on three sequential decisions |
CN111814737B (en) * | 2020-07-27 | 2022-02-18 | 西北工业大学 | Target intention identification method based on three sequential decisions |
CN112580785B (en) * | 2020-12-18 | 2022-04-05 | 河北工业大学 | Neural network topological structure optimization method based on three-branch decision |
CN112580785A (en) * | 2020-12-18 | 2021-03-30 | 河北工业大学 | Neural network topological structure optimization method based on three-branch decision |
Also Published As
Publication number | Publication date |
---|---|
CN109101108B (en) | 2021-06-18 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN109101108A (en) | Method and system based on three decision optimization intelligence cockpit human-computer interaction interfaces | |
US11379711B2 (en) | Video action detection method based on convolutional neural network | |
CN108875674B (en) | Driver behavior identification method based on multi-column fusion convolutional neural network | |
CN107273845B (en) | Facial expression recognition method based on confidence region and multi-feature weighted fusion | |
Alani et al. | Hand gesture recognition using an adapted convolutional neural network with data augmentation | |
CN112507777A (en) | Optical remote sensing image ship detection and segmentation method based on deep learning | |
WO2018028255A1 (en) | Image saliency detection method based on adversarial network | |
CN113158862B (en) | Multitasking-based lightweight real-time face detection method | |
CN112150821A (en) | Lightweight vehicle detection model construction method, system and device | |
Yamashita et al. | Hand posture recognition based on bottom-up structured deep convolutional neural network with curriculum learning | |
CN110796018A (en) | Hand motion recognition method based on depth image and color image | |
Ali et al. | Facial emotion detection using neural network | |
CN113487610B (en) | Herpes image recognition method and device, computer equipment and storage medium | |
CN113743505A (en) | Improved SSD target detection method based on self-attention and feature fusion | |
CN112381030A (en) | Satellite optical remote sensing image target detection method based on feature fusion | |
CN113255602A (en) | Dynamic gesture recognition method based on multi-modal data | |
CN114723010B (en) | Automatic learning enhancement method and system for asynchronous event data | |
CN114492634B (en) | Fine granularity equipment picture classification and identification method and system | |
CN114241458B (en) | Driver behavior recognition method based on attitude estimation feature fusion | |
CN111126155A (en) | Pedestrian re-identification method for generating confrontation network based on semantic constraint | |
Li et al. | Gadet: A geometry-aware x-ray prohibited items detector | |
CN117710841A (en) | Small target detection method and device for aerial image of unmanned aerial vehicle | |
CN117079095A (en) | Deep learning-based high-altitude parabolic detection method, system, medium and equipment | |
Gupta et al. | Progression modelling for online and early gesture detection | |
WO2022126367A1 (en) | Sequence processing for a dataset with frame dropping |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |