CN104834922B - Gesture identification method based on hybrid neural networks - Google Patents

Gesture identification method based on hybrid neural networks Download PDF

Info

Publication number
CN104834922B
CN104834922B CN201510280013.3A CN201510280013A CN104834922B CN 104834922 B CN104834922 B CN 104834922B CN 201510280013 A CN201510280013 A CN 201510280013A CN 104834922 B CN104834922 B CN 104834922B
Authority
CN
China
Prior art keywords
gesture
mrow
point
pixel
neural network
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related
Application number
CN201510280013.3A
Other languages
Chinese (zh)
Other versions
CN104834922A (en
Inventor
纪禄平
尹力
周龙
王强
卢鑫
黄青君
杨洁
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
University of Electronic Science and Technology of China
Original Assignee
University of Electronic Science and Technology of China
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by University of Electronic Science and Technology of China filed Critical University of Electronic Science and Technology of China
Priority to CN201510280013.3A priority Critical patent/CN104834922B/en
Publication of CN104834922A publication Critical patent/CN104834922A/en
Application granted granted Critical
Publication of CN104834922B publication Critical patent/CN104834922B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/20Movements or behaviour, e.g. gesture recognition
    • G06V40/28Recognition of hand or arm movements, e.g. recognition of deaf sign language
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/107Static hand or arm

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Multimedia (AREA)
  • Human Computer Interaction (AREA)
  • Biophysics (AREA)
  • Mathematical Physics (AREA)
  • Data Mining & Analysis (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Computational Linguistics (AREA)
  • Evolutionary Computation (AREA)
  • Software Systems (AREA)
  • Biomedical Technology (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Psychiatry (AREA)
  • Social Psychology (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a kind of gesture identification method based on hybrid neural networks, for images of gestures to be identified and images of gestures training sample, detect to obtain noise spot using Pulse Coupled Neural Network first, compound Denoising Algorithm is recycled to handle noise spot, then using the marginal point in cell neural network extraction images of gestures, marginal point according to extracting obtains connected region, finger tip detection is carried out to each connected region using curvature and obtains finger tip point undetermined, face part is excluded to disturb to obtain gesture area, then split according to gesture shape feature, the Fourier descriptor for retaining phase information is obtained according to the profile point of gesture area after segmentation, several Fourier descriptors are as gesture feature before selection;BP neural network is trained according to the gesture feature of images of gestures training sample, the gesture feature input BP neural network of images of gestures to be identified is identified.The present invention improves the accuracy rate to gesture identification by the utilization of a variety of neutral nets.

Description

Gesture identification method based on hybrid neural networks
Technical field
The invention belongs to technical field of hand gesture recognition, more specifically, is related to a kind of hand based on hybrid neural networks Gesture recognition methods.
Background technology
With advancing by leaps and bounds for computer technology, human-computer interaction technology is increasingly popularized in the life of people.Man-machine friendship Mutually (Human-Computer Interaction, HCI) technology refer between user and computer using certain mode of operation and Interaction between a kind of people performed and computer.Its development substantially experienced pure handwork stage, verbal order Control stage, user interface stage etc., however as the continuous development of the technologies such as artificial intelligence in recent years, gradually cause to man-machine The attention of interaction technique development.
Now with continuous expansion of the computer in terms of application field, existing man-machine interaction mode can not meet Higher level requirement of the people to daily demand, it is badly in need of a kind of more succinct, friendly new man-machine mode interacted. Because the final purpose of man-machine interaction is exchanged naturally between people and machine to realize, and in daily life person to person it Between be largely to pass on information by body language or facial expression, only sub-fraction is come complete by natural language Into, this has bigger advantage in terms of indicating that body language expression human emotion or intention.Due to working as in body language In, hand plays particularly important role, and therefore, the interactive mode based on gesture behavior is gesture Activity recognition system, namely Gesture recognition system gets more and more people's extensive concerning.
Generally, gesture recognition system is mainly made up of following components:Gesture pretreatment, Hand Gesture Segmentation, hand Gesture modeling, gesture feature extraction, gesture identification.Denoising operation for gesture pretreatment operation, mainly images of gestures, at present Common Denoising Algorithm includes:Mean filter, medium filtering, low pass spatial filtering, frequency domain low-pass ripple and pulse-couple god Through network etc., but for a variety of noises in the presence of, the noise removal capability of current algorithm, which can not all reach, good to go Make an uproar effect, therefore it is most important for the identification process in later stage to design a good Denoising Algorithm.Operated for Hand Gesture Segmentation, Currently used Hand Gesture Segmentation method has the dividing method based on Skin Color Information, the dividing method based on movable information and is based on The dividing method of marginal information.Because the dividing method based on Skin Color Information is easily disturbed by background information, based on edge The dividing method of information can not reach good segmentation effect again, therefore how to design a good effectively partitioning algorithm and be also It is vital.Extract and operate for gesture feature, most widely used at present is the feature extraction side based on Fourier descriptor Method, but because the rotational invariance of this method make it that this method is little for the changing features of the gesture after gesture rotation, Therefore it is also vital for how designing a Fourier descriptor without rotational invariance.Grasped for gesture identification Make, method common at present has template matching technique, SVMs, neural net method, hidden Markov model etc., therefore It is how equally most important for gesture recognition system from a good gesture identification method.
Neural net method refers to simulate people's brain neuron using some simple processing units, and these are simple Processing unit connects into network to realize a science to brain simulation in some way.Neural net method often have with Lower advantage:Parallel computation, distributed storage, robustness, nonlinear processing and good adaptivity and fault-tolerance ability. Therefore, neural net method can be applied under multiple scenes.Such as:Gesture identification, image segmentation, noise processed etc..
At present, neural net method has obtained increasing application in gesture Activity recognition field.It is however, neural Application of the network method in gesture Activity recognition field is also only limitted to this stage of gesture identification, for gesture Activity recognition The application in other stages is seldom.
The content of the invention
It is an object of the invention to overcome the deficiencies of the prior art and provide a kind of gesture identification based on hybrid neural networks Method, the denoising effect of images of gestures is improved using Pulse Coupled Neural Network, Hand Gesture Segmentation is carried out using cell neural network, Using the Fourier descriptor with rotational variability as gesture feature, gesture identification is carried out using BP neural network, so as to Improve the accuracy rate of gesture identification
For achieving the above object, the gesture identification method of the invention based on hybrid neural networks, comprises the following steps:
S1:The feature of images of gestures and images of gestures training sample to be identified is extracted, specific steps include:
S1.1:It is recommended that the Pulse-coupled Neural Network Model of gesture gray level image, by current each pixel of gesture gray level image The gray value of point is special using the granting of Pulse Coupled Neural Network as the input that neuron is corresponded in Pulse Coupled Neural Network Property detects to the pixel of images of gestures, if the output state of pixel is fired state, by testing result matrix In element corresponding to the pixel be set to 1, be otherwise set to 0;Each element of testing result matrix is traveled through, if element value is 1, Then the center using the element as noise reduction process window, the size of noise reduction process window are set according to actual conditions, are counted at noise reduction The value of the other elements in window in addition to central point element is managed, if value is more than predetermined threshold value for 0 number of elements, illustrates this Central point is noise spot, and other situations are not then noise spots;
Calculate two kinds of noise estimation value H (i, j) and the V (i, j) of noise spot as follows respectively:
H (i, j)=| a (i, j)-b (i, j) |
Wherein, a (i, j) is the gray value at pixel (i, j) place in image, and b (i, j) is that the pixel carries out medium filtering Intermediate value output gray level value afterwards;
Wherein, m1(i, j) and m2It is closest with a (i, j) gray value in neighborhood where (i, j) represents pixel (i, j) respectively Two points gray value;
If H (i, j) >=T1, and V (i, j) >=T2, then the noise spot is handled using medium filtering, otherwise adopted The noise spot is handled with mean filter;
S1.2:Histogram equalization is carried out to the gesture gray level image after step S1.1 denoisings;
S1.3:The Cellular Neural Networks of gesture gray level image are established, by each picture of gesture gray level image after equalization The gray value of vegetarian refreshments (i, j) is as the input u that cell is corresponded in Cellular Neural Networksij, according to the public affairs of state migration procedure Formula is iterated, and until the convergence of whole cell neural network, obtains the output y of each cellij(t);Traversal cell neutral net In cell member corresponding to each pixel output valve, when the output valve of some pixel is in the range of [0,1], if it is corresponding The pixel value of other pixels and more than predetermined threshold value in neighborhood, then this pixel is not edge pixel, is otherwise edge pixel point; When output valve [- 1,0) in the range of, be not edge pixel point;
S1.4:The edge pixel point obtained according to step S1.3 obtains connected region, and extraction obtains the profile of connected region, Finger tip detection is carried out respectively to each connected region, Fingertip Detection is:
Each wire-frame image vegetarian refreshments in connected region is traveled through, using the pixel as datum mark, coordinate is designated as p (px,py, 0) a distance constant L, is preset, the l-th point p before p points is taken along contour direction1(p1x,p1y, 0), take the l-th behind point p Point p2(p2x,p2y, 0), calculate vectorWith vectorBetween angle cosine value cos α, if cos α are more than preset curvature Threshold value T, then it is finger tip point undetermined to judge the point, otherwise not as finger tip point undetermined;
Fingertip location vector product is determined according to traversal directionSymbol, if according to gesture area overall profile Traversal clockwise when, vectorial product code should be negative, otherwise just, to calculate finger tip point undetermined vectorWith vectorBetween Vector productIf the symbol of the vector product is identical with symbol corresponding to fingertip location, finger tip undetermined is left Point, does not otherwise retain;
Judge in all finger tip points undetermined for being detected in the connected region, the maximum finger tip point undetermined of y-coordinate and y-coordinate Whether the y-coordinate difference of minimum finger tip point undetermined exceedes the half of face height, if it is, the connected region is not gesture area Domain, otherwise as gesture area undetermined;Whether finger tip point quantity undetermined surpasses in each gesture area undetermined further judged Default amount threshold is crossed, if it is, the connected region is gesture area, is not otherwise;
The principal direction of gesture area is asked for, is entered according to principal direction according to gesture length with width ratio for 2 pairs of gesture areas Row segmentation, the gesture area after being split;
S1.5:The gesture area that will be obtained after step S1.4 segmentations, by the profile point coordinates of gesture area with plural shape Formula represents, all profile point coordinates is formed into discrete series, note profile point quantity is n, and Fourier's change is carried out to the discrete series Change, obtain n Fourier coefficient z (k), k=0,1 ..., n-1, calculate Fourier descriptor
Wherein k '=1,2 ..., n-1,Represent the angle of gesture area principal direction and x-axis.
Q constitutive characteristic vector before being selected in Fourier descriptor;
S2:BP neural network, its corresponding hand are inputted using the characteristic vector of training sample images of gestures as training sample Output of the gesture image category as BP neural network, is trained to BP neural network;
S3:In the BP neural network that the characteristic vector input step S2 of images of gestures to be identified is trained, output identification Obtained images of gestures classification.
Gesture identification method of the invention based on hybrid neural networks, is trained for images of gestures to be identified and images of gestures Sample, the differentiation for being carried out noise spot and marginal point using Pulse Coupled Neural Network first are detected, and recycle compound Denoising Algorithm Noise spot is handled, the marginal point in images of gestures is then extracted using cell neural network, according to the edge extracted Point obtains connected region, and carrying out finger tip detection to each connected region using curvature obtains finger tip point undetermined, then excludes face The interference divided, obtains gesture area, is then split according to hand features of shape, the gesture area after being split;According to The profile point of gesture area obtains the Fourier descriptor for retaining phase information, and several Fourier descriptors are as hand before selecting Gesture feature;BP neural network is trained according to the gesture feature of images of gestures training sample, the gesture of images of gestures to be identified is special Sign input BP neural network is identified.
The invention has the advantages that:
(1) differentiation of noise spot and marginal point is carried out using Pulse Coupled Neural Network, with reference to compound Denoising Algorithm opponent Gesture image carries out denoising, can improve denoising effect;
(2) Hand Gesture Segmentation combines the coarse segmentation of cell neural network and the subdivision based on gesture shape feature is cut, can be with Improve the degree of accuracy of Hand Gesture Segmentation;
(3) gesture feature uses Fourier descriptor, remains phase information, can improve discrimination.
Brief description of the drawings
Fig. 1 is the flow chart of the gesture identification method of the invention based on hybrid neural networks
Fig. 2 is the flow chart of images of gestures feature extraction in the present invention
Fig. 3 is to combine gesture shape characteristic to carry out the flow chart that gesture subdivision is cut
Fig. 4 is the schematic diagram of finger tip detection of the present invention;
Fig. 5 is the exemplary plot of gesture coarse segmentation;
Fig. 6 is the exemplary plot that gesture subdivision is cut.
Embodiment
The embodiment of the present invention is described below in conjunction with the accompanying drawings, so as to those skilled in the art preferably Understand the present invention.Requiring particular attention is that in the following description, when known function and the detailed description of design perhaps When can desalinate the main contents of the present invention, these descriptions will be ignored herein.
Embodiment
Fig. 1 is the flow chart of the gesture identification method of the invention based on hybrid neural networks.As shown in figure 1, base of the present invention Comprise the following steps in the gesture identification method of hybrid neural networks:
S101:Extract the feature of sample and training sample to be identified:
Feature extraction is carried out firstly the need of to images of gestures to be identified and images of gestures training sample.Fig. 2 is in the present invention The flow chart of images of gestures feature extraction.As shown in Fig. 2 images of gestures feature extraction comprises the following steps in the present invention:
S201:Images of gestures noise suppression preprocessing:
The present invention is using based on Pulse Coupled Neural Network (PCNN-Pulse Coupled Neural Network) and again The denoising that the Denoising Algorithm that Denoising Algorithm is combined carries out gesture gray level image is closed, is first passed through using Pulse Coupled Neural Network pair Images of gestures carries out the differentiation detection of noise spot and marginal point, is carried out afterwards according to the type of noise spot using compound Denoising Algorithm Denoising operates, so as to reach the purpose that a variety of noises are removed on the premise of marginal information is retained.
Each neuron of Pulse Coupled Neural Network is made up of three parts:Receiving portion, modulating part and pulse production Raw device.Pulse Coupled Neural Network is a kind of common method of image noise reduction pretreatment, and its main function is that removing the spiced salt makes an uproar Sound.When Pulse Coupled Neural Network is used for image noise reduction field, it is possible to understand that network is connected into the local of two-dimension single layer, Neuron and the pixel in pending gray level image are one-to-one in this network, and between adjacent neurons It is the relation being connected with each other.During noise reduction process, the gray value of each pixel of pending image can be understood as god Feed back input through member, while the output of each neuron is only used as the input of adjacent neurons, and each neuron is defeated Do well only two kinds:Fired state and the state that misfires, can be designated as 1 and 0 respectively.Due to pixel corresponding to noise and surrounding Pixel difference it is larger, therefore can utilize the granting characteristic combination noise of Pulse Coupled Neural Network self-characteristic progress The judgement of noise spot, specific determination methods are as follows:
Establish the Pulse-coupled Neural Network Model of gesture gray level image;By the ash of current each pixel of gesture gray level image Then angle value utilizes the granting characteristic of Pulse Coupled Neural Network as the input that neuron is corresponded in Pulse Coupled Neural Network The pixel of whole image is detected, if the output state of pixel is fired state, by testing result matrix Element corresponding to the pixel is set to 1, is otherwise set to 0, it is seen that testing result matrix is identical with the size of pending image;Set Noise reduction process window size, it is 3 × 3 in the present embodiment;Each element of testing result matrix is traveled through, if element value is 1, It is fired state to be exactly, then the center using the element as noise reduction process window, counts in noise reduction process window and removes central point element The value (i.e. the testing results of other pixels beyond central point corresponding pixel points point) of other elements in addition, if be worth for 0 (i.e. Misfire state) number of elements be more than predetermined threshold value, it is noise spot to illustrate the central point, and then the central point is not other situations Noise spot.So as to reach the purpose for noise spot and marginal point judge differentiation.Amount threshold is usually noise reduction process window The half of middle number of elements.
Corresponding denoising operation is carried out after judging to obtain noise spot, then using compound Denoising Algorithm, its main method is:
Assuming that a (i, j) is the gray value at pixel (i, j) place in image, 1≤i≤M, 1≤j≤N, M represent gesture gray scale The pixel quantity (i.e. columns) that image is often gone, N represent the pixel quantity (i.e. line number) of gesture gray level image each column, b (i, j) The intermediate value output gray level value after medium filtering is carried out for the pixel.In order to reach the purpose to Gaussian noise, using noise spot Pixel value and intermediate value output gray level value difference as noise estimation value, as shown in following formula (1)
H (i, j)=| a (i, j)-b (i, j) | (1)
Because the type of noise is different, if merely using above-mentioned method of estimation, it is impossible to reach and distinguish a variety of noises Purpose, therefore on the basis of above-mentioned formula, introduce another noise estimation value V (i, j) again, the parameter is pixel The pixel value a (i, j) at (i, j) place and similar two points m1(i, j) and m2The average value of the gradient sum of (i, j), such as following formula (3) It is shown
Wherein, m1(i, j) and m2It is closest with a (i, j) gray value in neighborhood where (i, j) represents pixel (i, j) respectively Two points gray value.
Setting threshold value is T1And T2, then by the relation between above two noise estimation value and threshold value, realize and difference made an uproar The alignment processing of sound, specific method are:
If H (i, j) >=T1, and V (i, j) >=T2, then judge that the type of the noise spot is made an uproar for salt-pepper noise or pulse Sound, the noise spot is handled using medium filtering, will the gray value of the noise spot be revised as medium filtering output valve, such as Fruit H (i, j) < T1, or H (i, j) >=T1And V (i, j) < T2, then the noise type is judged for Gaussian noise, using average Filtering is handled the noise spot, will the gray value of the noise spot be revised as mean filter output valve.
In above-mentioned algorithm, threshold value T1And T2Selection to the fine or not most important of compound Denoising Algorithm result.It is wherein current Conventional Research on threshold selection is that mean absolute deviation algorithm is MAD algorithms.It can be seen from the algorithm, T1=3.5 δij, δijRepresent The mean absolute deviation of all pixels point in the denoising window of pixel (i, j).Threshold value T2Selection primarily directed to gesture figure The texture being likely to occur as in, according to MAD algorithms and experiment experience, T2Value be ordinarily selected to 6~10 integer.
S202:Histogram equalization:
Histogram equalization processing refers to the method being adjusted using image histogram to the contrast of image, so as to handle The grey level histogram of original image becomes to be uniformly distributed in global scope from some gray areas for comparing concentration.The present invention is right Gesture gray level image after step S201 denoisings carries out histogram equalization processing, be in order to expand images of gestures prospect and The difference of background part gray value.Histogram equalization is a kind of current method of conventional picture superposition, and its is specific Step will not be repeated here.
S203:Gesture coarse segmentation based on cell neural network:
As Pulse Coupled Neural Network, the neuron in cell neural network and the pixel in gesture gray level image Corresponding, the cells of note the i-th row jth row is C (i, j) (pixel (i, j) in corresponding gesture gray level image), cell C (i, J) formed by four parts:Input variable uij, state transfering variable xij, output variable yijAnd threshold values I.Cell neural network Cell between locally interconnect, cell C (i, j) only neighborhood N with itrCell in (i, j) is connected with each other, and and other Cell without direct annexation.Cell C (i, j) neighborhood Nr(i, j) can be defined as:
Nr(i, j)=C (k, l) | max (k-i, l-j)≤r (3)
Wherein, r is positive integer, 1≤i, k≤M, 1≤j, l≤N, and M represents the pixel quantity that gesture gray level image is often gone, N represents the pixel quantity of gesture gray level image each column.That is the neighborhood of cell C (i, j) is the length of side 2r centered on C (i, j) Scope included by+1 square.
The main formulas of cell neural network is:
State migration procedure:
Output equation:
Wherein, 1≤i, k≤M, 1≤j, l≤N;T represents iterations;A (k, l) represents the neighborhood residing for cell C (i, j) NrThe feedback weight of cell C (k, l) in (i, j);B (k, l) then represents the neighborhood N residing for cell C (i, j)rIt is thin in (i, j) Other elements in born of the same parents C (k, l) control weight, namely template B in addition to the element of center.Here the value of (k, l) according to According to neighborhood NrThe definition of (i, j) determines.
Feedback template A and control module B is (2r+1) × (2r+1) matrix, and I represents the threshold values of cell neural network Template, A, B and I value synthesis determine the input quantity u of cell neural networkij, output quantity yijAnd state transfer amount xijPair It should be related to.Therefore for Cellular Neural Networks, how correctly design of feedback template A, Control architecture B and threshold values I value is most important.
The template design method that the present invention uses is to be set based on algebraically knot with the template that forefathers' stencil design experience is combined Meter method, template A, B, I form are typically designed as follows:
I=-d (8)
Wherein, a, b, c, d are normal number.
Establish the Cellular Neural Networks of gesture gray level image, by each pixel of gesture gray level image after equalization (i, J) gray value is as the input u that cell is corresponded in Cellular Neural Networksij, changed according to the formula of state migration procedure In generation, until whole cell neural network is restrained, there is output y in each cellij(t).It can be seen from output equation, cytocidal action The output valve y of networkij(t) between 1 and -1, y is worked asij(t) be 1 when, represent completely black;Work as yij(t) when being -1, represent complete white.
Judge whether certain pixel is that the general principle of marginal point is:When some pixel value is completely black, when being+1, if Each pixel value and more than setting threshold parameter in its corresponding neighborhood, then this pixel is not edge pixel, now pixel Value tends to be complete white;, whereas if each pixel value and less than setting the threshold parameter in its corresponding neighborhood, then this pixel Edge pixel is represented, now pixel value tends to be completely black.When this pixel value is complete white, i.e., when -1, then no matter in its corresponding neighborhood respectively How is the value size of individual pixel, and this pixel value will all tend to be complete white.
According to principles above, judge whether certain pixel is that the method for marginal point is in the present invention:Traversal cell nerve net The output valve of each cell corresponding to pixel member in network, when the output valve of some pixel is in the range of [0,1], if its is right Answer the pixel value of other pixels neighborhood Nei and more than predetermined threshold value, then this pixel is not edge pixel, is otherwise edge pixel Point;When output valve [- 1,0) in the range of, be not edge pixel point.The threshold value of neighborhood territory pixel value sum is set according to actual conditions Put.
S204:Gesture subdivision is carried out with reference to gesture shape characteristic to cut:
Fig. 3 is to combine gesture shape characteristic to carry out the flow chart that gesture subdivision is cut.As shown in figure 3, gesture subdivision of the present invention Cut and comprise the following steps:
S301:Extract connected region and profile:
According to the edge pixel point obtained using cell neural network, connected region is asked for, so as to remove other backgrounds letter The interference of breath, only retain the hand and face area of people.Algorithm that connected region uses is asked in the present embodiment as two_pass Algorithm.Then the profile of connected region is extracted, the present embodiment is using search labeling method extraction profile, idiographic flow:To upper Image after face extraction connected region is systematically scanned, if some point in connected region is run into, with the point For starting point, its edge is then tracked, and the pixel above edge is marked.Completely closed when the profile of scanning reaches Close, then return to a position and continue to scan on, until finding new Pixel Information.Extract connected region and profile can also basis Need to select other method.
S302:Finger tip detection is carried out to each connected region:
Finger tip detection is carried out respectively for obtained each connected region, so as to determine whether gesture area.General feelings Under condition when carrying out gesture identification, finger is all separated, therefore can carry out finger tip detection by curvature estimation.Fig. 4 is The schematic diagram of finger tip detection of the present invention.As indicated at 4, the method for finger tip detection is:
Each wire-frame image vegetarian refreshments in connected region is traveled through, using the pixel as datum mark, coordinate is designated as p (px,py, 0), (px,py) two-dimensional coordinate of the datum mark in images of gestures is represented, a distance constant L is preset, p is taken along contour direction The l-th point p of point above1(p1x,p1y, 0), then point p and point p1Straight line is formed, is then taken along contour direction behind point p L-th point p2(p2x,p2y, 0), then point p and point p2Straight line can also be formed, meeting shape in an angle, should between this two Angle is designated as α;By vectorWith vectorBetween angle cosine value as the curvature result that will be calculated, i.e. curvature estimation Formula is:
If cos α are more than preset curvature threshold value T, it is finger tip point undetermined to judge the point.Threshold value T size be according to away from From constant L come what is set, when distance constant L is bigger, threshold value T is also bigger.Distance constant L generally also can not be too small or excessive, Set generally according to a quarter of finger average length to half.
For interference for the groove part of finger, vector can be passed throughWith vectorBetween vector product symbol Number determine.As seen in Figure 4, when point p is located at fingertip locationThe symbol of vector product is located at groove with point p During positionThe symbol of vector product is different, therefore can pass throughSymbol judge point p position.Exactly For this purpose, just by point p, p1And p2Coordinate represented in a manner of three-dimensional rectangular coordinate.Fingertip locationVector Long-pending symbol is relevant with traversal direction, when according to the traversal clockwise of gesture area overall profile, according to the right hand of vector product Rule, fingertip locationVector product is inside perpendicular to image, as negative, when according to the inverse of gesture area overall profile When hour hands travel through (traversal direction as shown in Figure 4), fingertip locationVector product is outside perpendicular to image, is Just.According to fingertip locationThe symbol of vector product, so as to remove the interference of groove part.Judge finger tip point undeterminedThe symbol of vector product, if with fingertip location corresponding to symbol it is identical, be left finger tip point undetermined, otherwise do not protect Stay.
S303:Judge gesture area:
After finger tip point is detected, it is also necessary to finger tip point is judged, so as to remove some parts of face part because It is more than the interference of threshold value for curvature caused by angle problem, judgement obtains gesture area.Present invention employs double judgement side Method:
The maximum finger tip point undetermined of i.e. y-coordinate for first determining whether to detect in connected region and the y-coordinate detected are minimum Finger tip point undetermined between y-coordinate difference whether exceed face height half, if it is, the connected region is not gesture area Domain, otherwise as gesture area undetermined.Why distance is dimensioned to the half of face height here, is surveyed by testing What examination was drawn, it on the premise of correct finger tip point is completely retained, can thus remove the interference of face part.
Whether finger tip point quantity undetermined exceedes default amount threshold in each gesture area undetermined further judged, If it is, the connected region is gesture area, otherwise it is not.The number for the finger tip point quantity that actual gesture area obtains with it is bent Rate threshold value T is relevant, therefore in actual applications, the threshold value of finger tip point quantity can pass through the reality to several gesture training samples Result is tested to be counted to obtain.
S304:Gesture area is split:
Operation eliminates the interference of other connected regions such as face more than, has obtained gesture area.But gesture area The palm portion for not merely including people is possible to inside domain, it is possible to also have the parts such as wrist.Generally, the gesture of people The information that effective information all concentrates on the parts such as the palm portion of people, wrist can be ignored substantially.Therefore in order that obtaining later stage spy Sign extraction and tracking efficient and effectively, it is necessary to split to gesture area, reaches and only retains finger and the mesh of palm portion 's.
According to the shape facility of human hand, the present invention is approximately equal to 2 come real according to the ratio of the length of gesture and the width of gesture Now to the segmentation of gesture., it is necessary to first know the principal direction of gesture area before being split, gesture master is asked in the present embodiment The method in direction is:The barycenter of gesture area is asked for, barycenter is then tried to achieve to the vector of each finger tip point, these vectors is carried out Average, the direction of the average vector is gesture area principal direction.Then gesture is carried out further according to the principal direction of gesture area Segmentation.The dividing method that the present embodiment uses for:The boundary rectangle of gesture area is obtained by gesture area principal direction, with principal direction Parallel place side is length, and the side vertical with principal direction is width, the broadside where finger tip point is selected, since the broadside, along long side Interception distance is the boundary rectangle of 2 times of width edge lengths, and the interior gesture area included of the boundary rectangle is to split to be obtained The gesture area for only retaining finger and palm portion.
S205:Gesture feature is extracted using the Fourier descriptor for retaining phase information:
The gesture area for splitting to obtain for step S204, the present invention devise a kind of Fourier for retaining phase information and retouched Son is stated to extract gesture feature information, so as to remove the rotational invariance of conventional Fourier description, reaches differentiation rotation gesture Purpose.
Discrete fourier coefficient z (k) can be expressed as:
Wherein, p (i) represents i-th of data in discrete series, and n represents the data bulk in discrete series, and e is represented certainly Right constant, j are imaginary unit.It is gesture profile due to needing into line translation, therefore discrete series p (i) is step in the present invention Rapid S104 splits the plural form of coordinate in obtained gesture area wire-frame image vegetarian refreshments.
Inverse Fourier transform can be expressed as:
According to the fundamental property z (k) of Fourier transformation=z*(n-k) remove in Fourier transformation form z from K+1 to n- K-1 HFS, wherein, z* here represents z conjugate complex number form;K span is:[0,n/2].Then it is right again The z for removing HFS carries out inverse Fourier transform, will obtain with the approximate curve of former Fourier transformation, but the curve becomes More smooth, this curve turns into the K curve of approximation of former Fourier's change curve.Wherein, above-mentioned described Fourier system Several subsets { z (k) n-K < k≤K } then seeks to the Fourier descriptor for extracting gesture feature.
The Fourier descriptor and yardstick of shape, the original position of direction and curve have certain relation.Therefore, in order to Ensure that recognizer has rotation, translation and scale invariance, then need Fourier descriptor to be normalized operation.According to The fundamental property of Fourier's change can prove, when representing profile with Fourier coefficient, coefficient amplitude | | z (k) | | there is rotation Consistency, translation invariance and start position independence, wherein, 0≤k≤n-1, and because Z [0] does not have translation invariant Property, therefore k span is arranged to [1, n-1].In order to realize the scale invariability of Fourier descriptor, Z [0] can will be removed The amplitude Z (k) of each coefficient in addition | | divided by | | Z (1) | |, so as to reach the characteristic of Scale invariant.Normalization operation it Fourier descriptor S [k '] afterwards can be expressed as:
Wherein, 1≤k '≤n-1;| | | | represent modulo operator.
The detailed description of unitary Fourier descriptor may refer to document, and " Song Rui China describes the hand of son based on Fourier Gesture recognizer [D] Xian Electronics Science and Technology University, 2008 ".
The present invention describes sub rotational invariance to remove conventional Fourier, remains the phase information after rotation, The normalized form of Fourier descriptor after improvementIt can be expressed as:
Wherein,The angle of gesture area principal direction and x-axis is represented, j is imaginary unit.Fourier descriptor S above [k '] remains the phase information of gesture rotation, therefore description does not have rotational invariance.Therefore the present invention is used and removed In additionFeature of the coefficient as gesture area.This feature has a translation and scale invariance, and with gesture profile The original position of curve is unrelated, while has rotational variability again, and this feature vector can reach the mesh distinguished to rotation gesture 's.Because the profile point quantity of different gesture areas is not necessarily identical, therefore in actual applications, only in Fourier descriptor Q constitutive characteristic vector, Q size can be determined according to actual conditions before unified selection.
S102:BP neural network is trained according to training sample:
BP neural network, its corresponding gesture figure are inputted using the characteristic vector of training sample images of gestures as training sample As output of the classification as BP neural network, BP neural network is trained.BP neural network is a kind of conventional nerve net Network, the specific composition and parameter and training method of its network, will not be repeated here.
S103:Gesture identification is carried out to the sample identified:
In the BP neural network that the characteristic vector input step S102 of images of gestures to be identified is trained, output identifies The images of gestures classification arrived.
In order to illustrate the technique effect of the present invention, experimental verification has been carried out to the present invention.The gesture training sample point of selection For gesture upward, gesture down, gesture towards left, gesture towards right four parts, the training samples number per part is 80, equally from Reselection test sample in this four classes image, per partial test sample size 40.In order to show conveniently, gesture court is only chosen herein Loading carries out implementation process explanation originally, and the size in sample per pictures is 256 × 256, gray level 256.
Image denoising is carried out firstly the need of to sample upward.Because the scale size of samples pictures is 256 × 256, due to When Pulse Coupled Neural Network is used for image noise reduction field, its neuron number is to correspond with pending image slices vegetarian refreshments , therefore the neuron number of Pulse Coupled Neural Network is arranged to 65536, the pulse coupled neural net that the present embodiment uses The parameter of network model is arranged to:Neuron iterations τ=10, neuron bonding strength β=3, dynamic threshold parameter θij=1, The amplification coefficient V of threshold value outputθ=20, the attenuation coefficient a of threshold function tableθ=0.2, then using providing characteristic to pulse-couple Neutral net is detected, then judges by testing result to obtain noise spot, then according to the type of noise spot, is gone using compound Algorithm of making an uproar carries out denoising operation, wherein the parameter of compound Denoising Algorithm is arranged to T1=3.5 δij, wherein SkRepresent noise window, The detection window of noise window size and Pulse Coupled Neural Network is in the same size, and size is 3 × 3, T2=8.
After carrying out histogram equalization to the images of gestures after denoising, detect to obtain images of gestures using cell neural network The edge of middle gesture, realizes the coarse segmentation to images of gestures, in the present embodiment, the neighborhood of each cell in cell neural network Size is 3*3, and used template is:
Fig. 5 is the exemplary plot of gesture coarse segmentation.
Images of gestures is finely divided then in conjunction with gesture shape feature and cut.Wherein constant L size is 80, curvature estimation Threshold value T size be 0.5.Fig. 6 is the exemplary plot that gesture subdivision is cut.It can be seen that after progress gesture subdivision is cut, can eliminate The influence in the regions such as face, obtain accurate gesture area.
Then the profile point coordinates for segmenting the gesture area for cutting to obtain is built into discrete series again, after carrying out Fourier transformation Fourier coefficient is obtained, is then normalized according to formula (13), 200 before being selected in the Fourier descriptor after normalization Form gesture feature vector.
BP neural network is trained using the gesture feature vector of training sample, wherein the input layer of BP neural network Number determines that the number of output layer is determined by gesture sample species by gesture feature vector, the input layer used of the invention Number is 200, and the number of hidden layer is 10, and the number of output layer is 4.Output result can by binary form 0001,0010, 0100,1000 represents, wherein 0001 represents gesture upward, 0010 represents gesture down, and 0100 represents that gesture represents towards a left side, 1000 Which kind of type is gesture belong to towards the right side according to the result judgement gesture that gesture exports.
In order to verify that what the present invention designed new is gone based on what Pulse Coupled Neural Network and compound Denoising Algorithm were combined Make an uproar algorithm noise reduction quality, the Denoising Algorithm that designs of the present invention and simple compound Denoising Algorithm and medium filtering are done Comparative analysis, the leading indicator of contrast is Y-PSNR PSNR.Table 1 is PSNR pair of Denoising Algorithm of the present invention and contrast algorithm According to table.
Table 1
As it can be seen from table 1 in the case of identical noise density, its PSNR of denoising method proposed by the present invention value Apparently higher than medium filtering and the value of simple compound Denoising Algorithm.As can be seen here, the combination pulse coupled neural that the present invention designs The Denoising Algorithm of network and compound Denoising Algorithm has good denoising effect.
In addition, also being contrasted using the recognition effect of traditional Fourier descriptor, contrast index is gesture sample Discrimination.Table 2 is the gesture sample recognition result statistical form of conventional Fourier description.Table 3 is Fourier descriptor of the present invention Gesture sample recognition result statistical form.
Table 2
Table 3
By contrast table 2 and the result of table 3, traditional Fourier descriptor can not identify that rotation is larger well Gesture, discrimination only have 71% or so, and discrimination is relatively low, therefore this method has the scene of different implications for rotating gesture When, effect is not fine.The improved Fourier descriptor of the present invention can rotate a certain angle in tolerance gesture, although at angle Degree rotation will be considered that it is two kinds of different images when excessive, but still reach 91% or so by the experimental verification present invention Discrimination, achieve good gesture identification effect.
Although the illustrative embodiment of the present invention is described above, in order to the technology of the art Personnel understand the present invention, it should be apparent that the invention is not restricted to the scope of embodiment, to the common skill of the art For art personnel, if various change in the spirit and scope of the present invention that appended claim limits and determines, these Change is it will be apparent that all utilize the innovation and creation of present inventive concept in the row of protection.

Claims (4)

1. a kind of gesture identification method based on hybrid neural networks, it is characterised in that comprise the following steps:
S1:The feature of images of gestures and images of gestures training sample to be identified is extracted, specific steps include:
S1.1:It is recommended that the Pulse-coupled Neural Network Model of gesture gray level image, by current each pixel of gesture gray level image Gray value utilizes the granting characteristic pair of Pulse Coupled Neural Network as the input that neuron is corresponded in Pulse Coupled Neural Network The pixel of images of gestures is detected, if the output state of pixel is fired state, will be somebody's turn to do in testing result matrix Element corresponding to pixel is set to 1, is otherwise set to 0;Each element of testing result matrix is traveled through, if element value is 1, with The element is the center of noise reduction process window, and the size of noise reduction process window is set according to actual conditions, counts noise reduction process window The value of other elements in mouthful in addition to central point element, if value is more than predetermined threshold value for 0 number of elements, illustrate the center Point is noise spot, and other situations are not then noise spots;
Calculate two kinds of noise estimation value H (i, j) and the V (i, j) of noise spot as follows respectively:
H (i, j)=| a (i, j)-b (i, j) |
Wherein, a (i, j) is the gray value at pixel (i, j) place in image, and b (i, j) is after the pixel carries out medium filtering Intermediate value output gray level value;
<mrow> <mi>V</mi> <mrow> <mo>(</mo> <mi>i</mi> <mo>,</mo> <mi>j</mi> <mo>)</mo> </mrow> <mo>=</mo> <mfrac> <mrow> <mo>|</mo> <msub> <mi>m</mi> <mn>1</mn> </msub> <mrow> <mo>(</mo> <mi>i</mi> <mo>,</mo> <mi>j</mi> <mo>)</mo> </mrow> <mo>-</mo> <mi>a</mi> <mrow> <mo>(</mo> <mi>i</mi> <mo>,</mo> <mi>j</mi> <mo>)</mo> </mrow> <mo>|</mo> <mo>+</mo> <mo>|</mo> <msub> <mi>m</mi> <mn>2</mn> </msub> <mrow> <mo>(</mo> <mi>i</mi> <mo>,</mo> <mi>j</mi> <mo>)</mo> </mrow> <mo>-</mo> <mi>a</mi> <mrow> <mo>(</mo> <mi>i</mi> <mo>,</mo> <mi>j</mi> <mo>)</mo> </mrow> <mo>|</mo> </mrow> <mn>2</mn> </mfrac> </mrow>
Wherein, m1(i, j) and m2(i, j) represent respectively in neighborhood where pixel (i, j) with a (i, j) gray value immediate two The gray value of individual point;
If H (i, j) >=T1, and V (i, j) >=T2, T1And T2The threshold value set is represented, then using medium filtering to the noise spot Handled, otherwise the noise spot is handled using mean filter;
S1.2:Histogram equalization is carried out to the gesture gray level image after step S1.1 denoisings;
S1.3:The Cellular Neural Networks of gesture gray level image are established, by each pixel of gesture gray level image after equalization The gray value of (i, j) is as the input u that cell is corresponded in Cellular Neural Networksij, enter according to the formula of state migration procedure Row iteration, until the convergence of whole cell neural network, obtain the output yi of each cellj(t);It is every in traversal cell neutral net The output valve of cell member corresponding to individual pixel, when the output valve of some pixel is in the range of [0,1], if its corresponding neighborhood The pixel value of other interior pixels and more than predetermined threshold value, then this pixel is not edge pixel, is otherwise edge pixel point;When defeated Go out value [- 1,0) in the range of, be not edge pixel point;
S1.4:The edge pixel point obtained according to step S1.3 obtains connected region, and extraction obtains the profile of connected region, to every Individual connected region carries out finger tip detection respectively, and Fingertip Detection is:
Each wire-frame image vegetarian refreshments in connected region is traveled through, using the pixel as datum mark, coordinate is designated as p (px,py, 0), in advance If a distance constant L, the l-th point p before p points is taken along contour direction1(p1x,p1y, 0), take the l-th point p behind point p2 (p2x,p2y, 0), calculate vectorWith vectorBetween angle cosine value cos α, if cos α are more than preset curvature threshold value T, then it is finger tip point undetermined to judge the point, otherwise not as finger tip point undetermined;
Fingertip location vector product is determined according to traversal directionSymbol, if according to the suitable of gesture area overall profile When hour hands travel through, vectorial product code should be negative, otherwise vectorial just, to calculate finger tip point undeterminedWith vectorBetween to Amount productIf the symbol of the vector product is identical with symbol corresponding to fingertip location, finger tip point undetermined is left, it is no Do not retain then;
Judge in all finger tip points undetermined for being detected in the connected region, the maximum finger tip point undetermined of y-coordinate and y-coordinate are minimum The y-coordinate difference of finger tip point undetermined whether exceed the half of face height, if it is, the connected region is not gesture area, Otherwise it is used as gesture area undetermined;Whether finger tip point quantity undetermined exceedes pre- in each gesture area undetermined further judged If amount threshold, if it is, the connected region is gesture area, be not otherwise;
The principal direction of gesture area is asked for, is divided according to principal direction according to gesture length and width ratio for 2 pairs of gesture areas Cut, the gesture area after being split;
S1.5:The gesture area that will be obtained after step S1.4 segmentations, by the profile point coordinates of gesture area with plural form table Show, all profile point coordinates are formed into discrete series, note profile point quantity is n, carries out Fourier transformation to the discrete series, obtains To n Fourier coefficient z (k), k=0,1 ..., n-1, Fourier descriptor is calculated
Wherein k '=1,2 ..., n-1,Represent the angle of gesture area principal direction and x-axis;
Q constitutive characteristic vector before being selected in Fourier descriptor;
S2:BP neural network, its corresponding gesture figure are inputted using the characteristic vector of training sample images of gestures as training sample As output of the classification as BP neural network, BP neural network is trained;
S3:In the BP neural network that the characteristic vector input step S2 of images of gestures to be identified is trained, output identification obtains Images of gestures classification.
2. gesture identification method according to claim 1, it is characterised in that threshold value T in step S1.11=3.5 δij, δijTable Show the mean absolute deviation of all pixels point in the denoising window of pixel (i, j);Threshold value T2For 6~10 integer.
3. gesture identification method according to claim 1, it is characterised in that in step S1.3, in cell neural network Feeding back template A is:
<mrow> <mi>A</mi> <mrow> <mo>(</mo> <mi>k</mi> <mo>,</mo> <mi>l</mi> <mo>)</mo> </mrow> <mo>=</mo> <mfenced open = "{" close = ""> <mtable> <mtr> <mtd> <mrow> <mi>a</mi> <mo>,</mo> </mrow> </mtd> <mtd> <mrow> <mi>k</mi> <mo>=</mo> <mi>i</mi> <mo>,</mo> <mi>l</mi> <mo>=</mo> <mi>j</mi> </mrow> </mtd> </mtr> <mtr> <mtd> <mrow> <mn>0</mn> <mo>,</mo> </mrow> </mtd> <mtd> <mrow> <mi>k</mi> <mo>&amp;NotEqual;</mo> <mi>i</mi> <mo>,</mo> <mi>l</mi> <mo>&amp;NotEqual;</mo> <mi>j</mi> </mrow> </mtd> </mtr> </mtable> </mfenced> <mo>,</mo> </mrow>
Control architecture B is:
<mrow> <mi>B</mi> <mrow> <mo>(</mo> <mi>k</mi> <mo>,</mo> <mi>l</mi> <mo>)</mo> </mrow> <mo>=</mo> <mfenced open = "{" close = ""> <mtable> <mtr> <mtd> <mrow> <mi>b</mi> <mo>,</mo> </mrow> </mtd> <mtd> <mrow> <mi>k</mi> <mo>=</mo> <mi>i</mi> <mo>,</mo> <mi>l</mi> <mo>=</mo> <mi>j</mi> </mrow> </mtd> </mtr> <mtr> <mtd> <mrow> <mo>-</mo> <mi>c</mi> <mo>,</mo> </mrow> </mtd> <mtd> <mrow> <mi>k</mi> <mo>&amp;NotEqual;</mo> <mi>i</mi> <mo>,</mo> <mi>l</mi> <mo>&amp;NotEqual;</mo> <mi>j</mi> </mrow> </mtd> </mtr> </mtable> </mfenced> <mo>,</mo> </mrow>
Threshold value I=-d,
Wherein (k, l) be using in cell neural network centered on cell C (i, j), the length of side as 2r+1 neighborhood NrPoint in (i, j), A, b, c, d are normal number.
4. gesture identification method according to claim 1, it is characterised in that in step S1.4, ask for the main side of gesture area To method be:The barycenter of gesture area is asked for, barycenter is then tried to achieve to the vector of each finger tip point, these vectors is put down , the direction of the average vector is gesture area principal direction.
CN201510280013.3A 2015-05-27 2015-05-27 Gesture identification method based on hybrid neural networks Expired - Fee Related CN104834922B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201510280013.3A CN104834922B (en) 2015-05-27 2015-05-27 Gesture identification method based on hybrid neural networks

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201510280013.3A CN104834922B (en) 2015-05-27 2015-05-27 Gesture identification method based on hybrid neural networks

Publications (2)

Publication Number Publication Date
CN104834922A CN104834922A (en) 2015-08-12
CN104834922B true CN104834922B (en) 2017-11-21

Family

ID=53812800

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201510280013.3A Expired - Fee Related CN104834922B (en) 2015-05-27 2015-05-27 Gesture identification method based on hybrid neural networks

Country Status (1)

Country Link
CN (1) CN104834922B (en)

Families Citing this family (28)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105487772A (en) * 2015-11-26 2016-04-13 上海斐讯数据通信技术有限公司 Information capturing method and apparatus
CN105373785B (en) * 2015-11-30 2019-08-02 北京地平线机器人技术研发有限公司 Gesture identification detection method and device based on deep neural network
CN106022343B (en) * 2016-05-19 2019-08-16 东华大学 A kind of clothes fashion recognition methods based on Fourier descriptor and BP neural network
CN106022297A (en) * 2016-06-01 2016-10-12 苏州大学 Gesture identification method and gesture identification device
CN108073979A (en) * 2016-11-14 2018-05-25 顾泽苍 A kind of ultra-deep study of importing artificial intelligence knows method for distinguishing for image
CN108229277B (en) * 2017-03-31 2020-05-01 北京市商汤科技开发有限公司 Gesture recognition method, gesture control method, multilayer neural network training method, device and electronic equipment
CN109101860B (en) * 2017-06-21 2022-05-13 富泰华工业(深圳)有限公司 Electronic equipment and gesture recognition method thereof
CN107894834B (en) * 2017-11-09 2021-04-02 上海交通大学 Control gesture recognition method and system in augmented reality environment
CN108230257A (en) * 2017-11-15 2018-06-29 北京市商汤科技开发有限公司 Image processing method, device, electronic equipment and storage medium
CN108052884A (en) * 2017-12-01 2018-05-18 华南理工大学 A kind of gesture identification method based on improvement residual error neutral net
CN108108024B (en) * 2018-01-02 2021-01-22 京东方科技集团股份有限公司 Dynamic gesture obtaining method and device and display device
CN108198567A (en) * 2018-02-22 2018-06-22 成都启英泰伦科技有限公司 A kind of novel voice is except system of making an uproar
CN109344689A (en) * 2018-08-07 2019-02-15 西安理工大学 A kind of sign language gesture identification method based on Kinect
CN109359538B (en) * 2018-09-14 2020-07-28 广州杰赛科技股份有限公司 Training method of convolutional neural network, gesture recognition method, device and equipment
CN109344793B (en) 2018-10-19 2021-03-16 北京百度网讯科技有限公司 Method, apparatus, device and computer readable storage medium for recognizing handwriting in the air
CN109697407A (en) * 2018-11-13 2019-04-30 北京物灵智能科技有限公司 A kind of image processing method and device
CN109612326B (en) * 2018-12-19 2021-08-24 西安建筑科技大学 Intelligent auxiliary teaching system for small arms shooting based on Internet of things
CN113033256B (en) * 2019-12-24 2024-06-11 武汉Tcl集团工业研究院有限公司 Training method and device for fingertip detection model
CN111259902A (en) * 2020-01-13 2020-06-09 上海眼控科技股份有限公司 Arc-shaped vehicle identification number detection method and device, computer equipment and medium
CN111216133B (en) * 2020-02-05 2022-11-22 广州中国科学院先进技术研究所 Robot demonstration programming method based on fingertip identification and hand motion tracking
CN112487981A (en) * 2020-11-30 2021-03-12 哈尔滨工程大学 MA-YOLO dynamic gesture rapid recognition method based on two-way segmentation
WO2022126367A1 (en) * 2020-12-15 2022-06-23 Qualcomm Incorporated Sequence processing for a dataset with frame dropping
CN112800954A (en) * 2021-01-27 2021-05-14 北京市商汤科技开发有限公司 Text detection method and device, electronic equipment and storage medium
CN113191361B (en) * 2021-04-19 2023-08-01 苏州大学 Shape recognition method
CN113449600B (en) * 2021-05-28 2023-07-04 宁波春建电子科技有限公司 Two-hand gesture segmentation algorithm based on 3D data
CN113792624A (en) * 2021-08-30 2021-12-14 河南林业职业学院 Early warning security monitoring method for bank ATM
CN114625333B (en) * 2022-03-08 2022-10-18 深圳康荣电子有限公司 Method and system capable of recording gesture instructions to control liquid crystal splicing LCD
CN117558068B (en) * 2024-01-11 2024-03-19 深圳市阿龙电子有限公司 Intelligent device gesture recognition method based on multi-source data fusion

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8620024B2 (en) * 2010-09-17 2013-12-31 Sony Corporation System and method for dynamic gesture recognition using geometric classification
CN103778407A (en) * 2012-10-23 2014-05-07 南开大学 Gesture recognition algorithm based on conditional random fields under transfer learning framework
CN104298354A (en) * 2014-10-11 2015-01-21 河海大学 Man-machine interaction gesture recognition method
CN104573621A (en) * 2014-09-30 2015-04-29 李文生 Dynamic gesture learning and identifying method based on Chebyshev neural network

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8620024B2 (en) * 2010-09-17 2013-12-31 Sony Corporation System and method for dynamic gesture recognition using geometric classification
CN103778407A (en) * 2012-10-23 2014-05-07 南开大学 Gesture recognition algorithm based on conditional random fields under transfer learning framework
CN104573621A (en) * 2014-09-30 2015-04-29 李文生 Dynamic gesture learning and identifying method based on Chebyshev neural network
CN104298354A (en) * 2014-10-11 2015-01-21 河海大学 Man-machine interaction gesture recognition method

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
基于神经网络的手势识别技术研究;江立 等;《北京交通大学学报》;20061031(第5期);正文32-36页 *

Also Published As

Publication number Publication date
CN104834922A (en) 2015-08-12

Similar Documents

Publication Publication Date Title
CN104834922B (en) Gesture identification method based on hybrid neural networks
CN110084173B (en) Human head detection method and device
CN107657279B (en) Remote sensing target detection method based on small amount of samples
Yuan et al. Factorization-based texture segmentation
Parvin et al. Iterative voting for inference of structural saliency and characterization of subcellular events
CN110837768B (en) Online detection and identification method for rare animal protection
CN112184752A (en) Video target tracking method based on pyramid convolution
CN108647588A (en) Goods categories recognition methods, device, computer equipment and storage medium
CN105205449B (en) Sign Language Recognition Method based on deep learning
CN105574534A (en) Significant object detection method based on sparse subspace clustering and low-order expression
CN112837344B (en) Target tracking method for generating twin network based on condition countermeasure
CN105069774B (en) The Target Segmentation method of optimization is cut based on multi-instance learning and figure
CN101615245A (en) Expression recognition method based on AVR and enhancing LBP
CN104038792B (en) For the video content analysis method and apparatus of IPTV supervision
CN102270308A (en) Facial feature location method based on five sense organs related AAM (Active Appearance Model)
CN105956570A (en) Lip characteristic and deep learning based smiling face recognition method
CN110991547A (en) Image significance detection method based on multi-feature optimal fusion
CN106611158A (en) Method and equipment for obtaining human body 3D characteristic information
CN111951283A (en) Medical image identification method and system based on deep learning
CN104361339B (en) Slap shape Graph Extraction and recognition methods
CN103927555A (en) Static sign language letter recognition system and method based on Kinect sensor
CN112949570A (en) Finger vein identification method based on residual attention mechanism
CN106778491B (en) The acquisition methods and equipment of face 3D characteristic information
CN102509308A (en) Motion segmentation method based on mixtures-of-dynamic-textures-based spatiotemporal saliency detection
CN109902692A (en) A kind of image classification method based on regional area depth characteristic coding

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
EXSB Decision made by sipo to initiate substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20171121

Termination date: 20200527

CF01 Termination of patent right due to non-payment of annual fee