CN102402289A - Mouse recognition method for gesture based on machine vision - Google Patents

Mouse recognition method for gesture based on machine vision Download PDF

Info

Publication number
CN102402289A
CN102402289A CN2011103743163A CN201110374316A CN102402289A CN 102402289 A CN102402289 A CN 102402289A CN 2011103743163 A CN2011103743163 A CN 2011103743163A CN 201110374316 A CN201110374316 A CN 201110374316A CN 102402289 A CN102402289 A CN 102402289A
Authority
CN
China
Prior art keywords
gesture
mouse
target
frame
target gesture
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN2011103743163A
Other languages
Chinese (zh)
Other versions
CN102402289B (en
Inventor
徐向民
孙骁
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
South China University of Technology SCUT
Original Assignee
South China University of Technology SCUT
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by South China University of Technology SCUT filed Critical South China University of Technology SCUT
Priority to CN201110374316.3A priority Critical patent/CN102402289B/en
Publication of CN102402289A publication Critical patent/CN102402289A/en
Application granted granted Critical
Publication of CN102402289B publication Critical patent/CN102402289B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Landscapes

  • Image Analysis (AREA)

Abstract

The invention discloses a mouse recognition method for gesture based on machine vision, comprising the following steps: (1) establishing an active shape model for gestures; (2) off-line training for gestures and getting a gesture features classifier; (3) collecting images; (4) extracting partial binary pattern characteristics for the images, searching the target gesture among the images through the gesture features classifier from step (2), then step (5); (5) positioning on the fingertips; (6) mapping of mouse. The invention is a natural and intuitive way of human-computer interaction, and has advantages of no need to carry other auxiliary equipments, finishing the mouse operation through natural hand and finger movement, small influence by light and background, and freely operation right switch when multi-users are operating the computer.

Description

A kind of gesture mouse recognition methods based on machine vision
Technical field
The present invention relates to machine vision technique and human-computer interaction technology, particularly a kind of gesture mouse recognition methods based on machine vision.
Background technology
The development of Along with computer technology, the interacting activity of people and computing machine become an important component part of human schedule life.Mouse mainly is used in the man-machine interaction of graphical interfaces; But in reciprocal process; Staff must be held a kind of mouse apparatus; On a plane, move, all there is certain limitation in this mode of operation at aspects such as the naturality of man-machine interaction and friendly, and therefore research meets the development trend that the human man-machine interaction mode of being accustomed to naturally becomes field of human-computer interaction.
Human-computer interaction technology based on gesture mainly changes the operation that judges need be carried out through identification user's gesture.It is divided into based on the receipt gloves with based on two types of machine vision.Based on the method for data glove, require the user to wear a kind of sensor of similar gloves, thereby obtain the position and the movable information of hand through computing machine, there are shortcomings such as burdensome, dumb in this method.Existing method based on machine, the method that mostly adopts the colour of skin to extract are come the dividing gesture zone, and this method receives illumination effect big, and require usually the zone of video acquisition be confined to hand region among a small circle in, this has limited the scope of activities of staff.
Summary of the invention
In order to overcome the deficiency of prior art, the object of the present invention is to provide a kind of natural, intuitively based on the gesture mouse recognition methods of machine vision.
The object of the invention is realized through following technical scheme:
A kind of gesture mouse recognition methods based on machine vision may further comprise the steps:
(1) sets up the moving shape model of gesture: at first the profile of the open gesture of the five fingers is carried out the basic gesture modeling of B batten; On the basis of basic gesture model, set up the moving shape model of the gesture of B batten then, shown in the following formula of its state space:
χ=(x,y,θ,s,λ,θ 1,l 1,θ 2,l 2,θ 3,l 3,θ 4,l 4,θ 5,θ 6)
Wherein, x, y are respectively the horizontal ordinate and the ordinate of the position of gesture; θ representes the gesture anglec of rotation planar; S represents the size of gesture; θ 1, l 1Represent angle and the little finger length of little finger respectively around its root rotation; θ 2, l 2The expression third finger is around the angle and the nameless length of its root rotation respectively; θ 2, l 3Represent angle and the middle finger length of middle finger respectively around its root rotation; θ 4, l 4Represent angle and the forefinger length of forefinger respectively around its root rotation; θ 6The expression thumb is around the angle of its root rotation; θ 5The expression thumb is around the angle of joint, center rotation;
(2) gesture is carried out off-line training, obtain the gesture feature sorter;
(3) images acquired;
(4) extract partial binary pattern (LBP) characteristic of image, the gesture feature sorter ferret out gesture in image that obtains through step (2); When searching the target gesture, carry out step (5);
(5) finger tip location:
(5-1) adopt the moving shape model of the gesture in the step (1) that the target gesture that searches is estimated match, obtain the state space initial value χ of target gesture 0, make the initial value p of posterior probability density 00| Z 0)=1;
(5-2) the target gesture is observed, utilized the particle filter tracking algorithm that the state of target gesture is carried out the iteration renewal, said iteration renewal process is specific as follows:
The observed quantity that defines the t-1 two field picture is Z T-1, then the state space of target gesture is χ during the t-1 frame T-1Posterior probability be p T-1T-1| Z T-1);
The optimal profile state of target gesture during according to the t-1 frame
Figure BDA0000111086070000021
The optimal profile state of target gesture during to the t frame
Figure BDA0000111086070000022
The estimation probability be: p tt| χ T-1); Order
Figure BDA0000111086070000023
The likelihood probability of the profile state of target gesture is p during the t frame t(Z t| χ t);
The state space of target gesture is χ when obtaining the t frame tPosterior probability be:
p tt|Z t)=p t(Z tt)*∫p ttt-1)*p t-1t-1|Z t-1)dχ t-1/p t(Z t);
The posterior probability density p of target gesture during according to the t frame tt| Z t) the optimal profile state of target gesture when estimating the t frame
Figure BDA0000111086070000024
(5-3) when the t frame, indicate the coordinate A of forefinger finger tip and the coordinate B of thumb finger tip on the optimal profile of target gesture;
(6) mouse mappings:
Make up right-angle triangle ABC: cross the A point and do vertical line, cross the B point and do horizontal line, intersection point is C;
If line segment AB and CB angle are α, the angle of line segment BA and CA is β, and α is less than threshold value formula α TBe mapped as the left mouse button click action, β is less than threshold value beta TBe mapped as and click action by mouse right button;
The mid point M of line taking section AB is mapped as the relative coordinate of mouse; If the forefinger fingertip location is A in the previous frame image *, the coordinate B of thumb finger tip *, line segment A *B *Mid point be M *, then directed line segment M*M is mapped as the relatively move vector of mouse on screen.
Step (2) is said carries out off-line training to gesture, obtains the gesture feature sorter, is specially:
Use 600 open one's eyes wide the mark gesture as positive sample and 1200 non-target gestures as negative sample; The opencv_traincascade program that use is increased income to be provided among the Flame Image Process class libraries OpenCV is trained.
The said gesture feature sorter ferret out gesture in image that obtains through step (2) of step (4) is specially:
All subwindows of 30 * 30 of intercepting entire image, each subwindow are eliminated non-target gesture subwindow step by step successively through 20 grades of gesture feature sorters, confirm as the target gesture through the subwindow of all 20 grades of gesture feature sorters; If in this layer search, do not find the target gesture, then with subwindow with 1.3 times of amplifications, detect through the gesture feature sorter again.
In the said ferret out gesture of step (2) process, if the previous frame image has searched the target gesture, then the hunting zone with current frame image is reduced into the zone around the target gesture region in the previous frame image, is specially:
The left margin distance is d to target gesture region if the center of the target gesture region that searches in the previous frame image is for some O, some O 1, some O is d to the right margin distance 2, distance is d to some O to the coboundary 3, some O is d to the lower boundary distance 4, the rectangular search frame is made for being the center with an O in then new region of search, and some O is 2*d to rectangular search frame left margin distance 1, some O is 2*d to rectangular search frame right margin distance 2, some O is 2*d to rectangular search upper frame edge circle distance 3, some O is 2*d to rectangular search frame lower boundary distance 4
Step (5-2) is said to be observed the target gesture, specifically observes through adaptive skin color segmentation method; Said adaptive skin color segmentation method may further comprise the steps:
(5-2-1) set up TSL space complexion model;
(5-2-2) carry out colour of skin filtering;
(5-2-3) bianry image that obtains after the colour of skin filtering is carried out dilation operation.
The said TSL of foundation of step (5-2-1) space complexion model is specially:
Is the TSL color space through following formula with the RGB color space conversion:
T = 1 2 π tan - 1 ( r ′ g ′ ) + 0.5 S = 9 5 ( r ′ 2 + g ′ 2 ) L = 0.299 * R + 0.587 * G + 0.114 * B
Wherein r ′ = ( r - 1 3 ) , g ′ = ( g - 1 3 )
r = R R + G + B , g = G R + G + B
R, G, B are respectively the RGB component under the rgb color model; T, S, L are respectively the TSL component under the TSL colour model.
Step (5-2-2) is said carries out colour of skin filtering, is specially:
500 face and hand region that comprise the image of area of skin color are sampled the equal value matrix E and the covariance matrix ∑ of the two-dimentional Gaussian distribution probability distribution parameters of T and S under the estimation TSL model;
Each pixel is detected, if the C=that the T of a pixel and S component are formed (T, S) vector is lower than threshold value Threshold with the mahalanobis distance of mean vector E, thinks that then this pixel belongs to area of skin color; Said mahalanobis distance d=(C-E) T-1(C-E).
Said threshold value Threshold is confirmed by following process:
According to equal value matrix E and covariance matrix ∑ estimation C=(T, S) distance of vector and mean vector E obtains initial value;
Calculate the degree of confidence confidence of each threshold value Threshold according to following formula:
Confidence = PosSkin MaskArea * 2 - NegSkin BgArea
Wherein PosSkin is meant the quantity of the pixel of the colour of skin in the colour of skin template zone, and MaskArea is meant the total area in colour of skin template zone, and NegSkin is meant the quantity of skin pixel point in the background template zone, and BgArea is meant the area in background template zone;
The threshold value Threshold that degree of confidence is maximum is updated to the TSL complexion model parameter of current the best, and upgrades TSL space complexion model.
Compared with prior art, the present invention has the following advantages and technique effect:
1, the present invention is a kind of natural, man-machine interaction mode intuitively, and the user need not carry other utility appliance, accomplishes mouse action through the hand and the finger motion of nature;
2, particle filter tracking model under comprehensive LBP characteristic of the present invention and the ASM carries out gesture zone location and profile and follows the tracks of, and receives illumination and background influence little;
3, the present invention has overcome conventional mouse the user has been operated the restriction of degree of freedom, can make the user before camera in the 2m middle distance computing machine is controlled;
4, the present invention can be when the multi-user operation computing machine, free blocked operation power.
Description of drawings
Fig. 1 is the process flow diagram that the present invention is based on the gesture mouse recognition methods of machine vision.
Fig. 2 is the synoptic diagram of the open gesture of the five fingers.
Fig. 3 is the synoptic diagram of target gesture.
Fig. 4 is the synoptic diagram in the ferret out gesture process.
Fig. 5 is provided with synoptic diagram for the template in the complexion model.
Fig. 6 is the synoptic diagram that makes up right-angle triangle ABC in the mouse mappings process.
Fig. 7 clicks synoptic diagram for left mouse button.
Fig. 8 is for clicking synoptic diagram by mouse right button.
The synoptic diagram that Fig. 9 moves for mouse.
Embodiment
Below in conjunction with embodiment and accompanying drawing, the present invention is done to specify further, but embodiment of the present invention is not limited thereto.
Embodiment
Gesture mouse of the present invention recognition methods is based on the gesture mouse system; The gesture mouse system of present embodiment is made up of image capture module, image processing module and mouse event respond module; Image capture module comprises camera, is responsible for gathering user's image in real time and being transferred in the image processing module.
Image processing module and mouse event respond module are accomplished by computing machine, and wherein image processing module moves various image processing algorithm real-time analysis user images, are the control corresponding instruction with the movement conversion of user's hand and finger.The mouse event respond module is accepted the steering order that image processing module sends, and converts corresponding instruction into the mouse response events.
As shown in Figure 1, the gesture mouse recognition methods based on machine of present embodiment may further comprise the steps:
(1) sets up the moving shape model of gesture: at first the profile of the open gesture of the five fingers (as shown in Figure 2) is carried out the basic gesture modeling of B batten; On the basis of basic gesture model, set up the moving shape model of the gesture of B batten then, shown in the following formula of its state space:
χ=(x,y,θ,s,λ,θ 1,l 1,θ 2,l 2,θ 3,l 3,θ 4,l 4,θ 5,θ 6)
Wherein, x, y are respectively the horizontal ordinate and the ordinate of the position of gesture; θ representes the gesture anglec of rotation planar; S represents the size of gesture; θ 1, l 1Represent angle and the little finger length of little finger respectively around its root rotation; θ 2, l 2The expression third finger is around the angle and the nameless length of its root rotation respectively; θ 3, l 3Represent angle and the middle finger length of middle finger respectively around its root rotation; θ 4, l 4Represent angle and the forefinger length of forefinger respectively around its root rotation; θ 5, θ 6The expression thumb is around the angle of its root rotation; θ 6The expression thumb is around the angle of its root rotation; θ 5The expression thumb is around the angle of joint, center rotation;
(2) off-line training is carried out in the staff zone, obtains the gesture feature sorter: use 600 open one's eyes wide the mark gesture as positive sample and with 1200 non-target gestures as negative sample; The opencv traincascade program that use is increased income to be provided among the Flame Image Process class libraries OpenCV is trained.
(3) images acquired;
(4) partial binary mode characteristic of extraction image is searched for target gesture as shown in Figure 3 through the gesture feature sorter that step (2) obtains in image; When searching the target gesture, carry out step (5);
The said gesture feature sorter ferret out gesture in image that obtains through step (2) is specially:
All subwindows of 30 * 30 of intercepting entire image; Each subwindow is successively through 20 grades of gesture feature sorters; Eliminate non-target gesture subwindow step by step, have only through the subwindow of all 20 grades of gesture feature sorters and just confirm as the target gesture, if in this layer search, do not find the target gesture; Then with subwindow with 1.3 times of amplifications, detect through the gesture feature sorter again.
In search procedure, in order to improve search efficiency, if the previous frame image has searched the target gesture, then the hunting zone with current frame image is reduced into the zone around the target gesture region in the previous frame image.As shown in Figure 4, suppose that the center, target gesture region (being the inside casing among Fig. 4) that in the previous frame image, searches is some O, some O is d to the left margin distance of target gesture region 1, some O is d to the right margin distance 2, distance is d to some O to the coboundary 3, some O is at d to the lower boundary distance 4, rectangular search frame (being the housing among Fig. 4) is made for being the center with an O in then new region of search, and mid point O is 2*d to rectangular search frame left margin distance 1, O is 2*d to rectangular search frame right margin distance 2, O is 2*d to rectangular search upper frame edge circle distance 3, O is 2*d to rectangular search frame lower boundary distance 4
(5) finger tip location:
(5-1) adopt the moving shape model of the gesture in the step (1) that the target gesture that searches is estimated match, obtain the state space initial value χ of target gesture 0, make the initial value p of posterior probability density 00| Z 0)=1;
(5-2) the target gesture is observed, utilized the particle filter tracking algorithm that the state of target gesture is carried out the iteration renewal through adaptive skin color segmentation method;
Said adaptive skin color segmentation method may further comprise the steps:
(5-2-1) set up TSL space complexion model:
Be the TSL color space with the RGB color space conversion at first through following formula:
T = 1 2 π tan - 1 ( r ′ g ′ ) + 0.5 S = 9 5 ( r ′ 2 + g ′ 2 ) L = 0.299 * R + 0.587 * G + 0.114 * B
Wherein r ′ = ( r - 1 3 ) , g ′ = ( g - 1 3 )
r = R R + G + B , g = G R + G + B
R, G, B are respectively the RGB component under the rgb color model; T, S, L are respectively the TSL component under the TSL colour model;
(5-2-2) carry out colour of skin filtering: face and hand region through comprising the image of area of skin color to 500 are sampled, the equal value matrix E and the covariance matrix ∑ of the two-dimentional Gaussian distribution probability distribution parameters of T and S under the estimation TSL model;
Each pixel is detected, if the C=that the T of a pixel and S component are formed (T, S) vector is lower than threshold value Threshold with the mahalanobis distance of mean vector E, thinks that then this pixel belongs to area of skin color; Said mahalanobis distance d=(C-E) T-1(C-E);
Said threshold value Threshold is confirmed by following process:
According to equal value matrix E and covariance matrix ∑ estimation C=(T, S) distance of vector and mean vector E obtains initial value;
Calculate the degree of confidence confidence of each threshold value Threshold according to following formula:
Confidence = PosSkin MaskArea * 2 - NegSkin BgArea
Template in the complexion model is provided with as shown in Figure 5.PosSkin is meant the quantity of the pixel of the colour of skin in the colour of skin template zone (being gesture zone among Fig. 5) in the following formula; MaskArea is meant the total area in colour of skin template zone; NegSkin is meant the quantity of skin pixel point in the background template zone (being the black background among Fig. 5), and BgArea is meant the area in background template zone.
Getting the maximum threshold value Threshold of degree of confidence is the TSL complexion model parameter of current the best, and upgrades TSL space complexion model;
(5-2-3) bianry image that obtains after the colour of skin filtering is carried out dilation operation, to reduce the cavity that colour of skin filtering causes.
Said iteration renewal process is specific as follows:
The observed quantity that defines the t-1 two field picture is Z T-1, then the state space of target gesture is χ during the t-1 frame T-1Posterior probability be p T-1T-1| Z T-1);
The optimal profile state of target gesture during according to the t-1 frame
Figure BDA0000111086070000081
The estimation probability of the optimal profile state of target gesture is during to the t frame: p tt| χ T-1); Order
Figure BDA0000111086070000082
The likelihood probability of the profile state of target gesture is p during the t frame t(Z t| χ t);
The state space of target gesture is χ when obtaining the t frame tPosterior probability be:
p tt|Z t)=p t(Z tt)*∫p ttt-1)*p t-1t-1|Z t-1)dχ t-1/p t(Z t);
The posterior probability density p of target gesture during according to the t frame tt| Z t) the optimal profile state of target gesture does when estimating the t frame
Figure BDA0000111086070000083
(5-3) when the t frame, indicate the coordinate A of forefinger finger tip and the coordinate B of thumb finger tip on the optimal profile of target gesture;
(6) mouse mappings:
As shown in Figure 6, make up right-angle triangle ABC: cross the A point and do vertical line, cross the B point and do horizontal line, intersection point is C, and establishing line segment AB and CB angle is α, and the angle of line segment BA and CA is β, and α is less than threshold alpha TBe mapped as the left mouse button click action, Fig. 7 clicks synoptic diagram for left mouse button.β is less than threshold value beta TBe mapped as and click action by mouse right button, Fig. 8 is for clicking synoptic diagram by mouse right button.
The mid point M of line taking section AB is mapped as the relative coordinate of mouse, establishes that the forefinger fingertip location is A in the previous frame image *, the coordinate B of thumb finger tip *, line segment A *B *Mid point be M *, then directed line segment M*M is mapped as the relatively move vector of mouse on screen, the synoptic diagram that Fig. 9 moves for mouse.
The foregoing description is a preferred implementation of the present invention; But embodiment of the present invention is not limited by the examples; Other any do not deviate from change, the modification done under spirit of the present invention and the principle, substitutes, combination, simplify; All should be the substitute mode of equivalence, be included within protection scope of the present invention.

Claims (8)

1. the gesture mouse recognition methods based on machine vision is characterized in that, may further comprise the steps:
(1) sets up the moving shape model of gesture: at first the profile of the open gesture of the five fingers is carried out the basic gesture modeling of B batten; On the basis of basic gesture model, set up the moving shape model of the gesture of B batten then, shown in the following formula of its state space:
χ=(x,y,θ,s,λ,θ 1,l 1,θ 2,l 2,θ 3,l 3,θ 4,l 4,θ 5,θ 6)
Wherein, x, y are respectively the horizontal ordinate and the ordinate of the position of gesture; θ representes the gesture anglec of rotation planar; S represents the size of gesture; θ 1, l 1Represent angle and the little finger length of little finger respectively around its root rotation; θ 2, l 2The expression third finger is around the angle and the nameless length of its root rotation respectively; θ 3, l 3Represent angle and the middle finger length of middle finger respectively around its root rotation; θ 4, l 4Represent angle and the forefinger length of forefinger respectively around its root rotation; θ 6The expression thumb is around the angle of its root rotation; θ 5The expression thumb is around the angle of joint, center rotation;
(2) gesture is carried out off-line training, obtain the gesture feature sorter;
(3) images acquired;
(4) extract the partial binary mode characteristic of image, the gesture feature sorter ferret out gesture in image that obtains through step (2); When searching the target gesture, carry out step (5);
(5) finger tip location:
(5-1) adopt the moving shape model of the gesture in the step (1) that the target gesture that searches is estimated match, obtain the state space initial value χ of target gesture 0, make the initial value p of posterior probability density 00| Z 0)=1;
(5-2) the target gesture is observed, utilized the particle filter tracking algorithm that the state of target gesture is carried out the iteration renewal, said iteration renewal process is specific as follows:
The observed quantity that defines the t-1 two field picture is Z T-1, then the state space of target gesture is χ during the t-1 frame T-1Posterior probability be p T-1T-1| Z T-1);
The optimal profile state of target gesture during according to the t-1 frame
Figure FDA0000111086060000011
The optimal profile state of target gesture during to the t frame
Figure FDA0000111086060000012
The estimation probability be: p tt| χ T-1); Order
Figure FDA0000111086060000013
The likelihood probability of the profile state of target gesture is p during the t frame t(Z t| χ t);
The state space of target gesture is χ during the t frame tPosterior probability be:
p tt|Z t)=p t(Z tt)*∫p ttt-1)*p t-1t-1|Z t-1)dχ t-1/p t(Z t);
The posterior probability density p of target gesture during according to the t frame tt| Z t) the optimal profile state of target gesture does when estimating the t frame
(5-3) when the t frame, indicate the coordinate A of forefinger finger tip and the coordinate B of thumb finger tip on the optimal profile of target gesture;
(6) mouse mappings:
Make up right-angle triangle ABC: cross the A point and do vertical line, cross the B point and do horizontal line, intersection point is C;
If line segment AB and CB angle are α, the angle of line segment BA and CA is β, and α is less than threshold alpha TBe mapped as the left mouse button click action, β is less than threshold value beta TBe mapped as and click action by mouse right button;
The mid point M of line taking section AB is mapped as the relative coordinate of mouse; If the forefinger fingertip location is A in the previous frame image *, the coordinate B of thumb finger tip *, line segment A *B *Mid point be M *, then directed line segment M*M is mapped as the relatively move vector of mouse on screen.
2. the gesture mouse recognition methods based on machine vision according to claim 1 is characterized in that, step (2) is said carries out off-line training to gesture, obtains the gesture feature sorter, is specially:
Use 600 open one's eyes wide the mark gesture as positive sample and 1200 non-target gestures as negative sample; The opencv_traincascade program that use is increased income to be provided among the Flame Image Process class libraries OpenCV is trained.
3. the gesture mouse recognition methods based on machine vision according to claim 1 is characterized in that, the said gesture feature sorter ferret out gesture in image that obtains through step (2) of step (4) is specially:
All subwindows of 30 * 30 of intercepting entire image, each subwindow are eliminated non-target gesture subwindow step by step successively through 20 grades of gesture feature sorters, confirm as the target gesture through the subwindow of all 20 grades of gesture feature sorters; If in this layer search, do not find the target gesture, then with subwindow with 1.3 times of amplifications, detect through the gesture feature sorter again.
4. the gesture mouse recognition methods based on machine vision according to claim 1 and 2; It is characterized in that; In the said ferret out gesture of step (2) process; If the previous frame image has searched the target gesture, then the hunting zone with current frame image is reduced into the zone around the target gesture region in the previous frame image, is specially:
The left margin distance is d to target gesture region if the center of the target gesture region that searches in the previous frame image is for some O, some O 1, some O is d to the right margin distance 2, distance is d to some O to the coboundary 3, some O is d to the lower boundary distance 4, the rectangular search frame is made for being the center with an O in then new region of search, and some O is 2*d to rectangular search frame left margin distance 1, some O is 2*d to rectangular search frame right margin distance 2, some O is 2*d to rectangular search upper frame edge circle distance 3, some O is 2*d to rectangular search frame lower boundary distance 4
5. the gesture mouse recognition methods based on machine vision according to claim 1 is characterized in that step (5-2) is said to be observed the target gesture, specifically observes through adaptive skin color segmentation method; Said adaptive skin color segmentation method may further comprise the steps:
(5-2-1) set up TSL space complexion model;
(5-2-2) carry out colour of skin filtering;
(5-2-3) bianry image that obtains after the colour of skin filtering is carried out dilation operation.
6. the gesture mouse recognition methods based on machine vision according to claim 5 is characterized in that, the said TSL of foundation of step (5-2-1) space complexion model is specially:
Is the TSL color space through following formula with the RGB color space conversion:
T = 1 2 π tan - 1 ( r ′ g ′ ) + 0.5 S = 9 5 ( r ′ 2 + g ′ 2 ) L = 0.299 * R + 0.587 * G + 0.114 * B
Wherein r ′ = ( r - 1 3 ) , g ′ = ( g - 1 3 )
r = R R + G + B , g = G R + G + B
R, G, B are respectively the RGB component under the rgb color model; T, S, L are respectively the TSL component under the TSL colour model.
7. the gesture mouse recognition methods based on machine vision according to claim 6 is characterized in that, step (5-2-2) is said carries out colour of skin filtering, is specially:
500 face and hand region that comprise the image of area of skin color are sampled the equal value matrix E and the covariance matrix ∑ of the two-dimentional Gaussian distribution probability distribution parameters of T and S under the estimation TSL model;
Each pixel is detected, if the C=that the T of a pixel and S component are formed (T, S) vector is lower than threshold value Threshold with the mahalanobis distance of mean vector E, thinks that then this pixel belongs to area of skin color; Said mahalanobis distance d=(C-E) T-1(C-E).
8. the gesture mouse recognition methods based on machine vision according to claim 7 is characterized in that, said threshold value Threshold is confirmed by following process:
According to equal value matrix E and covariance matrix ∑ estimation C=(T, S) distance of vector and mean vector E obtains initial value;
Calculate the degree of confidence confidence of each threshold value Threshold according to following formula:
Confidence = PosSkin MaskArea * 2 - NegSkin BgArea
Wherein PosSkin is meant the quantity of the pixel of the colour of skin in the colour of skin template zone, and MaskArea is meant the total area in colour of skin template zone, and NegSkin is meant the quantity of skin pixel point in the background template zone, and BgArea is meant the area in background template zone;
The threshold value Threshold that degree of confidence is maximum is updated to the TSL complexion model parameter of current the best, and upgrades TSL space complexion model.
CN201110374316.3A 2011-11-22 2011-11-22 Mouse recognition method for gesture based on machine vision Expired - Fee Related CN102402289B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201110374316.3A CN102402289B (en) 2011-11-22 2011-11-22 Mouse recognition method for gesture based on machine vision

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201110374316.3A CN102402289B (en) 2011-11-22 2011-11-22 Mouse recognition method for gesture based on machine vision

Publications (2)

Publication Number Publication Date
CN102402289A true CN102402289A (en) 2012-04-04
CN102402289B CN102402289B (en) 2014-09-10

Family

ID=45884573

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201110374316.3A Expired - Fee Related CN102402289B (en) 2011-11-22 2011-11-22 Mouse recognition method for gesture based on machine vision

Country Status (1)

Country Link
CN (1) CN102402289B (en)

Cited By (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102662471A (en) * 2012-04-09 2012-09-12 沈阳航空航天大学 Computer vision mouse
CN103809738A (en) * 2012-11-13 2014-05-21 联想(北京)有限公司 Information acquisition method and electronic device
CN103914149A (en) * 2014-04-01 2014-07-09 复旦大学 Gesture interaction method and gesture interaction system for interactive television
CN105261038A (en) * 2015-09-30 2016-01-20 华南理工大学 Bidirectional optical flow and perceptual hash based fingertip tracking method
CN106406518A (en) * 2016-08-26 2017-02-15 清华大学 Gesture control device and gesture recognition method
CN106778670A (en) * 2016-12-30 2017-05-31 上海集成电路研发中心有限公司 Gesture identifying device and recognition methods
CN107085438A (en) * 2017-04-28 2017-08-22 中国船舶重工集团公司第七0九研究所 Unmanned plane path modification method and system based on accurate uniform SPL
CN107390867A (en) * 2017-07-12 2017-11-24 武汉大学 A kind of man-machine interactive system based on Android wrist-watch
CN107608510A (en) * 2017-09-13 2018-01-19 华中师范大学 Method for building up, device and the electronic equipment in gesture model storehouse
CN109101872A (en) * 2018-06-20 2018-12-28 济南大学 A kind of generation method of 3D gesture mouse
CN109300351A (en) * 2017-07-25 2019-02-01 西门子保健有限责任公司 Tool is associated with gesture is picked up
CN110567441A (en) * 2019-07-29 2019-12-13 广东星舆科技有限公司 Particle filter-based positioning method, positioning device, mapping and positioning method
CN115578627A (en) * 2022-09-21 2023-01-06 凌度(广东)智能科技发展有限公司 Monocular image boundary identification method and device, medium and curtain wall robot

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040001113A1 (en) * 2002-06-28 2004-01-01 John Zipperer Method and apparatus for spline-based trajectory classification, gesture detection and localization
CN101344816A (en) * 2008-08-15 2009-01-14 华南理工大学 Human-machine interaction method and device based on sight tracing and gesture discriminating
CN101763515A (en) * 2009-09-23 2010-06-30 中国科学院自动化研究所 Real-time gesture interaction method based on computer vision
CN101901052A (en) * 2010-05-24 2010-12-01 华南理工大学 Target control method based on mutual reference of both hands

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040001113A1 (en) * 2002-06-28 2004-01-01 John Zipperer Method and apparatus for spline-based trajectory classification, gesture detection and localization
CN101344816A (en) * 2008-08-15 2009-01-14 华南理工大学 Human-machine interaction method and device based on sight tracing and gesture discriminating
CN101763515A (en) * 2009-09-23 2010-06-30 中国科学院自动化研究所 Real-time gesture interaction method based on computer vision
CN101901052A (en) * 2010-05-24 2010-12-01 华南理工大学 Target control method based on mutual reference of both hands

Cited By (23)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102662471B (en) * 2012-04-09 2015-02-18 沈阳航空航天大学 Computer vision mouse
CN102662471A (en) * 2012-04-09 2012-09-12 沈阳航空航天大学 Computer vision mouse
CN103809738A (en) * 2012-11-13 2014-05-21 联想(北京)有限公司 Information acquisition method and electronic device
CN103809738B (en) * 2012-11-13 2017-03-29 联想(北京)有限公司 A kind of information collecting method and electronic equipment
CN103914149A (en) * 2014-04-01 2014-07-09 复旦大学 Gesture interaction method and gesture interaction system for interactive television
CN103914149B (en) * 2014-04-01 2017-02-08 复旦大学 Gesture interaction method and gesture interaction system for interactive television
CN105261038B (en) * 2015-09-30 2018-02-27 华南理工大学 Finger tip tracking based on two-way light stream and perception Hash
CN105261038A (en) * 2015-09-30 2016-01-20 华南理工大学 Bidirectional optical flow and perceptual hash based fingertip tracking method
CN106406518A (en) * 2016-08-26 2017-02-15 清华大学 Gesture control device and gesture recognition method
CN106406518B (en) * 2016-08-26 2019-01-18 清华大学 Gesture control device and gesture identification method
CN106778670A (en) * 2016-12-30 2017-05-31 上海集成电路研发中心有限公司 Gesture identifying device and recognition methods
CN107085438A (en) * 2017-04-28 2017-08-22 中国船舶重工集团公司第七0九研究所 Unmanned plane path modification method and system based on accurate uniform SPL
CN107390867A (en) * 2017-07-12 2017-11-24 武汉大学 A kind of man-machine interactive system based on Android wrist-watch
CN107390867B (en) * 2017-07-12 2019-12-10 武汉大学 Man-machine interaction system based on android watch
CN109300351A (en) * 2017-07-25 2019-02-01 西门子保健有限责任公司 Tool is associated with gesture is picked up
US10802597B2 (en) 2017-07-25 2020-10-13 Siemens Healthcare Gmbh Assigning a tool to a pick-up gesture
CN107608510A (en) * 2017-09-13 2018-01-19 华中师范大学 Method for building up, device and the electronic equipment in gesture model storehouse
CN109101872A (en) * 2018-06-20 2018-12-28 济南大学 A kind of generation method of 3D gesture mouse
CN109101872B (en) * 2018-06-20 2023-04-18 济南大学 Method for generating 3D gesture mouse
CN110567441A (en) * 2019-07-29 2019-12-13 广东星舆科技有限公司 Particle filter-based positioning method, positioning device, mapping and positioning method
CN110567441B (en) * 2019-07-29 2021-09-28 广东星舆科技有限公司 Particle filter-based positioning method, positioning device, mapping and positioning method
CN115578627A (en) * 2022-09-21 2023-01-06 凌度(广东)智能科技发展有限公司 Monocular image boundary identification method and device, medium and curtain wall robot
CN115578627B (en) * 2022-09-21 2023-05-09 凌度(广东)智能科技发展有限公司 Monocular image boundary recognition method, monocular image boundary recognition device, medium and curtain wall robot

Also Published As

Publication number Publication date
CN102402289B (en) 2014-09-10

Similar Documents

Publication Publication Date Title
CN102402289B (en) Mouse recognition method for gesture based on machine vision
Mukherjee et al. Fingertip detection and tracking for recognition of air-writing in videos
CN102831404B (en) Gesture detecting method and system
CN105739702B (en) Multi-pose finger tip tracking for natural human-computer interaction
CN103257713B (en) A kind of gesture control method
CN104407694A (en) Man-machine interaction method and device combining human face and gesture control
RU2013154102A (en) FINGER RECOGNITION AND TRACKING SYSTEM
CN102981742A (en) Gesture interaction system based on computer visions
CN103226388A (en) Kinect-based handwriting method
CN107357414B (en) Click action recognition method and device
Wu et al. Vision-based fingertip tracking utilizing curvature points clustering and hash model representation
Vivek Veeriah et al. Robust hand gesture recognition algorithm for simple mouse control
Ranawat et al. Hand gesture recognition based virtual mouse events
CN109189219A (en) The implementation method of contactless virtual mouse based on gesture identification
CN108614988A (en) A kind of motion gesture automatic recognition system under complex background
Yousefi et al. 3D gesture-based interaction for immersive experience in mobile VR
Elakkiya et al. Intelligent system for human computer interface using hand gesture recognition
Czupryna et al. Real-time vision pointer interface
CN108108648A (en) A kind of new gesture recognition system device and method
ul Haq et al. New hand gesture recognition method for mouse operations
CN103400118B (en) The gestural control method that a kind of mapping relations are adaptively adjusted
Thomas et al. A comprehensive review on vision based hand gesture recognition technology
Shaker et al. Real-time finger tracking for interaction
Aggarwal et al. An Approach to Control the PC with Hand Gesture Recognition using Computer Vision Technique
Dhamanskar et al. Human computer interaction using hand gestures and voice

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20140910

Termination date: 20211122