CN107450715A - A kind of man-machine interaction multifunctional wrist strap terminal based on gesture identification - Google Patents

A kind of man-machine interaction multifunctional wrist strap terminal based on gesture identification Download PDF

Info

Publication number
CN107450715A
CN107450715A CN201610369667.8A CN201610369667A CN107450715A CN 107450715 A CN107450715 A CN 107450715A CN 201610369667 A CN201610369667 A CN 201610369667A CN 107450715 A CN107450715 A CN 107450715A
Authority
CN
China
Prior art keywords
gesture
man
wrist strap
matching
strap terminal
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201610369667.8A
Other languages
Chinese (zh)
Inventor
宋立新
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Datang Telecommunication Science & Technology Co Ltd
Datang Telecom Technology Co Ltd
Original Assignee
Datang Telecommunication Science & Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Datang Telecommunication Science & Technology Co Ltd filed Critical Datang Telecommunication Science & Technology Co Ltd
Priority to CN201610369667.8A priority Critical patent/CN107450715A/en
Publication of CN107450715A publication Critical patent/CN107450715A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/017Gesture based interaction, e.g. based on a set of recognized hand gestures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/20Movements or behaviour, e.g. gesture recognition
    • G06V40/28Recognition of hand or arm movements, e.g. recognition of deaf sign language

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Social Psychology (AREA)
  • Psychiatry (AREA)
  • Multimedia (AREA)
  • Health & Medical Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

The present invention relates to the field of human-computer interaction of wearable design, is based particularly on the man-machine interaction multifunctional wrist strap terminal of gesture identification, the wrist strap terminal is including dressing man-machine interactive system and gesture recognition process device.Wearing man-machine interactive system includes monocular cam and wearable operating system, and it is responsible for the collection of video flowing of the collection comprising gesture information and transmitting video-frequency flow data to gesture recognition processing module;Gesture recognition process device includes Hand Gesture Segmentation module and gesture classification and matching module, and it carries out Hand Gesture Segmentation, gesture analysis, gesture classification and matching to the gesture information in video flowing, finally obtains recognition result, returns to system and instruct accordingly.The implementation of the invention efficiently utilizes existing shooting collecting device in wrist strap terminal, without extra hardware spending, using efficient Gesture Recognition Algorithm, improves Consumer's Experience, the defects of overcoming the interaction design inconvenience such as wearable device screen is small.

Description

A kind of man-machine interaction multifunctional wrist strap terminal based on gesture identification
Technical field
The present invention relates to wearable technology field, particularly a kind of multifunctional wrist strap terminal based on gesture identification.
Background technology
Wearable device is the development trend of Intelligent hardware of future generation, and wearable device has certain computing function, can The existing Intelligent mobile equipment such as mobile phone and flat board is connected, the product of main flow includes the bracelet using wrist as support and wrist-watch system Row, and the eyes series using head as support, wearable device just occurs in the form of more enriching, but development is most at present The equipment of main flow be still wrist strap series product, according to the statistics of current major hardware vendor, wrist product be it is wearable go out Goods amount highest.The research contents of wearable device includes, I/O Interface, the calculating of continuous sensing data, Activity recognition, environment Context-aware etc., or even the research also comprising material and energy technology.Wearable device relate in substantial amounts of correlative technology field Hold, be the content that multidisciplinary field intersects.
With the development of wearable device, wearable wrist product has endurance poor, it is impossible to independent use or work( Can not entirely, the shortcomings of interactive screen is small.The problems such as screen of wearable wrist, has had a strong impact on the usage experience of user, but Interaction aspect, if screen removed, Consumer's Experience can be very poor, it is impossible to means that product function can not be complete with product direct interaction Entirety is existing, has so had a strong impact on that the data exchange between user and wearable device produces obstacle, then wearable device is just User can not be served, if a wearable device does not have software service, smart machine cannot be referred to as.
In order to solve the problems, such as above section, it is necessary to using a kind of more reasonable and advanced man-machine interaction mode, based on wearing The human-computer interaction technology of vision is worn, also known as wearable vision interacts.It is achieved in that on wearable device outfit shooting is first-class Imaging device, capture static state in interaction scenarios, multidate information at any time by visually-perceptible technology, fully understanding that context believes On the basis of breath, the intention of user is fully understood, so as to complete nature, efficient, reliable interactive task.Using monocular vision Gesture identification method carry out operation and control to wearable device.
The content of the invention
The present invention is directed to the drawbacks described above of prior art, it is proposed that a kind of multi-functional wrist of man-machine interaction based on gesture identification Tape terminal.The system includes wearing man-machine interactive system and gesture recognition process device;Wherein, described wearing man-machine interaction system System, including monocular cam and wearable operating system, it is responsible for the video for including gesture video of terminal monocular cam collection Stream, and give video stream to gesture recognition process device, the analysis and identification of gesture are carried out, finally sends the finger of specific function Order;The gesture recognition process device, including Hand Gesture Segmentation module and gesture classification and matching module, according to described by transmission Video flowing is handled by specific algorithm, by the processing of each processing module in gesture recognition process device, finally The specific function representated by the gesture of user is identified, sending function command to wrist terminal carries out specific action response. Wherein, the wearing man-machine interactive system and the gesture recognition process device in wrist strap terminal by being mounted in wrist strap terminal On operating system carry out data interaction.
Further, it is described wearing man-machine interactive system include wrist strap terminal on monocular cam and wrist strap terminal in itself The wearable operating system carried;Wherein, monocular cam is responsible for the collection of video, is by operating by the video data of collection System passes to gesture recognition process module and carries out the operation such as segmentation, analysis, matching of gesture.Wherein, the segmentation of the gesture is adopted It is the algorithm that Adaptive Gauss model is combined with MLP, the process directly influences the accuracy and precision of identification.The hand Gesture recognition process unit includes the geometrical normalization of images of gestures, removes arm interference module, also gesture matching module.
Further, in described gesture recognition process module, the Shape context based on interior distance has been used Gesture matching algorithm, calculate the characteristic vector in gesture profile point first in the algorithm, then found according to the characteristics of hand Optimum sampling point is matched, and finds the principal direction of palm center, then carries out corresponding polar coordinate transform, by polar Direction rotated to and coincided with arm regions principal direction beginning, is classified using MLP algorithms, is finally reached identification rapidly and efficiently Gesture.
Brief description of the drawings
The present invention is further detailed explanation with reference to the accompanying drawings and detailed description.
Fig. 1 is the entire block diagram of the present invention.
In figure:Including wearing man-machine interactive system, gesture recognition process device.
Fig. 2 is the system process chart of the present invention.
In figure:Including Hand Gesture Segmentation module and gesture classification and matching module.
Fig. 3 is the flow chart of Hand Gesture Segmentation module.
Fig. 4 is MLP iterative process figures.
In figure:Using the EM algorithms of classics.
Fig. 5 is Gesture Recognition Algorithm flow chart.
In figure:Contain Euclidean transformation, the matching algorithm based on interior distance.
Embodiment
As shown in figure 1, the present invention is the man-machine interaction multifunctional wrist strap terminal based on gesture identification, wearing man-machine interaction system System includes monocular cam and wearable operating system, wherein wearing man-machine interactive system is responsible for collection regarding comprising gesture information Frequency flow data, and pass to gesture recognition process device progress gesture identification, gesture recognition process dress by dressing operating system Put including Hand Gesture Segmentation module and gesture classification and matching module, be responsible for the identifying processing of gesture.
Specific system handling process is as shown in Fig. 2 wherein the video comprising gesture is imaged by the monocular in interactive system Head collection.After camera collects video flowing, Hand Gesture Segmentation, Hand Gesture Segmentation such as Fig. 3 are carried out for each two field picture of video It is shown.The video of input is split according to fixed threshold value first, single Gauss model then is established to each cut zone, Then variance is reduced with iterative method, obtains the likelihood figure of regional output, the segmentation of threshold value is fixed again, with reference to fixation The model that Threshold segmentation, single Gauss model are combined with gauss hybrid models, and the Gauss model with adaptivity, are solved Can not linear partition the problem of.For hand region, using geometry plavini and connected region is removed;For background area Domain, influence of the background to recognition accuracy is weakened using geometry etch.
When the change of background is less than the threshold value set, iteration convergence, Hand Gesture Segmentation terminates;Conversely, carry out MLP instructions Practice, until result restrains.Hand Gesture Segmentation it is not convergent in the case of, carry out MLP repetitive exercise such as Fig. 4, for background change Data, estimate the initial value of Gauss model parameter first, variance reduced according to EM algorithms, while also reduces the region of segmentation, Meet the threshold value that originally set until variance is sufficiently small, otherwise just iteration always.
As shown in figure 5, after splitting by image, after the profile for obtaining image, by Euclidean transformation, wherein Range formula in Euclidean transformation is:
ED=min (Di)
Try to achieve centre of the palm position.Palm is assumed for ellipse from human morphology, and major and minor axis is known and then calculates oval Parameter, its formula are as follows:
Spalm=max (EDxy)
Lpaim=1.5max (EDxy)
Palm and arm are removed by calculating elliptic parameter, then use the algorithmic match gesture based on interior distance, it is interior The algorithm of distance mainly has two steps:Initially set up the connection figure of sampled point.For configuration sampling point set P={ pi, i=1,2, 3...n }, if pi、pjConnecting line falls in profile, then connecting line is drawn in connection figure.Pay attention to, sample connection figure here and may be used also To show out the hole information in profile;Then the algorithm of shortest path planning is used sampled point connection figure, finally obtains knowledge Other result.Influence of the mode of interior distance metric point distance for non-rigid object swivel of hand and part-structure deformation is smaller, more Suitable for gesture identification.If accelerating the process of matching using finger fingertip as sampled point can in addition, answering for algorithm is reduced Miscellaneous to spend the problem of high, finger fingertip can be obtained by following mathematical method, with OcCentered on calculate hand profile and OcAway from From curve f (x) is obtained, f (x) can be converted into by least square method by continuous function f (x)=a0+a1x+...anxn's Form, then tries to achieve the position of extreme point, and then identifies the position of finger tip.
Wherein ε=0.5Lpalm
Although the foregoing describing the embodiment of the present invention, those skilled in the art should be appreciated that this Be merely illustrative of, various changes or modifications can be made to present embodiment, without departing from the present invention principle and essence, Protection scope of the present invention is only limited by the claims that follow.

Claims (5)

  1. A kind of 1. man-machine interaction multifunctional wrist strap terminal based on gesture identification, it is characterised in that:The human-computer interaction terminal includes Wearing man-machine interactive system and gesture recognition process device.
  2. 2. according to the wearing man-machine interactive system of the multifunctional wrist strap terminal based on gesture identification described in claim 1, it is special Sign is:
    Including monocular cam and wearable operating system, the video flowing containing gesture information is carried out using monocular cam, selected Selecting most popular equipment reduces the manufacturing cost of hardware, while on the premise of basic acquisition hardware determines, carries specific Wearable operating system, it is responsible for sending the video information of collection to gesture recognition process device.
  3. 3. according to the gesture recognition process device of the multifunctional wrist strap terminal based on gesture identification described in claim 1, it is special Sign is:
    Including Hand Gesture Segmentation module and gesture classification and matching module;Wherein Hand Gesture Segmentation module uses the side based on region segmentation Method and MLP training methods, and by establishing single Gauss model to regional and then reducing error by iterative method;Wherein, hand Gesture classification includes the gesture classification model of the context based on shape with matching module and the matching algorithm based on interior distance is adopted The matching of gesture is carried out with the matching algorithm based on interior distance, finally gives matching result.
  4. 4. MLP training methods in Hand Gesture Segmentation module according to claim 3, it is characterised in that:
    By the parameter of EM algorithm iterations progressively improved model, the likelihood probability of parameter and training sample is set gradually to increase, finally A maximal point is terminated at, EM algorithms is employed and reduces variance, improve the accuracy of identification.
  5. 5. the gesture classification model of the context according to claim 3 based on shape, it is characterised in that:
    The gesture classification model of context employs oval palm model and carries out image modeling, including Euclidean transformation, Ask for centre of the palm position and calculate three steps of elliptic parameter, palm is reduced to ellipse, is obtained and slapped by Euclidean transformation Heart position, then palm area is positioned by calculating oval major and minor axis, then remove the matching and knowledge that arm regions carry out gesture Not.
CN201610369667.8A 2016-05-31 2016-05-31 A kind of man-machine interaction multifunctional wrist strap terminal based on gesture identification Pending CN107450715A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201610369667.8A CN107450715A (en) 2016-05-31 2016-05-31 A kind of man-machine interaction multifunctional wrist strap terminal based on gesture identification

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201610369667.8A CN107450715A (en) 2016-05-31 2016-05-31 A kind of man-machine interaction multifunctional wrist strap terminal based on gesture identification

Publications (1)

Publication Number Publication Date
CN107450715A true CN107450715A (en) 2017-12-08

Family

ID=60485646

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201610369667.8A Pending CN107450715A (en) 2016-05-31 2016-05-31 A kind of man-machine interaction multifunctional wrist strap terminal based on gesture identification

Country Status (1)

Country Link
CN (1) CN107450715A (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108814572A (en) * 2018-05-28 2018-11-16 Oppo广东移动通信有限公司 Wearing state detection method and relevant device
CN111338461A (en) * 2018-12-18 2020-06-26 鸿合科技股份有限公司 Gesture operation method and device and electronic equipment

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104268514A (en) * 2014-09-17 2015-01-07 西安交通大学 Gesture detection method based on multi-feature fusion
CN104298354A (en) * 2014-10-11 2015-01-21 河海大学 Man-machine interaction gesture recognition method
CN204191757U (en) * 2014-09-24 2015-03-11 南京航空航天大学 A kind of gesture catches bracelet
CN104503582A (en) * 2014-12-29 2015-04-08 北京智谷睿拓技术服务有限公司 Wearable-equipment-based interaction method and interaction device, and wearable equipment
CN105354551A (en) * 2015-11-03 2016-02-24 北京英梅吉科技有限公司 Gesture recognition method based on monocular camera
CN105487673A (en) * 2016-01-04 2016-04-13 京东方科技集团股份有限公司 Man-machine interactive system, method and device

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104268514A (en) * 2014-09-17 2015-01-07 西安交通大学 Gesture detection method based on multi-feature fusion
CN204191757U (en) * 2014-09-24 2015-03-11 南京航空航天大学 A kind of gesture catches bracelet
CN104298354A (en) * 2014-10-11 2015-01-21 河海大学 Man-machine interaction gesture recognition method
CN104503582A (en) * 2014-12-29 2015-04-08 北京智谷睿拓技术服务有限公司 Wearable-equipment-based interaction method and interaction device, and wearable equipment
CN105354551A (en) * 2015-11-03 2016-02-24 北京英梅吉科技有限公司 Gesture recognition method based on monocular camera
CN105487673A (en) * 2016-01-04 2016-04-13 京东方科技集团股份有限公司 Man-machine interactive system, method and device

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
陈泽宁: "基于穿戴视觉的人机交互技术", 《中国知网硕士论文库》 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108814572A (en) * 2018-05-28 2018-11-16 Oppo广东移动通信有限公司 Wearing state detection method and relevant device
CN111338461A (en) * 2018-12-18 2020-06-26 鸿合科技股份有限公司 Gesture operation method and device and electronic equipment

Similar Documents

Publication Publication Date Title
Wang et al. Joint activity recognition and indoor localization with WiFi fingerprints
CN107765855A (en) A kind of method and system based on gesture identification control machine people motion
CN103984928B (en) Finger gesture recognition methods based on depth image
CN107908288A (en) A kind of quick human motion recognition method towards human-computer interaction
Chen et al. Rapid recognition of dynamic hand gestures using leap motion
CN111985265A (en) Image processing method and device
Qi et al. Computer vision-based hand gesture recognition for human-robot interaction: a review
CN110471526A (en) A kind of human body attitude estimates the unmanned aerial vehicle (UAV) control method in conjunction with gesture identification
CN104407694A (en) Man-machine interaction method and device combining human face and gesture control
CN104777775A (en) Two-wheeled self-balancing robot control method based on Kinect device
CN109800676A (en) Gesture identification method and system based on depth information
US20220262093A1 (en) Object detection method and system, and non-transitory computer-readable medium
CN109032337A (en) A kind of KEM Gesture Recognition Algorithm based on data glove
CN111209812A (en) Target face picture extraction method and device and terminal equipment
WO2023061037A1 (en) Air gesture recognition method and related device thereof
CN106814853A (en) A kind of eye control tracking based on machine learning
CN112036261A (en) Gesture recognition method and device, storage medium and electronic device
CN110807391A (en) Human body posture instruction identification method for human-unmanned aerial vehicle interaction based on vision
CN110569775A (en) Method, system, storage medium and electronic device for recognizing human body posture
Xu et al. Robust hand gesture recognition based on RGB-D Data for natural human–computer interaction
CN110866468A (en) Gesture recognition system and method based on passive RFID
Hu et al. Trajectory image based dynamic gesture recognition with convolutional neural networks
CN107450715A (en) A kind of man-machine interaction multifunctional wrist strap terminal based on gesture identification
CN111695408A (en) Intelligent gesture information recognition system and method and information data processing terminal
Li et al. Hand gesture tracking and recognition based human-computer interaction system and its applications

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20171208

RJ01 Rejection of invention patent application after publication