CN110208946A - A kind of wearable device and the exchange method based on wearable device - Google Patents

A kind of wearable device and the exchange method based on wearable device Download PDF

Info

Publication number
CN110208946A
CN110208946A CN201910470231.1A CN201910470231A CN110208946A CN 110208946 A CN110208946 A CN 110208946A CN 201910470231 A CN201910470231 A CN 201910470231A CN 110208946 A CN110208946 A CN 110208946A
Authority
CN
China
Prior art keywords
braille
wearable device
processing units
image
control signal
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201910470231.1A
Other languages
Chinese (zh)
Inventor
邵继洋
毕育欣
孙剑
訾峰
丁亚东
刘炳鑫
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
BOE Technology Group Co Ltd
Beijing BOE Optoelectronics Technology Co Ltd
Original Assignee
BOE Technology Group Co Ltd
Beijing BOE Optoelectronics Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by BOE Technology Group Co Ltd, Beijing BOE Optoelectronics Technology Co Ltd filed Critical BOE Technology Group Co Ltd
Priority to CN201910470231.1A priority Critical patent/CN110208946A/en
Publication of CN110208946A publication Critical patent/CN110208946A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • G02B27/017Head mounted
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/20Scenes; Scene-specific elements in augmented reality scenes
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • G02B27/017Head mounted
    • G02B2027/0178Eyeglass type

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Optics & Photonics (AREA)
  • Multimedia (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

The embodiment of the invention discloses a kind of wearable device and based on the exchange method of wearable device, in the wearable device, the camera being arranged in lens body carries out Image Acquisition for the scene where user, first processing units obtain the recognition result that target identification is carried out to image, braille control signal is converted by recognition result, braille control signal is sent to the second processing unit in cuff assemblies, the second processing unit by control can in array of protrusions can protrusion unit whether protrusion, export the corresponding braille of braille control signal;In this way, user can also be obtained to the target identification of scene image by the braille of perception cuff assemblies output as a result, the recognition result can assist user preferably to perceive ambient enviroment even if user's eyesight is poor;As it can be seen that this programme plays the booster action of the people poor to eyesight.

Description

A kind of wearable device and the exchange method based on wearable device
Technical field
The present invention relates to technical field of electronic equipment, particularly relate to a kind of wearable device and the interaction side based on wearable device Method.
Background technique
In daily life, the poor people of eyesight cannot preferably perceive the environment of surrounding, for example, they cannot accurately identify Moving target in environment, such as vehicle, pedestrian, can not accurately identify the static target in environment, such as barrier, traffic lights Deng.In general, these people usually require other people and accompany, or need the help of seeing-eye dog.It can be auxiliary currently, needing one kind These people are helped to carry out the scheme of daily routines.
Summary of the invention
In view of this, it is an object of the invention to propose a kind of wearable device and based on the exchange method of wearable device, with Play the booster action of the people poor to eyesight.
Based on above-mentioned purpose, the present invention provides a kind of wearable devices, comprising: eyeglasses assembly and cuff assemblies;Wherein, institute Stating eyeglasses assembly includes lens body and the camera being set in the lens body and first processing units;The finger Set component include fingerstall main body and be set in the fingerstall main body can array of protrusions and the second processing unit, it is described can Array of protrusions include it is multiple can protrusion unit;
The camera is sent to the first processing units for acquiring image, and by described image;
The first processing units obtain the identification knot that target identification is carried out to described image for receiving described image Acquired recognition result is converted braille control signal by fruit, and braille control signal that conversion obtains is sent to described the Two processing units;
Described the second processing unit controls signal for receiving the braille that the first processing units are sent, passes through control It is described can in array of protrusions can protrusion unit whether protrusion, export the received corresponding braille of braille control signal.
In some embodiments, described the second processing unit is also used to determine the corresponding braille point of braille control signal Number is counted according to the braille, it is described can determine in array of protrusions it is to be processed can protrusion unit, it is described wait locate by controlling Reason can protrusion unit whether protrusion, export the received corresponding braille of braille control signal.
In some embodiments, the eyeglasses assembly further includes the sensor being set in the lens body;It is described The posture information is sent to the first processing units for obtaining the posture information of the eyeglasses assembly by sensor;
The first processing units are also used to calibrate described image based on the posture information, be obtained to calibration Image afterwards carries out the recognition result of target identification.
In some embodiments, the first processing units are also used to based on image after the calibration and described The parameter of camera calculates the distance between target and the camera, as the recognition result.
In some embodiments, the first processing units are also used to continuous more by acquiring to the camera Frame image carries out target following, the movement tendency of target is obtained, as the recognition result.
In some embodiments, the wearable device further includes microphone, the microphone, for acquiring voice number According to the voice data is sent to the first processing units;
The first processing units, are also used to by parsing the voice data, obtain it is instructions to be performed, based on it is described to It executes instruction and obtains the recognition result.
In some embodiments, the eyeglasses assembly is connect with communication of mobile terminal, and the first processing units are also used In obtaining the interactive information with the mobile terminal, braille control signal is converted by the interactive information.
In some embodiments, the first processing units, are also used in the mobile terminal that there are notification messages In the case where, it identifies the type of the notification message and initiates object, as the interactive information.
In some embodiments, the camera is also used to acquire the interface image of the mobile terminal, by the boundary Face image is sent to the first processing units;
The first processing units are also used to identify present in the mobile terminal by analyzing the interface image The type and initiation object of notification message, as the interactive information.
In some embodiments, the cuff assemblies further include the vibration unit being set in the fingerstall main body, institute The second processing unit is stated, is also used to drive the vibration unit to vibrate after receiving the braille control signal.
In some embodiments, the wearable device further includes key;The key, for opening or closing described wear Wear equipment;And/or the key, for answering or refusing the incoming call in the mobile terminal.
In some embodiments, the wearable device further includes voice-output unit;
The first processing units are also used to after obtaining the recognition result for carrying out target identification to described image, by institute It states recognition result and is sent to the voice-output unit;
The voice-output unit is also used to export the recognition result in the form of speech.
On the other hand, the present invention also provides a kind of exchange methods based on wearable device, are applied to wearable device, described Wearable device includes: eyeglasses assembly and cuff assemblies;Wherein, the eyeglasses assembly includes lens body and is set to described Camera and first processing units in lens body;The cuff assemblies include fingerstall main body and are set to the fingerstall In main body can array of protrusions and the second processing unit, it is described can array of protrusions include it is multiple can protrusion unit;The method packet It includes:
The camera collection image, and described image is sent to the first processing units;
The first processing units receive described image, obtain the recognition result that target identification is carried out to described image, will Acquired recognition result is converted into braille control signal, and the braille control signal that conversion obtains is sent to the second processing Unit;
Described the second processing unit receives the braille control signal that the first processing units are sent, by can described in control In array of protrusions can protrusion unit whether protrusion, export the received corresponding braille of braille control signal.
Using the embodiment of the present invention, the camera being arranged in lens body carries out image for the scene where user and adopts Collection, first processing units obtain the recognition result that target identification is carried out to image, convert braille control signal for recognition result, Braille control signal is sent to the second processing unit in cuff assemblies, the second processing unit can be in array of protrusions by control Can protrusion unit whether protrusion, export the corresponding braille of braille control signal;In this way, even if user's eyesight is poor, user It can also be obtained to the target identification of scene image by the braille of perception cuff assemblies output as a result, the recognition result can be with Auxiliary user preferably perceives ambient enviroment;As it can be seen that this programme plays the booster action of the people poor to eyesight.
Detailed description of the invention
In order to more clearly explain the embodiment of the invention or the technical proposal in the existing technology, to embodiment or will show below There is attached drawing needed in technical description to be briefly described, it should be apparent that, the accompanying drawings in the following description is only this Some embodiments of invention for those of ordinary skill in the art without creative efforts, can be with It obtains other drawings based on these drawings.
Fig. 1 is the first schematic diagram of wearable device provided in an embodiment of the present invention;
Fig. 2 is second of schematic diagram of wearable device provided in an embodiment of the present invention;
Fig. 3 be one of wearable device provided in an embodiment of the present invention can protrusion display schematic diagram;
Fig. 4 is the third schematic diagram of wearable device provided in an embodiment of the present invention;
Fig. 5 is a kind of flow diagram of the exchange method based on wearable device provided in an embodiment of the present invention.
Specific embodiment
To make the objectives, technical solutions, and advantages of the present invention clearer, below in conjunction with specific embodiment, and reference Attached drawing, the present invention is described in more detail.
It should be noted that all statements for using " first " and " second " are for differentiation two in the embodiment of the present invention The non-equal entity of a same names or non-equal parameter, it is seen that " first " " second " only for the convenience of statement, does not answer It is interpreted as the restriction to the embodiment of the present invention, subsequent embodiment no longer illustrates this one by one.
In order to solve the above-mentioned technical problem, the embodiment of the invention provides a kind of wearable device and based on the friendship of wearable device Mutual method, first below describes to wearable device in detail.
Fig. 1 is the first structural schematic diagram of wearable device provided in an embodiment of the present invention, comprising: 100 He of eyeglasses assembly Cuff assemblies 200;Wherein, eyeglasses assembly 100 includes lens body 110 and the camera being set in lens body 110 120 and first processing units 130;Cuff assemblies 200 are including fingerstall main body 210 and are set to dashing forward in fingerstall main body 210 Rise array 220 and the second processing unit 230, can array of protrusions 220 include it is multiple can protrusion unit 221.
For example, camera 120 can be set on the mirror holder between two eyeglasses of lens body 110, in this way may be used To collect the preferable image in visual angle.First processing units 130 can be set on the leg of spectacles of lens body 110, right in this way User's sight influences smaller.The position of camera 120 and first processing units 130 in lens body 110 can be according to reality Situation is set, specifically without limitation.
For example, first processing units 130 and the second processing unit 230 can for IC (integrated circuit, Integrated circuit), or may be other processing chips, such as MCU (Microcontroller Unit, micro-control unit) etc. Deng the concrete type of processing unit is without limitation.
For example, can array of protrusions can for it is dotted can array of protrusions, or column can array of protrusions, specific shape Shape is without limitation.Can include in array of protrusions it is multiple can protrusion unit, these can protrusion unit can export braille.Currently, depositing In the braille, such as 6 brailles, 12 brailles, 16 brailles, etc. of some different points.It, can array of protrusions in a kind of situation In include can the quantity of protrusion unit can be identical as braille points;In this case, can protrusion unit quantity can be 6 (as shown in fig. 1) can export 6 brailles in this way;Can the quantity of protrusion unit may be 12, can export in this way at 12 points Braille;Can protrusion unit quantity can also be 16,16 brailles can be exported in this way.
In another case, can include in array of protrusions can the quantity of protrusion unit can be greater than braille and count;For example, Can protrusion unit quantity can be 16,6 brailles, 12 brailles and 16 brailles can be exported in this way;It for another example, can protrusion The quantity of unit can be 12, can export 6 brailles and 12 brailles in this way.
It continues to refer to figure 1 and is illustrated, image is sent to first processing units for acquiring image by camera 120 130;
First processing units 130 obtain the recognition result for carrying out target identification to image, identification are tied for receiving image Fruit is converted into braille control signal, and the braille control signal that conversion obtains is sent to the second processing unit 230;
The second processing unit 230 is used to receive the braille control signal of first processing units transmission, can protrusion by control In array 220 can protrusion unit 221 whether protrusion, export the received corresponding braille of braille control signal.
Eyeglasses assembly 100 and cuff assemblies 200 communicate to connect, for example, eyeglasses assembly 100 and cuff assemblies 200 it Between can be communicated by wired or wireless mode, communication can be Bluetooth communication, or may be its He, specifically without limitation.In this way, the first processing units 130 in eyeglasses assembly 100 can be in cuff assemblies 200 Two processing units 230 carry out data transmission.
For example, during the poor user of eyesight uses wearable device provided in an embodiment of the present invention, glasses group Camera 120 in part 100 carries out Image Acquisition for scene where user.Camera can acquire in real time according to specified frame per second Scene image, and the scene image of acquisition is stored.In a kind of situation, first collected scene image can be cut It takes, for example, interception intermediate region, or the image-region of interception pre-set dimension, then truncated picture is stored.Camera The image of storage is sent to first processing units 130.
As an implementation, the image that camera 120 is sent can be sent to cloud clothes by first processing units 130 Business device, cloud server carries out target identification to the image, and recognition result is returned to first processing units 130.
As another embodiment, target knowledge is carried out to the image that camera 120 is sent by first processing units 130 Not, for example, the information such as the type of target, color, position can be identified." target " mentioned here can be the movement in environment Target, such as vehicle, pedestrian, or may be the static target in environment, such as barrier, traffic lights.
For example, which can be filtered, to the progress color space conversion of filtered image, such as from RGB (Red, Green, Blue, RGB) color space converts that (" Y " indicates brightness, " U " and " V " expression coloration, dense to YUV Degree) color space, then gray scale extraction can be carried out to the yuv data being converted to, obtained gradation data will be extracted and inputted In the identification model obtained to preparatory training, the identification model can export target identification as a result, such as target type, face The information such as color, position.
In a kind of situation, it may include: to obtain sample data to be trained, sample that training, which obtains the process of the identification model, Data can be image, which can be the image of given resolution or aspect ratio, and sample data may be to image It is filtered, the data that color space is converted, gray scale obtains after extracting.The sample data is input to the nerve net of preset structure In network, output neural network based is iterated adjustment to the parameter in neural network, until meeting iteration termination condition When, the neural network of trained completion is just obtained, that is, obtained the identification model.The neural network of the preset structure can be with For the neural network based on YOLO (You Only Look Once, you only see once) algorithm, or, or other are used for The neural network of target identification, specifically without limitation.
In a kind of embodiment, as shown in Fig. 2, eyeglasses assembly 100 can also include the biography being set in lens body 110 Sensor 140;The posture information is sent to the first processing for obtaining the posture information of eyeglasses assembly 100 by sensor 140 Unit 130;First processing units 130 are also used to based on the posture information, and the image sent to camera is calibrated, and is obtained Take the recognition result that target identification is carried out to the image after calibration.
In a kind of situation, the image after calibration can be sent to cloud server and carry out target by first processing units 130 Identification, and receive the recognition result of cloud server return.In another case, can also be by 130 pairs of schools of first processing units Image after standard carries out target identification, obtains recognition result.
For example, sensor 140 can be set in apart from the closer position of camera 120, in this way, being based on sensor The image that the posture information of 140 acquisitions acquires camera 120 is calibrated, and can achieve preferable calibration result.Above-mentioned one In the case of kind, camera is set on the mirror holder between two eyeglasses of glasses, and in this case, sensor 140 can also be set It is placed on the mirror holder between two eyeglasses of glasses.Position of the sensor 140 in lens body 110 can be according to the actual situation It is set, specifically without limitation.
Sensor 140 can be acquired for the attitude transducers such as gyroscope, sensor 140 eyeglasses assembly rotation angle, The posture informations such as pitch angle, in this way, first processing units 130 can the postures letter such as rotation angle, pitch angle based on eyeglasses assembly Breath calibrates the image of camera acquisition, obtains the recognition result for carrying out target identification to the image after calibration, obtain Recognition result is more acurrate.
As an implementation, first processing units 130, image and camera 120 after being also based on calibration Parameter, calculate the distance between target and camera 120, as the recognition result.
In a kind of situation, camera 120 can be depth camera, for example have the camera of depth transducer, depth It include depth information in the depth image of camera acquisition, first processing units 130 detect the target in depth image, according to this The target can be calculated at a distance from camera in the depth information of target and the relevant parameter of depth camera.
In another case, camera can be binocular camera, first processing units 130 are acquired according to binocular camera Binocular image and binocular camera internal reference and outer ginseng, target can also be calculated at a distance from camera.
In a kind of embodiment, first processing units 130 carry out mesh by the continuous multiple frames image acquired to camera 120 Mark tracking, obtains the movement tendency of target, as the recognition result.
For example, first processing units 130 can carry out target detection in N (N is positive integer) frame image, determine The region is split as multiple images block, as image block to be processed by region of the target in nth frame image;Then in N+1 It is determining with each highest image block of image block similarity to be processed respectively in frame image, as matching image block;According to wait locate The moving direction and moving distance between the corresponding matching image block of image block are managed, the movement of target in the picture is obtained and becomes Gesture;The transformational relation being then based between image coordinate system and physical space coordinates system, by the movement tendency of target in the picture Be converted to movement tendency of the target in physical space.
As another example, in above embodiment, first processing units 130 can calculate target and camera shooting based on every frame image The distance of head in a kind of situation, can use this embodiment, will be carried out based on every frame image calculated distance pair Than to judge that target is to obtain the movement tendency of target close to or far from camera.
First processing units 130 can convert recognition result obtained above to braille control signal, by the braille control Signal processed is sent to the second processing unit 230 in cuff assemblies 200.The second processing unit 230 receives first processing units 130 The braille of transmission controls signal, by control can in array of protrusions 220 can protrusion unit 221 whether protrusion, output received The corresponding braille of braille control signal arrived.That is, the second processing unit 230, which is based on braille control signal control, to dash forward It plays array 220 and exports braille corresponding with recognition result.
For example, recognition result is the information such as type, color, the position of target in a kind of above-mentioned embodiment, then it can protrusion The braille that array 220 exports can be the information such as type, color, position;In a kind of above-mentioned embodiment, recognition result is target At a distance from camera, then can array of protrusions 220 export braille can for target at a distance from camera;A kind of above-mentioned implementation In mode, recognition result is the movement tendency of target, then can the braille that exports of array of protrusions 220 can become for the movement of target Gesture.The second processing unit 230 can control can array of protrusions 220 according to prefixed time interval, be sequentially output each braille lettering.
It is currently, there are the braille of miscellaneous editions, the braille of miscellaneous editions, user can be stored in advance in first processing units The braille suitable for itself can be selected in the braille of the miscellaneous editions, the braille that first processing units 130 convert controls letter Number be user selection the corresponding control signal of braille.Braille control signal can be understood as to can array of protrusions 220 drive Dynamic signal, the driving signal can drive can in array of protrusions can protrusion unit show the corresponding waveform of braille.
In a kind of above-mentioned situation, can include in array of protrusions can the quantity of protrusion unit can be greater than braille and count, this In the case of kind, the second processing unit 230 is also used to determine the corresponding braille points of braille control signal, according to the braille point Number, it is described can determine in array of protrusions it is to be processed can protrusion unit, by control it is described it is to be processed can protrusion unit be No protrusion exports the received corresponding braille of braille control signal.
As described above, the corresponding braille points of braille control signal can be 6,12,16.With reference to Fig. 3, it is assumed that can protrusion battle array Include in column can protrusion unit quantity be 16, Fig. 3 in can protrusion unit layout not to dashing forward in the present embodiment It plays array 220 and constitutes restriction.Assuming that the second processing unit 230 determines that the corresponding braille points of braille control signal are 6, a kind of feelings Under condition, can in array of protrusions 220 upper left 6 can protrusion unit (frame 1) be determined as it is to be processed can protrusion unit; In the case of other, can in array of protrusions 220 6 of lower right (frame 2) or upper right side or lower left can dash forward Rise unit be determined as it is to be processed can protrusion unit;In the case of other, can also determine 6 it is non-conterminous can protrusion unit, As it is to be processed can protrusion unit, it is to be processed can protrusion unit will namely carry out braille displaying can protrusion unit, It is to be processed can protrusion unit specific location without limitation.
Except it is to be processed can in addition to protrusion unit other can protrusion unit can be understood as temporarily leaving unused can protrusion list Member, it is temporarily idle can protrusion unit can be fitted closely with fingerstall, make user be not felt by these it is idle can protrusion unit Protrusion, such user can more accurately experience the braille of output.
As an implementation, wearable device can also include microphone 300, and microphone 300 is for acquiring voice number According to the voice data being sent to first processing units 130, first processing units 130 are obtained by parsing the voice data To instructions to be performed, the recognition result of target identification is carried out to described image based on instructions to be performed obtain.
As shown in figure 4, microphone 300 can be set in lens body 110, for example, can be set in lens body 110 Temple on, the specific location of microphone is without limitation.For example, user can set when by traffic intersection to wearing Standby inquiry " being red light or green light now ";Microphone collects the voice data, which is sent to the first processing Unit 130;First processing units 130 identify the voice data by speech recognition algorithm, obtain " identification traffic lights color " Instruction.First processing units 130 detect " traffic lights " this target in the image that video camera 120 is sent, and then further know The color of this other target, it is assumed that recognition result is red, then generates red corresponding braille control signal, which is controlled Signal is sent to the second processing unit 230.
In a kind of situation, user can first say it is default wake up word, microphone collects default after waking up word, wearable device Working condition is switched to by standby mode.Alternatively, judging the corresponding vocal print of wake-up word after microphone collects default wake-up word Whether matched with user's vocal print pre-stored in wearable device, if it does, wearable device is switched to work by standby mode State.
In a kind of embodiment, wearable device can also include key 400, as shown in figure 4, key 400 can be set in In eyeglasses assembly 100, or in other embodiments, key 400 also be can be set in cuff assemblies 200, alternatively, can also be with Different keys is set in eyeglasses assembly 100 and cuff assemblies 200, and the specific location of key is without limitation.
In a kind of embodiment, key 400 can be used for opening or closing the wearable device.For example, key 400 It can be mechanical key, or may be fingerprint extracting device etc., specifically without limitation.User can pass through click keys 400, wearable device is switched to working condition by standby mode, or wearable device is switched to standby shape by working condition State.Alternatively, user can also reset wearable device by click keys 400.
In a kind of embodiment, as shown in figure 4, wearable device can also include voice-output unit 500, a kind of situation Under, voice-output unit 500 can be earphone, or may be speaker etc., specifically without limitation.In above content, wear It wears after equipment is switched to working condition by standby mode, voice-output unit 500 can play the response language to above-mentioned wake-up word Sound promotes user experience to prompt user's wearable device to enter working condition.User, can be with after hearing the response voice The voice data for carrying instruction is issued to wearable device.
In a kind of embodiment, first processing units 130 carry out the recognition result of target identification obtaining to described image Afterwards, the recognition result can also be sent to voice-output unit 500;The shape that voice-output unit 500 is also used to voice Formula exports the recognition result.
In a kind of above-mentioned embodiment, wearable device can also include key 400, and in this embodiment, key 400 can Think volume button, user passes through the volume that the adjustable voice-output unit 500 of key 400 exports.
In a kind of situation, camera 120, sensor 140, microphone 300, voice-output unit 500, key 400 and One processing unit 130, which can integrate, to be set on the mirror holder of lens body 110.Alternatively, in another case, microphone 300, language Sound output unit 500, key 400 and first processing units 130, which can integrate, to be set in the temple of lens body 110.
In a kind of embodiment, eyeglasses assembly 100 can be connect with communication of mobile terminal, and first processing units 130 may be used also To obtain the interactive information with the mobile terminal, braille control signal is converted by the interactive information.
In a kind of situation, first processing units 130 can be known in the mobile terminal there are in the case where notification message The type and initiation object of the not described notification message, as the interactive information.
For example, mobile terminal and wearable device can carry out pairing connection in advance, and then wearable device can be examined The notification message in mobile terminal is measured, the type of the notification message can be incoming information, short message or communication software Prompt information, etc., specifically without limitation.
For by taking incoming information as an example, first processing units 130, can after detecting that mobile terminal receives incoming information To identify the initiation object of the incoming information, that is, incoming call connection people, and incite somebody to action " type: incoming information initiates object: so-and-so People " is used as interactive information.First processing units 130 can convert the interactive information to braille control signal and be sent at second Unit 230 is managed, the control of the second processing unit 230 can the corresponding braille of the output braille control signal of array of protrusions 220.Alternatively, The interactive information can also be sent directly to voice-output unit 500 by first processing units 130, voice-output unit 500 with The form of voice exports the interactive information.
In a kind of above-mentioned embodiment, wearable device can also include key 400, and in this embodiment, key 400 can To answer or refuse the incoming call in mobile terminal.That is, user can by click keys 400, select answer or Refuse the incoming call in mobile terminal.For example, " answering " key different with " refusal " two can be set, in a kind of situation, " connect Listen " " answering " corresponding braille can be set on key, " refusal " corresponding braille can be set on " refusal " key, this Sample can be convenient user and select.
Alternatively, user can use wearable device operation mobile terminal in the case of other.For example, user can be with Using the social software in wearable device operation mobile terminal, braille that wearable device is exported by the second processing unit 230 or The voice that person's voice-output unit 500 exports feeds back the operation of user.
For example, in above embodiment, after obtaining the location information of target, user can be by social software by the position Information is sent to the friend of user, in this case, mobile terminal can using each contact information in social software as Interactive information is sent to wearable device, and the first processing units 130 in wearable device can convert braille for the interactive information Control signal is sent to the second processing unit 230, and the control of the second processing unit 230 can the output braille control of array of protrusions 220 The corresponding braille of signal, so that user selects corresponding contact person, which is sent to selected by user by mobile terminal Contact person.
In another embodiment, camera 120 can acquire the interface image of mobile terminal, and the interface image is sent out It send to first processing units 130;First processing units 130 can also be identified described mobile whole by analyzing the interface image The type and initiation object of notification message present in end, as the interactive information.
In above embodiment, mobile terminal can be interacted with wearable device by way of pairing connection, and this In embodiment, the camera 120 in wearable device acquires the interface image of mobile terminal, by dividing the interface image Analysis, obtains the interactive information with mobile terminal.
It as an implementation, can also include being set to fingerstall main body 210 in cuff assemblies 200 with continued reference to Fig. 4 On vibration unit 240, the second processing unit 230 is also used to drive vibration unit after receiving braille control signal 240 vibrations.
The second processing unit 230 control can array of protrusions 220 export braille before, can first drive vibration unit 240 Vibration, to lift user note that cause user pay attention to and then control can array of protrusions 220 output braille, in this way can be more Effectively interacted with user.
For example, in the case that first processing units 130 have notification message in the terminal, the notice is identified The type and initiation object of message convert braille control signal for the interactive information, which are controlled as interactive information Signal is sent to the second processing unit 230;The second processing unit 230 first drives vibration unit 240 to vibrate, to lift user's note Meaning, then control can array of protrusions 220 export the corresponding braille of braille control signal, that is, the type and hair of output notification message Play object.The concrete form of vibration is without limitation.
It as an implementation, can also include temperature sensor or GPS (Global in wearable device Positioning System, global positioning system) sensor or smoke sensor device or other kinds of sensor, no It enumerates again.Collected heat transfer agent can be sent to first processing units 130, first processing units by these sensors 130, which can convert the heat transfer agent to braille control signal, is sent to the second processing unit 230, and the second processing unit 230 is controlled System can the corresponding braille of the output braille control signal of array of protrusions 220.Alternatively, first processing units 130 can also be by the biography Sense information is sent directly to voice-output unit 500, and voice-output unit 500 exports the heat transfer agent in the form of speech.It is logical Excess temperature sensor can prompt the temperature of user's local environment, by GPS sensor, can prompt the position where user Information can prompt user in the environment and whether there is more smog by smoke sensor device, in this way can user more Context aware is preferably carried out fastly.
Using the embodiment of the present invention, the camera being arranged in lens body carries out image for the scene where user and adopts Collection, first processing units obtain the recognition result that target identification is carried out to image, convert braille control signal for recognition result, Braille control signal is sent to the second processing unit in cuff assemblies, the second processing unit can be in array of protrusions by control Can protrusion unit whether protrusion, export the corresponding braille of braille control signal;In this way, even if user's eyesight is poor, user It can also be obtained to the target identification of scene image by the braille of perception cuff assemblies output as a result, the recognition result can be with Auxiliary user preferably perceives ambient enviroment;As it can be seen that this programme plays the booster action of the people poor to eyesight.
Wearable device provided in this embodiment can be understood as a kind of AR (Augmented Reality, augmented reality) and set Standby or a kind of VR (Virtual Reality, virtual reality) equipment, the wearable device can be used as user and external environment Communication media, the user that eyesight can be assisted poor more rapid and better carries out Context aware.For example, user can be assisted to understand Moving target, static target, temperature information, location information in environment etc. other environmental factor information.Furthermore it is also possible to auxiliary Help user using mobile terminal, it can also be by the social software in mobile terminal, by environmental factor information acquired in user It is shared with other contact persons of the user.
Corresponding with above equipment embodiment, the embodiment of the present invention also provides a kind of exchange method based on wearable device, This method is applied to wearable device, which includes: eyeglasses assembly and cuff assemblies;Wherein, eyeglasses assembly includes glasses Main body and the camera being set in lens body and first processing units;Cuff assemblies include fingerstall main body and setting In in fingerstall main body can array of protrusions and the second processing unit, can array of protrusions include it is multiple can protrusion unit.
For example, camera can be set on the mirror holder between two eyeglasses of lens body, can acquire in this way To the preferable image in visual angle.First processing units can be set on the leg of spectacles of lens body, influence in this way on user's sight It is smaller.The position of camera and first processing units in lens body can be set according to the actual situation, not done specifically It limits.
For example, first processing units and the second processing unit can be IC (integrated circuit, integrated electricity Road), or may be other processing chips, such as MCU (Microcontroller Unit, micro-control unit) etc., place Manage the concrete type of unit without limitation.
For example, can array of protrusions can for it is dotted can array of protrusions, or column can array of protrusions, specific shape Shape is without limitation.Can include in array of protrusions it is multiple can protrusion unit, these can protrusion unit can export braille.Currently, depositing In the braille, such as 6 brailles, 12 brailles, 16 brailles, etc. of some different points.It, can array of protrusions in a kind of situation In include can the quantity of protrusion unit can be identical as braille points;In this case, can protrusion unit quantity can be 6 (as shown in fig. 1) can export 6 brailles in this way;Can the quantity of protrusion unit may be 12, can export in this way at 12 points Braille;Can protrusion unit quantity can also be 16,16 brailles can be exported in this way.
In another case, can include in array of protrusions can the quantity of protrusion unit can be greater than braille and count;For example, Can protrusion unit quantity can be 16,6 brailles, 12 brailles and 16 brailles can be exported in this way;It for another example, can protrusion The quantity of unit can be 12, can export 6 brailles and 12 brailles in this way.
Fig. 5 is a kind of flow diagram of exchange method provided in an embodiment of the present invention, comprising:
S501: camera collection image, and the image is sent to the first processing units.
S502: first processing units receive the image, obtain the recognition result for carrying out target identification to the image, will be obtained The recognition result taken is converted into braille control signal, and the braille control signal that conversion obtains is sent to the second processing unit.
S503: the second processing unit receives the braille that first processing units are sent and controls signal, can protrusion by controlling this In array can protrusion unit whether protrusion, export the received corresponding braille of braille control signal.
It is illustrated with reference to Fig. 1, eyeglasses assembly and cuff assemblies communicate to connect, for example, eyeglasses assembly and fingerstall group It can be communicated by wired or wireless mode between part, communication can be Bluetooth communication, or can also be with For other, specifically without limitation.In this way, the first processing units in eyeglasses assembly can be with the second processing in cuff assemblies Unit carries out data transmission.
For example, during the poor user of eyesight uses wearable device provided in an embodiment of the present invention, glasses group Camera in part carries out Image Acquisition for scene where user.Camera can acquire scene figure according to specified frame per second in real time Picture, and the scene image of acquisition is stored.In a kind of situation, first collected scene image can be intercepted, than Such as, intermediate region, or the image-region of interception pre-set dimension are intercepted, then truncated picture is stored.Camera will be deposited The image of storage is sent to first processing units.
As an implementation, the image that camera is sent can be sent to cloud server by first processing units, Cloud server carries out target identification to the image, and recognition result is returned to first processing units.
As another embodiment, target identification is carried out to the image that camera is sent by first processing units, for example, It can identify the information such as the type, color, position of target." target " mentioned here can be the moving target in environment, such as Vehicle, pedestrian etc., or may be the static target in environment, such as barrier, traffic lights.
For example, which can be filtered, to the progress color space conversion of filtered image, such as from RGB (Red, Green, Blue, RGB) color space converts that (" Y " indicates brightness, " U " and " V " expression coloration, dense to YUV Degree) color space, then gray scale extraction can be carried out to the yuv data being converted to, obtained gradation data will be extracted and inputted In the identification model obtained to preparatory training, the identification model can export target identification as a result, such as target type, face The information such as color, position.
In a kind of situation, it may include: to obtain sample data to be trained, sample that training, which obtains the process of the identification model, Data can be image, which can be the image of given resolution or aspect ratio, and sample data may be to image It is filtered, the data that color space is converted, gray scale obtains after extracting.The sample data is input to the nerve net of preset structure In network, output neural network based is iterated adjustment to the parameter in neural network, until meeting iteration termination condition When, the neural network of trained completion is just obtained, that is, obtained the identification model.The neural network of the preset structure can be with For the neural network based on YOLO (You Only Look Once, you only see once) algorithm, or, or other are used for The neural network of target identification, specifically without limitation.
In a kind of embodiment, as shown in Fig. 2, eyeglasses assembly can also include the sensor being set in lens body; The method also includes: sensor obtains the posture information of eyeglasses assembly, and the posture information is sent to first processing units. " obtaining the recognition result that target identification is carried out to described image " in S502 may include: that first processing units are based on the appearance State information, the image sent to camera are calibrated, and the recognition result that target identification is carried out to the image after calibration is obtained.
In a kind of situation, the image after calibration can be sent to cloud server and carry out target knowledge by first processing units Not, and the recognition result that cloud server returns is received.In another case, can also by first processing units to calibration after Image carries out target identification, obtains recognition result.
For example, sensor can be set in apart from the closer position of camera, in this way, the appearance based on sensor acquisition The image that state information acquires camera is calibrated, and can achieve preferable calibration result.In a kind of above-mentioned situation, camera It is set on the mirror holder between two eyeglasses of glasses, in this case, two eyeglasses in glasses also can be set in sensor Between mirror holder on.Position of the sensor in lens body can be set according to the actual situation, specifically without limitation.
Sensor can acquire the rotation angle of eyeglasses assembly, pitch angle for attitude transducers, sensors such as gyroscopes Equal posture informations, in this way, first processing units can the posture informations such as rotation angle, pitch angle based on eyeglasses assembly, to camera shooting The image of head acquisition is calibrated, and the recognition result that target identification is carried out to the image after calibration, obtained recognition result are obtained It is more acurrate.
As an implementation, " the obtaining the recognition result that target identification is carried out to described image " in S502 can wrap Include: parameter of the first processing units based on image and camera after calibration calculates the distance between target and camera, makees For the recognition result.
In a kind of situation, camera can be depth camera, for example have the camera of depth transducer, depth camera It include depth information in the depth image of head acquisition, first processing units detect the target in depth image, according to the target The target can be calculated at a distance from camera in the relevant parameter of depth information and depth camera.
In another case, camera can be binocular camera, first processing units are acquired according to binocular camera The internal reference and outer ginseng of binocular image and binocular camera, can also be calculated target at a distance from camera.
As an implementation, " the obtaining the recognition result that target identification is carried out to described image " in S502 can wrap Include: first processing units carry out target following by the continuous multiple frames image acquired to camera, obtain the movement tendency of target, As the recognition result.
For example, first processing units can carry out target detection in N (N is positive integer) frame image, determine mesh It is marked on the region in nth frame image, which is split as multiple images block, as image block to be processed;Then in N+1 frame It is determining with each highest image block of image block similarity to be processed respectively in image, as matching image block;According to be processed Moving direction and moving distance between the corresponding matching image block of image block, obtain the movement of target in the picture and become Gesture;The transformational relation being then based between image coordinate system and physical space coordinates system, by the movement tendency of target in the picture Be converted to movement tendency of the target in physical space.
As another example, in above embodiment, first processing units can calculate target and camera based on every frame image Distance, in a kind of situation, this embodiment can be used, will be compared based on every frame image calculated distance, To judge that target is to obtain the movement tendency of target close to or far from camera.
First processing units can convert recognition result obtained above to braille control signal, which is controlled and is believed The second processing unit number being sent in cuff assemblies.The second processing unit receives the braille that first processing units are sent and controls letter Number, by control can in array of protrusions can protrusion unit whether protrusion, it is corresponding to export received braille control signal Braille.That is, the second processing unit be based on the braille control signal control can array of protrusions output it is opposite with recognition result The braille answered.
For example, recognition result is the information such as type, color, the position of target in a kind of above-mentioned embodiment, then it can protrusion The braille of array output can be the information such as type, color, position;In a kind of above-mentioned embodiment, recognition result be target with The distance of camera, then can array of protrusions output braille can for target at a distance from camera;A kind of above-mentioned embodiment In, recognition result be target movement tendency, then can array of protrusions output braille can be target movement tendency.At second Reason unit can control can array of protrusions according to prefixed time interval, be sequentially output each braille lettering.
It is currently, there are the braille of miscellaneous editions, the braille of miscellaneous editions, user can be stored in advance in first processing units The braille suitable for itself can be selected in the braille of the miscellaneous editions, the braille control signal of first processing units conversion is For the corresponding control signal of braille of user's selection.Braille control signal can be understood as to can array of protrusions driving signal, The driving signal can drive can in array of protrusions can protrusion unit show the corresponding waveform of braille.
In a kind of above-mentioned situation, can include in array of protrusions can the quantity of protrusion unit can be greater than braille and count, this In the case of kind, the method also includes: the second processing unit determines the corresponding braille points of braille control signal, according to described blind Text points, it is described can determine in array of protrusions it is to be processed can protrusion unit.In S503 " by control described in can protrusion battle array In column can protrusion unit whether protrusion, export the received corresponding braille of braille control signal " may include: to pass through control Make it is described it is to be processed can protrusion unit whether protrusion, export the received corresponding braille of braille control signal.
As described above, the corresponding braille points of braille control signal can be 6,12,16.With reference to Fig. 3, it is assumed that can protrusion battle array Include in column can protrusion unit quantity be 16, Fig. 3 in can protrusion unit layout not to dashing forward in the present embodiment It plays array and constitutes restriction.Assuming that the second processing unit determines that braille control signal corresponding braille points are 6, in a kind of situation, Can in array of protrusions upper left 6 can protrusion unit (frame 1) be determined as it is to be processed can protrusion unit;Other feelings Under condition, can in array of protrusions 6 of lower right (frame 2) or upper right side or lower left can protrusion unit be determined as To be processed can protrusion unit;In the case of other, can also determine 6 it is non-conterminous can protrusion unit, as to be processed Can protrusion unit, it is to be processed can protrusion unit will namely carry out braille displaying can protrusion unit, to be processed dashes forward Play the specific location of unit without limitation.
Except it is to be processed can in addition to protrusion unit other can protrusion unit can be understood as temporarily leaving unused can protrusion list Member, it is temporarily idle can protrusion unit can be fitted closely with fingerstall, make user be not felt by these it is idle can protrusion unit Protrusion, such user can more accurately experience the braille of output.
As an implementation, wearable device can also include microphone, the method also includes: microphone acquires language The voice data is sent to first processing units by sound data, and first processing units are obtained by parsing the voice data It is instructions to be performed;" obtaining the recognition result that target identification is carried out to described image " in S502 may include: based on it is described to Execute instruction the recognition result for obtaining and carrying out target identification to described image.
As shown in figure 4, microphone can be set in lens body, for example, can be set in the temple of lens body, The specific location of microphone is without limitation.For example, user can inquire " existing when by traffic intersection to wearable device It is being red light or green light ";Microphone collects the voice data, which is sent to first processing units;First Processing unit identifies the voice data by speech recognition algorithm, obtains the instruction of " identification traffic lights color ".First processing is single Member detects " traffic lights " this target, then further identifies the color of this target in the image that video camera is sent, it is assumed that Recognition result is red, then generates red corresponding braille control signal, and braille control signal is sent to second processing list Member.
In a kind of situation, user can first say it is default wake up word, microphone collects default after waking up word, wearable device Working condition is switched to by standby mode.Alternatively, judging the corresponding vocal print of wake-up word after microphone collects default wake-up word Whether matched with user's vocal print pre-stored in wearable device, if it does, wearable device is switched to work by standby mode State.
In a kind of embodiment, wearable device can also include key, as shown in figure 4, key can be set in glasses group In part, or in other embodiments, key also be can be set in cuff assemblies, alternatively, in eyeglasses assembly and can also refer to Different keys is set in set component, the specific location of key is without limitation.
In a kind of embodiment, key can be used for opening or closing the wearable device.For example, key can be Mechanical key, or may be fingerprint extracting device etc., specifically without limitation.User can will be worn by click keys It wears equipment and working condition is switched to by standby mode, or wearable device is switched to standby mode by working condition.Alternatively, with Family can also reset wearable device by click keys 400.
In a kind of embodiment, as shown in figure 4, wearable device can also include voice-output unit, in a kind of situation, language Sound output unit can be earphone, or may be speaker etc., specifically without limitation.In above content, wearable device by After standby mode is switched to working condition, voice-output unit can play the response voice to above-mentioned wake-up word, to prompt to use Family wearable device has entered working condition, promotes user experience.User can send out after hearing the response voice to wearable device The voice data of instruction is carried out.
In a kind of embodiment, the method also includes: first processing units to carry out target knowledge to described image obtaining After other recognition result, the recognition result is sent to voice-output unit;Voice-output unit exports in the form of speech The recognition result.
In a kind of above-mentioned embodiment, wearable device can also include key, and in this embodiment, key can be sound Measure key, the volume that user passes through the adjustable voice-output unit output of key.
In a kind of situation, camera, sensor, microphone, voice-output unit, key and first processing units can collect At on the mirror holder for being set to lens body.Alternatively, in another case, microphone, voice-output unit, key and the first processing Unit, which can integrate, to be set in the temple of lens body.
In a kind of embodiment, eyeglasses assembly can be connect with communication of mobile terminal, the method also includes: the first processing Unit obtains the interactive information with the mobile terminal, converts braille control signal for the interactive information.In this way, in S501 The braille control signal that interactive information converts can also be sent to described the second processing unit.
In a kind of situation, it may include: the first processing that first processing units, which obtain the interactive information with the mobile terminal, Unit can there are the type in the case where notification message, identifying the notification message and initiations pair in the mobile terminal As the interactive information.
For example, mobile terminal and wearable device can carry out pairing connection in advance, and then wearable device can be examined The notification message in mobile terminal is measured, the type of the notification message can be incoming information, short message or communication software Prompt information, etc., specifically without limitation.
For by taking incoming information as an example, first processing units, can be with after detecting that mobile terminal receives incoming information It identifies the initiation object of the incoming information, that is, incoming call connection people, and incites somebody to action " type: incoming information initiates object: someone " As interactive information.First processing units can convert the interactive information to braille control signal and be sent to second processing list Member, the second processing unit control can array of protrusions export the corresponding braille of braille control signal.Alternatively, first processing units The interactive information can be sent directly to voice-output unit, voice-output unit exports interaction letter in the form of speech Breath.
In a kind of above-mentioned embodiment, wearable device can also include key, and in this embodiment, key can be answered Or the incoming call in refusal mobile terminal.That is, user can select to answer or refuse mobile whole by click keys Incoming call in end.For example, " answering " key different with " refusal " two can be set, in a kind of situation, on " answering " key " answering " corresponding braille can be set, " refusal " corresponding braille can be set on " refusal " key, in this way, can be square Just user selects.
Alternatively, user can use wearable device operation mobile terminal in the case of other.For example, user can be with Using the social software in wearable device operation mobile terminal, the braille or language that wearable device is exported by the second processing unit The voice of sound output unit output feeds back the operation of user.
For example, in above embodiment, after obtaining the location information of target, user can be by social software by the position Information is sent to the friend of user, in this case, mobile terminal can using each contact information in social software as Interactive information is sent to wearable device, and the first processing units in wearable device can convert the interactive information to braille control Signal is sent to the second processing unit, the second processing unit control can array of protrusions to export braille control signal corresponding blind Text, so that user selects corresponding contact person, which is sent to the selected contact person of user by mobile terminal.
In another embodiment, the method also includes: camera acquires the interface image of mobile terminal, by the boundary Face image is sent to first processing units;" first processing units obtain the interactive information with the mobile terminal " may include: First processing units can also identify the type of notification message present in the mobile terminal by analyzing the interface image And object is initiated, as the interactive information.
In above embodiment, mobile terminal can be interacted with wearable device by way of pairing connection, and this In embodiment, the interface image of the camera acquisition mobile terminal in wearable device, by analyzing the interface image, Obtain the interactive information with mobile terminal.
It as an implementation, can also include the vibration being set in fingerstall main body in cuff assemblies with continued reference to Fig. 4 Moving cell;The method also includes: the second processing unit driving vibration unit vibration after receiving the braille control signal.
The second processing unit control can array of protrusions output braille before, first vibration unit can be driven to vibrate, to mention Rise user note that cause user pay attention to and then control can array of protrusions export braille, in this way can more effectively with user It interacts.
For example, in the case that first processing units have notification message in the terminal, identify that the notice disappears The type and initiation object of breath convert braille control signal for the interactive information, which are controlled and is believed as interactive information Number it is sent to the second processing unit;The second processing unit first drives vibration unit to vibrate, to lift user note that controlling can dash forward again Play the corresponding braille of array output braille control signal, that is, the type and initiation object of output notification message.The tool of vibration Body form is without limitation.
It as an implementation, can also include temperature sensor or GPS (Global in wearable device Positioning System, global positioning system) sensor or smoke sensor device or other kinds of sensor, no It enumerates again.Collected heat transfer agent can be sent to first processing units by these sensors, and first processing units can It is sent to the second processing unit to convert braille control signal for the heat transfer agent, the second processing unit control can array of protrusions Export the corresponding braille of braille control signal.Alternatively, the heat transfer agent can also be sent directly to language by first processing units Sound output unit, voice-output unit export the heat transfer agent in the form of speech.By temperature sensor, user can be prompted The temperature of local environment can prompt the location information where user by GPS sensor, pass through smoke sensor device, Ke Yiti Show user in the environment and whether there is more smog, in this way can user more rapid and better carry out Context aware.
Using the embodiment of the present invention, the camera being arranged in lens body carries out image for the scene where user and adopts Collection, first processing units obtain the recognition result that target identification is carried out to image, convert braille control signal for recognition result, Braille control signal is sent to the second processing unit in cuff assemblies, the second processing unit can be in array of protrusions by control Can protrusion unit whether protrusion, export the corresponding braille of braille control signal;In this way, even if user's eyesight is poor, user It can also be obtained to the target identification of scene image by the braille of perception cuff assemblies output as a result, the recognition result can be with Auxiliary user preferably perceives ambient enviroment;As it can be seen that this programme plays the booster action of the people poor to eyesight.
Wearable device provided in this embodiment can be understood as a kind of AR (Augmented Reality, augmented reality) and set Standby or a kind of VR (Virtual Reality, virtual reality) equipment, the wearable device can be used as user and external environment Communication media, the user that eyesight can be assisted poor more rapid and better carries out Context aware.For example, user can be assisted to understand Moving target, static target, temperature information, location information in environment etc. other environmental factor information.Furthermore it is also possible to auxiliary Help user using mobile terminal, it can also be by the social software in mobile terminal, by environmental factor information acquired in user It is shared with other contact persons of the user.
It should be understood by those ordinary skilled in the art that: the discussion of any of the above embodiment is exemplary only, not It is intended to imply that the scope of the present disclosure (including claim) is limited to these examples;Under thinking of the invention, above embodiments Or can also be combined between the technical characteristic in different embodiments, step can be realized with random order, and be existed such as Many other variations of the upper different aspect of the invention, for simplicity, they are not provided in details.
In addition, to simplify explanation and discussing, and in order not to obscure the invention, it can in provided attached drawing It is connect with showing or can not show with the well known power ground of integrated circuit (IC) chip and other components.Furthermore, it is possible to Device is shown in block diagram form, to avoid obscuring the invention, and this has also contemplated following facts, i.e., about this The details of the embodiment of a little block diagram arrangements be height depend on will implementing platform of the invention (that is, these details should It is completely within the scope of the understanding of those skilled in the art).Elaborating that detail (for example, circuit) is of the invention to describe In the case where exemplary embodiment, it will be apparent to those skilled in the art that can be in these no details In the case where or implement the present invention in the case that these details change.Therefore, these descriptions should be considered as explanation Property rather than it is restrictive.
Although having been incorporated with specific embodiments of the present invention, invention has been described, according to retouching for front It states, many replacements of these embodiments, modifications and variations will be apparent for those of ordinary skills.Example Such as, discussed embodiment can be used in other memory architectures (for example, dynamic ram (DRAM)).
The embodiment of the present invention be intended to cover fall into all such replacements within the broad range of appended claims, Modifications and variations.Therefore, all within the spirits and principles of the present invention, any omission, modification, equivalent replacement, the improvement made Deng should all be included in the protection scope of the present invention.

Claims (13)

1. a kind of wearable device characterized by comprising eyeglasses assembly and cuff assemblies;Wherein, the eyeglasses assembly includes eye Mirror main body and the camera being set in the lens body and first processing units;The cuff assemblies include fingerstall master Body and be set in the fingerstall main body can array of protrusions and the second processing unit, it is described can array of protrusions include multiple It can protrusion unit;
The camera is sent to the first processing units for acquiring image, and by described image;
The first processing units, for receiving described image, acquisition carries out the recognition result of target identification to described image, will Acquired recognition result is converted into braille control signal, and the braille control signal that conversion obtains is sent to the second processing Unit;
Described the second processing unit controls signal for receiving the braille that the first processing units are sent, by described in control Can in array of protrusions can protrusion unit whether protrusion, export the received corresponding braille of braille control signal.
2. wearable device according to claim 1, which is characterized in that described the second processing unit is also used to determine braille Control signal corresponding braille points, counted according to the braille, it is described can determine in array of protrusions it is to be processed can protrusion Unit, by control it is described it is to be processed can protrusion unit whether protrusion, it is corresponding to export received braille control signal Braille.
3. wearable device according to claim 1, which is characterized in that the eyeglasses assembly further includes being set to the glasses Sensor in main body;The posture information is sent to by the sensor for obtaining the posture information of the eyeglasses assembly The first processing units;
The first processing units are also used to calibrate described image based on the posture information, after obtaining to calibration The recognition result of image progress target identification.
4. wearable device according to claim 3, which is characterized in that the first processing units are also used to based on described The parameter of image and the camera after calibration calculates the distance between target and the camera, as the identification As a result.
5. wearable device according to claim 1, which is characterized in that the first processing units are also used to by institute The continuous multiple frames image for stating camera acquisition carries out target following, the movement tendency of target is obtained, as the recognition result.
6. wearable device according to claim 1, which is characterized in that the wearable device further includes microphone, the wheat The voice data is sent to the first processing units for acquiring voice data by gram wind;
The first processing units, are also used to by parsing the voice data, obtain it is instructions to be performed, based on described pending Instruction obtains the recognition result.
7. wearable device according to claim 1, which is characterized in that the eyeglasses assembly is connect with communication of mobile terminal, The first processing units are also used to obtain the interactive information with the mobile terminal, convert braille for the interactive information Control signal.
8. wearable device according to claim 7, which is characterized in that the first processing units are also used in the shifting There are the type in the case where notification message, identifying the notification message and initiation objects in dynamic terminal, as the interaction letter Breath.
9. wearable device according to claim 8, which is characterized in that the camera is also used to acquire described mobile whole The interface image is sent to the first processing units by the interface image at end;
The first processing units are also used to identify notice present in the mobile terminal by analyzing the interface image The type and initiation object of message, as the interactive information.
10. wearable device according to claim 1, which is characterized in that the cuff assemblies further include being set to the finger The vibration unit in main body is covered, described the second processing unit is also used to after receiving the braille control signal described in driving Vibration unit vibration.
11. wearable device according to claim 7, which is characterized in that the wearable device further includes key;
The key, for opening or closing the wearable device;
And/or the key, for answering or refusing the incoming call in the mobile terminal.
12. wearable device according to claim 1, which is characterized in that the wearable device further includes voice-output unit;
The first processing units are also used to after obtaining the recognition result for carrying out target identification to described image, by the knowledge Other result is sent to the voice-output unit;
The voice-output unit is also used to export the recognition result in the form of speech.
13. a kind of exchange method based on wearable device, which is characterized in that be applied to wearable device, the wearable device includes: Eyeglasses assembly and cuff assemblies;Wherein, the eyeglasses assembly includes lens body and is set to taking the photograph in the lens body As head and first processing units;The cuff assemblies include fingerstall main body and be set in the fingerstall main body can protrusion Array and the second processing unit, it is described can array of protrusions include it is multiple can protrusion unit;The described method includes:
The camera collection image, and described image is sent to the first processing units;
The first processing units receive described image, obtain the recognition result for carrying out target identification to described image, will be obtained The recognition result taken is converted into braille control signal, and the braille control signal that conversion obtains is sent to the second processing list Member;
Described the second processing unit receives the braille control signal that the first processing units are sent, by can protrusion described in control In array can protrusion unit whether protrusion, export the received corresponding braille of braille control signal.
CN201910470231.1A 2019-05-31 2019-05-31 A kind of wearable device and the exchange method based on wearable device Pending CN110208946A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910470231.1A CN110208946A (en) 2019-05-31 2019-05-31 A kind of wearable device and the exchange method based on wearable device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910470231.1A CN110208946A (en) 2019-05-31 2019-05-31 A kind of wearable device and the exchange method based on wearable device

Publications (1)

Publication Number Publication Date
CN110208946A true CN110208946A (en) 2019-09-06

Family

ID=67790068

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910470231.1A Pending CN110208946A (en) 2019-05-31 2019-05-31 A kind of wearable device and the exchange method based on wearable device

Country Status (1)

Country Link
CN (1) CN110208946A (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110728886A (en) * 2019-10-30 2020-01-24 京东方科技集团股份有限公司 Braille learning system, fingertip sensor and forming method thereof
CN110989832A (en) * 2019-11-21 2020-04-10 维沃移动通信有限公司 Control method and electronic equipment
CN111942285A (en) * 2020-07-10 2020-11-17 夏牧谣 Intelligent vision-impaired person service method and system based on vehicle-mounted glass vibration feedback
CN113449549A (en) * 2020-03-25 2021-09-28 中移(成都)信息通信科技有限公司 Prompt message generation method, device, equipment and storage medium
CN113534715A (en) * 2021-07-21 2021-10-22 歌尔科技有限公司 Intelligent wearable device, and control method and system of target device

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103514450A (en) * 2012-06-29 2014-01-15 华为技术有限公司 Image feature extracting method and image correcting method and equipment
CN106506954A (en) * 2016-10-25 2017-03-15 深圳市高巨创新科技开发有限公司 Image pickup method and system based on multi-rotor aerocraft
CN107328424A (en) * 2017-07-12 2017-11-07 三星电子(中国)研发中心 Air navigation aid and device
CN108279780A (en) * 2018-03-01 2018-07-13 京东方科技集团股份有限公司 Wearable device and control method
CN108616744A (en) * 2017-01-23 2018-10-02 上海爱观视觉科技有限公司 A kind of bionical binocular vision calibration system and calibration method
CN109674628A (en) * 2019-01-29 2019-04-26 桂林电子科技大学 A kind of intelligent glasses
KR101975761B1 (en) * 2018-07-02 2019-05-07 이단경 Braille input and output system for mobile device

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103514450A (en) * 2012-06-29 2014-01-15 华为技术有限公司 Image feature extracting method and image correcting method and equipment
CN106506954A (en) * 2016-10-25 2017-03-15 深圳市高巨创新科技开发有限公司 Image pickup method and system based on multi-rotor aerocraft
CN108616744A (en) * 2017-01-23 2018-10-02 上海爱观视觉科技有限公司 A kind of bionical binocular vision calibration system and calibration method
CN107328424A (en) * 2017-07-12 2017-11-07 三星电子(中国)研发中心 Air navigation aid and device
CN108279780A (en) * 2018-03-01 2018-07-13 京东方科技集团股份有限公司 Wearable device and control method
KR101975761B1 (en) * 2018-07-02 2019-05-07 이단경 Braille input and output system for mobile device
CN109674628A (en) * 2019-01-29 2019-04-26 桂林电子科技大学 A kind of intelligent glasses

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110728886A (en) * 2019-10-30 2020-01-24 京东方科技集团股份有限公司 Braille learning system, fingertip sensor and forming method thereof
CN110728886B (en) * 2019-10-30 2021-10-08 京东方科技集团股份有限公司 Braille learning system, fingertip sensor and forming method thereof
CN110989832A (en) * 2019-11-21 2020-04-10 维沃移动通信有限公司 Control method and electronic equipment
CN113449549A (en) * 2020-03-25 2021-09-28 中移(成都)信息通信科技有限公司 Prompt message generation method, device, equipment and storage medium
CN111942285A (en) * 2020-07-10 2020-11-17 夏牧谣 Intelligent vision-impaired person service method and system based on vehicle-mounted glass vibration feedback
CN111942285B (en) * 2020-07-10 2022-11-18 夏牧谣 Intelligent vision-impaired person service method and system based on vehicle-mounted glass vibration feedback
CN113534715A (en) * 2021-07-21 2021-10-22 歌尔科技有限公司 Intelligent wearable device, and control method and system of target device
CN113534715B (en) * 2021-07-21 2023-01-24 歌尔科技有限公司 Intelligent wearable device, and control method and system of target device

Similar Documents

Publication Publication Date Title
CN110208946A (en) A kind of wearable device and the exchange method based on wearable device
CN108416235B (en) The anti-peeping method, apparatus of display interface, storage medium and terminal device
CN107231529A (en) Image processing method, mobile terminal and storage medium
CN103839054B (en) Multi-functional mobile intelligent terminal sensor supporting iris recognition
CN109804357A (en) Server, client, control method and storage medium
CN108701214A (en) Image processing method, device and equipment
CN108700982A (en) Information processing equipment, information processing method and program
CN109240572A (en) A kind of method obtaining picture, method and device that picture is handled
CN107302662A (en) A kind of method, device and mobile terminal taken pictures
WO2022227393A1 (en) Image photographing method and apparatus, electronic device, and computer readable storage medium
CN111601088A (en) Sitting posture monitoring system based on monocular camera sitting posture identification technology
CN108989678A (en) A kind of image processing method, mobile terminal
CN105763829A (en) Image processing method and electronic device
US11837233B2 (en) Information processing device to automatically detect a conversation
US10843299B2 (en) Object recognition and presentation for the visually impaired
US20230079582A1 (en) Image processing method and apparatus, terminal, and storage medium
CN110533651B (en) Image processing method and device
CN104535059A (en) Indoor positioning system specific to totally blind population
CN110807405A (en) Detection method of candid camera device and electronic equipment
CN110222567B (en) Image processing method and device
CN112286347A (en) Eyesight protection method, device, storage medium and terminal
CN116030512A (en) Gaze point detection method and device
CN105227828B (en) Filming apparatus and method
CN109639981A (en) A kind of image capturing method and mobile terminal
CN110602384B (en) Exposure control method and electronic device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination