CN110208946A - A kind of wearable device and the exchange method based on wearable device - Google Patents
A kind of wearable device and the exchange method based on wearable device Download PDFInfo
- Publication number
- CN110208946A CN110208946A CN201910470231.1A CN201910470231A CN110208946A CN 110208946 A CN110208946 A CN 110208946A CN 201910470231 A CN201910470231 A CN 201910470231A CN 110208946 A CN110208946 A CN 110208946A
- Authority
- CN
- China
- Prior art keywords
- braille
- wearable device
- processing units
- image
- control signal
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 24
- 238000012545 processing Methods 0.000 claims abstract description 210
- 238000000429 assembly Methods 0.000 claims abstract description 33
- 230000000712 assembly Effects 0.000 claims abstract description 33
- 230000002452 interceptive effect Effects 0.000 claims description 35
- 230000000977 initiatory effect Effects 0.000 claims description 11
- 238000004891 communication Methods 0.000 claims description 10
- 238000006243 chemical reaction Methods 0.000 claims description 9
- 239000011521 glass Substances 0.000 claims description 9
- 230000003993 interaction Effects 0.000 claims description 3
- 241000209140 Triticum Species 0.000 claims 1
- 235000021307 Triticum Nutrition 0.000 claims 1
- 230000004438 eyesight Effects 0.000 abstract description 14
- 230000009471 action Effects 0.000 abstract description 7
- 230000008447 perception Effects 0.000 abstract description 4
- 230000036544 posture Effects 0.000 description 13
- 238000013528 artificial neural network Methods 0.000 description 12
- 238000010586 diagram Methods 0.000 description 9
- 239000011551 heat transfer agent Substances 0.000 description 7
- 230000003068 static effect Effects 0.000 description 5
- 230000003190 augmentative effect Effects 0.000 description 4
- 230000005540 biological transmission Effects 0.000 description 4
- 230000007613 environmental effect Effects 0.000 description 4
- 230000004044 response Effects 0.000 description 4
- 239000000779 smoke Substances 0.000 description 4
- 238000012549 training Methods 0.000 description 4
- 230000001755 vocal effect Effects 0.000 description 4
- 230000004888 barrier function Effects 0.000 description 3
- 230000004048 modification Effects 0.000 description 3
- 238000012986 modification Methods 0.000 description 3
- 241000208340 Araliaceae Species 0.000 description 2
- 235000005035 Panax pseudoginseng ssp. pseudoginseng Nutrition 0.000 description 2
- 235000003140 Panax quinquefolius Nutrition 0.000 description 2
- 238000000151 deposition Methods 0.000 description 2
- 238000001514 detection method Methods 0.000 description 2
- 235000013399 edible fruits Nutrition 0.000 description 2
- 238000000605 extraction Methods 0.000 description 2
- 235000008434 ginseng Nutrition 0.000 description 2
- 210000004218 nerve net Anatomy 0.000 description 2
- 230000008569 process Effects 0.000 description 2
- 230000000007 visual effect Effects 0.000 description 2
- 230000002618 waking effect Effects 0.000 description 2
- 230000008901 benefit Effects 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 230000004069 differentiation Effects 0.000 description 1
- 230000005611 electricity Effects 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 238000010191 image analysis Methods 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/017—Head mounted
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/20—Scenes; Scene-specific elements in augmented reality scenes
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/017—Head mounted
- G02B2027/0178—Eyeglass type
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Human Computer Interaction (AREA)
- Optics & Photonics (AREA)
- Multimedia (AREA)
- User Interface Of Digital Computer (AREA)
Abstract
The embodiment of the invention discloses a kind of wearable device and based on the exchange method of wearable device, in the wearable device, the camera being arranged in lens body carries out Image Acquisition for the scene where user, first processing units obtain the recognition result that target identification is carried out to image, braille control signal is converted by recognition result, braille control signal is sent to the second processing unit in cuff assemblies, the second processing unit by control can in array of protrusions can protrusion unit whether protrusion, export the corresponding braille of braille control signal;In this way, user can also be obtained to the target identification of scene image by the braille of perception cuff assemblies output as a result, the recognition result can assist user preferably to perceive ambient enviroment even if user's eyesight is poor;As it can be seen that this programme plays the booster action of the people poor to eyesight.
Description
Technical field
The present invention relates to technical field of electronic equipment, particularly relate to a kind of wearable device and the interaction side based on wearable device
Method.
Background technique
In daily life, the poor people of eyesight cannot preferably perceive the environment of surrounding, for example, they cannot accurately identify
Moving target in environment, such as vehicle, pedestrian, can not accurately identify the static target in environment, such as barrier, traffic lights
Deng.In general, these people usually require other people and accompany, or need the help of seeing-eye dog.It can be auxiliary currently, needing one kind
These people are helped to carry out the scheme of daily routines.
Summary of the invention
In view of this, it is an object of the invention to propose a kind of wearable device and based on the exchange method of wearable device, with
Play the booster action of the people poor to eyesight.
Based on above-mentioned purpose, the present invention provides a kind of wearable devices, comprising: eyeglasses assembly and cuff assemblies;Wherein, institute
Stating eyeglasses assembly includes lens body and the camera being set in the lens body and first processing units;The finger
Set component include fingerstall main body and be set in the fingerstall main body can array of protrusions and the second processing unit, it is described can
Array of protrusions include it is multiple can protrusion unit;
The camera is sent to the first processing units for acquiring image, and by described image;
The first processing units obtain the identification knot that target identification is carried out to described image for receiving described image
Acquired recognition result is converted braille control signal by fruit, and braille control signal that conversion obtains is sent to described the
Two processing units;
Described the second processing unit controls signal for receiving the braille that the first processing units are sent, passes through control
It is described can in array of protrusions can protrusion unit whether protrusion, export the received corresponding braille of braille control signal.
In some embodiments, described the second processing unit is also used to determine the corresponding braille point of braille control signal
Number is counted according to the braille, it is described can determine in array of protrusions it is to be processed can protrusion unit, it is described wait locate by controlling
Reason can protrusion unit whether protrusion, export the received corresponding braille of braille control signal.
In some embodiments, the eyeglasses assembly further includes the sensor being set in the lens body;It is described
The posture information is sent to the first processing units for obtaining the posture information of the eyeglasses assembly by sensor;
The first processing units are also used to calibrate described image based on the posture information, be obtained to calibration
Image afterwards carries out the recognition result of target identification.
In some embodiments, the first processing units are also used to based on image after the calibration and described
The parameter of camera calculates the distance between target and the camera, as the recognition result.
In some embodiments, the first processing units are also used to continuous more by acquiring to the camera
Frame image carries out target following, the movement tendency of target is obtained, as the recognition result.
In some embodiments, the wearable device further includes microphone, the microphone, for acquiring voice number
According to the voice data is sent to the first processing units;
The first processing units, are also used to by parsing the voice data, obtain it is instructions to be performed, based on it is described to
It executes instruction and obtains the recognition result.
In some embodiments, the eyeglasses assembly is connect with communication of mobile terminal, and the first processing units are also used
In obtaining the interactive information with the mobile terminal, braille control signal is converted by the interactive information.
In some embodiments, the first processing units, are also used in the mobile terminal that there are notification messages
In the case where, it identifies the type of the notification message and initiates object, as the interactive information.
In some embodiments, the camera is also used to acquire the interface image of the mobile terminal, by the boundary
Face image is sent to the first processing units;
The first processing units are also used to identify present in the mobile terminal by analyzing the interface image
The type and initiation object of notification message, as the interactive information.
In some embodiments, the cuff assemblies further include the vibration unit being set in the fingerstall main body, institute
The second processing unit is stated, is also used to drive the vibration unit to vibrate after receiving the braille control signal.
In some embodiments, the wearable device further includes key;The key, for opening or closing described wear
Wear equipment;And/or the key, for answering or refusing the incoming call in the mobile terminal.
In some embodiments, the wearable device further includes voice-output unit;
The first processing units are also used to after obtaining the recognition result for carrying out target identification to described image, by institute
It states recognition result and is sent to the voice-output unit;
The voice-output unit is also used to export the recognition result in the form of speech.
On the other hand, the present invention also provides a kind of exchange methods based on wearable device, are applied to wearable device, described
Wearable device includes: eyeglasses assembly and cuff assemblies;Wherein, the eyeglasses assembly includes lens body and is set to described
Camera and first processing units in lens body;The cuff assemblies include fingerstall main body and are set to the fingerstall
In main body can array of protrusions and the second processing unit, it is described can array of protrusions include it is multiple can protrusion unit;The method packet
It includes:
The camera collection image, and described image is sent to the first processing units;
The first processing units receive described image, obtain the recognition result that target identification is carried out to described image, will
Acquired recognition result is converted into braille control signal, and the braille control signal that conversion obtains is sent to the second processing
Unit;
Described the second processing unit receives the braille control signal that the first processing units are sent, by can described in control
In array of protrusions can protrusion unit whether protrusion, export the received corresponding braille of braille control signal.
Using the embodiment of the present invention, the camera being arranged in lens body carries out image for the scene where user and adopts
Collection, first processing units obtain the recognition result that target identification is carried out to image, convert braille control signal for recognition result,
Braille control signal is sent to the second processing unit in cuff assemblies, the second processing unit can be in array of protrusions by control
Can protrusion unit whether protrusion, export the corresponding braille of braille control signal;In this way, even if user's eyesight is poor, user
It can also be obtained to the target identification of scene image by the braille of perception cuff assemblies output as a result, the recognition result can be with
Auxiliary user preferably perceives ambient enviroment;As it can be seen that this programme plays the booster action of the people poor to eyesight.
Detailed description of the invention
In order to more clearly explain the embodiment of the invention or the technical proposal in the existing technology, to embodiment or will show below
There is attached drawing needed in technical description to be briefly described, it should be apparent that, the accompanying drawings in the following description is only this
Some embodiments of invention for those of ordinary skill in the art without creative efforts, can be with
It obtains other drawings based on these drawings.
Fig. 1 is the first schematic diagram of wearable device provided in an embodiment of the present invention;
Fig. 2 is second of schematic diagram of wearable device provided in an embodiment of the present invention;
Fig. 3 be one of wearable device provided in an embodiment of the present invention can protrusion display schematic diagram;
Fig. 4 is the third schematic diagram of wearable device provided in an embodiment of the present invention;
Fig. 5 is a kind of flow diagram of the exchange method based on wearable device provided in an embodiment of the present invention.
Specific embodiment
To make the objectives, technical solutions, and advantages of the present invention clearer, below in conjunction with specific embodiment, and reference
Attached drawing, the present invention is described in more detail.
It should be noted that all statements for using " first " and " second " are for differentiation two in the embodiment of the present invention
The non-equal entity of a same names or non-equal parameter, it is seen that " first " " second " only for the convenience of statement, does not answer
It is interpreted as the restriction to the embodiment of the present invention, subsequent embodiment no longer illustrates this one by one.
In order to solve the above-mentioned technical problem, the embodiment of the invention provides a kind of wearable device and based on the friendship of wearable device
Mutual method, first below describes to wearable device in detail.
Fig. 1 is the first structural schematic diagram of wearable device provided in an embodiment of the present invention, comprising: 100 He of eyeglasses assembly
Cuff assemblies 200;Wherein, eyeglasses assembly 100 includes lens body 110 and the camera being set in lens body 110
120 and first processing units 130;Cuff assemblies 200 are including fingerstall main body 210 and are set to dashing forward in fingerstall main body 210
Rise array 220 and the second processing unit 230, can array of protrusions 220 include it is multiple can protrusion unit 221.
For example, camera 120 can be set on the mirror holder between two eyeglasses of lens body 110, in this way may be used
To collect the preferable image in visual angle.First processing units 130 can be set on the leg of spectacles of lens body 110, right in this way
User's sight influences smaller.The position of camera 120 and first processing units 130 in lens body 110 can be according to reality
Situation is set, specifically without limitation.
For example, first processing units 130 and the second processing unit 230 can for IC (integrated circuit,
Integrated circuit), or may be other processing chips, such as MCU (Microcontroller Unit, micro-control unit) etc.
Deng the concrete type of processing unit is without limitation.
For example, can array of protrusions can for it is dotted can array of protrusions, or column can array of protrusions, specific shape
Shape is without limitation.Can include in array of protrusions it is multiple can protrusion unit, these can protrusion unit can export braille.Currently, depositing
In the braille, such as 6 brailles, 12 brailles, 16 brailles, etc. of some different points.It, can array of protrusions in a kind of situation
In include can the quantity of protrusion unit can be identical as braille points;In this case, can protrusion unit quantity can be 6
(as shown in fig. 1) can export 6 brailles in this way;Can the quantity of protrusion unit may be 12, can export in this way at 12 points
Braille;Can protrusion unit quantity can also be 16,16 brailles can be exported in this way.
In another case, can include in array of protrusions can the quantity of protrusion unit can be greater than braille and count;For example,
Can protrusion unit quantity can be 16,6 brailles, 12 brailles and 16 brailles can be exported in this way;It for another example, can protrusion
The quantity of unit can be 12, can export 6 brailles and 12 brailles in this way.
It continues to refer to figure 1 and is illustrated, image is sent to first processing units for acquiring image by camera 120
130;
First processing units 130 obtain the recognition result for carrying out target identification to image, identification are tied for receiving image
Fruit is converted into braille control signal, and the braille control signal that conversion obtains is sent to the second processing unit 230;
The second processing unit 230 is used to receive the braille control signal of first processing units transmission, can protrusion by control
In array 220 can protrusion unit 221 whether protrusion, export the received corresponding braille of braille control signal.
Eyeglasses assembly 100 and cuff assemblies 200 communicate to connect, for example, eyeglasses assembly 100 and cuff assemblies 200 it
Between can be communicated by wired or wireless mode, communication can be Bluetooth communication, or may be its
He, specifically without limitation.In this way, the first processing units 130 in eyeglasses assembly 100 can be in cuff assemblies 200
Two processing units 230 carry out data transmission.
For example, during the poor user of eyesight uses wearable device provided in an embodiment of the present invention, glasses group
Camera 120 in part 100 carries out Image Acquisition for scene where user.Camera can acquire in real time according to specified frame per second
Scene image, and the scene image of acquisition is stored.In a kind of situation, first collected scene image can be cut
It takes, for example, interception intermediate region, or the image-region of interception pre-set dimension, then truncated picture is stored.Camera
The image of storage is sent to first processing units 130.
As an implementation, the image that camera 120 is sent can be sent to cloud clothes by first processing units 130
Business device, cloud server carries out target identification to the image, and recognition result is returned to first processing units 130.
As another embodiment, target knowledge is carried out to the image that camera 120 is sent by first processing units 130
Not, for example, the information such as the type of target, color, position can be identified." target " mentioned here can be the movement in environment
Target, such as vehicle, pedestrian, or may be the static target in environment, such as barrier, traffic lights.
For example, which can be filtered, to the progress color space conversion of filtered image, such as from
RGB (Red, Green, Blue, RGB) color space converts that (" Y " indicates brightness, " U " and " V " expression coloration, dense to YUV
Degree) color space, then gray scale extraction can be carried out to the yuv data being converted to, obtained gradation data will be extracted and inputted
In the identification model obtained to preparatory training, the identification model can export target identification as a result, such as target type, face
The information such as color, position.
In a kind of situation, it may include: to obtain sample data to be trained, sample that training, which obtains the process of the identification model,
Data can be image, which can be the image of given resolution or aspect ratio, and sample data may be to image
It is filtered, the data that color space is converted, gray scale obtains after extracting.The sample data is input to the nerve net of preset structure
In network, output neural network based is iterated adjustment to the parameter in neural network, until meeting iteration termination condition
When, the neural network of trained completion is just obtained, that is, obtained the identification model.The neural network of the preset structure can be with
For the neural network based on YOLO (You Only Look Once, you only see once) algorithm, or, or other are used for
The neural network of target identification, specifically without limitation.
In a kind of embodiment, as shown in Fig. 2, eyeglasses assembly 100 can also include the biography being set in lens body 110
Sensor 140;The posture information is sent to the first processing for obtaining the posture information of eyeglasses assembly 100 by sensor 140
Unit 130;First processing units 130 are also used to based on the posture information, and the image sent to camera is calibrated, and is obtained
Take the recognition result that target identification is carried out to the image after calibration.
In a kind of situation, the image after calibration can be sent to cloud server and carry out target by first processing units 130
Identification, and receive the recognition result of cloud server return.In another case, can also be by 130 pairs of schools of first processing units
Image after standard carries out target identification, obtains recognition result.
For example, sensor 140 can be set in apart from the closer position of camera 120, in this way, being based on sensor
The image that the posture information of 140 acquisitions acquires camera 120 is calibrated, and can achieve preferable calibration result.Above-mentioned one
In the case of kind, camera is set on the mirror holder between two eyeglasses of glasses, and in this case, sensor 140 can also be set
It is placed on the mirror holder between two eyeglasses of glasses.Position of the sensor 140 in lens body 110 can be according to the actual situation
It is set, specifically without limitation.
Sensor 140 can be acquired for the attitude transducers such as gyroscope, sensor 140 eyeglasses assembly rotation angle,
The posture informations such as pitch angle, in this way, first processing units 130 can the postures letter such as rotation angle, pitch angle based on eyeglasses assembly
Breath calibrates the image of camera acquisition, obtains the recognition result for carrying out target identification to the image after calibration, obtain
Recognition result is more acurrate.
As an implementation, first processing units 130, image and camera 120 after being also based on calibration
Parameter, calculate the distance between target and camera 120, as the recognition result.
In a kind of situation, camera 120 can be depth camera, for example have the camera of depth transducer, depth
It include depth information in the depth image of camera acquisition, first processing units 130 detect the target in depth image, according to this
The target can be calculated at a distance from camera in the depth information of target and the relevant parameter of depth camera.
In another case, camera can be binocular camera, first processing units 130 are acquired according to binocular camera
Binocular image and binocular camera internal reference and outer ginseng, target can also be calculated at a distance from camera.
In a kind of embodiment, first processing units 130 carry out mesh by the continuous multiple frames image acquired to camera 120
Mark tracking, obtains the movement tendency of target, as the recognition result.
For example, first processing units 130 can carry out target detection in N (N is positive integer) frame image, determine
The region is split as multiple images block, as image block to be processed by region of the target in nth frame image;Then in N+1
It is determining with each highest image block of image block similarity to be processed respectively in frame image, as matching image block;According to wait locate
The moving direction and moving distance between the corresponding matching image block of image block are managed, the movement of target in the picture is obtained and becomes
Gesture;The transformational relation being then based between image coordinate system and physical space coordinates system, by the movement tendency of target in the picture
Be converted to movement tendency of the target in physical space.
As another example, in above embodiment, first processing units 130 can calculate target and camera shooting based on every frame image
The distance of head in a kind of situation, can use this embodiment, will be carried out based on every frame image calculated distance pair
Than to judge that target is to obtain the movement tendency of target close to or far from camera.
First processing units 130 can convert recognition result obtained above to braille control signal, by the braille control
Signal processed is sent to the second processing unit 230 in cuff assemblies 200.The second processing unit 230 receives first processing units 130
The braille of transmission controls signal, by control can in array of protrusions 220 can protrusion unit 221 whether protrusion, output received
The corresponding braille of braille control signal arrived.That is, the second processing unit 230, which is based on braille control signal control, to dash forward
It plays array 220 and exports braille corresponding with recognition result.
For example, recognition result is the information such as type, color, the position of target in a kind of above-mentioned embodiment, then it can protrusion
The braille that array 220 exports can be the information such as type, color, position;In a kind of above-mentioned embodiment, recognition result is target
At a distance from camera, then can array of protrusions 220 export braille can for target at a distance from camera;A kind of above-mentioned implementation
In mode, recognition result is the movement tendency of target, then can the braille that exports of array of protrusions 220 can become for the movement of target
Gesture.The second processing unit 230 can control can array of protrusions 220 according to prefixed time interval, be sequentially output each braille lettering.
It is currently, there are the braille of miscellaneous editions, the braille of miscellaneous editions, user can be stored in advance in first processing units
The braille suitable for itself can be selected in the braille of the miscellaneous editions, the braille that first processing units 130 convert controls letter
Number be user selection the corresponding control signal of braille.Braille control signal can be understood as to can array of protrusions 220 drive
Dynamic signal, the driving signal can drive can in array of protrusions can protrusion unit show the corresponding waveform of braille.
In a kind of above-mentioned situation, can include in array of protrusions can the quantity of protrusion unit can be greater than braille and count, this
In the case of kind, the second processing unit 230 is also used to determine the corresponding braille points of braille control signal, according to the braille point
Number, it is described can determine in array of protrusions it is to be processed can protrusion unit, by control it is described it is to be processed can protrusion unit be
No protrusion exports the received corresponding braille of braille control signal.
As described above, the corresponding braille points of braille control signal can be 6,12,16.With reference to Fig. 3, it is assumed that can protrusion battle array
Include in column can protrusion unit quantity be 16, Fig. 3 in can protrusion unit layout not to dashing forward in the present embodiment
It plays array 220 and constitutes restriction.Assuming that the second processing unit 230 determines that the corresponding braille points of braille control signal are 6, a kind of feelings
Under condition, can in array of protrusions 220 upper left 6 can protrusion unit (frame 1) be determined as it is to be processed can protrusion unit;
In the case of other, can in array of protrusions 220 6 of lower right (frame 2) or upper right side or lower left can dash forward
Rise unit be determined as it is to be processed can protrusion unit;In the case of other, can also determine 6 it is non-conterminous can protrusion unit,
As it is to be processed can protrusion unit, it is to be processed can protrusion unit will namely carry out braille displaying can protrusion unit,
It is to be processed can protrusion unit specific location without limitation.
Except it is to be processed can in addition to protrusion unit other can protrusion unit can be understood as temporarily leaving unused can protrusion list
Member, it is temporarily idle can protrusion unit can be fitted closely with fingerstall, make user be not felt by these it is idle can protrusion unit
Protrusion, such user can more accurately experience the braille of output.
As an implementation, wearable device can also include microphone 300, and microphone 300 is for acquiring voice number
According to the voice data being sent to first processing units 130, first processing units 130 are obtained by parsing the voice data
To instructions to be performed, the recognition result of target identification is carried out to described image based on instructions to be performed obtain.
As shown in figure 4, microphone 300 can be set in lens body 110, for example, can be set in lens body 110
Temple on, the specific location of microphone is without limitation.For example, user can set when by traffic intersection to wearing
Standby inquiry " being red light or green light now ";Microphone collects the voice data, which is sent to the first processing
Unit 130;First processing units 130 identify the voice data by speech recognition algorithm, obtain " identification traffic lights color "
Instruction.First processing units 130 detect " traffic lights " this target in the image that video camera 120 is sent, and then further know
The color of this other target, it is assumed that recognition result is red, then generates red corresponding braille control signal, which is controlled
Signal is sent to the second processing unit 230.
In a kind of situation, user can first say it is default wake up word, microphone collects default after waking up word, wearable device
Working condition is switched to by standby mode.Alternatively, judging the corresponding vocal print of wake-up word after microphone collects default wake-up word
Whether matched with user's vocal print pre-stored in wearable device, if it does, wearable device is switched to work by standby mode
State.
In a kind of embodiment, wearable device can also include key 400, as shown in figure 4, key 400 can be set in
In eyeglasses assembly 100, or in other embodiments, key 400 also be can be set in cuff assemblies 200, alternatively, can also be with
Different keys is set in eyeglasses assembly 100 and cuff assemblies 200, and the specific location of key is without limitation.
In a kind of embodiment, key 400 can be used for opening or closing the wearable device.For example, key 400
It can be mechanical key, or may be fingerprint extracting device etc., specifically without limitation.User can pass through click keys
400, wearable device is switched to working condition by standby mode, or wearable device is switched to standby shape by working condition
State.Alternatively, user can also reset wearable device by click keys 400.
In a kind of embodiment, as shown in figure 4, wearable device can also include voice-output unit 500, a kind of situation
Under, voice-output unit 500 can be earphone, or may be speaker etc., specifically without limitation.In above content, wear
It wears after equipment is switched to working condition by standby mode, voice-output unit 500 can play the response language to above-mentioned wake-up word
Sound promotes user experience to prompt user's wearable device to enter working condition.User, can be with after hearing the response voice
The voice data for carrying instruction is issued to wearable device.
In a kind of embodiment, first processing units 130 carry out the recognition result of target identification obtaining to described image
Afterwards, the recognition result can also be sent to voice-output unit 500;The shape that voice-output unit 500 is also used to voice
Formula exports the recognition result.
In a kind of above-mentioned embodiment, wearable device can also include key 400, and in this embodiment, key 400 can
Think volume button, user passes through the volume that the adjustable voice-output unit 500 of key 400 exports.
In a kind of situation, camera 120, sensor 140, microphone 300, voice-output unit 500, key 400 and
One processing unit 130, which can integrate, to be set on the mirror holder of lens body 110.Alternatively, in another case, microphone 300, language
Sound output unit 500, key 400 and first processing units 130, which can integrate, to be set in the temple of lens body 110.
In a kind of embodiment, eyeglasses assembly 100 can be connect with communication of mobile terminal, and first processing units 130 may be used also
To obtain the interactive information with the mobile terminal, braille control signal is converted by the interactive information.
In a kind of situation, first processing units 130 can be known in the mobile terminal there are in the case where notification message
The type and initiation object of the not described notification message, as the interactive information.
For example, mobile terminal and wearable device can carry out pairing connection in advance, and then wearable device can be examined
The notification message in mobile terminal is measured, the type of the notification message can be incoming information, short message or communication software
Prompt information, etc., specifically without limitation.
For by taking incoming information as an example, first processing units 130, can after detecting that mobile terminal receives incoming information
To identify the initiation object of the incoming information, that is, incoming call connection people, and incite somebody to action " type: incoming information initiates object: so-and-so
People " is used as interactive information.First processing units 130 can convert the interactive information to braille control signal and be sent at second
Unit 230 is managed, the control of the second processing unit 230 can the corresponding braille of the output braille control signal of array of protrusions 220.Alternatively,
The interactive information can also be sent directly to voice-output unit 500 by first processing units 130, voice-output unit 500 with
The form of voice exports the interactive information.
In a kind of above-mentioned embodiment, wearable device can also include key 400, and in this embodiment, key 400 can
To answer or refuse the incoming call in mobile terminal.That is, user can by click keys 400, select answer or
Refuse the incoming call in mobile terminal.For example, " answering " key different with " refusal " two can be set, in a kind of situation, " connect
Listen " " answering " corresponding braille can be set on key, " refusal " corresponding braille can be set on " refusal " key, this
Sample can be convenient user and select.
Alternatively, user can use wearable device operation mobile terminal in the case of other.For example, user can be with
Using the social software in wearable device operation mobile terminal, braille that wearable device is exported by the second processing unit 230 or
The voice that person's voice-output unit 500 exports feeds back the operation of user.
For example, in above embodiment, after obtaining the location information of target, user can be by social software by the position
Information is sent to the friend of user, in this case, mobile terminal can using each contact information in social software as
Interactive information is sent to wearable device, and the first processing units 130 in wearable device can convert braille for the interactive information
Control signal is sent to the second processing unit 230, and the control of the second processing unit 230 can the output braille control of array of protrusions 220
The corresponding braille of signal, so that user selects corresponding contact person, which is sent to selected by user by mobile terminal
Contact person.
In another embodiment, camera 120 can acquire the interface image of mobile terminal, and the interface image is sent out
It send to first processing units 130;First processing units 130 can also be identified described mobile whole by analyzing the interface image
The type and initiation object of notification message present in end, as the interactive information.
In above embodiment, mobile terminal can be interacted with wearable device by way of pairing connection, and this
In embodiment, the camera 120 in wearable device acquires the interface image of mobile terminal, by dividing the interface image
Analysis, obtains the interactive information with mobile terminal.
It as an implementation, can also include being set to fingerstall main body 210 in cuff assemblies 200 with continued reference to Fig. 4
On vibration unit 240, the second processing unit 230 is also used to drive vibration unit after receiving braille control signal
240 vibrations.
The second processing unit 230 control can array of protrusions 220 export braille before, can first drive vibration unit 240
Vibration, to lift user note that cause user pay attention to and then control can array of protrusions 220 output braille, in this way can be more
Effectively interacted with user.
For example, in the case that first processing units 130 have notification message in the terminal, the notice is identified
The type and initiation object of message convert braille control signal for the interactive information, which are controlled as interactive information
Signal is sent to the second processing unit 230;The second processing unit 230 first drives vibration unit 240 to vibrate, to lift user's note
Meaning, then control can array of protrusions 220 export the corresponding braille of braille control signal, that is, the type and hair of output notification message
Play object.The concrete form of vibration is without limitation.
It as an implementation, can also include temperature sensor or GPS (Global in wearable device
Positioning System, global positioning system) sensor or smoke sensor device or other kinds of sensor, no
It enumerates again.Collected heat transfer agent can be sent to first processing units 130, first processing units by these sensors
130, which can convert the heat transfer agent to braille control signal, is sent to the second processing unit 230, and the second processing unit 230 is controlled
System can the corresponding braille of the output braille control signal of array of protrusions 220.Alternatively, first processing units 130 can also be by the biography
Sense information is sent directly to voice-output unit 500, and voice-output unit 500 exports the heat transfer agent in the form of speech.It is logical
Excess temperature sensor can prompt the temperature of user's local environment, by GPS sensor, can prompt the position where user
Information can prompt user in the environment and whether there is more smog by smoke sensor device, in this way can user more
Context aware is preferably carried out fastly.
Using the embodiment of the present invention, the camera being arranged in lens body carries out image for the scene where user and adopts
Collection, first processing units obtain the recognition result that target identification is carried out to image, convert braille control signal for recognition result,
Braille control signal is sent to the second processing unit in cuff assemblies, the second processing unit can be in array of protrusions by control
Can protrusion unit whether protrusion, export the corresponding braille of braille control signal;In this way, even if user's eyesight is poor, user
It can also be obtained to the target identification of scene image by the braille of perception cuff assemblies output as a result, the recognition result can be with
Auxiliary user preferably perceives ambient enviroment;As it can be seen that this programme plays the booster action of the people poor to eyesight.
Wearable device provided in this embodiment can be understood as a kind of AR (Augmented Reality, augmented reality) and set
Standby or a kind of VR (Virtual Reality, virtual reality) equipment, the wearable device can be used as user and external environment
Communication media, the user that eyesight can be assisted poor more rapid and better carries out Context aware.For example, user can be assisted to understand
Moving target, static target, temperature information, location information in environment etc. other environmental factor information.Furthermore it is also possible to auxiliary
Help user using mobile terminal, it can also be by the social software in mobile terminal, by environmental factor information acquired in user
It is shared with other contact persons of the user.
Corresponding with above equipment embodiment, the embodiment of the present invention also provides a kind of exchange method based on wearable device,
This method is applied to wearable device, which includes: eyeglasses assembly and cuff assemblies;Wherein, eyeglasses assembly includes glasses
Main body and the camera being set in lens body and first processing units;Cuff assemblies include fingerstall main body and setting
In in fingerstall main body can array of protrusions and the second processing unit, can array of protrusions include it is multiple can protrusion unit.
For example, camera can be set on the mirror holder between two eyeglasses of lens body, can acquire in this way
To the preferable image in visual angle.First processing units can be set on the leg of spectacles of lens body, influence in this way on user's sight
It is smaller.The position of camera and first processing units in lens body can be set according to the actual situation, not done specifically
It limits.
For example, first processing units and the second processing unit can be IC (integrated circuit, integrated electricity
Road), or may be other processing chips, such as MCU (Microcontroller Unit, micro-control unit) etc., place
Manage the concrete type of unit without limitation.
For example, can array of protrusions can for it is dotted can array of protrusions, or column can array of protrusions, specific shape
Shape is without limitation.Can include in array of protrusions it is multiple can protrusion unit, these can protrusion unit can export braille.Currently, depositing
In the braille, such as 6 brailles, 12 brailles, 16 brailles, etc. of some different points.It, can array of protrusions in a kind of situation
In include can the quantity of protrusion unit can be identical as braille points;In this case, can protrusion unit quantity can be 6
(as shown in fig. 1) can export 6 brailles in this way;Can the quantity of protrusion unit may be 12, can export in this way at 12 points
Braille;Can protrusion unit quantity can also be 16,16 brailles can be exported in this way.
In another case, can include in array of protrusions can the quantity of protrusion unit can be greater than braille and count;For example,
Can protrusion unit quantity can be 16,6 brailles, 12 brailles and 16 brailles can be exported in this way;It for another example, can protrusion
The quantity of unit can be 12, can export 6 brailles and 12 brailles in this way.
Fig. 5 is a kind of flow diagram of exchange method provided in an embodiment of the present invention, comprising:
S501: camera collection image, and the image is sent to the first processing units.
S502: first processing units receive the image, obtain the recognition result for carrying out target identification to the image, will be obtained
The recognition result taken is converted into braille control signal, and the braille control signal that conversion obtains is sent to the second processing unit.
S503: the second processing unit receives the braille that first processing units are sent and controls signal, can protrusion by controlling this
In array can protrusion unit whether protrusion, export the received corresponding braille of braille control signal.
It is illustrated with reference to Fig. 1, eyeglasses assembly and cuff assemblies communicate to connect, for example, eyeglasses assembly and fingerstall group
It can be communicated by wired or wireless mode between part, communication can be Bluetooth communication, or can also be with
For other, specifically without limitation.In this way, the first processing units in eyeglasses assembly can be with the second processing in cuff assemblies
Unit carries out data transmission.
For example, during the poor user of eyesight uses wearable device provided in an embodiment of the present invention, glasses group
Camera in part carries out Image Acquisition for scene where user.Camera can acquire scene figure according to specified frame per second in real time
Picture, and the scene image of acquisition is stored.In a kind of situation, first collected scene image can be intercepted, than
Such as, intermediate region, or the image-region of interception pre-set dimension are intercepted, then truncated picture is stored.Camera will be deposited
The image of storage is sent to first processing units.
As an implementation, the image that camera is sent can be sent to cloud server by first processing units,
Cloud server carries out target identification to the image, and recognition result is returned to first processing units.
As another embodiment, target identification is carried out to the image that camera is sent by first processing units, for example,
It can identify the information such as the type, color, position of target." target " mentioned here can be the moving target in environment, such as
Vehicle, pedestrian etc., or may be the static target in environment, such as barrier, traffic lights.
For example, which can be filtered, to the progress color space conversion of filtered image, such as from
RGB (Red, Green, Blue, RGB) color space converts that (" Y " indicates brightness, " U " and " V " expression coloration, dense to YUV
Degree) color space, then gray scale extraction can be carried out to the yuv data being converted to, obtained gradation data will be extracted and inputted
In the identification model obtained to preparatory training, the identification model can export target identification as a result, such as target type, face
The information such as color, position.
In a kind of situation, it may include: to obtain sample data to be trained, sample that training, which obtains the process of the identification model,
Data can be image, which can be the image of given resolution or aspect ratio, and sample data may be to image
It is filtered, the data that color space is converted, gray scale obtains after extracting.The sample data is input to the nerve net of preset structure
In network, output neural network based is iterated adjustment to the parameter in neural network, until meeting iteration termination condition
When, the neural network of trained completion is just obtained, that is, obtained the identification model.The neural network of the preset structure can be with
For the neural network based on YOLO (You Only Look Once, you only see once) algorithm, or, or other are used for
The neural network of target identification, specifically without limitation.
In a kind of embodiment, as shown in Fig. 2, eyeglasses assembly can also include the sensor being set in lens body;
The method also includes: sensor obtains the posture information of eyeglasses assembly, and the posture information is sent to first processing units.
" obtaining the recognition result that target identification is carried out to described image " in S502 may include: that first processing units are based on the appearance
State information, the image sent to camera are calibrated, and the recognition result that target identification is carried out to the image after calibration is obtained.
In a kind of situation, the image after calibration can be sent to cloud server and carry out target knowledge by first processing units
Not, and the recognition result that cloud server returns is received.In another case, can also by first processing units to calibration after
Image carries out target identification, obtains recognition result.
For example, sensor can be set in apart from the closer position of camera, in this way, the appearance based on sensor acquisition
The image that state information acquires camera is calibrated, and can achieve preferable calibration result.In a kind of above-mentioned situation, camera
It is set on the mirror holder between two eyeglasses of glasses, in this case, two eyeglasses in glasses also can be set in sensor
Between mirror holder on.Position of the sensor in lens body can be set according to the actual situation, specifically without limitation.
Sensor can acquire the rotation angle of eyeglasses assembly, pitch angle for attitude transducers, sensors such as gyroscopes
Equal posture informations, in this way, first processing units can the posture informations such as rotation angle, pitch angle based on eyeglasses assembly, to camera shooting
The image of head acquisition is calibrated, and the recognition result that target identification is carried out to the image after calibration, obtained recognition result are obtained
It is more acurrate.
As an implementation, " the obtaining the recognition result that target identification is carried out to described image " in S502 can wrap
Include: parameter of the first processing units based on image and camera after calibration calculates the distance between target and camera, makees
For the recognition result.
In a kind of situation, camera can be depth camera, for example have the camera of depth transducer, depth camera
It include depth information in the depth image of head acquisition, first processing units detect the target in depth image, according to the target
The target can be calculated at a distance from camera in the relevant parameter of depth information and depth camera.
In another case, camera can be binocular camera, first processing units are acquired according to binocular camera
The internal reference and outer ginseng of binocular image and binocular camera, can also be calculated target at a distance from camera.
As an implementation, " the obtaining the recognition result that target identification is carried out to described image " in S502 can wrap
Include: first processing units carry out target following by the continuous multiple frames image acquired to camera, obtain the movement tendency of target,
As the recognition result.
For example, first processing units can carry out target detection in N (N is positive integer) frame image, determine mesh
It is marked on the region in nth frame image, which is split as multiple images block, as image block to be processed;Then in N+1 frame
It is determining with each highest image block of image block similarity to be processed respectively in image, as matching image block;According to be processed
Moving direction and moving distance between the corresponding matching image block of image block, obtain the movement of target in the picture and become
Gesture;The transformational relation being then based between image coordinate system and physical space coordinates system, by the movement tendency of target in the picture
Be converted to movement tendency of the target in physical space.
As another example, in above embodiment, first processing units can calculate target and camera based on every frame image
Distance, in a kind of situation, this embodiment can be used, will be compared based on every frame image calculated distance,
To judge that target is to obtain the movement tendency of target close to or far from camera.
First processing units can convert recognition result obtained above to braille control signal, which is controlled and is believed
The second processing unit number being sent in cuff assemblies.The second processing unit receives the braille that first processing units are sent and controls letter
Number, by control can in array of protrusions can protrusion unit whether protrusion, it is corresponding to export received braille control signal
Braille.That is, the second processing unit be based on the braille control signal control can array of protrusions output it is opposite with recognition result
The braille answered.
For example, recognition result is the information such as type, color, the position of target in a kind of above-mentioned embodiment, then it can protrusion
The braille of array output can be the information such as type, color, position;In a kind of above-mentioned embodiment, recognition result be target with
The distance of camera, then can array of protrusions output braille can for target at a distance from camera;A kind of above-mentioned embodiment
In, recognition result be target movement tendency, then can array of protrusions output braille can be target movement tendency.At second
Reason unit can control can array of protrusions according to prefixed time interval, be sequentially output each braille lettering.
It is currently, there are the braille of miscellaneous editions, the braille of miscellaneous editions, user can be stored in advance in first processing units
The braille suitable for itself can be selected in the braille of the miscellaneous editions, the braille control signal of first processing units conversion is
For the corresponding control signal of braille of user's selection.Braille control signal can be understood as to can array of protrusions driving signal,
The driving signal can drive can in array of protrusions can protrusion unit show the corresponding waveform of braille.
In a kind of above-mentioned situation, can include in array of protrusions can the quantity of protrusion unit can be greater than braille and count, this
In the case of kind, the method also includes: the second processing unit determines the corresponding braille points of braille control signal, according to described blind
Text points, it is described can determine in array of protrusions it is to be processed can protrusion unit.In S503 " by control described in can protrusion battle array
In column can protrusion unit whether protrusion, export the received corresponding braille of braille control signal " may include: to pass through control
Make it is described it is to be processed can protrusion unit whether protrusion, export the received corresponding braille of braille control signal.
As described above, the corresponding braille points of braille control signal can be 6,12,16.With reference to Fig. 3, it is assumed that can protrusion battle array
Include in column can protrusion unit quantity be 16, Fig. 3 in can protrusion unit layout not to dashing forward in the present embodiment
It plays array and constitutes restriction.Assuming that the second processing unit determines that braille control signal corresponding braille points are 6, in a kind of situation,
Can in array of protrusions upper left 6 can protrusion unit (frame 1) be determined as it is to be processed can protrusion unit;Other feelings
Under condition, can in array of protrusions 6 of lower right (frame 2) or upper right side or lower left can protrusion unit be determined as
To be processed can protrusion unit;In the case of other, can also determine 6 it is non-conterminous can protrusion unit, as to be processed
Can protrusion unit, it is to be processed can protrusion unit will namely carry out braille displaying can protrusion unit, to be processed dashes forward
Play the specific location of unit without limitation.
Except it is to be processed can in addition to protrusion unit other can protrusion unit can be understood as temporarily leaving unused can protrusion list
Member, it is temporarily idle can protrusion unit can be fitted closely with fingerstall, make user be not felt by these it is idle can protrusion unit
Protrusion, such user can more accurately experience the braille of output.
As an implementation, wearable device can also include microphone, the method also includes: microphone acquires language
The voice data is sent to first processing units by sound data, and first processing units are obtained by parsing the voice data
It is instructions to be performed;" obtaining the recognition result that target identification is carried out to described image " in S502 may include: based on it is described to
Execute instruction the recognition result for obtaining and carrying out target identification to described image.
As shown in figure 4, microphone can be set in lens body, for example, can be set in the temple of lens body,
The specific location of microphone is without limitation.For example, user can inquire " existing when by traffic intersection to wearable device
It is being red light or green light ";Microphone collects the voice data, which is sent to first processing units;First
Processing unit identifies the voice data by speech recognition algorithm, obtains the instruction of " identification traffic lights color ".First processing is single
Member detects " traffic lights " this target, then further identifies the color of this target in the image that video camera is sent, it is assumed that
Recognition result is red, then generates red corresponding braille control signal, and braille control signal is sent to second processing list
Member.
In a kind of situation, user can first say it is default wake up word, microphone collects default after waking up word, wearable device
Working condition is switched to by standby mode.Alternatively, judging the corresponding vocal print of wake-up word after microphone collects default wake-up word
Whether matched with user's vocal print pre-stored in wearable device, if it does, wearable device is switched to work by standby mode
State.
In a kind of embodiment, wearable device can also include key, as shown in figure 4, key can be set in glasses group
In part, or in other embodiments, key also be can be set in cuff assemblies, alternatively, in eyeglasses assembly and can also refer to
Different keys is set in set component, the specific location of key is without limitation.
In a kind of embodiment, key can be used for opening or closing the wearable device.For example, key can be
Mechanical key, or may be fingerprint extracting device etc., specifically without limitation.User can will be worn by click keys
It wears equipment and working condition is switched to by standby mode, or wearable device is switched to standby mode by working condition.Alternatively, with
Family can also reset wearable device by click keys 400.
In a kind of embodiment, as shown in figure 4, wearable device can also include voice-output unit, in a kind of situation, language
Sound output unit can be earphone, or may be speaker etc., specifically without limitation.In above content, wearable device by
After standby mode is switched to working condition, voice-output unit can play the response voice to above-mentioned wake-up word, to prompt to use
Family wearable device has entered working condition, promotes user experience.User can send out after hearing the response voice to wearable device
The voice data of instruction is carried out.
In a kind of embodiment, the method also includes: first processing units to carry out target knowledge to described image obtaining
After other recognition result, the recognition result is sent to voice-output unit;Voice-output unit exports in the form of speech
The recognition result.
In a kind of above-mentioned embodiment, wearable device can also include key, and in this embodiment, key can be sound
Measure key, the volume that user passes through the adjustable voice-output unit output of key.
In a kind of situation, camera, sensor, microphone, voice-output unit, key and first processing units can collect
At on the mirror holder for being set to lens body.Alternatively, in another case, microphone, voice-output unit, key and the first processing
Unit, which can integrate, to be set in the temple of lens body.
In a kind of embodiment, eyeglasses assembly can be connect with communication of mobile terminal, the method also includes: the first processing
Unit obtains the interactive information with the mobile terminal, converts braille control signal for the interactive information.In this way, in S501
The braille control signal that interactive information converts can also be sent to described the second processing unit.
In a kind of situation, it may include: the first processing that first processing units, which obtain the interactive information with the mobile terminal,
Unit can there are the type in the case where notification message, identifying the notification message and initiations pair in the mobile terminal
As the interactive information.
For example, mobile terminal and wearable device can carry out pairing connection in advance, and then wearable device can be examined
The notification message in mobile terminal is measured, the type of the notification message can be incoming information, short message or communication software
Prompt information, etc., specifically without limitation.
For by taking incoming information as an example, first processing units, can be with after detecting that mobile terminal receives incoming information
It identifies the initiation object of the incoming information, that is, incoming call connection people, and incites somebody to action " type: incoming information initiates object: someone "
As interactive information.First processing units can convert the interactive information to braille control signal and be sent to second processing list
Member, the second processing unit control can array of protrusions export the corresponding braille of braille control signal.Alternatively, first processing units
The interactive information can be sent directly to voice-output unit, voice-output unit exports interaction letter in the form of speech
Breath.
In a kind of above-mentioned embodiment, wearable device can also include key, and in this embodiment, key can be answered
Or the incoming call in refusal mobile terminal.That is, user can select to answer or refuse mobile whole by click keys
Incoming call in end.For example, " answering " key different with " refusal " two can be set, in a kind of situation, on " answering " key
" answering " corresponding braille can be set, " refusal " corresponding braille can be set on " refusal " key, in this way, can be square
Just user selects.
Alternatively, user can use wearable device operation mobile terminal in the case of other.For example, user can be with
Using the social software in wearable device operation mobile terminal, the braille or language that wearable device is exported by the second processing unit
The voice of sound output unit output feeds back the operation of user.
For example, in above embodiment, after obtaining the location information of target, user can be by social software by the position
Information is sent to the friend of user, in this case, mobile terminal can using each contact information in social software as
Interactive information is sent to wearable device, and the first processing units in wearable device can convert the interactive information to braille control
Signal is sent to the second processing unit, the second processing unit control can array of protrusions to export braille control signal corresponding blind
Text, so that user selects corresponding contact person, which is sent to the selected contact person of user by mobile terminal.
In another embodiment, the method also includes: camera acquires the interface image of mobile terminal, by the boundary
Face image is sent to first processing units;" first processing units obtain the interactive information with the mobile terminal " may include:
First processing units can also identify the type of notification message present in the mobile terminal by analyzing the interface image
And object is initiated, as the interactive information.
In above embodiment, mobile terminal can be interacted with wearable device by way of pairing connection, and this
In embodiment, the interface image of the camera acquisition mobile terminal in wearable device, by analyzing the interface image,
Obtain the interactive information with mobile terminal.
It as an implementation, can also include the vibration being set in fingerstall main body in cuff assemblies with continued reference to Fig. 4
Moving cell;The method also includes: the second processing unit driving vibration unit vibration after receiving the braille control signal.
The second processing unit control can array of protrusions output braille before, first vibration unit can be driven to vibrate, to mention
Rise user note that cause user pay attention to and then control can array of protrusions export braille, in this way can more effectively with user
It interacts.
For example, in the case that first processing units have notification message in the terminal, identify that the notice disappears
The type and initiation object of breath convert braille control signal for the interactive information, which are controlled and is believed as interactive information
Number it is sent to the second processing unit;The second processing unit first drives vibration unit to vibrate, to lift user note that controlling can dash forward again
Play the corresponding braille of array output braille control signal, that is, the type and initiation object of output notification message.The tool of vibration
Body form is without limitation.
It as an implementation, can also include temperature sensor or GPS (Global in wearable device
Positioning System, global positioning system) sensor or smoke sensor device or other kinds of sensor, no
It enumerates again.Collected heat transfer agent can be sent to first processing units by these sensors, and first processing units can
It is sent to the second processing unit to convert braille control signal for the heat transfer agent, the second processing unit control can array of protrusions
Export the corresponding braille of braille control signal.Alternatively, the heat transfer agent can also be sent directly to language by first processing units
Sound output unit, voice-output unit export the heat transfer agent in the form of speech.By temperature sensor, user can be prompted
The temperature of local environment can prompt the location information where user by GPS sensor, pass through smoke sensor device, Ke Yiti
Show user in the environment and whether there is more smog, in this way can user more rapid and better carry out Context aware.
Using the embodiment of the present invention, the camera being arranged in lens body carries out image for the scene where user and adopts
Collection, first processing units obtain the recognition result that target identification is carried out to image, convert braille control signal for recognition result,
Braille control signal is sent to the second processing unit in cuff assemblies, the second processing unit can be in array of protrusions by control
Can protrusion unit whether protrusion, export the corresponding braille of braille control signal;In this way, even if user's eyesight is poor, user
It can also be obtained to the target identification of scene image by the braille of perception cuff assemblies output as a result, the recognition result can be with
Auxiliary user preferably perceives ambient enviroment;As it can be seen that this programme plays the booster action of the people poor to eyesight.
Wearable device provided in this embodiment can be understood as a kind of AR (Augmented Reality, augmented reality) and set
Standby or a kind of VR (Virtual Reality, virtual reality) equipment, the wearable device can be used as user and external environment
Communication media, the user that eyesight can be assisted poor more rapid and better carries out Context aware.For example, user can be assisted to understand
Moving target, static target, temperature information, location information in environment etc. other environmental factor information.Furthermore it is also possible to auxiliary
Help user using mobile terminal, it can also be by the social software in mobile terminal, by environmental factor information acquired in user
It is shared with other contact persons of the user.
It should be understood by those ordinary skilled in the art that: the discussion of any of the above embodiment is exemplary only, not
It is intended to imply that the scope of the present disclosure (including claim) is limited to these examples;Under thinking of the invention, above embodiments
Or can also be combined between the technical characteristic in different embodiments, step can be realized with random order, and be existed such as
Many other variations of the upper different aspect of the invention, for simplicity, they are not provided in details.
In addition, to simplify explanation and discussing, and in order not to obscure the invention, it can in provided attached drawing
It is connect with showing or can not show with the well known power ground of integrated circuit (IC) chip and other components.Furthermore, it is possible to
Device is shown in block diagram form, to avoid obscuring the invention, and this has also contemplated following facts, i.e., about this
The details of the embodiment of a little block diagram arrangements be height depend on will implementing platform of the invention (that is, these details should
It is completely within the scope of the understanding of those skilled in the art).Elaborating that detail (for example, circuit) is of the invention to describe
In the case where exemplary embodiment, it will be apparent to those skilled in the art that can be in these no details
In the case where or implement the present invention in the case that these details change.Therefore, these descriptions should be considered as explanation
Property rather than it is restrictive.
Although having been incorporated with specific embodiments of the present invention, invention has been described, according to retouching for front
It states, many replacements of these embodiments, modifications and variations will be apparent for those of ordinary skills.Example
Such as, discussed embodiment can be used in other memory architectures (for example, dynamic ram (DRAM)).
The embodiment of the present invention be intended to cover fall into all such replacements within the broad range of appended claims,
Modifications and variations.Therefore, all within the spirits and principles of the present invention, any omission, modification, equivalent replacement, the improvement made
Deng should all be included in the protection scope of the present invention.
Claims (13)
1. a kind of wearable device characterized by comprising eyeglasses assembly and cuff assemblies;Wherein, the eyeglasses assembly includes eye
Mirror main body and the camera being set in the lens body and first processing units;The cuff assemblies include fingerstall master
Body and be set in the fingerstall main body can array of protrusions and the second processing unit, it is described can array of protrusions include multiple
It can protrusion unit;
The camera is sent to the first processing units for acquiring image, and by described image;
The first processing units, for receiving described image, acquisition carries out the recognition result of target identification to described image, will
Acquired recognition result is converted into braille control signal, and the braille control signal that conversion obtains is sent to the second processing
Unit;
Described the second processing unit controls signal for receiving the braille that the first processing units are sent, by described in control
Can in array of protrusions can protrusion unit whether protrusion, export the received corresponding braille of braille control signal.
2. wearable device according to claim 1, which is characterized in that described the second processing unit is also used to determine braille
Control signal corresponding braille points, counted according to the braille, it is described can determine in array of protrusions it is to be processed can protrusion
Unit, by control it is described it is to be processed can protrusion unit whether protrusion, it is corresponding to export received braille control signal
Braille.
3. wearable device according to claim 1, which is characterized in that the eyeglasses assembly further includes being set to the glasses
Sensor in main body;The posture information is sent to by the sensor for obtaining the posture information of the eyeglasses assembly
The first processing units;
The first processing units are also used to calibrate described image based on the posture information, after obtaining to calibration
The recognition result of image progress target identification.
4. wearable device according to claim 3, which is characterized in that the first processing units are also used to based on described
The parameter of image and the camera after calibration calculates the distance between target and the camera, as the identification
As a result.
5. wearable device according to claim 1, which is characterized in that the first processing units are also used to by institute
The continuous multiple frames image for stating camera acquisition carries out target following, the movement tendency of target is obtained, as the recognition result.
6. wearable device according to claim 1, which is characterized in that the wearable device further includes microphone, the wheat
The voice data is sent to the first processing units for acquiring voice data by gram wind;
The first processing units, are also used to by parsing the voice data, obtain it is instructions to be performed, based on described pending
Instruction obtains the recognition result.
7. wearable device according to claim 1, which is characterized in that the eyeglasses assembly is connect with communication of mobile terminal,
The first processing units are also used to obtain the interactive information with the mobile terminal, convert braille for the interactive information
Control signal.
8. wearable device according to claim 7, which is characterized in that the first processing units are also used in the shifting
There are the type in the case where notification message, identifying the notification message and initiation objects in dynamic terminal, as the interaction letter
Breath.
9. wearable device according to claim 8, which is characterized in that the camera is also used to acquire described mobile whole
The interface image is sent to the first processing units by the interface image at end;
The first processing units are also used to identify notice present in the mobile terminal by analyzing the interface image
The type and initiation object of message, as the interactive information.
10. wearable device according to claim 1, which is characterized in that the cuff assemblies further include being set to the finger
The vibration unit in main body is covered, described the second processing unit is also used to after receiving the braille control signal described in driving
Vibration unit vibration.
11. wearable device according to claim 7, which is characterized in that the wearable device further includes key;
The key, for opening or closing the wearable device;
And/or the key, for answering or refusing the incoming call in the mobile terminal.
12. wearable device according to claim 1, which is characterized in that the wearable device further includes voice-output unit;
The first processing units are also used to after obtaining the recognition result for carrying out target identification to described image, by the knowledge
Other result is sent to the voice-output unit;
The voice-output unit is also used to export the recognition result in the form of speech.
13. a kind of exchange method based on wearable device, which is characterized in that be applied to wearable device, the wearable device includes:
Eyeglasses assembly and cuff assemblies;Wherein, the eyeglasses assembly includes lens body and is set to taking the photograph in the lens body
As head and first processing units;The cuff assemblies include fingerstall main body and be set in the fingerstall main body can protrusion
Array and the second processing unit, it is described can array of protrusions include it is multiple can protrusion unit;The described method includes:
The camera collection image, and described image is sent to the first processing units;
The first processing units receive described image, obtain the recognition result for carrying out target identification to described image, will be obtained
The recognition result taken is converted into braille control signal, and the braille control signal that conversion obtains is sent to the second processing list
Member;
Described the second processing unit receives the braille control signal that the first processing units are sent, by can protrusion described in control
In array can protrusion unit whether protrusion, export the received corresponding braille of braille control signal.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910470231.1A CN110208946A (en) | 2019-05-31 | 2019-05-31 | A kind of wearable device and the exchange method based on wearable device |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910470231.1A CN110208946A (en) | 2019-05-31 | 2019-05-31 | A kind of wearable device and the exchange method based on wearable device |
Publications (1)
Publication Number | Publication Date |
---|---|
CN110208946A true CN110208946A (en) | 2019-09-06 |
Family
ID=67790068
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201910470231.1A Pending CN110208946A (en) | 2019-05-31 | 2019-05-31 | A kind of wearable device and the exchange method based on wearable device |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN110208946A (en) |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110728886A (en) * | 2019-10-30 | 2020-01-24 | 京东方科技集团股份有限公司 | Braille learning system, fingertip sensor and forming method thereof |
CN110989832A (en) * | 2019-11-21 | 2020-04-10 | 维沃移动通信有限公司 | Control method and electronic equipment |
CN111942285A (en) * | 2020-07-10 | 2020-11-17 | 夏牧谣 | Intelligent vision-impaired person service method and system based on vehicle-mounted glass vibration feedback |
CN113449549A (en) * | 2020-03-25 | 2021-09-28 | 中移(成都)信息通信科技有限公司 | Prompt message generation method, device, equipment and storage medium |
CN113534715A (en) * | 2021-07-21 | 2021-10-22 | 歌尔科技有限公司 | Intelligent wearable device, and control method and system of target device |
Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103514450A (en) * | 2012-06-29 | 2014-01-15 | 华为技术有限公司 | Image feature extracting method and image correcting method and equipment |
CN106506954A (en) * | 2016-10-25 | 2017-03-15 | 深圳市高巨创新科技开发有限公司 | Image pickup method and system based on multi-rotor aerocraft |
CN107328424A (en) * | 2017-07-12 | 2017-11-07 | 三星电子(中国)研发中心 | Air navigation aid and device |
CN108279780A (en) * | 2018-03-01 | 2018-07-13 | 京东方科技集团股份有限公司 | Wearable device and control method |
CN108616744A (en) * | 2017-01-23 | 2018-10-02 | 上海爱观视觉科技有限公司 | A kind of bionical binocular vision calibration system and calibration method |
CN109674628A (en) * | 2019-01-29 | 2019-04-26 | 桂林电子科技大学 | A kind of intelligent glasses |
KR101975761B1 (en) * | 2018-07-02 | 2019-05-07 | 이단경 | Braille input and output system for mobile device |
-
2019
- 2019-05-31 CN CN201910470231.1A patent/CN110208946A/en active Pending
Patent Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103514450A (en) * | 2012-06-29 | 2014-01-15 | 华为技术有限公司 | Image feature extracting method and image correcting method and equipment |
CN106506954A (en) * | 2016-10-25 | 2017-03-15 | 深圳市高巨创新科技开发有限公司 | Image pickup method and system based on multi-rotor aerocraft |
CN108616744A (en) * | 2017-01-23 | 2018-10-02 | 上海爱观视觉科技有限公司 | A kind of bionical binocular vision calibration system and calibration method |
CN107328424A (en) * | 2017-07-12 | 2017-11-07 | 三星电子(中国)研发中心 | Air navigation aid and device |
CN108279780A (en) * | 2018-03-01 | 2018-07-13 | 京东方科技集团股份有限公司 | Wearable device and control method |
KR101975761B1 (en) * | 2018-07-02 | 2019-05-07 | 이단경 | Braille input and output system for mobile device |
CN109674628A (en) * | 2019-01-29 | 2019-04-26 | 桂林电子科技大学 | A kind of intelligent glasses |
Cited By (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110728886A (en) * | 2019-10-30 | 2020-01-24 | 京东方科技集团股份有限公司 | Braille learning system, fingertip sensor and forming method thereof |
CN110728886B (en) * | 2019-10-30 | 2021-10-08 | 京东方科技集团股份有限公司 | Braille learning system, fingertip sensor and forming method thereof |
CN110989832A (en) * | 2019-11-21 | 2020-04-10 | 维沃移动通信有限公司 | Control method and electronic equipment |
CN113449549A (en) * | 2020-03-25 | 2021-09-28 | 中移(成都)信息通信科技有限公司 | Prompt message generation method, device, equipment and storage medium |
CN111942285A (en) * | 2020-07-10 | 2020-11-17 | 夏牧谣 | Intelligent vision-impaired person service method and system based on vehicle-mounted glass vibration feedback |
CN111942285B (en) * | 2020-07-10 | 2022-11-18 | 夏牧谣 | Intelligent vision-impaired person service method and system based on vehicle-mounted glass vibration feedback |
CN113534715A (en) * | 2021-07-21 | 2021-10-22 | 歌尔科技有限公司 | Intelligent wearable device, and control method and system of target device |
CN113534715B (en) * | 2021-07-21 | 2023-01-24 | 歌尔科技有限公司 | Intelligent wearable device, and control method and system of target device |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110208946A (en) | A kind of wearable device and the exchange method based on wearable device | |
CN108416235B (en) | The anti-peeping method, apparatus of display interface, storage medium and terminal device | |
CN107231529A (en) | Image processing method, mobile terminal and storage medium | |
CN103839054B (en) | Multi-functional mobile intelligent terminal sensor supporting iris recognition | |
CN109804357A (en) | Server, client, control method and storage medium | |
CN108701214A (en) | Image processing method, device and equipment | |
CN108700982A (en) | Information processing equipment, information processing method and program | |
CN109240572A (en) | A kind of method obtaining picture, method and device that picture is handled | |
CN107302662A (en) | A kind of method, device and mobile terminal taken pictures | |
WO2022227393A1 (en) | Image photographing method and apparatus, electronic device, and computer readable storage medium | |
CN111601088A (en) | Sitting posture monitoring system based on monocular camera sitting posture identification technology | |
CN108989678A (en) | A kind of image processing method, mobile terminal | |
CN105763829A (en) | Image processing method and electronic device | |
US11837233B2 (en) | Information processing device to automatically detect a conversation | |
US10843299B2 (en) | Object recognition and presentation for the visually impaired | |
US20230079582A1 (en) | Image processing method and apparatus, terminal, and storage medium | |
CN110533651B (en) | Image processing method and device | |
CN104535059A (en) | Indoor positioning system specific to totally blind population | |
CN110807405A (en) | Detection method of candid camera device and electronic equipment | |
CN110222567B (en) | Image processing method and device | |
CN112286347A (en) | Eyesight protection method, device, storage medium and terminal | |
CN116030512A (en) | Gaze point detection method and device | |
CN105227828B (en) | Filming apparatus and method | |
CN109639981A (en) | A kind of image capturing method and mobile terminal | |
CN110602384B (en) | Exposure control method and electronic device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination |