WO2015174597A1 - 음성제어 영상표시 장치 및 영상표시 장치의 음성제어 방법 - Google Patents
음성제어 영상표시 장치 및 영상표시 장치의 음성제어 방법 Download PDFInfo
- Publication number
- WO2015174597A1 WO2015174597A1 PCT/KR2014/011197 KR2014011197W WO2015174597A1 WO 2015174597 A1 WO2015174597 A1 WO 2015174597A1 KR 2014011197 W KR2014011197 W KR 2014011197W WO 2015174597 A1 WO2015174597 A1 WO 2015174597A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- voice
- control
- identification
- voice data
- user
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 52
- 230000010365 information processing Effects 0.000 claims abstract description 37
- 238000013507 mapping Methods 0.000 claims 1
- 238000010586 diagram Methods 0.000 description 3
- 239000011159 matrix material Substances 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 238000013459 approach Methods 0.000 description 1
- 230000015572 biosynthetic process Effects 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 239000004984 smart glass Substances 0.000 description 1
- 238000003786 synthesis reaction Methods 0.000 description 1
- 230000001755 vocal effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0481—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0484—Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
- G06F3/04842—Selection of displayed objects or displayed text elements
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0484—Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
- G06F3/04845—Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range for image manipulation, e.g. dragging, rotation, expansion or change of colour
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/16—Sound input; Sound output
- G06F3/167—Audio in a user interface, e.g. using voice commands for navigating, audio feedback
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L17/00—Speaker identification or verification techniques
- G10L17/06—Decision making techniques; Pattern matching strategies
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L17/00—Speaker identification or verification techniques
- G10L17/22—Interactive procedures; Man-machine interfaces
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2203/00—Indexing scheme relating to G06F3/00 - G06F3/048
- G06F2203/048—Indexing scheme relating to G06F3/048
- G06F2203/04806—Zoom, i.e. interaction techniques or interactors for controlling the zooming operation
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L17/00—Speaker identification or verification techniques
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
- G10L2015/223—Execution procedure of a spoken command
Definitions
- the present invention relates to a voice control image display apparatus and a voice control method of a video display apparatus, and more particularly, to a voice control apparatus and a video display apparatus which compares identification voice data assigned for each execution unit area on a screen displayed through a display unit, When there is identification voice data corresponding to the voice of the video display device, the input voice signal is generated in the execution unit area to which the identification voice data is assigned, and a voice control method of the video display device.
- a variety of input methods have been developed for controlling the image display device, such as a mouse, a keyboard, a touch pad, and a button-type remote controller, Remote control, touch screen, and so on.
- a voice control method for controlling a video display device by recognizing a voice of a user in order to control the video display device more easily has recently been spotlighted.
- the voice control method for controlling the video display device by recognizing the voice uttered by the user is disadvantageous in that the recognition rate is lowered due to the difference in oral structure and pronunciation of each user, and the user has to learn the voice command stored in the database Have been pointed out as problems. That is, a voice control method satisfactory level in terms of the user's convenience has not been shown yet.
- the present invention compares identification voice data allocated to each execution unit area on a screen displayed through a display unit with a voice of an input user, and when there is identification voice data corresponding to a voice of a user,
- a voice control image display device configured to generate an input signal in an area of a voice display area and to enable convenience and intuitiveness of a user experience UX of a conventional touch screen control scheme to be applied to voice control, The purpose is to provide.
- an image display apparatus having a display unit and capable of voice control, wherein a database in which identification voice data is allocated and mapped by execution unit areas on the screen displayed through the display unit A stored memory unit; A voice recognition unit for receiving a voice of a user; An information processing unit for searching the database when the voice recognition unit receives the user's voice and determining whether there is identification voice data corresponding to the voice of the user; And a control unit for generating an input signal in the execution unit area to which the identification voice data is assigned when identification sound data corresponding to the voice of the user exists as a result of the determination by the information processing unit A display device is provided.
- the display unit may also display identification sound data allocated for each execution unit area on the screen when the screen is displayed.
- the database may be characterized in that identification sound data is allocated and mapped to each of two or more different screens displayed on the display unit for each execution unit area on each screen.
- the data base further stores control voice data corresponding to a control command for performing a specific screen control on the basis of the execution unit area to which the identification voice data is allocated when being combined with the identification voice data,
- the voice recognition unit receives the voice of the user, searches the database to determine whether there exists identification voice data and control voice data corresponding to the voice of the user, and the control unit determines When the identification voice data and the control voice data corresponding to the voice are present, generates an execution unit area input signal to which the identification voice data is allocated, And It can be made.
- the present invention provides a voice control method for a video display device performed in the voice control video display device, the voice control method comprising the steps of: (a) assigning identification voice data for each execution unit area on a screen displayed through the display unit, Storing the database; (b) receiving the voice of the user through the voice recognition unit; (c) the information processing unit searches the database to determine whether identification voice data corresponding to the user's voice exists; And (d) generating an input signal in the execution unit area to which the identification voice data is allocated if the identification voice data corresponding to the voice of the user exists as a result of the determination by the information processing unit And a voice control method of the video display device.
- the step (b) may be performed in a state in which the voice data allocated to each execution unit area on the screen displayed on the display unit is visible.
- the step (a) may further include control voice data corresponding to a control command for performing a specific screen control based on the execution unit area to which the identification voice data is allocated when the memory unit is used in combination with the identification voice data
- the information processing unit searches the database and determines whether identification voice data and control voice data corresponding to the voice of the user are present, If the identification sound data and the control sound data corresponding to the user's voice exist as a result of the determination by the information processing unit, the control unit generates an execution unit area input signal to which the identification sound data is allocated, To the control speech data on the basis of the execution unit area generating the control speech data Which it can be characterized in that adapted to be performed in such a manner as to execute the control command.
- the voice control image display apparatus and the voice control method of the image display apparatus according to the present invention have the following effects.
- the input control is performed by comparing the voice data allocated to the execution unit area displayed on the display unit with the voice of the input user so that the existing touch screen input control method is directly applied to the voice control method To provide a simple and intuitive voice control.
- FIG. 1 is a schematic diagram of a touch screen.
- FIG. 2 is a general home screen of an Android smartphone displayed through a display unit of a voice control video display device according to the present invention.
- FIG. 3 is an application screen which is displayed when 'apps' (2) are touched on the home screen of Fig.
- FIG 4 is an exemplary configuration of an execution unit area on the screen displayed through the display unit of the audio control video display device according to the present invention.
- FIG. 5 is a diagram illustrating an example of a database stored in the memory unit of the audio control video display apparatus according to the present invention.
- FIG. 6 is a diagram illustrating an example of an audio-visual display apparatus according to an embodiment of the present invention. Referring to FIG. 6, Is assigned as identification voice data unique to each execution unit area.
- FIG. 7 is an embodiment in which identification voice data and control voice data are combined and used in a voice control video display device according to the present invention.
- FIG. 8 is a flowchart of a method of controlling a sound of an image display apparatus according to the present invention.
- a video display device having a display unit and capable of voice control, comprising: a memory unit for storing identification data of voice data assigned to each execution unit area displayed on the display unit and storing a mapped database; A voice recognition unit for receiving a voice of a user; An information processing unit for searching the database when the voice recognition unit receives the user's voice and determining whether there is identification voice data corresponding to the voice of the user; And a control unit for generating an input signal in the execution unit area to which the identification voice data is assigned when identification sound data corresponding to the voice of the user exists as a result of the determination by the information processing unit As a display device,
- the data base further stores control voice data corresponding to a control command for performing a specific screen control on the basis of the execution unit area to which the identification voice data is allocated when being combined with the identification voice data,
- the voice recognition unit receives the voice of the user, searches the database to determine whether there exists identification voice data and control voice data corresponding to the voice of the user, and the control unit determines that the voice of the user
- the control unit generates an execution unit area input signal to which the identification voice data is assigned when the identification voice data and the control voice data corresponding to the control voice data are present, Characterized in that It is a voice controlled video display device.
- a method for controlling a voice display of an image display device performed by the voice control video display device comprising the steps of: (a) assigning identification voice data to a memory unit of a display unit through the display unit; Storing; (b) receiving the voice of the user through the voice recognition unit; (c) the information processing unit searches the database to determine whether identification voice data corresponding to the user's voice exists; And (d) if an identification voice data corresponding to the voice of the user exists as a result of the determination by the information processing unit, the control unit generates an input signal in the execution unit area to which the identification voice data is allocated ,
- step (a) further comprises adding control voice data corresponding to a control command for performing a specific screen control based on the execution unit area to which the identification voice data is allocated when the memory unit is used in combination with the identification voice data
- step (c) is performed by the information processing unit searching the database to determine whether identification voice data and control voice data corresponding to the voice of the user exist, if the identification voice data and the control voice data corresponding to the user's voice exist as a result of the determination by the information processing unit, the control unit generates an execution unit area input signal to which the identification voice data is allocated, And generating a control speech data corresponding to the control speech data based on the generated execution unit area And the control command is executed in a manner of executing the control command.
- a voice control video display apparatus includes a display unit, a memory unit for storing a database mapped with identification voice data allocated to execution unit areas on the screen displayed through the display unit, An information processing unit for searching the database and determining whether identification voice data corresponding to the user's voice exists when the voice recognition unit receives the voice of the user; And a control unit for generating an input signal in the execution unit area to which the identified voice data is assigned when the identified voice data exists.
- the audio control video display device having the above-described structure can be applied to a wide variety of devices such as a smart phone, a tablet PC, a smart TV, a navigation device, a wearable device such as a smart glass, a smart watch and a virtual reality headset (VR device) Devices, and a beam projector equipped with a mobile operating system and capable of voice control, can be implemented in all image display devices capable of implementing voice control.
- devices such as a smart phone, a tablet PC, a smart TV, a navigation device, a wearable device such as a smart glass, a smart watch and a virtual reality headset (VR device) Devices, and a beam projector equipped with a mobile operating system and capable of voice control, can be implemented in all image display devices capable of implementing voice control.
- input control methods of touch screens widely used in smart phones, tablet PCs, and the like are vacuum pressure type and electrostatic type, and the pressure value of the pressure applied portion of the touch screen is measured
- the electrostatic type is a method in which a sensor attached to the four corners of the touch screen senses an electronic change of a touch portion and measures coordinate values and generates an input signal at the corresponding portion.
- the touch screen method is an intuitive input method in a GUI (Graphic User Interface) environment and is highly user-friendly.
- the present invention is characterized in that the advantages of the touch screen method can be applied to voice control through an entirely different approach from the conventional voice control method performed in a manner of 1: 1 correspondence between voice commands and specific execution contents .
- the execution unit area refers to a touch screen input method in which, as a concept corresponding to a contact surface where a touch screen and a touch means (for example, a finger, an electrostatic pen, etc.) And the range in which the execution signal is generated.
- it means a predetermined area consisting of a large number of pixels, and one icon arrangement area that can be regarded as an area that gives the same result regardless of which pixel on the area has an input signal or an execution signal, And so on.
- the matrix-type lattice regions in which the shortcut icons of various applications are arranged on the screen displayed on the display unit of the smartphone are exemplified, Is a variable concept that can vary in size, number, shape, and arrangement.
- the memory unit is implemented as a memory chip incorporated in a voice control image display device implemented by a smart phone, a tablet PC, or the like.
- the database is mapped with identification sound data allocated to each execution unit area on the screen displayed through the display unit.
- the database includes unique coordinate information assigned to each area recognized as the same execution unit area on the screen .
- the identification voice data may utilize data recorded by the user in order to improve the voice recognition rate in consideration of the oral structure and vocal characteristics of the user.
- the memory unit may previously store a format for each distribution pattern of the execution unit area of the basic screen displayed through the display unit, so that the user may select a specific format.
- the voice recognition unit is implemented as a microphone device and a voice recognition circuit built in a voice control video display device implemented as a smart phone, a tablet PC, or the like as a part for receiving voice of a user.
- the information processing unit and the control unit are implemented as a control circuit unit including a CPU and a RAM incorporated in a voice control video display device implemented as a smart phone, a tablet PC, or the like.
- the voice recognition unit receives the voice of the user
- the information processing unit searches the database to determine whether there is identification voice data corresponding to the voice of the user. Specifically, And when the identification voice data exists, the unique coordinate information of the execution unit area to which the identification voice data is assigned is detected.
- the control unit plays an role of generating an input signal in the execution unit area to which the identification sound data is allocated, And generates an input signal in an area on the screen having the detected coordinate information.
- the result of the generation of the input signal depends on the content of the execution unit area. If a shortcut icon of a specific application exists in the execution unit area, the application will be executed. If a specific character of the virtual keyboard is present in the execution unit area, the specific character will be input. If a command is specified, the command will be executed, and in some cases there may be no execution.
- FIG. 2 is a general home screen of the Android smartphone.
- FIG. 3 is an application screen displayed when the 'apps' (2) are touched on the home screen. If you want to run 'abc' (1) application on the application screen instead of the home screen by touch screen operation, touch the 'apps' (2) on the bottom right of the home screen, and if the application screen is displayed as' abc '(3).
- such a process can be implemented by a voice control method.
- an execution unit area on the screen displayed through the display unit is divided, and the database includes an execution unit for each screen including a home screen and an application screen, as shown in FIG. 5
- ID sound data is mapped to the execution unit area 'F4' on the home screen displayed by the library 1, and the execution unit area 'C1' on the application screen displayed by the library 2 is mapped.
- the home screen is displayed on the display unit and the voice of the user named 'apps' is inputted through the voice recognition unit, the information processing unit displays the data on the home screen Searches for a base and determines whether there exists identification voice data corresponding to the voice of the user " abs " .
- the control unit When the information processing unit searches for 'apps' which are identification voice data corresponding to the user's voice 'apps', the control unit generates an input signal to 'F4' which is the execution unit area to which the identification voice data is allocated. As a result, the application screen is executed.
- the information processor searches the database for the application screen and displays the voice of the user 'abc' It is determined whether the corresponding identification voice data exists.
- the control unit When the information processing unit searches for 'abc' which is the identification voice data corresponding to the voice of the user 'abc', the control unit generates an input signal to the execution unit area 'C1' to which the identification voice data is allocated. As a result, an application called abc is executed.
- the database is mapped by assigning identification voice data to each of two or more different screens displayed on the display unit by execution unit areas on each screen.
- the structure of the database in this manner is preferable when the name of the icon displayed in the execution unit area on each screen is clear and the name of the icon is used as identification voice data.
- the same database may be used. For example, as in the embodiments of FIGS. 2 to 5, when each screen displayed through the display unit has a 6 ⁇ 4 matrix of execution units, It is conceivable to assign alphabetical characters in alphabetical order from the execution unit area existing in the execution unit area to the identification voice data unique to each execution unit area.
- This type of database configuration is preferable when it is effective to have consistent identification voice data regardless of the change of the screen, for example, because the name of the icon displayed in the execution unit area on each screen is not clear.
- the display unit is preferably configured to display identification sound data allocated for each execution unit area on the screen when the screen is displayed. Specifically, a method of blurring the unique identification voice data of each execution unit area on the screen as a background may be considered.
- the data base further stores control voice data corresponding to a control command for performing a specific screen control on the basis of the execution unit area to which the identification voice data is allocated when being combined with the identification voice data,
- the voice recognition unit receives the voice of the user, searches the database to determine whether there exists identification voice data and control voice data corresponding to the voice of the user, and the control unit determines that the voice of the user
- the control unit generates an execution unit area input signal to which the identification voice data is assigned when the identification voice data and the control voice data corresponding to the control voice data are present, Characterized in that Can.
- FIG. 7 shows a specific embodiment in which identification voice data and control voice data are used in combination.
- the screen displayed through the display unit is divided into execution unit areas each having a matrix of 6.times.4.
- identification sound data in alphabetical order from the execution unit area at the upper right of the screen Quot ;
- the control voice data for 'Zoom-In' is additionally stored in the database as a control command for enlarging the screen.
- the control unit enlarges and displays the execution unit area F (the portion corresponding to the second row and the second column) do.
- the order of inputting the identification voice data and the control voice data may be set arbitrarily.
- the audio control video display device may be the first device, and another device which is incapable of voice control or inconvenient may be the second device, and mirroring may be performed. Through the mirroring, the voice control system implemented in the voice control video display device can be used for controlling other devices.
- the second device may be a connected car infotainment system installed in a car, a smart TV, or the like.
- the information processing unit when it is difficult to identify the information of the second device in displaying the control interface of the second device through the audio control video display device as the first device, only the text of the control information among the video signal of the second device and the control information
- the information processing unit generates each text as identification voice data through text-based speech synthesis and maps the identification voice data for each execution unit area to generate a database Only the text of the control information is displayed on the screen displayed through the display unit, so that the user can use the text of the control information displayed on the display unit as a voice command word.
- the information of the second device is scaled when it is sent to the voice control video display device as the first device, Level information can be transmitted.
- a voice control method for an image display device performed in the voice control video display device comprising the steps of: (a) Storing a database; (b) receiving the voice of the user through the voice recognition unit; (c) the information processing unit searches the database to determine whether identification voice data corresponding to the user's voice exists; And (d) generating an input signal in the execution unit area to which the identification voice data is allocated if the identification voice data corresponding to the voice of the user exists as a result of the determination by the information processing unit And a voice control method of the video display device.
- the voice control method of the video display device is assumed to be performed in the voice control video display device according to the present invention described above.
- FIG. 8 is a flowchart of a method of controlling a sound of a video display device according to the present invention.
- the memory unit is configured to construct a database, and the database is mapped with identification voice data allocated to each execution unit area on the screen displayed on the display unit.
- the identification voice data includes unique coordinate information assigned for each area recognized as the same execution unit area on the screen.
- the identification voice data includes user's direct recorded data .
- the memory unit may previously store a format for each distribution pattern of the execution unit area of a basic screen displayed through the display unit, so that a user can select a specific format.
- the speech recognition unit receives the user's voice.
- This step is performed in a state where the voice control video display device is switched to the voice recognition mode.
- This step is preferably performed in a state in which the identification voice data allocated to each execution unit area on the screen displayed on the display unit is visible so that the user can efficiently recognize the identification voice data.
- the information processing unit searches the database and determines whether identification voice data corresponding to the user's voice exists. Specifically, the information processing unit detects the unique coordinate information of the execution unit area to which the identification voice data is assigned when the identification voice data corresponding to the user's voice exists.
- the controller when the identification sound data corresponding to the user's voice exists as a result of the determination, the controller generates an input signal in the execution unit area to which the identification sound data is allocated.
- the control unit plays an role of generating an input signal in the execution unit area to which the identification sound data is allocated, And generates an input signal in a region on the screen having the coordinate information detected by the coordinate information.
- the result of the generation of the input signal depends on the content of the execution unit area. If a shortcut icon of a specific application exists in the execution unit area, the application will be executed. If a specific character of the virtual keyboard is present in the execution unit area, the specific character will be input. If a command is specified, the command will be executed, and in some cases there may be no execution.
- step (a) when the memory unit is combined with the identification voice data, a specific screen control is performed based on the execution unit area to which the identification voice data is allocated
- the information processing unit searches the database for identification data corresponding to the voice of the user
- the control unit searches the database for the voice data corresponding to the voice of the user
- (d) determining whether the control voice data exists and the identification voice data and the control voice data corresponding to the voice of the user exist as a result of the determination by the information processing unit, Generates an execution unit area input signal to which data is assigned, And executing the control command corresponding to the control speech data based on the execution unit area.
- a specific embodiment related to this is as described with reference to FIG.
- the voice control method of the present invention is applied to a voice control video display apparatus and a video display apparatus so that input control is performed by a method of comparing voice data allocated to each execution unit area on a screen displayed through a display unit, It is industrially applicable because it is a technology that enables simple and accurate voice control by applying existing touch screen type input control method to voice control method as it is.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Multimedia (AREA)
- Acoustics & Sound (AREA)
- Computational Linguistics (AREA)
- General Health & Medical Sciences (AREA)
- Business, Economics & Management (AREA)
- Game Theory and Decision Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- User Interface Of Digital Computer (AREA)
- Controls And Circuits For Display Device (AREA)
Abstract
Description
Claims (7)
- 디스플레이부를 구비하고 음성제어가 가능한 영상표시 장치로서,상기 디스플레이부를 통해 표시되는 화면상의 실행단위영역별로 식별음성데이터가 할당되어 매핑(mapping)된 데이터 베이스가 저장된 메모리부;사용자의 음성을 입력받는 음성인식부;상기 음성인식부가 사용자의 음성을 수신한 경우 상기 데이터 베이스를 검색하여 상기 사용자의 음성과 대응되는 식별음성데이터가 존재하는지 판단하는 정보처리부; 및상기 정보처리부의 판단 결과 상기 사용자의 음성과 대응되는 식별음성데이터가 존재하는 경우 해당 식별음성데이터가 할당된 실행단위영역에 입력신호를 발생시키는 제어부;를 포함하여 구성된 것을 특징으로 하는 음성제어 영상표시 장치.
- 제1항에서,상기 디스플레이부는 화면을 표시할 때 화면상의 실행단위영역별로 할당된 식별음성데이터도 함께 현시하여 주도록 구성된 것을 특징으로 하는 음성제어 영상표시 장치.
- 제1항에서,상기 데이터 베이스는 상기 디스플레이부를 통해 표시되는 서로 다른 2이상의 화면에 대해 각 화면상의 실행단위영역별로 식별음성데이터가 할당되어 매핑된 것을 특징으로 하는 음성제어 영상표시 장치.
- 제1항에서,상기 데이터 베이스에는 식별음성데이터와 결합되어 사용될 경우 식별음성데이터가 할당된 실행단위영역을 기준으로 특정의 화면제어를 수행하도록 해주는 제어명령에 대응되는 제어음성데이터가 추가적으로 저장되어 있고,상기 정보처리부는 상기 음성인식부가 사용자의 음성을 수신한 경우 상기 데이터 베이스를 검색하여 상기 사용자의 음성과 대응되는 식별음성데이터 및 제어음성데이터가 존재하는지 판단하고,상기 제어부는 상기 정보처리부의 판단 결과 상기 사용자의 음성과 대응되는 식별음성데이터 및 제어음성데이터가 존재하는 경우 해당 식별음성데이터가 할당된 실행단위영역 입력신호를 발생시키되 입력신호를 발생시킨 실행단위영역을 기준으로 해당 제어음성데이터에 대응되는 제어명령을 실행하도록 구성된 것을 특징으로 하는 음성제어 영상표시 장치.
- 제1항 내지 제4항 중 어느 하나의 음성제어 영상표시 장치에서 수행되는 영상표시 장치의 음성제어 방법으로서,(a) 상기 메모리부가 상기 디스플레이부를 통해 표시되는 화면상의 실행단위영역별로 식별음성데이터가 할당되어 매핑(mapping)된 데이터 베이스를 저장하는 단계;(b) 상기 음성인식부가 사용자의 음성을 입력받는 단계;(c) 상기 정보처리부가 상기 데이터 베이스를 검색하여 상기 사용자의 음성과 대응되는 식별음성데이터가 존재하는지 판단하는 단계; 및(d) 상기 정보처리부의 판단 결과 상기 사용자의 음성과 대응되는 식별음성데이터가 존재하는 경우 상기 제어부가 해당 식별음성데이터가 할당된 실행단위영역에 입력신호를 발생시키는 단계;를 포함하여 구성된 것을 특징으로 하는 영상표시 장치의 음성제어 방법.
- 제5항에서,상기 (b)단계는 상기 디스플레이부에 표시되는 화면상의 실행단위영역별로 할당된 음성데이터가 현시되는 상태에서 수행되는 것을 특징으로 하는 영상표시 장치의 음성제어 방법.
- 제5항에서,상기 (a)단계는 상기 메모리부가 식별음성데이터와 결합되어 사용될 경우 식별음성데이터가 할당된 실행단위영역을 기준으로 특정의 화면제어를 수행하도록 해주는 제어명령에 대응되는 제어음성데이터를 추가적으로 포함하는 데이터 베이스를 저장하는 방식으로 수행되고,상기 (c)단계는 상기 정보처리부가 상기 데이터 베이스를 검색하여 상기 사용자의 음성과 대응되는 식별음성데이터 및 제어음성데이터가 존재하는지 판단하는 방식으로 수행되며,상기 (d)단계는 상기 정보처리부의 판단 결과 상기 사용자의 음성과 대응되는 식별음성데이터 및 제어음성데이터가 존재하는 경우 상기 제어부가 해당 식별음성데이터가 할당된 실행단위영역 입력신호를 발생시키되 입력신호를 발생시킨 실행단위영역을 기준으로 상기 제어음성데이터에 대응되는 제어명령을 실행하는 방식으로 수행되도록 구성된 것을 특징으로 하는 영상표시 장치의 음성제어 방법.
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US15/306,487 US20170047065A1 (en) | 2014-05-13 | 2014-11-20 | Voice-controllable image display device and voice control method for image display device |
CN201480078665.2A CN106462379A (zh) | 2014-05-13 | 2014-11-20 | 语音控制图像显示装置及图像显示装置的语音控制方法 |
JP2016566809A JP2017521692A (ja) | 2014-05-13 | 2014-11-20 | 音声制御映像表示装置及び映像表示装置の音声制御方法 |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR20140056992 | 2014-05-13 | ||
KR10-2014-0056992 | 2014-05-13 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2015174597A1 true WO2015174597A1 (ko) | 2015-11-19 |
Family
ID=54480113
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/KR2014/011197 WO2015174597A1 (ko) | 2014-05-13 | 2014-11-20 | 음성제어 영상표시 장치 및 영상표시 장치의 음성제어 방법 |
Country Status (4)
Country | Link |
---|---|
US (1) | US20170047065A1 (ko) |
JP (1) | JP2017521692A (ko) |
CN (1) | CN106462379A (ko) |
WO (1) | WO2015174597A1 (ko) |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108231073A (zh) * | 2016-12-16 | 2018-06-29 | 深圳富泰宏精密工业有限公司 | 语音控制装置、***及控制方法 |
US10448762B2 (en) | 2017-09-15 | 2019-10-22 | Kohler Co. | Mirror |
US10663938B2 (en) | 2017-09-15 | 2020-05-26 | Kohler Co. | Power operation of intelligent devices |
US10887125B2 (en) | 2017-09-15 | 2021-01-05 | Kohler Co. | Bathroom speaker |
US11093554B2 (en) | 2017-09-15 | 2021-08-17 | Kohler Co. | Feedback for water consuming appliance |
US11099540B2 (en) | 2017-09-15 | 2021-08-24 | Kohler Co. | User identity in household appliances |
Families Citing this family (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
GB2543019A (en) * | 2015-07-23 | 2017-04-12 | Muzaffar Saj | Virtual reality headset user input system |
CN107832036B (zh) * | 2017-11-22 | 2022-01-18 | 北京小米移动软件有限公司 | 语音控制方法、装置及计算机可读存储介质 |
CN109102808A (zh) * | 2018-10-25 | 2018-12-28 | 珠海格力电器股份有限公司 | 一种基于显示交互的本地语音识别***及其方法 |
JP7263919B2 (ja) * | 2019-05-22 | 2023-04-25 | コニカミノルタ株式会社 | 画像処理装置およびプログラム |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20050131700A1 (en) * | 2003-09-10 | 2005-06-16 | General Electric Company | Voice control of a generic input device for an ultrasound system |
KR20050108649A (ko) * | 2004-05-12 | 2005-11-17 | 엘지전자 주식회사 | 음성을 이용한 이동통신 단말기 조작방법 |
KR20110006509A (ko) * | 2009-07-14 | 2011-01-20 | 주식회사대성엘텍 | 음성을 이용한 화면의 디스플레이 상태 제어장치 및 그 제어방법 |
KR101227875B1 (ko) * | 2011-05-30 | 2013-01-31 | 김호진 | 사용자 동작 기반 디스플레이 장치 |
Family Cites Families (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP0747807B1 (en) * | 1995-04-11 | 2002-03-06 | Dragon Systems Inc. | Moving an element shown on a computer display |
JP2004029933A (ja) * | 2002-06-21 | 2004-01-29 | Mitsubishi Heavy Ind Ltd | 表示制御装置および表示制御方法 |
KR100632400B1 (ko) * | 2005-11-11 | 2006-10-11 | 한국전자통신연구원 | 음성 인식을 이용한 입출력 장치 및 그 방법 |
KR20120080069A (ko) * | 2011-01-06 | 2012-07-16 | 삼성전자주식회사 | 디스플레이 장치 및 그 음성 제어 방법 |
CN102752442A (zh) * | 2011-04-21 | 2012-10-24 | 英业达股份有限公司 | 手持通信装置及其通信方法 |
CN102622085A (zh) * | 2012-04-11 | 2012-08-01 | 北京航空航天大学 | 多维感官人机交互***及交互方法 |
US9836192B2 (en) * | 2014-02-25 | 2017-12-05 | Evan Glenn Katsuranis | Identifying and displaying overlay markers for voice command user interface |
-
2014
- 2014-11-20 CN CN201480078665.2A patent/CN106462379A/zh active Pending
- 2014-11-20 US US15/306,487 patent/US20170047065A1/en not_active Abandoned
- 2014-11-20 WO PCT/KR2014/011197 patent/WO2015174597A1/ko active Application Filing
- 2014-11-20 JP JP2016566809A patent/JP2017521692A/ja active Pending
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20050131700A1 (en) * | 2003-09-10 | 2005-06-16 | General Electric Company | Voice control of a generic input device for an ultrasound system |
KR20050108649A (ko) * | 2004-05-12 | 2005-11-17 | 엘지전자 주식회사 | 음성을 이용한 이동통신 단말기 조작방법 |
KR20110006509A (ko) * | 2009-07-14 | 2011-01-20 | 주식회사대성엘텍 | 음성을 이용한 화면의 디스플레이 상태 제어장치 및 그 제어방법 |
KR101227875B1 (ko) * | 2011-05-30 | 2013-01-31 | 김호진 | 사용자 동작 기반 디스플레이 장치 |
Cited By (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108231073A (zh) * | 2016-12-16 | 2018-06-29 | 深圳富泰宏精密工业有限公司 | 语音控制装置、***及控制方法 |
CN108231073B (zh) * | 2016-12-16 | 2021-02-05 | 深圳富泰宏精密工业有限公司 | 语音控制装置、***及控制方法 |
US10448762B2 (en) | 2017-09-15 | 2019-10-22 | Kohler Co. | Mirror |
US10663938B2 (en) | 2017-09-15 | 2020-05-26 | Kohler Co. | Power operation of intelligent devices |
US10887125B2 (en) | 2017-09-15 | 2021-01-05 | Kohler Co. | Bathroom speaker |
US11093554B2 (en) | 2017-09-15 | 2021-08-17 | Kohler Co. | Feedback for water consuming appliance |
US11099540B2 (en) | 2017-09-15 | 2021-08-24 | Kohler Co. | User identity in household appliances |
US11314214B2 (en) | 2017-09-15 | 2022-04-26 | Kohler Co. | Geographic analysis of water conditions |
US11314215B2 (en) | 2017-09-15 | 2022-04-26 | Kohler Co. | Apparatus controlling bathroom appliance lighting based on user identity |
US11892811B2 (en) | 2017-09-15 | 2024-02-06 | Kohler Co. | Geographic analysis of water conditions |
US11921794B2 (en) | 2017-09-15 | 2024-03-05 | Kohler Co. | Feedback for water consuming appliance |
US11949533B2 (en) | 2017-09-15 | 2024-04-02 | Kohler Co. | Sink device |
Also Published As
Publication number | Publication date |
---|---|
CN106462379A (zh) | 2017-02-22 |
JP2017521692A (ja) | 2017-08-03 |
US20170047065A1 (en) | 2017-02-16 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2015174597A1 (ko) | 음성제어 영상표시 장치 및 영상표시 장치의 음성제어 방법 | |
WO2013172607A1 (en) | Method of operating a display unit and a terminal supporting the same | |
WO2016080713A1 (ko) | 음성제어 영상표시 장치 및 영상표시 장치의 음성제어 방법 | |
WO2015088263A1 (ko) | 터치 입력의 압력 상태에 따라 동작하는 전자 장치 및 그 방법 | |
WO2015099293A1 (en) | Device and method for displaying user interface of virtual input device based on motion recognition | |
WO2014003365A1 (en) | Method and apparatus for processing multiple inputs | |
WO2014107005A1 (en) | Mouse function provision method and terminal implementing the same | |
WO2014129828A1 (en) | Method for providing a feedback in response to a user input and a terminal implementing the same | |
WO2012033345A1 (en) | Motion control touch screen method and apparatus | |
WO2016208808A1 (ko) | 디스플레이 디바이스 | |
WO2013125804A1 (en) | Method and apparatus for moving contents in terminal | |
WO2013070024A1 (en) | Method and apparatus for designating entire area using partial area touch in a portable equipment | |
WO2013048131A2 (en) | Method and apparatus for providing user interface in portable device | |
EP2561429A2 (en) | Method for providing graphical user interface and mobile device adapted thereto | |
WO2021031843A1 (zh) | 对象位置调整方法及电子设备 | |
WO2015156539A2 (en) | Computing apparatus, method for controlling computing apparatus thereof, and multi-display system | |
WO2015102325A1 (ko) | 디스플레이 장치 및 이의 추천 문자 제공 방법 | |
KR20150043272A (ko) | 영상표시 장치의 음성제어 방법 | |
CN107783669A (zh) | 光标产生***、方法及计算机程序产品 | |
WO2011081354A2 (en) | Korean input method and apparatus using touch screen, and portable terminal including key input apparatus | |
WO2011145788A1 (ko) | 시각 장애인을 위한 터치스크린 장치 및 사용자 인터페이스 | |
WO2018117518A1 (en) | Display apparatus and controlling method thereof | |
WO2023033570A1 (ko) | 리모컨 및 그 리모컨의 키를 설정하는 디스플레이 장치와 그 방법들 | |
KR101517738B1 (ko) | 음성제어 영상표시 장치 및 영상표시 장치의 음성제어 방법 | |
WO2021025369A1 (ko) | 인터랙션 스크롤 제어 방법, 장치, 프로그램 및 컴퓨터 판독가능 기록매체 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 14891981 Country of ref document: EP Kind code of ref document: A1 |
|
DPE1 | Request for preliminary examination filed after expiration of 19th month from priority date (pct application filed from 20040101) | ||
WWE | Wipo information: entry into national phase |
Ref document number: 15306487 Country of ref document: US |
|
ENP | Entry into the national phase |
Ref document number: 2016566809 Country of ref document: JP Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 14891981 Country of ref document: EP Kind code of ref document: A1 |