WO2020122564A1 - Procédé et dispositif de reconnaissance de la parole - Google Patents

Procédé et dispositif de reconnaissance de la parole Download PDF

Info

Publication number
WO2020122564A1
WO2020122564A1 PCT/KR2019/017401 KR2019017401W WO2020122564A1 WO 2020122564 A1 WO2020122564 A1 WO 2020122564A1 KR 2019017401 W KR2019017401 W KR 2019017401W WO 2020122564 A1 WO2020122564 A1 WO 2020122564A1
Authority
WO
WIPO (PCT)
Prior art keywords
content
speech recognition
information
speech
message
Prior art date
Application number
PCT/KR2019/017401
Other languages
English (en)
Korean (ko)
Inventor
마르친키에비츠피오트르
파벨 그루지악그제고즈
Original Assignee
삼성전자 주식회사
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 삼성전자 주식회사 filed Critical 삼성전자 주식회사
Publication of WO2020122564A1 publication Critical patent/WO2020122564A1/fr

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/02Feature extraction for speech recognition; Selection of recognition unit
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/26Speech to text systems
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L21/00Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
    • G10L21/02Speech enhancement, e.g. noise reduction or echo cancellation
    • G10L21/0208Noise filtering

Definitions

  • the present invention relates to a method and apparatus for recognizing speech, and more particularly, to a method and apparatus for reducing noise from other electronic devices for speech recognition.
  • the electronic device can recognize the voice and perform a specific operation according to the content of the voice. For example, the electronic device may perform various operations, such as executing a specific application, controlling the operation of the application, performing a search, or turning on/off the power, based on speech recognition.
  • Some mobile terminals, tablet computers or other electronic devices employ a voice assistant such as Siri, Cortana, Alexa, or Bixby. These voice assistants can perform various operations based on voice commands. For the efficient operation of the voice assistant, clear recognition of voice commands is required. In order to efficiently control not only the voice assistant of the mobile terminal but also other electronic devices operating based on speech recognition, it is necessary to increase the accuracy of speech recognition.
  • a voice assistant such as Siri, Cortana, Alexa, or Bixby.
  • Sounds detected by the electronic device in addition to the desired speech may act as noise for speech recognition.
  • audio output from another electronic device may act as noise, and when the audio includes voice, it may cause confusion with a desired voice, thereby significantly reducing the efficiency of voice detection.
  • An object of the present invention is to provide a method and apparatus for effectively canceling noise due to audio output from another electronic device for efficient speech recognition.
  • a method for recognizing a voice by a first device for solving the above problems includes receiving information regarding content output by the second device from a second device, and And performing voice recognition by erasing data obtained from the output content based on the information on the content.
  • Method for supporting speech recognition by a first device according to another embodiment of the present invention, transmitting information about the content output by the second device, and outputting the content Including a step, the information on the content is used by the first device to erase data obtained from the content based on the information on the content for speech recognition.
  • a first device for recognizing voice includes a transceiver configured to receive information on content output by the second device from a second device, and a voice configured to detect a voice signal Input method; And a controller configured to perform speech recognition based on the detected speech signal by erasing data obtained from the output content based on information about the content.
  • a second device for supporting speech recognition by a first device includes a transceiver configured to transmit information on content output by the second device, and an audio signal of the content It includes an audio output device configured to output, the information on the content is used by the first device to erase the data obtained from the content based on the information on the content for speech recognition.
  • FIG. 1 shows an exemplary speech recognition environment of the present invention.
  • FIG. 2 is a flowchart illustrating a voice recognition method according to an embodiment of the present invention.
  • FIG. 3 is a flowchart illustrating a voice recognition method according to another embodiment of the present invention.
  • FIG. 4 is a block diagram of a first device according to an embodiment of the present invention.
  • FIG. 5 is a block diagram of a second device according to an embodiment of the present invention.
  • the first device 110 may recognize speech.
  • the first device 110 can detect the voice signal S1 from the user 130 and recognize the voice therefrom.
  • the first device 110 recognizes the voice and performs an operation according to the recognized voice Can be.
  • an operation corresponding to the keyword eg, execution of an application, termination of an application, or on/off of a specific function
  • the first device 110 may be a smart phone, a cellular phone, a tablet computer, a personal computer, a smart speaker, or a household electronic device (for example, a refrigerator, a washing machine, an air cleaner, and a robot cleaner), but is not limited thereto. It may not be one of various electronic devices that can be operated through voice.
  • the second device 120 may be an electronic device capable of outputting the audio signal S2.
  • the second device 120 may be a household electronic device capable of outputting an audio signal, such as a television or radio, or a smart phone, a cellular phone, a tablet computer, or a personal computer.
  • the second device 120 is not limited to these types, and may be any electronic device capable of outputting an audio signal and communicating with other devices or networks.
  • the audio signal S2 output from the second device 120 may act as noise for speech recognition performed by the first device 110.
  • the second device 120 may not output an audio signal, but may be a device capable of generating noise for speech recognition of the first device 110 by an operation.
  • the second device 120 may transmit information on content being played or content to be played to the first device 110 (210 ).
  • Information about the content may be referred to as metadata about the content.
  • the second device 120 may broadcast information on the content, and the first device 110 may receive information on the broadcasted content.
  • the second device 120 may transmit information on content using a communication protocol such as Bluetooth or Wi-Fi, but the available communication protocol is not limited thereto.
  • a communication protocol such as Bluetooth or Wi-Fi, but the available communication protocol is not limited thereto.
  • the first device 110 and the second device 120 are paired before the second device 120 transmits the content being played or the content to be played to the first device 110.
  • the first device 110 and the second device 120 may be close to each other such that an audio signal from the second device 120 affects speech recognition to the first device 110. Whether the first device 110 and the second device 120 are in proximity may be made by measuring audio volume or measuring Bluetooth signal strength by the first device 110 and/or the second device 120. Pairing of the first device 110 and the second device 120 may be performed only when the first device 110 and the second device 120 are in close proximity. In this way, the first device 110 and the second device 120 can be paired without user intervention. According to some embodiments, the first device 110 and the second device 120 may be paired through user manipulation.
  • Content includes audio.
  • the content may additionally include experiential content such as video and sound, smell, and vibration.
  • the information on the content may include at least one of power levels for different frequencies of the audio signal included in the content, subtitles of the audio signal, lyrics of the song when the audio signal includes a song, and identification information about the content. Can be.
  • Information about the content may be referred to as metadata for the content.
  • information about the content may include an electronic program guide (EPG) associated with the content.
  • EPG electronic program guide
  • the second device 120 can play the content (220).
  • the audio signal S2 included in the reproduced content may act as noise for speech recognition performed by the first device 110.
  • FIG. 2 it is illustrated that playing the content 220 is performed after transmitting the information about the content 210, but this is only exemplary, and according to some embodiments, information about the content May be transmitted simultaneously with the start of playback of the content, or may be transmitted after the start of playback of the content.
  • the first device 110 may perform speech recognition using information on the received content (230).
  • the first device 110 can effectively eliminate noise due to the audio signal S2 included in the content to be reproduced using the information on the received content, thereby improving the accuracy of speech recognition.
  • the first device 110 is configured to block a signal of a specific frequency with respect to the detected audio signal by using power levels for different frequencies of the audio signal included in the content included in the information about the content. Filters can be applied.
  • the first device 110 may remove the text included in the subtitles from the text detected through speech recognition by using the subtitles of the audio signal included in the information about the content.
  • the first device 110 may remove the text included in the lyrics from the text detected through speech recognition by using the lyrics of the song included in the audio signal included in the information about the content.
  • the first device 110 may acquire subtitles or lyrics related to the content through a server or a network (for example, the Internet) by using identification information about the content included in the information about the content, and perform voice recognition. Text included in the subtitle or text included in the lyrics may be removed from the detected text. According to some embodiments, the first device 110 uses speech inference algorithms used in machine learning for speech recognition modified to ignore words present in subtitles for audio noise, thereby recognizing speech. Text included in the subtitle or text included in the lyrics may be removed from the detected text.
  • the first device 110 may use a cloud to perform speech recognition.
  • the first device 110 may transmit data on the detected voice signal and information on content to the cloud.
  • the cloud is obtained by removing text about the content included in the information about the content or text about the content (for example, EPG) from the text extracted from the data on the detected speech signal.
  • the text can be transmitted to the first device 110 as text for the recognized voice.
  • the present invention it is possible to easily and efficiently perform noise removal by providing information on content reproduced by a noise source device and using information on content provided by a speech recognition device. Accordingly, the accuracy of speech recognition by the speech recognition device can be increased.
  • the first device 110 may start speech recognition (310 ).
  • Speech recognition in step 310 may involve removing noise using information about the content received from the second device 230 as in step 230 of FIG. 2, but is not limited thereto.
  • the speech recognition in step 310 may be for detecting keywords.
  • the first device 110 may detect a keyword through voice recognition (320).
  • the keyword may be a voice command composed of at least one word for causing the first device 110 to perform a specific operation.
  • the first device 110 executes an application requiring detection of a subsequent voice in response to detection of the keyword, or the first device 110 detects a subsequent voice in response to detection of the keyword and You can enter a state that requires recognition.
  • the subsequent detection of speech in response to the detection of the keyword may be detection and recognition of additional keywords or detection and recognition of free speech.
  • the first device 110 may use a cloud for detection and recognition of pre-speech. Subsequent voice detection in response to the detection of the keyword may involve noise removal using information on the content received from the second device 120 as in step 230 of FIG. 2.
  • the first device 110 may transmit a first message to the second device 120 in response to the detection of the keyword (340 ). According to some embodiments, the first device 110 may broadcast the first message. The first device broadcasts the first message and delivers it to any devices, thereby allowing any devices to activate the audio filter, as described below. The first device 110 may transmit the first message using a communication protocol such as Bluetooth or Wi-Fi, but this is only an example, and the communication protocol for transmitting the first message is not limited thereto. The first message may be for the second device 120 to activate the audio filter. According to some embodiments, the first message may include a command for the second device 120 to activate the audio filter.
  • a communication protocol such as Bluetooth or Wi-Fi
  • the first message only includes information indicating that the first device 110 has detected the keyword or information that the first device 110 requires subsequent voice input, and the second device 120 ) May not contain a direct command to activate the audio filter.
  • the second device 120 may determine to activate the audio filter based on the information included in the first message.
  • the first message may be for causing the second device 120 to perform an operation to reduce noise in addition to activating the audio filter.
  • the second device 120 Before receiving the first message 340, the second device 120 may be playing content including an audio signal that may act as noise for speech recognition of the first device 110 (330 ). However, playing the content 330 does not necessarily have to start before receiving the first message (340), after receiving the first message (340) or after activating (350) the audio filter. The playback of content may start.
  • the second device 120 can activate the audio filter in response to receiving the first message (350 ).
  • the audio filter 350 may be a notch filter or a bandstop filter for blocking or attenuating an audio signal in a specific frequency band.
  • the frequency band blocked or attenuated by the audio filter may at least partially overlap with the frequency band of the human voice.
  • the second device 120 may activate an audio filter to reduce noise in a frequency band of speech, thereby supporting efficient speech recognition by the first device 110.
  • the second device 120 may perform other actions instead of activating the audio filter in response to receiving the first message. For example, the second device 120 may mute the sound output or decrease the volume of the sound. According to some embodiments, the second device 120 may maintain the volume of the sound when the volume of the sound is below a threshold.
  • the second device 120 may be a device (eg, a vacuum cleaner, a robot cleaner, a laundry dryer) that may generate noise for speech recognition of the first device 110 in addition to sound output. have.
  • the second device 120 may perform an operation for reducing noise in response to receiving the first message.
  • the operation for reducing noise may include, for example, stopping the operation of the second device 120 or reducing the output of the second device 120.
  • the operations described above by the second device 120 in response to receiving the first message i.e. activating the audio filter, muting the sound output, reducing the volume of the sound, Interrupting the operation of the two devices, reducing the output of the second device
  • an operation for reducing noise i.e. activating the audio filter, muting the sound output, reducing the volume of the sound, Interrupting the operation of the two devices, reducing the output of the second device
  • an operation for reducing different noise may be performed in response to receiving the first message according to the type of the noise source device, such as the second device 120.
  • the noise source device When the noise source device is of a type that is adjacent to the first device 110 and generates a volume of noise greater than a threshold, the noise source device may mute the audio output in response to receiving the first message. If the noise source device is of a type that generates average noise (for example, the size of the noise or audio output is within a certain range), the noise source device responds to receiving the first message, and the volume of the audio output You can decrease the level. For a type where the noise source device is very far away or generates noise below a threshold, the noise source device may not change the audio output even upon receiving the first message.
  • the first device In response to detecting a keyword, 110 may determine a device to be silenced among the plurality of noise source devices (ie, the second device 120).
  • the first message When the first message is broadcast, the first message may include identification information about the device to be silenced.
  • Each of the devices receiving the broadcasted first message may perform an operation for reducing noise in response to the reception of the first message when the first message includes its own identification information.
  • the first device 110 may be configured to receive electromagnetic energy from the second device 120 to estimate the distance from the second device 120 or the influence of noise from the second device 120.
  • the propagation of sound waves and the propagation of electromagnetic waves have a similar aspect in that the intensity decreases in inverse proportion to the square of the distance, and the obstacle between the first device 110 and the second device 120 is that of both sound waves and electromagnetic waves. Decrease the intensities.
  • the first device 120 may estimate the influence of distance or noise from the second device 120 through measurement of the intensity of the electromagnetic signal from the second device 120.
  • the first device 120 may estimate a distance or an effect of noise from each of the plurality of noise source devices, and determine a device to perform an operation for reducing noise in response to receiving the first message, and noise It is possible to determine the type of operation (eg, muting, volume reduction, or turn-off) to reduce.
  • the first device 110 may measure the intensity of the Bluetooth signal from the second device 120 to estimate the effect of distance or noise from the second device 120.
  • the second device 120 measures the intensity of the electromagnetic signal, including the Bluetooth signal from the first device 110, to determine the distance to the first device 110 or the first device ( The influence of noise on 110) can be estimated.
  • the second device 12 is based on the estimated distance to the first device 110 or the effect of noise on the first device 110, whether to perform an operation for reducing noise, and noise to perform. You can decide the type to reduce the.
  • the first device 110 can use information about the spectral characteristics of the noise received from the second device 120 to estimate the effect of the distance or noise from the second device 120. have.
  • Information about the spectral characteristics of the noise may be included in information about the content transmitted in step 210.
  • the first device 110 can estimate the effect of the distance or noise from the second device 120 by comparing the information on the noise spectrum characteristics and the power level according to the spectrum of the noise received from the second device 120. have.
  • the first device 110 may end speech recognition (360).
  • the first device 110 may end speech recognition when a voice signal is not detected for a specific period of time.
  • termination of speech recognition in step 360 only means termination of recognition for pre-speech, and first device 110 continues to attempt speech recognition for keywords after step 360 can do.
  • the first device 110 may transmit a second message to the second device based on the end of speech recognition (370 ).
  • the first device 110 may broadcast the second message.
  • the first device 110 may broadcast a second message to allow arbitrary devices to receive the second message to disable the audio filter.
  • the second message may be for the second device 120 to deactivate the audio filter.
  • the second message can include a command to cause the second device 120 to deactivate the audio filter.
  • the second message only includes information indicating that the first device 110 has ended speech recognition, and has issued an explicit command to cause the second device 120 to deactivate the audio filter. May not contain.
  • the second device 120 may determine to deactivate the audio filter based on the information included in the second message.
  • the second message may be to cause the second device 120 to perform an interruption of the operation to reduce noise, including or excluding the audio filter.
  • the second device 120 can deactivate the audio filter in response to receiving the second message (380). According to some embodiments, even if the second device 120 does not receive the second message, the audio filter may be deactivated after activating the audio filter or when a specific time has elapsed since receiving the first message. have. Through this, even if the second device 120 does not receive the broadcast second message, it is possible to prevent the audio filter from being continuously activated. According to some embodiments, second device 120 may perform an interruption in operation to reduce noise other than deactivating the audio filter in response to receiving the second message. Is a block diagram of a first device according to an embodiment of the present invention.
  • the first device 110 may include a voice input device 410, a transceiver 420 and a controller 430.
  • the voice input unit 410 may detect audio signals including voice as well as voice.
  • the voice input device 410 may be a microphone.
  • the voice input device 410 may include an audio filter.
  • the first device 110 uses an audio filter to block a signal of a specific frequency with respect to the detected audio signal by using power levels for different frequencies of the audio signal included in the content included in the information about the content Can be.
  • the audio filter may be configured separately from the voice input device 410.
  • the audio filter can be implemented in software.
  • the first device 110 may communicate with other devices or networks through the transceiver 420.
  • the transceiver 420 may support at least one of known short-range or telecommunication protocols such as Bluetooth, infrared communication, LTE, CDMA, Wi-Fi, Wi-Max, and Wi-Bro.
  • the transceiver may receive information about the content in step 210 and transmit the first message in step 340 and the second message in step 370.
  • the controller 430 may control operations of other components of the first device 110 (eg, the voice input device 410 and the transceiver 420 ).
  • the controller 430 may perform operations required for the operation of the first device 110.
  • the controller 430 may be implemented with at least one processor.
  • the controller 430 may identify a voice signal from the audio signal detected through the voice input device 410 and detect text corresponding to the voice signal.
  • the controller 430 may increase the accuracy of speech recognition by removing text of subtitles or lyrics obtained based on information about content provided by the second device 120 from text detected during speech recognition.
  • the controller 430 may perform voice recognition using a cloud.
  • the first device 110 may further include a memory 440, a display 450, and an audio output device 460.
  • the memory 440 may store instructions for operating the controller 430.
  • the memory 440 may store temporary or permanent data generated or required during the operation of the controller 430.
  • the display 450 may be configured to reproduce an image.
  • the display 450 may be constructed using at least one of known display elements such as a liquid crystal display (LCD), an organic light emitting display (OLED), or an electrophoresis display, and employable display elements are listed above It is not limited.
  • the audio output device 460 may be configured to output an audio signal.
  • the audio output device 460 may be a speaker or an audio terminal.
  • the operation of the memory 440, the display 450, and the audio writer 460 may be controlled by the controller 430.
  • the second device 120 includes a transceiver 510, an audio output unit 520, a controller 530 and an audio filter 540.
  • the second device 120 may communicate with other devices or networks through the transceiver 510.
  • the transceiver 510 may support at least one of known short-range or telecommunication protocols such as Bluetooth, infrared communication, LTE, CDMA, Wi-Fi, Wi-Max, and Wi-Bro.
  • the transceiver 510 may transmit information on the content in step 210 and receive the first message in step 340 and the second message in step 370.
  • the audio output device 520 may be configured to output an audio signal.
  • the audio output device 520 may be a speaker.
  • the audio output device 520 may include an interface (eg, an audio terminal) connectable to a speaker and a wired interface, or an interface (eg, a Bluetooth interface) connectable wirelessly.
  • the interface that can be wirelessly connected to the speaker may be included in the transceiver 510 or at least partially overlap with the transceiver 520.
  • the second device 120 may not include an audio output device 520.
  • the controller 530 may control operations of other components of the second device 120 (eg, the transceiver 510, the audio output unit 520, and the audio filter 540 ).
  • the controller 530 may perform operations necessary for the operation of the second device 120.
  • the controller 530 may be implemented with at least one processor.
  • the controller 530 may determine whether to perform an operation for reducing noise based on the first message received from the first device. For example, the controller 530 may determine whether to activate or deactivate the audio filter 540 based on the first message received from the first device.
  • the audio filter 540 may be a filter for blocking or attenuating an audio signal in a specific frequency band (eg, voice band).
  • a specific frequency band of the audio signal output from the audio output device 520 may be blocked or attenuated.
  • the audio filter 540 may be configured in hardware or software.
  • second device 120 may not include audio filter 540. If the second device 120 does not include the audio filter 540, the second device 120 may not also include the audio writer 520, but is not necessarily so.
  • the second device 120 is a memory 550 and a display 560 may be further included. Descriptions of the memory 550 and the display 560 are the same as the descriptions of the memory 440 and the display 450 of the first device 110, respectively, and thus will be omitted.

Landscapes

  • Engineering & Computer Science (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Quality & Reliability (AREA)
  • Signal Processing (AREA)
  • Telephonic Communication Services (AREA)
  • Telephone Function (AREA)

Abstract

L'invention concerne un procédé de reconnaissance de la parole au moyen d'un premier dispositif. Le procédé ci-dessus comprend : une étape de réception, à partir d'un second dispositif, d'informations concernant un contenu émis par le second dispositif ; et une étape de réalisation d'une reconnaissance de la parole par effacement de données obtenues à partir du contenu émis en fonction des informations concernant le contenu.
PCT/KR2019/017401 2018-12-14 2019-12-10 Procédé et dispositif de reconnaissance de la parole WO2020122564A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
KR10-2018-0162182 2018-12-14
KR1020180162182A KR20200074349A (ko) 2018-12-14 2018-12-14 음성을 인식하기 위한 방법 및 장치

Publications (1)

Publication Number Publication Date
WO2020122564A1 true WO2020122564A1 (fr) 2020-06-18

Family

ID=71077348

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/KR2019/017401 WO2020122564A1 (fr) 2018-12-14 2019-12-10 Procédé et dispositif de reconnaissance de la parole

Country Status (2)

Country Link
KR (1) KR20200074349A (fr)
WO (1) WO2020122564A1 (fr)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112037792A (zh) * 2020-08-20 2020-12-04 北京字节跳动网络技术有限公司 一种语音识别方法、装置、电子设备及存储介质

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20150007422A (ko) * 2013-07-11 2015-01-21 삼성전자주식회사 전기기기 및 그 제어 방법
KR101551968B1 (ko) * 2013-12-30 2015-09-09 현대자동차주식회사 차량에 적용된 미디어를 통한 음원 정보 제공 방법
JP2016127300A (ja) * 2014-12-26 2016-07-11 アイシン精機株式会社 音声処理装置
US20180160189A1 (en) * 2016-12-06 2018-06-07 The Directv Group, Inc. Search for content data in content
WO2018211983A1 (fr) * 2017-05-16 2018-11-22 Sony Corporation Amélioration de la parole pour des applications de reconnaissance de la parole dans des environnements de diffusion
KR20180127100A (ko) * 2017-05-19 2018-11-28 엘지전자 주식회사 홈 어플라이언스 및 음성 인식 서버 시스템의 동작 방법

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20150007422A (ko) * 2013-07-11 2015-01-21 삼성전자주식회사 전기기기 및 그 제어 방법
KR101551968B1 (ko) * 2013-12-30 2015-09-09 현대자동차주식회사 차량에 적용된 미디어를 통한 음원 정보 제공 방법
JP2016127300A (ja) * 2014-12-26 2016-07-11 アイシン精機株式会社 音声処理装置
US20180160189A1 (en) * 2016-12-06 2018-06-07 The Directv Group, Inc. Search for content data in content
WO2018211983A1 (fr) * 2017-05-16 2018-11-22 Sony Corporation Amélioration de la parole pour des applications de reconnaissance de la parole dans des environnements de diffusion
KR20180127100A (ko) * 2017-05-19 2018-11-28 엘지전자 주식회사 홈 어플라이언스 및 음성 인식 서버 시스템의 동작 방법

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112037792A (zh) * 2020-08-20 2020-12-04 北京字节跳动网络技术有限公司 一种语音识别方法、装置、电子设备及存储介质

Also Published As

Publication number Publication date
KR20200074349A (ko) 2020-06-25

Similar Documents

Publication Publication Date Title
KR102019361B1 (ko) 무선 오디오 출력 디바이스
US8314720B2 (en) Terminal and method for executing function using human body communication
WO2014069798A1 (fr) Appareil de reconnaissance vocale et procédé de reconnaissance vocale associé
US20160275961A1 (en) Structure for multi-microphone speech enhancement system
JP2018517167A (ja) 音声信号を処理するための方法及び装置
CN110677768A (zh) 一种无线耳机控制方法、装置及无线耳机和存储介质
CN106528545B (zh) 一种语音信息的处理方法及装置
JP2019128586A (ja) 音声処理方法、装置及びコンピュータ読み取り可能な記憶媒体
CN108024128B (zh) 蓝牙音乐播放的控制方法、装置、终端设备及存储介质
CN106940997B (zh) 一种向语音识别***发送语音信号的方法和装置
CN109473097B (zh) 一种智能语音设备及其控制方法
JP7254936B2 (ja) 情報の受信方法、送信方法、端末及びネットワーク側機器
JP7271711B2 (ja) Ssb伝送指示方法、装置、端末、機器及び媒体
WO2022083620A1 (fr) Procédé et appareil de transmission de signal, et dispositif terminal, dispositif intelligent et dispositif électronique
WO2020048175A1 (fr) Procédé de traitement d'effet sonore, dispositif, dispositif électronique, et support de stockage
WO2020122564A1 (fr) Procédé et dispositif de reconnaissance de la parole
CN110475312B (zh) 一种接入控制方法、消息广播方法及相关设备
CN110035506B (zh) 一种通信方法及相关设备
CN107911777B (zh) 一种耳返功能的处理方法、装置及移动终端
US10540974B2 (en) Detection and disposition of commands methods and apparatuses
WO2020080638A1 (fr) Dispositif électronique et son procédé de contrôle
US20210152965A1 (en) Method for detecting blocking of microphone and related products
CN105450880A (zh) 一种语音信号的处理方法和装置
CN106782614B (zh) 音质检测方法及装置
WO2021246795A1 (fr) Procédé et système de reconnaissance de gestes, et support d'enregistrement lisible par ordinateur non transitoire

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 19896428

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 19896428

Country of ref document: EP

Kind code of ref document: A1