CN107919123B - Multi-voice assistant control method, device and computer readable storage medium - Google Patents

Multi-voice assistant control method, device and computer readable storage medium Download PDF

Info

Publication number
CN107919123B
CN107919123B CN201711287860.8A CN201711287860A CN107919123B CN 107919123 B CN107919123 B CN 107919123B CN 201711287860 A CN201711287860 A CN 201711287860A CN 107919123 B CN107919123 B CN 107919123B
Authority
CN
China
Prior art keywords
voice
voice assistant
assistant
target
task
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201711287860.8A
Other languages
Chinese (zh)
Other versions
CN107919123A (en
Inventor
钱庄
陈维扬
胥亚伟
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Xiaomi Mobile Software Co Ltd
Original Assignee
Beijing Xiaomi Mobile Software Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Xiaomi Mobile Software Co Ltd filed Critical Beijing Xiaomi Mobile Software Co Ltd
Priority to CN201711287860.8A priority Critical patent/CN107919123B/en
Publication of CN107919123A publication Critical patent/CN107919123A/en
Application granted granted Critical
Publication of CN107919123B publication Critical patent/CN107919123B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/16Sound input; Sound output
    • G06F3/167Audio in a user interface, e.g. using voice commands for navigating, audio feedback
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/46Multiprogramming arrangements
    • G06F9/48Program initiating; Program switching, e.g. by interrupt
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • G10L2015/223Execution procedure of a spoken command

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Health & Medical Sciences (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Software Systems (AREA)
  • General Health & Medical Sciences (AREA)
  • Computational Linguistics (AREA)
  • Acoustics & Sound (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

The disclosure relates to a multi-voice assistant control method, a device and a computer readable storage medium, which are used for solving the technical problems that when the voice assistant is switched by the existing electronic equipment, the voice assistant needs to be manually enabled, and the switching mode is not flexible enough. The multi-voice assistant control method is applied to electronic equipment integrating a plurality of voice assistants, and comprises the following steps: when a trigger event for waking up the voice assistant is detected, determining a target voice assistant from a plurality of voice assistants according to the trigger event, wherein the waking functions of the plurality of voice assistants are all in an enabling state; the target voice assistant is awakened.

Description

Multi-voice assistant control method, device and computer readable storage medium
Technical Field
The present disclosure relates to the field of communications technologies, and in particular, to a multi-voice assistant control method, apparatus, and computer-readable storage medium.
Background
With the development of speech recognition technology and the increasing demand of people for speech interaction, an intelligent application of speech assistant is produced. In the related art, the voice assistant may implement an intelligent dialog or an instant question and answer with the user to help the user solve a problem. However, the conventional electronic device usually only has one voice assistant built therein, even if a plurality of voice assistants are built therein, only one of the voice assistants is in a wakenable state at the same time, the plurality of voice assistants cannot be turned on simultaneously, and the voice assistant cannot be wakened up if the voice assistant is not turned on.
Disclosure of Invention
The present disclosure provides a multi-voice assistant control method, an apparatus and a computer-readable storage medium, which are used to solve the technical problem that when an existing electronic device switches a voice assistant, the voice assistant needs to be manually enabled, and the switching manner is not flexible enough.
In order to achieve the above object, a first aspect of the embodiments of the present disclosure provides a multi-voice assistant control method applied to an electronic device integrating multiple voice assistants, the method including:
when a trigger event for waking up the voice assistant is detected, determining a target voice assistant from a plurality of voice assistants according to the trigger event, wherein the wake-up functions of the plurality of voice assistants are all in an enabled state;
waking the target voice assistant.
Optionally, before the determining a target voice assistant from the plurality of voice assistants according to the trigger event, the method includes:
collecting input first voice information;
judging whether the first voice message comprises a keyword which is consistent with a preset awakening word or not;
and if the first voice message comprises a keyword which is consistent with the preset awakening word, determining that the trigger event occurs.
Optionally, the triggering event includes collecting second voice information for waking up a voice assistant, and the determining a target voice assistant from the plurality of voice assistants according to the triggering event includes:
determining whether keywords associated with the identity of any of the voice assistants are included in the second voice message;
and if the second voice message comprises the keywords associated with the identification of the first voice assistant, determining that the first voice assistant is the target voice assistant.
Optionally, before the determining a target voice assistant from the plurality of voice assistants according to the trigger event, the method includes:
acquiring a first voice task input by a user through a second voice assistant awakened by the electronic equipment at present;
determining that the triggering event occurs when the first voice task instructs the second voice assistant to wake up the first voice assistant.
Optionally, the method further comprises:
acquiring a second voice task input by a user through a first voice assistant which is awakened currently by the electronic equipment, wherein the second voice task comprises target task information indicating the execution of the second voice assistant;
calling an application interface of the second voice assistant through the target task information based on the first voice assistant to acquire response information of the second voice assistant for the target task information;
presenting the response information by the first voice assistant.
In a second aspect of the embodiments of the present disclosure, a multi-voice assistant control apparatus is provided, where the apparatus is applied to an electronic device integrating multiple voice assistants, and includes:
the voice assistant determination module is configured to determine a target voice assistant from a plurality of voice assistants according to a trigger event when the trigger event for waking up the voice assistant is detected, wherein the wake-up functions of the plurality of voice assistants are all in an enabled state;
a voice assistant wake-up module configured to wake up the target voice assistant.
Optionally, the apparatus comprises:
a voice information collection module configured to collect input first voice information before the target voice assistant is determined from the plurality of voice assistants according to the trigger event;
the keyword judgment module is configured to judge whether the first voice message comprises a keyword which is consistent with a preset awakening word;
a trigger event determination module configured to determine that the trigger event occurs if the first voice message includes a keyword consistent with the preset wake-up word.
Optionally, the triggering event includes acquiring second voice information for waking up a voice assistant, and the voice assistant determining module includes:
a keyword judgment sub-module configured to judge whether a keyword associated with any of the voice assistant identifiers is included in the second voice message;
and the voice assistant determining sub-module is configured to determine that the first voice assistant is the target voice assistant if the second voice message comprises a keyword associated with the identifier of the first voice assistant.
Optionally, the apparatus comprises:
a first voice task obtaining module configured to obtain a first voice task input by a user through a second voice assistant which is already woken up currently by the electronic equipment before the target voice assistant is determined from the plurality of voice assistants according to the trigger event;
a trigger event determination module configured to determine that the trigger event occurs when the first voice task instructs the second voice assistant to wake up the first voice assistant.
Optionally, the apparatus further comprises:
the second voice task acquisition module is configured to acquire a second voice task input by a user through a first voice assistant which is awakened currently by the electronic equipment, and the second voice task comprises target task information indicating execution of the second voice assistant;
a response information acquisition module configured to call an application interface of the second voice assistant through the target task information based on the first voice assistant to acquire response information of the second voice assistant for the target task information;
a response information presentation module configured to present the response information by the first voice assistant.
In a third aspect of the embodiments of the present disclosure, a multi-voice assistant control apparatus is provided, where the apparatus is applied to an electronic device integrating multiple voice assistants, and the apparatus includes:
a processor;
a memory for storing processor-executable instructions;
wherein the processor is configured to:
when a trigger event for waking up a voice assistant is detected, determining a target voice assistant from a plurality of voice assistants according to the trigger event, wherein the wake-up functions of the plurality of voice assistants are all in an enabled state;
waking the target voice assistant.
In a fourth aspect of the embodiments of the present disclosure, a computer-readable storage medium is provided, on which computer program instructions are stored, which when executed by a processor, implement the steps of the multi-voice assistant control method provided in the first aspect of the embodiments of the present disclosure.
The technical scheme provided by the embodiment of the disclosure can have the following beneficial effects:
by the technical scheme, for the electronic equipment integrated with the voice assistants, when a trigger event for waking up the voice assistants is detected, the target voice assistant can be determined from the voice assistants of which the wake-up functions are all in the enabled state according to the trigger event, and the target voice assistant is woken up. Therefore, each voice assistant awakening function is kept in an enabled state, different voice assistants can be awakened directly through different trigger events, and the technical problem that when the voice assistants are switched by the existing electronic equipment, the voice assistants need to be manually enabled, and the switching mode is not flexible enough is solved.
It is to be understood that both the foregoing general description and the following detailed description are exemplary and explanatory only and are not restrictive of the disclosure.
Drawings
The accompanying drawings, which are incorporated in and constitute a part of this specification, illustrate embodiments consistent with the present disclosure and together with the description, serve to explain the principles of the disclosure.
FIG. 1 is a flow diagram illustrating a multi-voice assistant control method in accordance with an exemplary embodiment.
FIG. 2 is a flowchart illustrating a multi-voice assistant control method according to an example embodiment.
FIG. 3 is a flowchart illustrating a multi-voice assistant control method according to an example embodiment.
FIG. 4 is a flow diagram illustrating a multi-voice assistant control method in accordance with an exemplary embodiment.
FIG. 5 is a flow diagram illustrating a multi-voice assistant control method in accordance with an exemplary embodiment.
FIG. 6 is a block diagram illustrating a multi-voice assistant control apparatus according to an example embodiment.
FIG. 7 is a block diagram illustrating a multi-voice assistant control apparatus according to an example embodiment.
FIG. 8 is a block diagram illustrating a multi-voice assistant control apparatus according to an example embodiment.
FIG. 9 is a block diagram illustrating a multi-voice assistant control apparatus according to an example embodiment.
FIG. 10 is a block diagram illustrating a multi-voice assistant control apparatus according to an example embodiment.
FIG. 11 is a block diagram illustrating a multi-voice assistant control apparatus according to an example embodiment.
Detailed Description
Reference will now be made in detail to the exemplary embodiments, examples of which are illustrated in the accompanying drawings. When the following description refers to the accompanying drawings, like numbers in different drawings represent the same or similar elements unless otherwise indicated. The implementations described in the exemplary embodiments below are not intended to represent all implementations consistent with the present disclosure. Rather, they are merely examples of apparatus and methods consistent with certain aspects of the present disclosure, as detailed in the appended claims.
FIG. 1 is a flow chart illustrating a multi-voice assistant control method applied to an electronic device integrating a plurality of voice assistants according to an exemplary embodiment, where the method includes the following steps:
step 101, when a trigger event for waking up a voice assistant is detected, determining a target voice assistant from a plurality of voice assistants according to the trigger event;
step 102, wake up the target voice assistant.
The triggering event includes voice information for waking up the voice assistant, a key operation or a touch screen operation, where the touch screen operation may be a user performing a click operation or a sliding operation in a designated direction based on a designated area of a touch screen of the electronic device, for example, double-clicking a menu key or drawing a preset wake-up gesture path, and the electronic device determines an operation object, that is, a target voice assistant, according to a preset corresponding relationship between specific content of the triggering event and the voice assistant. After determining the target voice assistant, the electronic device may wake up the target voice assistant by turning off the screen or in a customized manner.
It should be noted that, a plurality of voice assistants in the electronic device are all in an enabled state, and can be awakened accordingly in case of a trigger event, and the plurality of voice assistants can coexist simultaneously.
By adopting the method, each voice assistant awakening function keeps the enabled state, different voice assistants can be awakened directly through different trigger events, and the technical problem that the voice assistants are required to be manually enabled when the existing electronic equipment switches the voice assistants, and the switching mode is not flexible enough is solved.
FIG. 2 is a flowchart illustrating a multi-voice assistant control method applied to an electronic device integrating a plurality of voice assistants according to an exemplary embodiment, as shown in FIG. 2, the method comprising the steps of:
step 201, collecting input first voice information;
step 202, judging whether the first voice message comprises a keyword which is consistent with a preset awakening word;
step 203, if the first voice message includes a keyword conforming to a preset awakening word, determining that a trigger event occurs;
step 204, determining a target voice assistant from a plurality of voice assistants according to a trigger event;
step 205, the target voice assistant is awakened.
In one possible implementation, the triggering event may be the acquisition of voice information by the electronic device to wake up a voice assistant. For example, the electronic device may collect voice information from the surrounding environment in real time through a recording function, input the collected voice information as first voice information into the system for recognition, and perform step 202 on the recognition result.
In step 202, the recognition result of the first voice message obtained in step 201 may be determined to determine whether the result includes a keyword consistent with the preset wake-up word. For example, the keyword of the preset wakeup word may be a name of each voice assistant, or may be a word or a sentence customized by the user according to actual needs. For example, the electronic device includes three voice assistants, wherein the name of voice assistant a is small a, the name of voice assistant B is small B, the name of voice assistant C is small C, and if the recognition result of the first voice message includes a keyword of small a, small B, or small C, the determination result in step 202 is yes. If the preset awakening words customized by the user are words such as "hello" and "out bar", the determination result in the step 202 may be determined to be "yes" when the recognition result of the first voice message includes the words.
It should be noted that the preset wake-up word customized by the user may or may not specifically specify the voice assistant. For example, if a preset corresponding relationship exists between a preset wake-up word and a voice assistant, for example, the preset wake-up word "wake-up voice assistant" corresponds to the voice assistant a, "hello" corresponds to the voice assistant B, "out-of-the-bar" corresponds to the voice assistant C, it indicates that the preset wake-up word of the user can specifically specify which voice assistant to wake up, and the electronic device can determine the target voice assistant by performing keyword judgment once; if the preset corresponding relation does not exist between the preset awakening words and the voice assistant, the preset awakening words 'awakening the voice assistant' and 'hello' defined by the user do not specifically specify which voice assistant to awaken, and the system can directly awaken the default voice assistant or prompt the user to select which voice assistant to awaken again.
In the case that the determination result in step 202 is "yes", it is determined that a trigger event occurs, and the target voice assistant determined according to the content of the trigger event is awakened.
By adopting the method, whether the acquired voice information comprises the keyword which is consistent with the preset awakening word or not can be judged, and the target voice assistant is awakened correspondingly according to the judgment result. In addition, the plurality of voice assistant awakening functions are kept in the enabled state, different voice assistants can be directly awakened through different trigger events, preset awakening words can be defined by a user, the use method of the voice assistants is enriched, the single role of the voice assistants is avoided, and the technical problem that when the voice assistants are switched by the existing electronic equipment, the voice assistants need to be manually enabled, and the switching mode is not flexible is solved.
FIG. 3 is a flowchart illustrating a multi-voice assistant control method according to an exemplary embodiment, applied to an electronic device integrating multiple voice assistants, as shown in FIG. 3, the method comprising the steps of:
step 301, judging whether the second voice message includes a keyword associated with the identifier of any voice assistant;
step 302, if the second voice message includes a keyword associated with the identifier of the first voice assistant, determining that the first voice assistant is a target voice assistant;
step 303 wakes up the target voice assistant.
In another possible implementation, when the first voice message collected and input by the electronic device includes a keyword matching a preset wake-up word, but does not specifically specify which voice assistant to wake up, the system may prompt the user to select which voice assistant to wake up again. In this case, the electronic device may collect second voice information for waking up the voice assistant from the surrounding environment through the recording function, input the collected second voice information into the system for recognition, and perform step 301 on the recognition result.
It may be determined whether a keyword associated with an identifier of any of the voice assistants is included in the recognition result of the second voice information through step 301. The identification of the voice assistant may include the name or number of the voice assistant, and may also be a user-defined keyword, where there is a preset correspondence between the keyword and the voice assistant role. For example, the name of the first voice assistant is small a, the number is 1, the user-defined keyword is "intern", and if the collected second voice information includes a word associated with the identifier of the first voice assistant, it may be determined that the first voice assistant is the target voice assistant, and the electronic device performs an operation of waking up the first voice assistant.
By adopting the method, the target voice assistant can be determined again by collecting and judging the keywords in the second voice message under the condition that the specific voice assistant can not be awakened according to the first voice message, so that the technical problem that the existing electronic equipment needs to manually enable the voice assistant when switching the voice assistant and the switching mode is not flexible is solved.
FIG. 4 is a flowchart illustrating a multi-voice assistant control method applied to an electronic device integrating a plurality of voice assistants according to an exemplary embodiment, as shown in FIG. 4, the method comprising the steps of:
step 401, acquiring a first voice task input by a user through a second voice assistant which is awakened currently by the electronic equipment;
step 402, when the first voice task instructs the second voice assistant to wake up the first voice assistant, determining that a trigger event occurs;
at step 403, the target voice assistant is awakened.
In the embodiment of the disclosure, the second voice assistant is a woken voice assistant, when the second voice assistant acquires a first voice task input by a user and the task instructs the second voice assistant to wake up the first voice assistant, it is determined that a trigger event occurs, and according to the trigger event, it may be determined that the first voice assistant is a target voice assistant, and a wake-up operation is performed on the first voice assistant. Exemplarily, the voice assistant B is in an awakened state, when a voice task acquired by the voice assistant B includes a preset keyword for awakening the voice assistant a, it is determined that a trigger event occurs, and it is determined that the voice assistant a is the target voice assistant, at this time, the electronic device awakens the voice assistant a according to the voice task, and the voice assistant a performs subsequent interaction with the user.
By adopting the method, when one voice assistant is in an awakening state, other voice assistants can be awakened in a chained manner by releasing the voice task, so that the problem that a user awakens the voice assistant in a single role is avoided, the interestingness of man-machine interaction is improved, and the technical problem that the voice assistant needs to be manually enabled when the existing electronic equipment switches the voice assistants and the switching manner is not flexible enough is solved.
FIG. 5 is a flowchart illustrating a multi-voice assistant control method applied to an electronic device integrating multiple voice assistants according to an exemplary embodiment, as shown in FIG. 5, the method comprising the steps of:
step 501, when a trigger event for waking up a voice assistant is detected, determining that a target voice assistant is a first voice assistant;
step 502, waking up the target voice assistant;
step 503, acquiring a second voice task input by the user through a first voice assistant which is awakened by the electronic equipment currently;
step 504, based on the first voice assistant, calling an application interface of the second voice assistant through the target task information to obtain response information of the second voice assistant for the target task information;
step 505, the response message is presented by the first voice assistant.
In another possible implementation manner, the voice assistant is awakened through a preset awakening word, the first voice assistant is determined to be the target voice assistant, and after the first voice assistant is awakened, a second voice task input by a user can be acquired through the first voice assistant.
For example, the first voice assistant is small a, the second voice assistant is small B, and in the case that the small voice assistant a is awakened, the second voice task input by the user is "small a, ask how small B hides the Excel cell". At this time, the voice assistant widget A recognizes that the voice task needs to call an application interface of the second voice assistant widget B so as to transmit the query of the user to the second voice assistant widget B, and the processing result of the second voice assistant widget B is displayed on the human-computer interaction interface through the first voice assistant widget A.
By adopting the method, based on the first voice assistant system, other voice assistants can be called to serve the user according to the voice task, and the technical problem that the existing electronic equipment needs to manually enable the voice assistant when switching the voice assistant and the switching mode is not flexible is solved.
Fig. 6 is a block diagram of a multi-voice assistant control apparatus 600 applied to an electronic device integrating a plurality of voice assistants according to an exemplary embodiment, and referring to fig. 6, the apparatus includes:
the voice assistant determination module 610 is configured to determine a target voice assistant from a plurality of voice assistants according to a trigger event when the trigger event for waking up the voice assistant is detected, wherein the wake-up functions of the plurality of voice assistants are all in an enabled state;
a voice assistant wake module 620 configured to wake the target voice assistant.
By adopting the device, each voice assistant awakening function keeps the enabled state, and different voice assistants can be directly awakened through the voice assistant determining module 610 according to different trigger events, so that the technical problem that the voice assistants are required to be manually enabled when the existing electronic equipment switches the voice assistants, and the switching mode is not flexible enough is solved.
Fig. 7 is a block diagram of a multi-voice assistant control apparatus 600 applied to an electronic device integrating a plurality of voice assistants according to an exemplary embodiment, and referring to fig. 7, the apparatus includes:
a voice information collection module 630 configured to collect the input first voice information before determining the target voice assistant from the plurality of voice assistants according to the triggering event;
a keyword judgment module 640 configured to judge whether the first voice message includes a keyword consistent with a preset wake-up word;
a trigger event determination module 650 configured to determine that a trigger event occurs if the first voice message includes a keyword corresponding to a preset wakeup word;
the voice assistant determination module 610 is configured to determine a target voice assistant from a plurality of voice assistants according to a trigger event when the trigger event for waking up the voice assistant is detected, wherein the wake-up functions of the plurality of voice assistants are all in an enabled state;
a voice assistant wake module 620 configured to wake the target voice assistant.
Optionally, the triggering event includes collecting a second voice message for waking up the voice assistant, and referring to fig. 8, the voice assistant determining module 610 may include:
a keyword determination sub-module 611 configured to determine whether the second voice message includes a keyword associated with an identifier of any voice assistant;
a voice assistant determination sub-module 612 configured to determine that the first voice assistant is the target voice assistant if the second voice message includes a keyword associated with the identifier of the first voice assistant.
By adopting the device, whether the acquired voice information comprises the keyword which is consistent with the preset awakening word or not can be judged, and the target voice assistant is awakened correspondingly according to the judgment result. In addition, the plurality of voice assistant awakening functions are kept in the enabled state, different voice assistants can be directly awakened through different trigger events, preset awakening words can be defined by a user, the use method of the voice assistants is enriched, the single role of the voice assistants is avoided, and the technical problem that when the voice assistants are switched by the existing electronic equipment, the voice assistants need to be manually enabled, and the switching mode is not flexible is solved.
Fig. 9 is a block diagram of a multi-voice assistant control apparatus 600 applied to an electronic device integrating a plurality of voice assistants according to an exemplary embodiment, and referring to fig. 9, the apparatus includes:
a first voice task obtaining module 660 configured to obtain a first voice task input by a user through a second voice assistant that is currently awakened by the electronic device before a target voice assistant is determined from the plurality of voice assistants according to a trigger event;
a trigger event determination module 650 configured to determine that a trigger event occurs when the first voice task instructs the second voice assistant to wake up the first voice assistant;
the voice assistant determination module 610 is configured to determine a target voice assistant from a plurality of voice assistants according to a trigger event when the trigger event for waking up the voice assistant is detected, wherein the wake-up functions of the plurality of voice assistants are all in an enabled state;
a voice assistant wake module 620 configured to wake the target voice assistant.
By adopting the device, when one voice assistant is in an awakening state, other voice assistants can be awakened in a chained mode of issuing voice tasks, the condition that a user awakens the voice assistant to have a single interaction role is avoided, the interestingness of man-machine interaction is improved, and the technical problem that the voice assistant is required to be manually enabled when the existing electronic equipment switches the voice assistants, and the switching mode is not flexible enough is solved.
Fig. 10 is a block diagram of a multi-voice assistant control apparatus 600 applied to an electronic device integrating a plurality of voice assistants according to an exemplary embodiment, and referring to fig. 10, the apparatus includes:
the voice assistant determination module 610 is configured to determine a target voice assistant from a plurality of voice assistants according to a trigger event when the trigger event for waking up the voice assistant is detected, wherein the wake-up functions of the plurality of voice assistants are all in an enabled state;
a voice assistant wake-up module 620 configured to wake up the target voice assistant;
a second voice task obtaining module 670 configured to obtain, by the first voice assistant that is currently woken up by the electronic device, a second voice task input by the user, where the second voice task includes target task information indicating that the second voice assistant executes;
a response information obtaining module 680 configured to call an application interface of the second voice assistant through the target task information based on the first voice assistant to obtain response information of the second voice assistant for the target task information;
a response information presentation module 690 configured to present the response information via the first voice assistant.
By adopting the device, based on the first voice assistant system, other voice assistants can be called to serve the user according to the voice task, and the technical problem that the existing electronic equipment needs to manually enable the voice assistant when switching the voice assistant and the switching mode is not flexible is solved.
With regard to the apparatus in the above-described embodiment, the specific manner in which each module performs the operation has been described in detail in the embodiment related to the method, and will not be elaborated here.
The present disclosure also provides a computer-readable storage medium having stored thereon computer program instructions which, when executed by a processor, implement the steps of the multi-voice assistant control method provided by the present disclosure.
FIG. 11 is a block diagram illustrating a multi-voice assistant control apparatus 1100 according to an example embodiment. For example, the apparatus 1100 may be a smartphone, tablet device, personal digital assistant, or the like.
Referring to fig. 11, apparatus 1100 may include one or more of the following components: processing component 1101, memory 1102, power component 1103, multimedia component 1104, audio component 1105, input/output (I/O) interface 1106, sensor component 1107, and communications component 1108.
The processing component 1101 generally controls the overall operation of the device 1100, such as operations associated with display, touch operations, and recording operations. The processing component 1101 may include one or more processors 1109 to execute instructions to perform all or a portion of the steps of the multi-voice assistant control method described above. Further, the processing component 1101 can include one or more modules that facilitate interaction between the processing component 1101 and other components. For example, the processing component 1101 can include a multimedia module to facilitate interaction between the multimedia component 1104 and the processing component 1101.
The memory 1102 is configured to store various types of data to support operation at the device 1100. Examples of such data include instructions for any application or method operating on device 1100, contact data, phonebook data, messages, pictures, videos, and so forth. The memory 1102 may be implemented by any type or combination of volatile or non-volatile memory devices such as Static Random Access Memory (SRAM), electrically erasable programmable read-only memory (EEPROM), erasable programmable read-only memory (EPROM), programmable read-only memory (PROM), read-only memory (ROM), magnetic memory, flash memory, magnetic or optical disks.
The power components 1103 provide power to the various components of the device 1100. The power components 1103 may include a power management system, one or more power sources, and other components associated with generating, managing, and distributing power for the apparatus 1100.
The multimedia component 1104 includes a screen that provides an output interface between the device 1100 and the user. In some embodiments, the screen may include a Liquid Crystal Display (LCD) and a Touch Panel (TP). If the screen includes a touch panel, the screen may be implemented as a touch screen to receive an input signal from a user. The touch panel includes one or more touch sensors to sense touch, slide, and gestures on the touch panel. The touch sensor may not only sense the boundary of a touch or slide action, but also detect the duration and pressure associated with the touch or slide operation. In some embodiments, the multimedia component 1104 includes a front facing camera and/or a rear facing camera. The front camera and/or the rear camera may receive external multimedia data when the device 1100 is in an operating mode, such as a shooting mode or a video mode. Each front camera and rear camera may be a fixed optical lens system or have a focal length and optical zoom capability.
Audio component 1105 is configured to output and/or input audio signals. For example, audio component 1105 may include a Microphone (MIC) configured to receive external audio signals when apparatus 1100 is in an operational mode, such as a call mode, recording mode, and voice recognition mode. The received audio signals may further be stored in memory 1102 or transmitted via communications component 1108. In some embodiments, audio component 1105 further includes a speaker for outputting audio signals.
The I/O interface 1106 provides an interface between the processing component 1101 and peripheral interface modules, which may be keyboards, click wheels, buttons, etc. These buttons may include, but are not limited to: a home button, a volume button, a start button, and a lock button.
Sensor component 1107 includes one or more sensors to provide various aspects of state estimation for apparatus 1100. For example, sensor component 1107 may detect the open/closed state of device 1100, the relative positioning of components, such as a display and keypad of device 1100, sensor component 1107 may also detect a change in the position of device 1100 or a component of device 1100, the presence or absence of user contact with device 1100, an orientation or acceleration/deceleration of device 1100, and a change in the temperature of device 1100. Sensor assembly 1107 may include a proximity sensor configured to detect the presence of a nearby object in the absence of any physical contact. Sensor assembly 1107 may also include a photosensor, such as a CMOS or CCD image sensor, for use in imaging applications. In some embodiments, the sensor assembly 1107 may also include an acceleration sensor, a gyroscope sensor, a magnetic sensor, a pressure sensor, or a temperature sensor.
The communication component 1108 is configured to facilitate communications between the apparatus 1100 and other devices in a wired or wireless manner. The apparatus 1100 may access a Wireless network based on a communication standard, such as Wi-Fi (Wireless-Fidelity), 2G or 3G, or a combination thereof. In an exemplary embodiment, the communication component 1108 receives a broadcast signal or broadcast related information from an external broadcast management system via a broadcast channel. In an exemplary embodiment, the communication component 1108 also includes a Near Field Communication (NFC) module to facilitate short-range communications. For example, the NFC module may be implemented based on Radio Frequency Identification (RFID) technology, infrared data association (IrDA) technology, Ultra Wideband (UWB) technology, Bluetooth (BT) technology, and other technologies.
In an exemplary embodiment, the apparatus 1100 may be implemented by one or more Application Specific Integrated Circuits (ASICs), Digital Signal Processors (DSPs), Digital Signal Processing Devices (DSPDs), Programmable Logic Devices (PLDs), Field Programmable Gate Arrays (FPGAs), controllers, micro-controllers, microprocessors or other electronic components for performing the multi-voice assistant control method described above.
In an exemplary embodiment, a non-transitory computer-readable storage medium comprising instructions, such as memory 1102 comprising instructions, executable by a processor of apparatus 1100 to perform the multi-voice assistant control method described above is also provided. For example, the non-transitory computer readable storage medium may be a ROM, a Random Access Memory (RAM), a CD-ROM, a magnetic tape, a floppy disk, an optical data storage device, and the like.
Other embodiments of the disclosure will be apparent to those skilled in the art from consideration of the specification and practice of the disclosure. This application is intended to cover any variations, uses, or adaptations of the disclosure following, in general, the principles of the disclosure and including such departures from the present disclosure as come within known or customary practice within the art to which the disclosure pertains. It is intended that the specification and examples be considered as exemplary only, with a true scope and spirit of the disclosure being indicated by the following claims.
It will be understood that the present disclosure is not limited to the precise arrangements described above and shown in the drawings and that various modifications and changes may be made without departing from the scope thereof. The scope of the present disclosure is limited only by the appended claims.

Claims (10)

1. A multi-voice assistant control method applied to an electronic device integrating a plurality of voice assistants, the method comprising:
when a trigger event for awakening the voice assistant is detected, determining a first voice assistant as a target voice assistant according to a preset corresponding relation between the trigger event and the voice assistant, wherein the first voice assistant is at least one voice assistant in a plurality of voice assistants of which awakening functions are all in an enabling state;
waking the target voice assistant;
acquiring a second voice task input by a user through the first voice assistant which is awakened currently by the electronic equipment, wherein the second voice task comprises target task information indicating the execution of the second voice assistant;
calling an application interface of the second voice assistant through the target task information based on the first voice assistant to acquire response information of the second voice assistant for the target task information;
presenting the response information by the first voice assistant.
2. The method of claim 1, wherein prior to the determining the first voice assistant as the target voice assistant based on the predetermined correspondence between the triggering event and the voice assistant, comprising:
collecting input first voice information;
judging whether the first voice message comprises a keyword which is consistent with a preset awakening word or not;
and if the first voice message comprises a keyword which is consistent with the preset awakening word, determining that the trigger event occurs.
3. The method of claim 1, wherein the triggering event includes collecting second voice information for waking up a voice assistant, and wherein determining the first voice assistant as the target voice assistant according to a preset correspondence between the triggering event and the voice assistant comprises:
determining whether keywords associated with the identification of any of the voice assistants are included in the second voice message;
and if the second voice message comprises the keywords associated with the identification of the first voice assistant, determining that the first voice assistant is the target voice assistant.
4. The method of claim 1, wherein prior to the determining the first voice assistant as the target voice assistant based on the predetermined correspondence between the triggering event and the voice assistant, comprising:
acquiring a first voice task input by a user through a third voice assistant awakened by the electronic equipment at present;
determining that the triggering event occurs when the first voice task instructs the third voice assistant to wake up the first voice assistant.
5. A multi-voice assistant control apparatus applied to an electronic device integrating a plurality of voice assistants, the apparatus comprising:
the voice assistant determining module is configured to determine a first voice assistant as a target voice assistant according to a preset corresponding relation between a trigger event and the voice assistant when the trigger event for waking up the voice assistant is detected, wherein the first voice assistant is at least one of a plurality of voice assistants of which the wake-up functions are all in an enabling state;
a voice assistant wake-up module configured to wake up the target voice assistant;
the device further comprises:
a second voice task obtaining module configured to obtain, by the first voice assistant that is currently woken up by the electronic device, a second voice task input by a user, where the second voice task includes target task information indicating that the second voice assistant executes;
a response information acquisition module configured to call an application interface of the second voice assistant through the target task information based on the first voice assistant to acquire response information of the second voice assistant for the target task information;
a response information presentation module configured to present the response information by the first voice assistant.
6. The apparatus of claim 5, comprising:
a voice information collecting module configured to collect the input first voice information before the first voice assistant is determined as the target voice assistant according to the triggering event;
the keyword judgment module is configured to judge whether the first voice message comprises a keyword which is consistent with a preset awakening word;
a trigger event determination module configured to determine that the trigger event occurs if the first voice message includes a keyword consistent with the preset wake-up word.
7. The apparatus of claim 5, wherein the triggering event comprises a collection of second voice information for waking a voice assistant, and wherein the voice assistant determination module comprises:
a keyword judgment sub-module configured to judge whether a keyword associated with any of the voice assistant identifiers is included in the second voice message;
and the voice assistant determining sub-module is configured to determine that the first voice assistant is the target voice assistant if the second voice message comprises a keyword associated with the identifier of the first voice assistant.
8. The apparatus of claim 5, comprising:
a first voice task obtaining module configured to obtain a first voice task input by a user through a third voice assistant which is already woken up currently by the electronic equipment before the first voice assistant is determined as a target voice assistant according to the trigger event;
a trigger event determination module configured to determine that the trigger event occurs when the first voice task instructs the third voice assistant to wake up the first voice assistant.
9. A multi-voice assistant control apparatus for use with an electronic device that integrates multiple voice assistants, comprising:
a processor;
a memory for storing processor-executable instructions;
wherein the processor is configured to:
when a trigger event for awakening the voice assistant is detected, determining a first voice assistant as a target voice assistant according to a preset corresponding relation between the trigger event and the voice assistant, wherein the first voice assistant is at least one voice assistant in a plurality of voice assistants of which awakening functions are all in an enabling state;
waking the target voice assistant;
acquiring a second voice task input by a user through the first voice assistant which is awakened currently by the electronic equipment, wherein the second voice task comprises target task information indicating the execution of the second voice assistant;
calling an application interface of the second voice assistant through the target task information based on the first voice assistant to acquire response information of the second voice assistant for the target task information;
presenting the response information by the first voice assistant.
10. A computer-readable storage medium, on which computer program instructions are stored, which program instructions, when executed by a processor, carry out the steps of the method according to any one of claims 1 to 4.
CN201711287860.8A 2017-12-07 2017-12-07 Multi-voice assistant control method, device and computer readable storage medium Active CN107919123B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201711287860.8A CN107919123B (en) 2017-12-07 2017-12-07 Multi-voice assistant control method, device and computer readable storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201711287860.8A CN107919123B (en) 2017-12-07 2017-12-07 Multi-voice assistant control method, device and computer readable storage medium

Publications (2)

Publication Number Publication Date
CN107919123A CN107919123A (en) 2018-04-17
CN107919123B true CN107919123B (en) 2022-06-03

Family

ID=61898470

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201711287860.8A Active CN107919123B (en) 2017-12-07 2017-12-07 Multi-voice assistant control method, device and computer readable storage medium

Country Status (1)

Country Link
CN (1) CN107919123B (en)

Families Citing this family (183)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8677377B2 (en) 2005-09-08 2014-03-18 Apple Inc. Method and apparatus for building an intelligent automated assistant
US9318108B2 (en) 2010-01-18 2016-04-19 Apple Inc. Intelligent automated assistant
US8977255B2 (en) 2007-04-03 2015-03-10 Apple Inc. Method and system for operating a multi-function portable electronic device using voice-activation
US8676904B2 (en) 2008-10-02 2014-03-18 Apple Inc. Electronic devices with voice command and contextual data processing capabilities
US20120309363A1 (en) 2011-06-03 2012-12-06 Apple Inc. Triggering notifications associated with tasks items that represent tasks to perform
US10276170B2 (en) 2010-01-18 2019-04-30 Apple Inc. Intelligent automated assistant
US8682667B2 (en) 2010-02-25 2014-03-25 Apple Inc. User profiling for selecting user specific voice input processing information
US9262612B2 (en) 2011-03-21 2016-02-16 Apple Inc. Device access using voice authentication
US10417037B2 (en) 2012-05-15 2019-09-17 Apple Inc. Systems and methods for integrating third party services with a digital assistant
US10199051B2 (en) 2013-02-07 2019-02-05 Apple Inc. Voice trigger for a digital assistant
US10652394B2 (en) 2013-03-14 2020-05-12 Apple Inc. System and method for processing voicemail
US10748529B1 (en) 2013-03-15 2020-08-18 Apple Inc. Voice activated device for use with a voice-based digital assistant
US10176167B2 (en) 2013-06-09 2019-01-08 Apple Inc. System and method for inferring user intent from speech inputs
KR101922663B1 (en) 2013-06-09 2018-11-28 애플 인크. Device, method, and graphical user interface for enabling conversation persistence across two or more instances of a digital assistant
US10296160B2 (en) 2013-12-06 2019-05-21 Apple Inc. Method for extracting salient dialog usage from live data
US10170123B2 (en) 2014-05-30 2019-01-01 Apple Inc. Intelligent assistant for home automation
US9715875B2 (en) 2014-05-30 2017-07-25 Apple Inc. Reducing the need for manual start/end-pointing and trigger phrases
US9633004B2 (en) 2014-05-30 2017-04-25 Apple Inc. Better resolution when referencing to concepts
US9430463B2 (en) 2014-05-30 2016-08-30 Apple Inc. Exemplar-based natural language processing
CN110797019B (en) 2014-05-30 2023-08-29 苹果公司 Multi-command single speech input method
US9338493B2 (en) 2014-06-30 2016-05-10 Apple Inc. Intelligent automated assistant for TV user interactions
US10074360B2 (en) 2014-09-30 2018-09-11 Apple Inc. Providing an indication of the suitability of speech recognition
US9668121B2 (en) 2014-09-30 2017-05-30 Apple Inc. Social reminders
US10127911B2 (en) 2014-09-30 2018-11-13 Apple Inc. Speaker identification and unsupervised speaker adaptation techniques
US10152299B2 (en) 2015-03-06 2018-12-11 Apple Inc. Reducing response latency of intelligent automated assistants
US9721566B2 (en) 2015-03-08 2017-08-01 Apple Inc. Competing devices responding to voice triggers
US9886953B2 (en) 2015-03-08 2018-02-06 Apple Inc. Virtual assistant activation
US10460227B2 (en) 2015-05-15 2019-10-29 Apple Inc. Virtual assistant in a communication session
US10200824B2 (en) 2015-05-27 2019-02-05 Apple Inc. Systems and methods for proactively identifying and surfacing relevant content on a touch-sensitive device
US10083688B2 (en) 2015-05-27 2018-09-25 Apple Inc. Device voice control for selecting a displayed affordance
US9578173B2 (en) 2015-06-05 2017-02-21 Apple Inc. Virtual assistant aided communication with 3rd party service in a communication session
US20160378747A1 (en) 2015-06-29 2016-12-29 Apple Inc. Virtual assistant for media playback
US10747498B2 (en) 2015-09-08 2020-08-18 Apple Inc. Zero latency digital assistant
US10671428B2 (en) 2015-09-08 2020-06-02 Apple Inc. Distributed personal assistant
US10331312B2 (en) 2015-09-08 2019-06-25 Apple Inc. Intelligent automated assistant in a media environment
US10740384B2 (en) 2015-09-08 2020-08-11 Apple Inc. Intelligent automated assistant for media search and playback
US10691473B2 (en) 2015-11-06 2020-06-23 Apple Inc. Intelligent automated assistant in a messaging environment
US10956666B2 (en) 2015-11-09 2021-03-23 Apple Inc. Unconventional virtual assistant interactions
US10223066B2 (en) 2015-12-23 2019-03-05 Apple Inc. Proactive assistance based on dialog communication between devices
US10509626B2 (en) 2016-02-22 2019-12-17 Sonos, Inc Handling of loss of pairing between networked devices
US9965247B2 (en) 2016-02-22 2018-05-08 Sonos, Inc. Voice controlled media playback system based on user profile
US10743101B2 (en) 2016-02-22 2020-08-11 Sonos, Inc. Content mixing
US10095470B2 (en) 2016-02-22 2018-10-09 Sonos, Inc. Audio response playback
US9947316B2 (en) 2016-02-22 2018-04-17 Sonos, Inc. Voice control of a media playback system
US10264030B2 (en) 2016-02-22 2019-04-16 Sonos, Inc. Networked microphone device control
US11227589B2 (en) 2016-06-06 2022-01-18 Apple Inc. Intelligent list reading
US9978390B2 (en) 2016-06-09 2018-05-22 Sonos, Inc. Dynamic player selection for audio signal processing
US10586535B2 (en) 2016-06-10 2020-03-10 Apple Inc. Intelligent digital assistant in a multi-tasking environment
DK179415B1 (en) 2016-06-11 2018-06-14 Apple Inc Intelligent device arbitration and control
DK201670540A1 (en) 2016-06-11 2018-01-08 Apple Inc Application integration with a digital assistant
US10152969B2 (en) 2016-07-15 2018-12-11 Sonos, Inc. Voice detection by multiple devices
US10134399B2 (en) 2016-07-15 2018-11-20 Sonos, Inc. Contextualization of voice inputs
US10115400B2 (en) 2016-08-05 2018-10-30 Sonos, Inc. Multiple voice services
US10474753B2 (en) 2016-09-07 2019-11-12 Apple Inc. Language identification using recurrent neural networks
US9942678B1 (en) 2016-09-27 2018-04-10 Sonos, Inc. Audio playback settings for voice interaction
US9743204B1 (en) 2016-09-30 2017-08-22 Sonos, Inc. Multi-orientation playback device microphones
US10181323B2 (en) 2016-10-19 2019-01-15 Sonos, Inc. Arbitration-based voice recognition
US11204787B2 (en) 2017-01-09 2021-12-21 Apple Inc. Application integration with a digital assistant
US11183181B2 (en) 2017-03-27 2021-11-23 Sonos, Inc. Systems and methods of multiple voice services
US10417266B2 (en) 2017-05-09 2019-09-17 Apple Inc. Context-aware ranking of intelligent response suggestions
DK201770383A1 (en) 2017-05-09 2018-12-14 Apple Inc. User interface for correcting recognition errors
DK180048B1 (en) 2017-05-11 2020-02-04 Apple Inc. MAINTAINING THE DATA PROTECTION OF PERSONAL INFORMATION
US10395654B2 (en) 2017-05-11 2019-08-27 Apple Inc. Text normalization based on a data-driven learning network
US10726832B2 (en) 2017-05-11 2020-07-28 Apple Inc. Maintaining privacy of personal information
DK201770428A1 (en) 2017-05-12 2019-02-18 Apple Inc. Low-latency intelligent automated assistant
DK179745B1 (en) 2017-05-12 2019-05-01 Apple Inc. SYNCHRONIZATION AND TASK DELEGATION OF A DIGITAL ASSISTANT
US11301477B2 (en) 2017-05-12 2022-04-12 Apple Inc. Feedback analysis of a digital assistant
DK179496B1 (en) 2017-05-12 2019-01-15 Apple Inc. USER-SPECIFIC Acoustic Models
DK179549B1 (en) 2017-05-16 2019-02-12 Apple Inc. Far-field extension for digital assistant services
US10311144B2 (en) 2017-05-16 2019-06-04 Apple Inc. Emoji word sense disambiguation
US20180336275A1 (en) 2017-05-16 2018-11-22 Apple Inc. Intelligent automated assistant for media exploration
US20180336892A1 (en) 2017-05-16 2018-11-22 Apple Inc. Detecting a trigger of a digital assistant
US10475449B2 (en) 2017-08-07 2019-11-12 Sonos, Inc. Wake-word detection suppression
US10048930B1 (en) 2017-09-08 2018-08-14 Sonos, Inc. Dynamic computation of system response volume
US10446165B2 (en) 2017-09-27 2019-10-15 Sonos, Inc. Robust short-time fourier transform acoustic echo cancellation during audio playback
US10621981B2 (en) 2017-09-28 2020-04-14 Sonos, Inc. Tone interference cancellation
US10051366B1 (en) 2017-09-28 2018-08-14 Sonos, Inc. Three-dimensional beam forming with a microphone array
US10482868B2 (en) 2017-09-28 2019-11-19 Sonos, Inc. Multi-channel acoustic echo cancellation
US10466962B2 (en) 2017-09-29 2019-11-05 Sonos, Inc. Media playback system with voice assistance
US10880650B2 (en) 2017-12-10 2020-12-29 Sonos, Inc. Network microphone devices with automatic do not disturb actuation capabilities
US10818290B2 (en) 2017-12-11 2020-10-27 Sonos, Inc. Home graph
WO2019152722A1 (en) 2018-01-31 2019-08-08 Sonos, Inc. Device designation of playback and network microphone device arrangements
US10592604B2 (en) 2018-03-12 2020-03-17 Apple Inc. Inverse text normalization for automatic speech recognition
US10818288B2 (en) 2018-03-26 2020-10-27 Apple Inc. Natural assistant interaction
CN108712566B (en) * 2018-04-27 2020-10-30 维沃移动通信有限公司 Voice assistant awakening method and mobile terminal
US11145294B2 (en) 2018-05-07 2021-10-12 Apple Inc. Intelligent automated assistant for delivering content from user experiences
US10928918B2 (en) 2018-05-07 2021-02-23 Apple Inc. Raise to speak
US11175880B2 (en) 2018-05-10 2021-11-16 Sonos, Inc. Systems and methods for voice-assisted media content selection
US10847178B2 (en) 2018-05-18 2020-11-24 Sonos, Inc. Linear filtering for noise-suppressed speech detection
US10959029B2 (en) 2018-05-25 2021-03-23 Sonos, Inc. Determining and adapting to changes in microphone performance of playback devices
WO2019227370A1 (en) * 2018-05-31 2019-12-05 Dong Xuezhang Method, apparatus and system for controlling multiple voice assistants, and computer-readable storage medium
DK180639B1 (en) 2018-06-01 2021-11-04 Apple Inc DISABILITY OF ATTENTION-ATTENTIVE VIRTUAL ASSISTANT
US10892996B2 (en) 2018-06-01 2021-01-12 Apple Inc. Variable latency device coordination
DK201870355A1 (en) 2018-06-01 2019-12-16 Apple Inc. Virtual assistant operation in multi-device environments
DK179822B1 (en) 2018-06-01 2019-07-12 Apple Inc. Voice interaction at a primary device to access call functionality of a companion device
US10944859B2 (en) 2018-06-03 2021-03-09 Apple Inc. Accelerated task performance
US10681460B2 (en) 2018-06-28 2020-06-09 Sonos, Inc. Systems and methods for associating playback devices with voice assistant services
CN110868648A (en) * 2018-08-27 2020-03-06 杭州海康威视数字技术股份有限公司 Intelligent voice realization method of indoor intercom device and indoor intercom device
US11076035B2 (en) 2018-08-28 2021-07-27 Sonos, Inc. Do not disturb feature for audio notifications
US10461710B1 (en) 2018-08-28 2019-10-29 Sonos, Inc. Media playback system with maximum volume setting
US10878811B2 (en) 2018-09-14 2020-12-29 Sonos, Inc. Networked devices, systems, and methods for intelligently deactivating wake-word engines
US10587430B1 (en) 2018-09-14 2020-03-10 Sonos, Inc. Networked devices, systems, and methods for associating playback devices based on sound codes
US11024331B2 (en) 2018-09-21 2021-06-01 Sonos, Inc. Voice detection optimization using sound metadata
US10811015B2 (en) 2018-09-25 2020-10-20 Sonos, Inc. Voice detection optimization based on selected voice assistant service
US11010561B2 (en) 2018-09-27 2021-05-18 Apple Inc. Sentiment prediction from textual data
US11170166B2 (en) 2018-09-28 2021-11-09 Apple Inc. Neural typographical error modeling via generative adversarial networks
US10839159B2 (en) 2018-09-28 2020-11-17 Apple Inc. Named entity normalization in a spoken dialog system
US11100923B2 (en) 2018-09-28 2021-08-24 Sonos, Inc. Systems and methods for selective wake word detection using neural network models
US11462215B2 (en) 2018-09-28 2022-10-04 Apple Inc. Multi-modal inputs for voice commands
US10692518B2 (en) 2018-09-29 2020-06-23 Sonos, Inc. Linear filtering for noise-suppressed speech detection via multiple network microphone devices
CN109147786A (en) * 2018-09-30 2019-01-04 联想(北京)有限公司 A kind of information processing method and electronic equipment
KR20200044175A (en) 2018-10-05 2020-04-29 삼성전자주식회사 Electronic apparatus and assistant service providing method thereof
US11899519B2 (en) 2018-10-23 2024-02-13 Sonos, Inc. Multiple stage network microphone device with reduced power consumption and processing load
US11475898B2 (en) 2018-10-26 2022-10-18 Apple Inc. Low-latency multi-speaker speech recognition
CN109325337A (en) * 2018-11-05 2019-02-12 北京小米移动软件有限公司 Unlocking method and device
EP3654249A1 (en) 2018-11-15 2020-05-20 Snips Dilated convolutions and gating for efficient keyword spotting
CN109658927A (en) * 2018-11-30 2019-04-19 北京小米移动软件有限公司 Wake-up processing method, device and the management equipment of smart machine
US11183183B2 (en) 2018-12-07 2021-11-23 Sonos, Inc. Systems and methods of operating media playback systems having multiple voice assistant services
US11132989B2 (en) 2018-12-13 2021-09-28 Sonos, Inc. Networked microphone devices, systems, and methods of localized arbitration
US10602268B1 (en) 2018-12-20 2020-03-24 Sonos, Inc. Optimization of network microphone devices using noise classification
US11638059B2 (en) 2019-01-04 2023-04-25 Apple Inc. Content playback on multiple devices
US10867604B2 (en) 2019-02-08 2020-12-15 Sonos, Inc. Devices, systems, and methods for distributed voice processing
US11315556B2 (en) 2019-02-08 2022-04-26 Sonos, Inc. Devices, systems, and methods for distributed voice processing by transmitting sound data associated with a wake word to an appropriate device for identification
CN110310641B (en) * 2019-02-26 2022-08-26 杭州蓦然认知科技有限公司 Method and device for voice assistant
US11348573B2 (en) 2019-03-18 2022-05-31 Apple Inc. Multimodality in digital assistant systems
CN110164433A (en) * 2019-04-03 2019-08-23 美国乐歌有限公司 A kind of speech control system and method for lifting column
US11120794B2 (en) 2019-05-03 2021-09-14 Sonos, Inc. Voice assistant persistence across multiple network microphone devices
DK201970509A1 (en) 2019-05-06 2021-01-15 Apple Inc Spoken notifications
US11423908B2 (en) 2019-05-06 2022-08-23 Apple Inc. Interpreting spoken requests
US11475884B2 (en) 2019-05-06 2022-10-18 Apple Inc. Reducing digital assistant latency when a language is incorrectly determined
US11307752B2 (en) 2019-05-06 2022-04-19 Apple Inc. User configurable task triggers
US11140099B2 (en) 2019-05-21 2021-10-05 Apple Inc. Providing message response suggestions
US11189279B2 (en) 2019-05-22 2021-11-30 Microsoft Technology Licensing, Llc Activation management for multiple voice assistants
US11496600B2 (en) 2019-05-31 2022-11-08 Apple Inc. Remote execution of machine-learned models
DK201970510A1 (en) 2019-05-31 2021-02-11 Apple Inc Voice identification in digital assistant systems
US11289073B2 (en) 2019-05-31 2022-03-29 Apple Inc. Device text to speech
DK180129B1 (en) 2019-05-31 2020-06-02 Apple Inc. User activity shortcut suggestions
US11227599B2 (en) 2019-06-01 2022-01-18 Apple Inc. Methods and user interfaces for voice-based control of electronic devices
US11360641B2 (en) 2019-06-01 2022-06-14 Apple Inc. Increasing the relevance of new available information
US11361756B2 (en) 2019-06-12 2022-06-14 Sonos, Inc. Conditional wake word eventing based on environment
US11200894B2 (en) 2019-06-12 2021-12-14 Sonos, Inc. Network microphone device with command keyword eventing
US10586540B1 (en) 2019-06-12 2020-03-10 Sonos, Inc. Network microphone device with command keyword conditioning
CN110310636A (en) * 2019-06-24 2019-10-08 歌尔股份有限公司 Interaction control method, device, equipment and audio frequency apparatus
CN110335601A (en) * 2019-07-10 2019-10-15 三星电子(中国)研发中心 Voice assistant equipment and its voice awakening method
US10871943B1 (en) 2019-07-31 2020-12-22 Sonos, Inc. Noise classification for event detection
US11138975B2 (en) 2019-07-31 2021-10-05 Sonos, Inc. Locally distributed keyword detection
US11138969B2 (en) 2019-07-31 2021-10-05 Sonos, Inc. Locally distributed keyword detection
CN112466300A (en) * 2019-09-09 2021-03-09 百度在线网络技术(北京)有限公司 Interaction method, electronic device, intelligent device and readable storage medium
CN110718218B (en) * 2019-09-12 2022-08-23 百度在线网络技术(北京)有限公司 Voice processing method, device, equipment and computer storage medium
CN110718219B (en) 2019-09-12 2022-07-22 百度在线网络技术(北京)有限公司 Voice processing method, device, equipment and computer storage medium
US11488406B2 (en) 2019-09-25 2022-11-01 Apple Inc. Text detection using global geometry estimators
CN110659361B (en) * 2019-10-11 2023-01-17 卢卡(北京)智能科技有限公司 Conversation method, device, equipment and medium
US11189286B2 (en) 2019-10-22 2021-11-30 Sonos, Inc. VAS toggle based on device orientation
CN111091814A (en) * 2019-12-13 2020-05-01 晶晨半导体(深圳)有限公司 Method for constructing multi-voice assistant
US11200900B2 (en) 2019-12-20 2021-12-14 Sonos, Inc. Offline voice control
CN111161714B (en) * 2019-12-25 2023-07-21 联想(北京)有限公司 Voice information processing method, electronic equipment and storage medium
CN110933500B (en) * 2019-12-30 2022-07-29 深圳Tcl新技术有限公司 Voice triggering method, device, equipment and computer storage medium
CN111096680B (en) * 2019-12-31 2022-02-01 广东美的厨房电器制造有限公司 Cooking equipment, electronic equipment, voice server, voice control method and device
US11562740B2 (en) 2020-01-07 2023-01-24 Sonos, Inc. Voice verification for media playback
CN111261160B (en) * 2020-01-20 2023-09-19 联想(北京)有限公司 Signal processing method and device
KR20210094251A (en) * 2020-01-21 2021-07-29 삼성전자주식회사 Display apparatus and controlling method thereof
US11556307B2 (en) 2020-01-31 2023-01-17 Sonos, Inc. Local voice data processing
US11308958B2 (en) 2020-02-07 2022-04-19 Sonos, Inc. Localized wakeword verification
US11061543B1 (en) 2020-05-11 2021-07-13 Apple Inc. Providing relevant data items based on context
US11038934B1 (en) 2020-05-11 2021-06-15 Apple Inc. Digital assistant hardware abstraction
US11755276B2 (en) 2020-05-12 2023-09-12 Apple Inc. Reducing description length based on confidence
US11482224B2 (en) 2020-05-20 2022-10-25 Sonos, Inc. Command keywords with input detection windowing
US11308962B2 (en) 2020-05-20 2022-04-19 Sonos, Inc. Input detection windowing
US11727919B2 (en) 2020-05-20 2023-08-15 Sonos, Inc. Memory allocation for keyword spotting engines
CN111667827B (en) * 2020-05-28 2023-10-17 北京小米松果电子有限公司 Voice control method and device for application program and storage medium
US11490204B2 (en) 2020-07-20 2022-11-01 Apple Inc. Multi-device audio adjustment coordination
US11438683B2 (en) 2020-07-21 2022-09-06 Apple Inc. User identification using headphones
US11698771B2 (en) 2020-08-25 2023-07-11 Sonos, Inc. Vocal guidance engines for playback devices
CN112165640B (en) * 2020-09-21 2023-04-14 Vidaa美国公司 Display device
CN112291431B (en) * 2020-10-23 2021-10-01 北京蓦然认知科技有限公司 Method and device for synchronously controlling call
CN112261232B (en) * 2020-10-23 2022-03-01 杭州蓦然认知科技有限公司 Method for processing voice instruction of calling party by voice assistant and voice assistant
CN112291436B (en) * 2020-10-23 2022-03-01 杭州蓦然认知科技有限公司 Method and device for scheduling calling subscriber
US11984123B2 (en) 2020-11-12 2024-05-14 Sonos, Inc. Network device interaction by range
US11551700B2 (en) 2021-01-25 2023-01-10 Sonos, Inc. Systems and methods for power-efficient keyword detection
CN112817556A (en) * 2021-02-04 2021-05-18 青岛海信传媒网络技术有限公司 Switching method of voice scheme on display equipment, display equipment and control device
WO2022160911A1 (en) * 2021-01-29 2022-08-04 青岛海信传媒网络技术有限公司 Switching method for voice solution on display device, display device, and control apparatus
CN113496703A (en) * 2021-07-23 2021-10-12 北京百度网讯科技有限公司 Method, device and program product for controlling program in voice mode
CN114115620B (en) * 2021-10-27 2023-10-24 青岛海尔科技有限公司 Prompt box response method and device, storage medium and electronic device

Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104219388A (en) * 2014-08-28 2014-12-17 小米科技有限责任公司 Voice control method and device
US20150186156A1 (en) * 2013-12-31 2015-07-02 Next It Corporation Virtual assistant conversations
CN105408919A (en) * 2013-07-03 2016-03-16 埃森哲环球服务有限公司 Query response device
CN106462617A (en) * 2014-06-30 2017-02-22 苹果公司 Intelligent automated assistant for tv user interactions
CN106469040A (en) * 2015-08-19 2017-03-01 华为终端(东莞)有限公司 Communication means, server and equipment
CN106898352A (en) * 2017-02-27 2017-06-27 联想(北京)有限公司 Sound control method and electronic equipment
CN107122179A (en) * 2017-03-31 2017-09-01 阿里巴巴集团控股有限公司 The function control method and device of voice
US20170269975A1 (en) * 2016-03-17 2017-09-21 Nuance Communications, Inc. Session processing interaction between two or more virtual assistants
CN107393548A (en) * 2017-07-05 2017-11-24 青岛海信电器股份有限公司 The processing method and processing device of the voice messaging of multiple voice assistant equipment collections
CN110741347A (en) * 2017-10-03 2020-01-31 谷歌有限责任公司 Multiple digital assistant coordination in a vehicle environment

Family Cites Families (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP4377361B2 (en) * 2005-09-12 2009-12-02 株式会社ナビタイムジャパン Navigation method and navigation system using a plurality of portable terminals
CN101916166B (en) * 2010-08-19 2013-12-04 中兴通讯股份有限公司 Method for starting application program and mobile terminal
US9172747B2 (en) * 2013-02-25 2015-10-27 Artificial Solutions Iberia SL System and methods for virtual assistant networks
US10311482B2 (en) * 2013-11-11 2019-06-04 At&T Intellectual Property I, Lp Method and apparatus for adjusting a digital assistant persona
EP2881898A1 (en) * 2013-12-09 2015-06-10 Accenture Global Services Limited Virtual assistant interactivity platform
CN106648289A (en) * 2016-09-14 2017-05-10 依偎科技(南昌)有限公司 Method for switching application program, switching device and terminal

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105408919A (en) * 2013-07-03 2016-03-16 埃森哲环球服务有限公司 Query response device
US20150186156A1 (en) * 2013-12-31 2015-07-02 Next It Corporation Virtual assistant conversations
CN106462617A (en) * 2014-06-30 2017-02-22 苹果公司 Intelligent automated assistant for tv user interactions
CN104219388A (en) * 2014-08-28 2014-12-17 小米科技有限责任公司 Voice control method and device
CN106469040A (en) * 2015-08-19 2017-03-01 华为终端(东莞)有限公司 Communication means, server and equipment
US20170269975A1 (en) * 2016-03-17 2017-09-21 Nuance Communications, Inc. Session processing interaction between two or more virtual assistants
CN106898352A (en) * 2017-02-27 2017-06-27 联想(北京)有限公司 Sound control method and electronic equipment
CN107122179A (en) * 2017-03-31 2017-09-01 阿里巴巴集团控股有限公司 The function control method and device of voice
CN107393548A (en) * 2017-07-05 2017-11-24 青岛海信电器股份有限公司 The processing method and processing device of the voice messaging of multiple voice assistant equipment collections
CN110741347A (en) * 2017-10-03 2020-01-31 谷歌有限责任公司 Multiple digital assistant coordination in a vehicle environment

Also Published As

Publication number Publication date
CN107919123A (en) 2018-04-17

Similar Documents

Publication Publication Date Title
CN107919123B (en) Multi-voice assistant control method, device and computer readable storage medium
CN108804010B (en) Terminal control method, device and computer readable storage medium
CN110554815B (en) Icon awakening method, electronic device and storage medium
CN106970754B (en) Screen capture processing method and device
KR101837333B1 (en) Method and apparatus for awakening electronic device
US10241637B2 (en) Method for reducing power consumption in electronic devices
EP3185160B1 (en) Screen unlocking method and apparatus, terminal
EP3113549A1 (en) Method and device for waking up mcu chip
US20180159809A1 (en) Method and device for message reading
CN105677460A (en) Application processing method and apparatus
EP3112979B1 (en) Method and device for waking up mcu
CN106357934B (en) Screen locking control method and device
EP3933570A1 (en) Method and apparatus for controlling a voice assistant, and computer-readable storage medium
EP4184506A1 (en) Audio processing
CN106357721B (en) Timing method and device
CN111063354B (en) Man-machine interaction method and device
EP3076745A1 (en) Methods and apparatuses for controlling wireless access point
CN110413148B (en) False touch prevention detection method, device, equipment and storage medium
CN107506246A (en) Information displaying method, device and computer-readable recording medium
CN106775377B (en) Gesture recognition device, equipment and control method of gesture recognition device
CN108874450B (en) Method and device for waking up voice assistant
CN108766427B (en) Voice control method and device
CN113361305B (en) Fingerprint acquisition method, electronic device and computer readable storage medium
CN105786561B (en) Method and device for calling process
CN109521923B (en) Floating window control method and device and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant