CN111276135A - Network voice recognition method, network service interaction method and intelligent earphone - Google Patents

Network voice recognition method, network service interaction method and intelligent earphone Download PDF

Info

Publication number
CN111276135A
CN111276135A CN201811465464.4A CN201811465464A CN111276135A CN 111276135 A CN111276135 A CN 111276135A CN 201811465464 A CN201811465464 A CN 201811465464A CN 111276135 A CN111276135 A CN 111276135A
Authority
CN
China
Prior art keywords
charging box
voice command
earphone
headset
cloud server
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201811465464.4A
Other languages
Chinese (zh)
Other versions
CN111276135B (en
Inventor
龚树强
仇存收
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Huawei Device Co Ltd
Original Assignee
Huawei Device Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Huawei Device Co Ltd filed Critical Huawei Device Co Ltd
Priority to CN201811465464.4A priority Critical patent/CN111276135B/en
Priority to PCT/CN2019/115873 priority patent/WO2020114181A1/en
Publication of CN111276135A publication Critical patent/CN111276135A/en
Application granted granted Critical
Publication of CN111276135B publication Critical patent/CN111276135B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L65/00Network arrangements, protocols or services for supporting real-time applications in data packet communication
    • H04L65/40Support for services or applications
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/26Speech to text systems
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/28Constructional details of speech recognition systems
    • G10L15/30Distributed recognition, e.g. in client-server systems, for mobile phones or network applications
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L19/00Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/02Protocols based on web technology, e.g. hypertext transfer protocol [HTTP]
    • H04L67/025Protocols based on web technology, e.g. hypertext transfer protocol [HTTP] for remote control or remote monitoring of applications
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/14Session management
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R1/00Details of transducers, loudspeakers or microphones
    • H04R1/10Earpieces; Attachments therefor ; Earphones; Monophonic headphones
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04WWIRELESS COMMUNICATION NETWORKS
    • H04W88/00Devices specially adapted for wireless communication networks, e.g. terminals, base stations or access point devices
    • H04W88/02Terminal devices
    • H04W88/06Terminal devices adapted for operation in multiple networks or having at least two operational modes, e.g. multi-mode terminals
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • G10L2015/223Execution procedure of a spoken command

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Telephonic Communication Services (AREA)
  • Charge And Discharge Circuits For Batteries Or The Like (AREA)

Abstract

The application discloses a network voice recognition method, a network service interaction method and an intelligent headset. The intelligent earphone is in communication connection with the cloud server through the charging box in the intelligent earphone, the charging box sends the voice command sent by the earphone to the cloud server, and the cloud server performs voice command recognition and network service interaction with the cloud server, so that the intelligent earphone can achieve network voice recognition and network service interaction.

Description

Network voice recognition method, network service interaction method and intelligent earphone
Technical Field
The application relates to the technical field of intelligent terminals, in particular to a network voice recognition method, a network service interaction method and an intelligent headset.
Background
As shown in fig. 1, the wireless smart headset 100 includes two components, namely a wireless headset 11 and a charging box 12. The wireless headset 11 includes an audio capture/play/codec unit 111 and a wireless audio unit 122. The audio collecting/playing/encoding/decoding unit 111 is configured to collect audio data and play the received audio data, and the wireless audio unit 122 is configured to implement wireless communication with the intelligent terminal, establish an audio data transmission channel, and transmit the audio data collected by the wireless headset and the audio data to be played. The charging box 12 includes a charging/key control unit 121 and an energy storage/power supply unit 122, and the energy storage/power supply unit 122 is used for realizing battery energy storage, supplying power to the earphone, and the like.
At present, network voice recognition or network service interaction is popular with consumers, and solutions for supporting network voice recognition or network service interaction by intelligent terminals are sought. At present, a smart speaker/mobile phone can already support network voice recognition and service interaction, however, the smart speaker is an external audio unit and has poor privacy. The size of the smart phone is limited by the screen, and the smart phone cannot meet the experience requirements of a scene with a carrying requirement, such as a sports scene. The intelligent headset has good privacy and is convenient to carry, however, no intelligent headset can realize network voice recognition or network service interaction at present.
Therefore, how to implement network voice recognition or network service interaction by the intelligent headset becomes a problem to be solved urgently at present.
Disclosure of Invention
The application provides a network voice recognition method, a network service interaction method and an intelligent headset, so that the intelligent headset can realize network voice recognition or network service interaction.
One aspect of the present application provides a network voice recognition method applied to an intelligent headset, where the intelligent headset includes a headset and a charging box, and the method includes: the charging box receives voice commands from the headset; the charging box sends the voice command to a cloud server; the charging box receives a voice command recognition result from the cloud server; the charging box sends the voice command recognition result to the earphone, or the charging box executes the voice command and sends a voice command execution result to the earphone; and the earphone plays the voice command recognition result or the voice command execution result.
In the aspect, the charging box sends the voice command received from the earphone to the cloud server, and then receives the voice command recognition result sent by the cloud server, so that the intelligent earphone can realize network voice recognition, and the charging box can execute the recognized voice command.
In one implementation, the method further comprises: the earphone collects the voice command; and the earphone sends the collected voice command to the charging box.
In another implementation, the method further comprises: the earphone establishes a communication connection with the charging box; and the charging box establishes a communication connection with the cloud server.
In yet another implementation, after the charging box sends the voice command recognition result to the headset, the method further includes: the headset executes the voice command.
In this implementation, the recognized voice command is executed by the headset.
Another aspect of the present application provides a network service interaction method, which is applied to an intelligent headset, where the intelligent headset includes a headset and a charging box, and the method includes: the charging box receives a voice command from the earphone, wherein the voice command is used for indicating to acquire audio data; the charging box sends the voice command to a cloud server; the charging box receives audio data from the cloud server; the charging box sends the audio data to the headset; and the earphone decodes and plays the audio data.
In the aspect, the voice instruction for acquiring the audio data sent by the earphone is received by the charging box and sent to the cloud server, the cloud server identifies the voice instruction and executes the voice instruction, the audio data is sent to the charging box, and the charging box sends the audio data to the earphone for decoding and playing, so that the intelligent earphone can realize network service interaction.
In one implementation, the method further comprises: the charging box decodes audio data received from the cloud server; and the charging box sending the audio data to the headset, including: the charging box sends the decoded audio data to the headset.
In this implementation, when the formats of the audio data sent between the charging box and the headset are inconsistent, the charging box further decodes the audio data received from the cloud server, and sends the decoded audio data to the headset for re-decoding.
In another implementation, after the charging box sends the voice command to a cloud server, the method further includes: the charging box receives a voice command recognition result from the cloud server; and the charging box sends the voice command recognition result to the earphone.
In the implementation mode, when the voice recognition and the network service interaction are not integrated in the cloud server, the charging box also receives a voice command recognition result sent by the cloud server and sends the voice command recognition result to the earphone.
In yet another implementation, the method further comprises: the earphone collects the voice command; and the earphone sends the collected voice command to the charging box.
In yet another implementation, the method further comprises: the earphone establishes a communication connection with the charging box; and the charging box establishes a communication connection with the cloud server.
Yet another aspect of the present application provides a network service interaction method applied to a smart headset, where the smart headset includes a headset and a charging box, and the method includes: the charging box receives voice commands from the headset; the charging box sends the voice command to a cloud server; the charging box receives a voice command execution result from the cloud server; the charging box sends the voice command execution result to the earphone; and the earphone plays the voice command execution result.
In the aspect, the charging box receives the voice command sent by the earphone and sends the voice command to the cloud server, and the cloud server recognizes and executes the voice command and returns a voice command execution result to the charging box, so that the intelligent earphone realizes network service interaction.
Yet another aspect of the present application provides a smart headset comprising a headset and a charging box for receiving voice commands from the headset; the charging box is further used for sending the voice command to a cloud server; the charging box is further used for receiving a voice command recognition result from the cloud server; the charging box is also used for sending the voice command recognition result to the earphone, or the charging box is also used for executing the voice command and sending a voice command execution result to the earphone; and the earphone is used for playing the voice command recognition result or the voice command execution result.
In one implementation, the headset is further configured to capture the voice command; and the earphone is also used for sending the collected voice command to the charging box.
In another implementation, the headset is further configured to establish a communication connection with the charging box; and the charging box is also used for establishing communication connection with the cloud server.
In yet another implementation, the headset is further configured to execute the voice command.
Yet another aspect of the present application provides a smart headset comprising a headset and a charging box for receiving a voice command from the headset, the voice command being indicative of acquiring audio data; the charging box is further used for sending the voice command to a cloud server; the charging box is also used for receiving audio data from the cloud server; the charging box is further used for sending the audio data to the earphone; and the earphone is used for decoding and playing the audio data.
In one implementation, the charging box is further configured to decode audio data received from the cloud server; and the charging box is further configured to send the decoded audio data to the headset.
In another implementation, the charging box is further configured to receive a voice command recognition result from the cloud server; and the charging box is also used for sending the voice command recognition result to the earphone.
In yet another implementation, the headset is further configured to capture the voice command; and the earphone is also used for sending the collected voice command to the charging box.
In yet another implementation, the headset is further configured to establish a communication connection with the charging box; and the charging box is also used for establishing communication connection with the cloud server.
Yet another aspect of the present application provides a smart headset comprising a headset and a charging box for receiving voice commands from the headset; the charging box is further used for sending the voice command to a cloud server; the charging box is also used for receiving a voice command execution result from the cloud server; the charging box is also used for sending the voice command execution result to the earphone; and the earphone is used for playing the voice command execution result.
Yet another aspect of the present application provides a computer-readable storage medium having stored therein instructions, which when executed on a computer, cause the computer to perform the method of the above-described aspects.
Yet another aspect of the present application provides a computer program product containing instructions which, when run on a computer, cause the computer to perform the method of the above-described aspects.
Drawings
Fig. 1 is a schematic structural diagram of a conventional smart headset;
fig. 2 is a general structural schematic diagram of an intelligent headset according to an embodiment of the present disclosure;
fig. 3 is a schematic flowchart of a network speech recognition method according to an embodiment of the present application;
fig. 4 is a schematic flowchart of a network service interaction provided in an embodiment of the present application;
fig. 5 is a schematic flow chart of another network service interaction provided in the embodiment of the present application.
Detailed Description
The embodiments of the present application will be described below with reference to the drawings.
The network voice recognition is a function that the terminal collects a user voice command, transmits the user voice command to the cloud server by using a network, recognizes the user voice command, returns the user voice command to the terminal, and is executed by the terminal.
The network interaction service means that a terminal sends a service requirement to a cloud server, and the cloud server responds to the terminal requirement and sends an execution result to the terminal.
In the application, the network voice recognition service and the network interaction service may be integrated in one server, or may be executed by different servers.
Referring to fig. 2, fig. 2 is a schematic diagram of a general structure of an intelligent headset according to an embodiment of the present disclosure, where the intelligent headset includes a headset 21 and a charging box 22. Generally, the earphone 21 is wirelessly connected with the charging box 22. The structure of the headphone 21 is the same as that shown in fig. 1, and includes an audio capture/play/codec unit 211 and a wireless audio unit 212. The audio capturing/playing/encoding/decoding unit 211 further specifically includes functions of voice wakeup, local audio playing, audio capturing, audio encoding and decoding, and the like. It should be noted that the audio capture/play/codec unit 211 may include different functions for different network functions. For example, when the smart headset is used to implement network voice recognition, the audio acquisition/playing/codec unit 211 may include functions such as voice wakeup, audio acquisition, and audio playing; when the smart headset is used to implement network interaction services, such as acquiring audio data, the audio collecting/playing/codec unit 211 includes functions of voice wakeup, audio playing, audio collecting, audio codec, and the like. When the smart headset is used to implement both the network voice recognition and the network interactive service, the audio collection/play/codec unit 211 may include all the above functions. The wireless audio unit 212 can perform Bluetooth (BT) communication or the like with the wireless audio unit 223 of the charging box 22.
The charging box 22 includes an internet service unit 221, an audio codec unit 222, a wireless audio unit 223, a wireless network unit 224, a charging/key control unit 225, and a power storage/supply 226, etc. Since the charging box 22 has a function of connecting to a cloud server, it may also be referred to as a networking charging box. The internet service unit 221 includes a streaming media service Software Development Kit (SDK) and a speech engine cloud SDK. The streaming media service SDK is streaming media software for connecting the cloud server, and the voice engine cloud end SDK is voice software for connecting the cloud server. According to the function implementation of the smart headset, the two SDKs can be included, or one SDK can be included. The audio codec unit 222 includes functions of local playing, streaming media playing, and audio codec. The audio codec unit 222 is optional according to the functional implementation of the smart headset. The wireless network unit 224 is configured to implement communication connection with the cloud server, and may establish communication connection with the server in a communication connection manner such as 4G/3G/2G, Wi-Fi and TCP/IP. The functions of the charging/key control unit 225 and the power storing/supplying unit 226 are the same as those of the charging/key control unit 121 and the power storing/supplying unit 122 of the embodiment shown in fig. 1, respectively.
Referring to fig. 3, fig. 3 is a schematic flowchart of a network voice recognition method according to an embodiment of the present application, and is applied to the smart headset shown in fig. 2. The method comprises the following steps:
s101, the earphone sends a voice command to the charging box.
In this step, after the user wears the earphone, the user gives a voice command by waking up the earphone through voice or manually. The headset collects the voice command of the user through the microphone, and optionally before S101, the method further includes: the earphone collects the voice command; and the earphone sends the collected voice command to the charging box.
The headset may also retrieve local voice commands.
The headset itself cannot recognize the voice command. Therefore, after the headset collects the voice command, the voice command is converted into voice command data, and the voice command data is transmitted to the wireless audio unit 223 of the charging box through the wireless audio unit 212. The voice command may be a control command or may be another command.
Optionally, before S101, the method further includes:
the earphone establishes a communication connection with the charging box;
the charging box establishes a communication connection with the cloud server.
In particular, the headset establishes a communication connection, which may be a bluetooth connection, for example, with the wireless audio unit 223 of the charging box through the wireless audio unit 212.
The charging box establishes a communication connection with the cloud server through the wireless network unit 224. For example, communication connection modes such as 4G/3G/2G, Wi-Fi and TCP/IP can be adopted.
And S102, after receiving the voice command, the charging box sends the voice command to a cloud server.
After the wireless audio unit 223 of the charging box receives the voice command transmitted from the wireless audio unit 212 of the headset, it cannot recognize the voice command by itself, and transmits the voice command received from the headset to the cloud server through the wireless network unit 224.
In another embodiment, the charging box may also have a local language command recognition capability, and after receiving the voice command sent by the headset, the charging box performs local recognition, if the voice command cannot be locally recognized, and forwards the voice command to the cloud server for network recognition.
And S103, after receiving the voice command, the cloud server identifies the voice command.
The cloud server has a voice recognition function, and after receiving the voice command sent by the charging box, recognizes the voice command to obtain a voice command recognition result. The speech recognition process can refer to the existing speech recognition technology, and is not described in detail herein.
And S104, the cloud server sends a voice command recognition result to the charging box.
And the cloud server sends a voice command recognition result to the charging box through a wireless network.
As one implementation, the voice command may be executed by the charging box when the voice command is a control command. And after S104, proceeding to S105, after the charging box receives the voice command recognition result, executing the voice command.
For example, the voice command is "play song" love Nostol ", and the charging box calls the song" love Nostol "from the memory or acquires the song" love Nostol "from the server after receiving the voice command recognition result sent by the cloud server.
And S106, the charging box sends a voice command execution result to the earphone.
As shown in the above example, the result of the voice command execution may be audio data for the song.
And S107, after receiving the voice command execution result, the earphone plays the voice command execution result.
As shown in the above example, the headset plays the song after receiving the audio data for the song.
Optionally, after S107, the method further includes: the headset executes the voice command.
As another implementation, the voice command may also be executed by the headset when the voice command is a control command. After S104, proceeding to S106, after the charging box receives the voice command recognition result, the charging box sends the voice command recognition result to the headset.
For example, the voice command is "play song" love noun ", and the charging box sends the voice command recognition result to the headset" — play song "love noun".
And S107, after receiving the voice command recognition result, the earphone plays the voice command recognition result.
As shown in the above example, after receiving the voice command recognition result, "play song" love noun ", the headset acquires the audio data of the song from the charging box through the wireless audio unit 212, or acquires the audio data of the song from the local memory, and plays the song.
Due to the limitation of the size and the power supply of the earphone, the network voice recognition function of the earphone and the cloud server cannot be realized, and in the embodiment, the network voice recognition function of the earphone and the cloud server can be realized by integrating the charging box with the unit. The intelligent earphone with the integrated network voice recognition function can be realized by combining the original earphone.
According to the network voice recognition method provided by the embodiment of the application, the charging box sends the voice command received from the earphone to the cloud server, and then the charging box receives the voice command recognition result sent by the cloud server, so that the intelligent earphone can realize network voice recognition, and the charging box can execute the recognized voice command.
Referring to fig. 4, fig. 4 is a schematic flowchart of a network service interaction method according to an embodiment of the present application, and is applied to the smart headset shown in fig. 2. The method comprises the following steps:
s201, the earphone sends a voice command to the charging box, and the voice command is used for indicating to acquire audio data.
In this step, after the user wears the earphone, the user gives a voice command by waking up the earphone through voice or manually. The headset collects the voice command of the user through the microphone, and optionally before S201, the method further includes: the earphone collects the voice command; and the earphone sends the collected voice command to the charging box.
The headset may also retrieve local voice commands.
The headset itself cannot recognize the voice command. Therefore, after the headset collects the voice command, the voice command is converted into voice command data, and the voice command data is transmitted to the wireless audio unit 223 of the charging box through the wireless audio unit 212. The voice command is used to indicate data audio data. For example, audio data of song "noun of love" is acquired.
Optionally, before S201, the method further includes:
the earphone establishes a communication connection with the charging box;
the charging box establishes a communication connection with the cloud server.
In particular, the headset establishes a communication connection, which may be a bluetooth connection, for example, with the wireless audio unit 223 of the charging box through the wireless audio unit 212.
The charging box establishes a communication connection with the cloud server through the wireless network unit 224. For example, communication connection modes such as 4G/3G/2G, Wi-Fi and TCP/IP can be adopted.
And S202, after receiving the voice command, the charging box sends the voice command to a cloud server.
After the wireless audio unit 223 of the charging box receives the voice command transmitted from the wireless audio unit 212 of the headset, it cannot recognize the voice command by itself, and transmits the voice command received from the headset to the cloud server through the wireless network unit 224.
In another embodiment, the charging box has a function of storing a local media resource file, and when a command for playing the local media resource is identified (local identification or network identification), the local media resource file is read, decoded and sent to the earphone for playing.
S203, the cloud server identifies the voice command and acquires the audio data indicated by the voice command.
The cloud server has a voice recognition function and a network audio service function. Alternatively, the voice recognition function and the network audio service function may be integrated in one server, or may be implemented by different servers.
When the voice recognition function and the network audio service function are integrated in one server, the cloud server recognizes the voice command after receiving the voice command sent by the charging box, and acquires audio data indicated by the voice command.
When the voice recognition function and the network audio service function are realized by different servers, after recognizing the voice command, the cloud server sends a request for acquiring audio data to another server with the network audio service function, and receives audio data sent by another server. Optionally, after S203, the method further includes: the cloud server sends a voice command recognition result to the charging box; and after receiving the voice command recognition result from the cloud server, the charging box sends the voice command recognition result to the earphone.
S204, the cloud server sends the audio data to the charging box.
S205, after receiving the audio data sent by the cloud server, the charging box sends the audio data to the earphone.
Specifically, the wireless network unit 224 of the charging box receives audio data transmitted from the cloud server and transmits the audio data to the wireless audio unit 212 of the headset through the wireless audio unit 223.
S206, the earphone decodes and plays the audio data.
When the format of the audio data sent by the charging box is consistent with the format of the audio data played by the earphone, the charging box can directly send the audio data to the earphone after receiving the audio data sent by the cloud server, and the audio data is decoded and played by the earphone.
Optionally, when the format of the audio data sent by the charging box is not consistent with the format of the audio data played by the earphone, after S204, the method further includes: the charging box decodes audio data received from the cloud server.
S205 specifically comprises: the charging box sends the decoded audio data to the headset.
Specifically, since the format of the audio data sent by the charging box is inconsistent with the format of the audio data played by the headset, the charging box needs to decode the audio data received from the cloud server and send the decoded audio data to the headset. And after receiving the audio data decoded by the charging box, the earphone decodes the audio data according to the playing format of the earphone and plays the audio data decoded by the earphone.
According to the network service interaction method provided by the embodiment of the application, the charging box receives the voice instruction for acquiring the audio data sent by the earphone and sends the voice instruction to the cloud server, the cloud server recognizes the voice instruction and executes the voice instruction, the audio data are sent to the charging box, and the charging box sends the audio data to the earphone for decoding and playing, so that the intelligent earphone can realize network service interaction.
Referring to fig. 5, fig. 5 is a schematic flowchart of a network service interaction method according to an embodiment of the present application, and is applied to the smart headset shown in fig. 2. The method comprises the following steps:
s301, the earphone sends a voice command to the charging box.
Unlike the embodiment shown in fig. 4, the voice command may be a network traffic service command instructing the cloud server to proceed. For example, the user's daily number of athletic steps is uploaded to the cloud server. When the user walks quickly at the current moment, the user sends a voice instruction to the earphone, and the cloud server is instructed to prompt when the exercise step number of the user exceeds the average value of the exercise step numbers of the user every day. For example, the user sends a voice command "do i exceed average daily steps today? "
S302, after the charging box receives the voice command from the earphone, the charging box sends the voice command to a cloud server.
After the wireless audio unit 223 of the charging box receives the voice command transmitted from the wireless audio unit 212 of the headset, it cannot recognize the voice command by itself, and transmits the voice command received from the headset to the cloud server through the wireless network unit 224.
And S303, after receiving the voice command, the cloud server identifies the voice command and executes the voice command.
The cloud server has a voice recognition function, and recognizes a voice command of a user, "do i exceed the average number of steps per day today? The cloud server executes the voice command, compares the current step number of the user uploaded by the earphone or the charging box with the daily average step number of the user, and obtains a voice command execution result that the current step number of the user exceeds the daily average step number of the user, namely that the current step number of the user exceeds the daily average step number of the user! "
S304, the cloud server sends a voice command execution result to the charging box.
S305, after the charging box receives a voice command execution result from the cloud server, the charging box sends the voice command execution result to the earphone.
Specifically, the wireless network unit 224 of the charging box receives the voice command execution result transmitted from the cloud server and transmits the voice command execution result to the wireless audio unit 212 of the headset through the wireless audio unit 223.
S306, after the earphone receives the voice command execution result, playing the voice command execution result.
The wireless audio unit 212 of the headset plays the voice command execution result after receiving the voice command execution result.
Of course, after the cloud server executes the voice command, the command execution result may be in another prompting manner, for example, the charging box performs vibration prompting after receiving the command execution result.
According to the network service interaction method provided by the embodiment of the application, the charging box receives the voice command sent by the earphone and sends the voice command to the cloud server, and the cloud server identifies and executes the voice command and returns the voice command execution result to the charging box, so that the intelligent earphone realizes network service interaction.
The embodiment of the application also provides an intelligent headset, and the structure of the intelligent headset is shown in fig. 2. Specifically, the method comprises the following steps:
the charging box is used for receiving voice commands from the earphone;
the charging box is further used for sending the voice command to a cloud server;
the charging box is further used for receiving a voice command recognition result from the cloud server;
the charging box is also used for sending the voice command recognition result to the earphone, or the charging box is also used for executing the voice command and sending a voice command execution result to the earphone;
the earphone is used for playing the voice command recognition result or the voice command execution result.
Optionally, the headset is further configured to capture the voice command; and the earphone is also used for sending the collected voice command to the charging box.
Optionally, the headset is further configured to establish a communication connection with the charging box; and the charging box is also used for establishing communication connection with the cloud server.
Optionally, the headset is further configured to execute the voice command.
Specific functional implementation can refer to the embodiment shown in fig. 3.
According to the intelligent earphone provided by the embodiment of the application, the charging box sends the voice command received from the earphone to the cloud server, and then the charging box receives the voice command recognition result sent by the cloud server, so that the intelligent earphone can realize network voice recognition, and the charging box can execute the recognized voice command.
The embodiment of the application also provides another intelligent headset, and the structure of the intelligent headset is shown in fig. 2. Specifically, the method comprises the following steps:
the charging box is used for receiving a voice command from the earphone, and the voice command is used for indicating to acquire audio data;
the charging box is further used for sending the voice command to a cloud server;
the charging box is also used for receiving audio data from the cloud server;
the charging box is further used for sending the audio data to the earphone;
the earphone is used for decoding and playing the audio data.
Optionally, the charging box is further configured to decode audio data received from the cloud server; and the charging box is further configured to send the decoded audio data to the headset.
Optionally, the charging box is further configured to receive a voice command recognition result from the cloud server; and the charging box is also used for sending the voice command recognition result to the earphone.
Optionally, the headset is further configured to capture the voice command; and the earphone is also used for sending the collected voice command to the charging box.
Optionally, the headset is further configured to establish a communication connection with the charging box; and the charging box is also used for establishing communication connection with the cloud server.
Specific functional implementation can refer to the embodiment shown in fig. 4.
According to the intelligent earphone provided by the embodiment of the application, the voice instruction for acquiring the audio data sent by the earphone is received by the charging box and sent to the cloud server, the cloud server identifies the voice instruction and executes the voice instruction, the audio data is sent to the charging box, and the charging box sends the audio data to the earphone for decoding and playing, so that the intelligent earphone can realize network service interaction.
The embodiment of the application also provides another intelligent headset, and the structure of the intelligent headset is shown in fig. 2. Specifically, the method comprises the following steps:
the charging box is used for receiving voice commands from the earphone;
the charging box is further used for sending the voice command to a cloud server;
the charging box is also used for receiving a voice command execution result from the cloud server;
the charging box is also used for sending the voice command execution result to the earphone;
the earphone is used for playing the voice command execution result.
Specific functional implementation can refer to the embodiment shown in fig. 5.
According to the intelligent earphone provided by the embodiment of the application, the charging box receives the voice command sent by the earphone and sends the voice command to the cloud server, the cloud server identifies and executes the voice command, and the voice command execution result is returned to the charging box, so that the intelligent earphone realizes network service interaction.
Embodiments of the present application further provide a computer-readable storage medium, on which a computer program is stored, and when the computer program is executed by a processor, the computer program implements the method performed by the smart headset in the embodiments shown in fig. 3 to 5.
It is clear to those skilled in the art that, for convenience and brevity of description, the specific working processes of the above-described systems, apparatuses and units may refer to the corresponding processes in the foregoing method embodiments, and are not described herein again.
In the several embodiments provided in the present application, it should be understood that the disclosed system, apparatus and method may be implemented in other ways. For example, the division of the unit is only one logical function division, and other division may be implemented in practice, for example, a plurality of units or components may be combined or integrated into another system, or some features may be omitted, or not executed. The shown or discussed mutual coupling, direct coupling or communication connection may be an indirect coupling or communication connection of devices or units through some interfaces, and may be in an electrical, mechanical or other form.
Units described as separate parts may or may not be physically separate, and parts displayed as units may or may not be physical units, may be located in one place, or may be distributed on a plurality of network units. Some or all of the units can be selected according to actual needs to achieve the purpose of the solution of the embodiment.
In the above embodiments, the implementation may be wholly or partially realized by software, hardware, firmware, or any combination thereof. When implemented in software, may be implemented in whole or in part in the form of a computer program product. The computer program product includes one or more computer instructions. The procedures or functions according to the embodiments of the present application are wholly or partially generated when the computer program instructions are loaded and executed on a computer. The computer may be a general purpose computer, a special purpose computer, a network of computers, or other programmable device. The computer instructions may be stored on or transmitted over a computer-readable storage medium. The computer instructions may be transmitted from one website, computer, server, or data center to another website, computer, server, or data center by wire (e.g., coaxial cable, fiber optic, Digital Subscriber Line (DSL)), or wirelessly (e.g., infrared, wireless, microwave, etc.). The computer-readable storage medium can be any available medium that can be accessed by a computer or a data storage device, such as a server, a data center, etc., that includes one or more of the available media. The usable medium may be a read-only memory (ROM), or a Random Access Memory (RAM), or a magnetic medium, such as a floppy disk, a hard disk, a magnetic tape, a magnetic disk, or an optical medium, such as a Digital Versatile Disk (DVD), or a semiconductor medium, such as a Solid State Disk (SSD).

Claims (21)

1. A network voice recognition method is applied to an intelligent headset, the intelligent headset comprises a headset body and a charging box, and the method comprises the following steps:
the charging box receives voice commands from the headset;
the charging box sends the voice command to a cloud server;
the charging box receives a voice command recognition result from the cloud server;
the charging box sends the voice command recognition result to the earphone, or the charging box executes the voice command and sends a voice command execution result to the earphone;
and the earphone plays the voice command recognition result or the voice command execution result.
2. The method of claim 1, wherein the method further comprises:
the earphone collects the voice command;
and the earphone sends the collected voice command to the charging box.
3. The method of claim 1 or 2, wherein the method further comprises:
the earphone establishes a communication connection with the charging box;
the charging box establishes a communication connection with the cloud server.
4. The method of claim 1, wherein after the charging box sends the voice command recognition result to the headset, the method further comprises:
the headset executes the voice command.
5. A network service interaction method is applied to an intelligent headset, wherein the intelligent headset comprises a headset body and a charging box, and the method comprises the following steps:
the charging box receives a voice command from the earphone, wherein the voice command is used for indicating to acquire audio data;
the charging box sends the voice command to a cloud server;
the charging box receives audio data from the cloud server;
the charging box sends the audio data to the headset;
the headphones decode and play the audio data.
6. The method of claim 5, wherein the method further comprises:
the charging box decodes audio data received from the cloud server;
the charging box sending the audio data to the headset, comprising:
the charging box sends the decoded audio data to the headset.
7. The method of claim 5, wherein after the charging box sends the voice command to a cloud server, the method further comprises:
the charging box receives a voice command recognition result from the cloud server;
the charging box sends the voice command recognition result to the headset.
8. The method of any one of claims 5 to 7, further comprising:
the earphone collects the voice command;
and the earphone sends the collected voice command to the charging box.
9. The method of any one of claims 5 to 8, further comprising:
the earphone establishes a communication connection with the charging box;
the charging box establishes a communication connection with the cloud server.
10. A network service interaction method is applied to an intelligent headset, wherein the intelligent headset comprises a headset body and a charging box, and the method comprises the following steps:
the charging box receives voice commands from the headset;
the charging box sends the voice command to a cloud server;
the charging box receives a voice command execution result from the cloud server;
the charging box sends the voice command execution result to the earphone;
and the earphone plays the voice command execution result.
11. The utility model provides an intelligent earphone, intelligent earphone includes the earphone and charges the box, its characterized in that:
the charging box is used for receiving voice commands from the earphone;
the charging box is further used for sending the voice command to a cloud server;
the charging box is further used for receiving a voice command recognition result from the cloud server;
the charging box is also used for sending the voice command recognition result to the earphone, or the charging box is also used for executing the voice command and sending a voice command execution result to the earphone;
the earphone is used for playing the voice command recognition result or the voice command execution result.
12. The smart headset as recited in claim 11, wherein:
the earphone is also used for collecting the voice command;
the earphone is also used for sending the collected voice command to the charging box.
13. The smart headset as recited in claim 11 or 12, wherein:
the earphone is also used for establishing communication connection with the charging box;
the charging box is also used for establishing communication connection with the cloud server.
14. The smart headset as recited in claim 11, wherein:
the headset is also configured to execute the voice command.
15. The utility model provides an intelligent earphone, intelligent earphone includes the earphone and charges the box, its characterized in that:
the charging box is used for receiving a voice command from the earphone, and the voice command is used for indicating to acquire audio data;
the charging box is further used for sending the voice command to a cloud server;
the charging box is also used for receiving audio data from the cloud server;
the charging box is further used for sending the audio data to the earphone;
the earphone is used for decoding and playing the audio data.
16. The smart headset as recited in claim 15, wherein:
the charging box is further used for decoding audio data received from the cloud server;
the charging box is further configured to send the decoded audio data to the headset.
17. The smart headset as recited in claim 15, wherein:
the charging box is further used for receiving a voice command recognition result from the cloud server;
the charging box is further configured to send the voice command recognition result to the headset.
18. The smart headset as recited in any one of claims 15-17, wherein:
the earphone is also used for collecting the voice command;
the earphone is also used for sending the collected voice command to the charging box.
19. The smart headset as recited in any one of claims 15-18, wherein:
the earphone is also used for establishing communication connection with the charging box;
the charging box is also used for establishing communication connection with the cloud server.
20. The utility model provides an intelligent earphone, intelligent earphone includes the earphone and charges the box, its characterized in that:
the charging box is used for receiving voice commands from the earphone;
the charging box is further used for sending the voice command to a cloud server;
the charging box is also used for receiving a voice command execution result from the cloud server;
the charging box is also used for sending the voice command execution result to the earphone;
the earphone is used for playing the voice command execution result.
21. A computer-readable storage medium, on which a computer program is stored, which program, when being executed by a processor, carries out the method of any one of claims 1 to 4, or the method of any one of claims 5 to 9, or the method of claim 10.
CN201811465464.4A 2018-12-03 2018-12-03 Network voice recognition method, network service interaction method and intelligent earphone Active CN111276135B (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN201811465464.4A CN111276135B (en) 2018-12-03 2018-12-03 Network voice recognition method, network service interaction method and intelligent earphone
PCT/CN2019/115873 WO2020114181A1 (en) 2018-12-03 2019-11-06 Network voice recognition method, network service interaction method and intelligent earphone

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201811465464.4A CN111276135B (en) 2018-12-03 2018-12-03 Network voice recognition method, network service interaction method and intelligent earphone

Publications (2)

Publication Number Publication Date
CN111276135A true CN111276135A (en) 2020-06-12
CN111276135B CN111276135B (en) 2023-06-20

Family

ID=70974054

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201811465464.4A Active CN111276135B (en) 2018-12-03 2018-12-03 Network voice recognition method, network service interaction method and intelligent earphone

Country Status (2)

Country Link
CN (1) CN111276135B (en)
WO (1) WO2020114181A1 (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113286212A (en) * 2021-05-20 2021-08-20 北京明略软件***有限公司 Wearing type audio acquisition assembly
CN113380251A (en) * 2021-06-22 2021-09-10 紫优科技(深圳)有限公司 Mobile voice interaction method and device based on intelligent earphone
CN113411709A (en) * 2021-06-28 2021-09-17 紫优科技(深圳)有限公司 Design method and system of cloud intelligent earphone system
CN113421570A (en) * 2021-06-21 2021-09-21 紫优科技(深圳)有限公司 Intelligent earphone identity authentication method and device

Citations (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP3119248U (en) * 2005-12-06 2006-02-16 ▲シウ▼瑩企業有限公司 Wireless earphone device and charging base assembly
CN102594988A (en) * 2012-02-10 2012-07-18 深圳市中兴移动通信有限公司 Method and system capable of achieving automatic pairing connection of Bluetooth earphones by speech recognition
US20140119554A1 (en) * 2012-10-25 2014-05-01 Elwha Llc Methods and systems for non-volatile memory in wireless headsets
WO2014197635A2 (en) * 2013-06-07 2014-12-11 Apple Inc. Intelligent automated assistant
US20170110124A1 (en) * 2015-10-20 2017-04-20 Bragi GmbH Wearable Earpiece Voice Command Control System and Method
CN106850847A (en) * 2017-03-10 2017-06-13 上海斐讯数据通信技术有限公司 Voice messaging sharing method and its intelligent earphone based on cloud platform
CN107333201A (en) * 2017-07-24 2017-11-07 歌尔科技有限公司 One kind translation earphone storage box, wireless translation earphone and wireless translation system
CN206977651U (en) * 2017-06-05 2018-02-06 广东朝阳电子科技股份有限公司 The circuit structure of TWS Bluetooth earphone devices with WiFi function
CN207518810U (en) * 2017-11-20 2018-06-19 深圳市胜兴旺精密科技有限公司 Charging box
CN108509428A (en) * 2018-02-26 2018-09-07 深圳市百泰实业股份有限公司 Earphone interpretation method and system
CN108550367A (en) * 2018-05-18 2018-09-18 深圳傲智天下信息科技有限公司 A kind of portable intelligent interactive voice control device, method and system
CN108549206A (en) * 2018-07-12 2018-09-18 深圳傲智天下信息科技有限公司 A kind of band has the smartwatch of voice interactive function earphone
CN108564949A (en) * 2018-05-18 2018-09-21 深圳傲智天下信息科技有限公司 A kind of TWS earphones, Wrist belt-type AI voice interaction devices and system
CN108900945A (en) * 2018-09-29 2018-11-27 上海与德科技有限公司 Bluetooth headset box and audio recognition method, server and storage medium

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9402120B2 (en) * 2014-09-05 2016-07-26 Epickal AB Wireless earbuds

Patent Citations (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP3119248U (en) * 2005-12-06 2006-02-16 ▲シウ▼瑩企業有限公司 Wireless earphone device and charging base assembly
CN102594988A (en) * 2012-02-10 2012-07-18 深圳市中兴移动通信有限公司 Method and system capable of achieving automatic pairing connection of Bluetooth earphones by speech recognition
US20140119554A1 (en) * 2012-10-25 2014-05-01 Elwha Llc Methods and systems for non-volatile memory in wireless headsets
WO2014197635A2 (en) * 2013-06-07 2014-12-11 Apple Inc. Intelligent automated assistant
US20170110124A1 (en) * 2015-10-20 2017-04-20 Bragi GmbH Wearable Earpiece Voice Command Control System and Method
CN106850847A (en) * 2017-03-10 2017-06-13 上海斐讯数据通信技术有限公司 Voice messaging sharing method and its intelligent earphone based on cloud platform
CN206977651U (en) * 2017-06-05 2018-02-06 广东朝阳电子科技股份有限公司 The circuit structure of TWS Bluetooth earphone devices with WiFi function
CN107333201A (en) * 2017-07-24 2017-11-07 歌尔科技有限公司 One kind translation earphone storage box, wireless translation earphone and wireless translation system
CN207518810U (en) * 2017-11-20 2018-06-19 深圳市胜兴旺精密科技有限公司 Charging box
CN108509428A (en) * 2018-02-26 2018-09-07 深圳市百泰实业股份有限公司 Earphone interpretation method and system
CN108550367A (en) * 2018-05-18 2018-09-18 深圳傲智天下信息科技有限公司 A kind of portable intelligent interactive voice control device, method and system
CN108564949A (en) * 2018-05-18 2018-09-21 深圳傲智天下信息科技有限公司 A kind of TWS earphones, Wrist belt-type AI voice interaction devices and system
CN108549206A (en) * 2018-07-12 2018-09-18 深圳傲智天下信息科技有限公司 A kind of band has the smartwatch of voice interactive function earphone
CN108900945A (en) * 2018-09-29 2018-11-27 上海与德科技有限公司 Bluetooth headset box and audio recognition method, server and storage medium

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113286212A (en) * 2021-05-20 2021-08-20 北京明略软件***有限公司 Wearing type audio acquisition assembly
CN113286212B (en) * 2021-05-20 2022-07-12 北京明略软件***有限公司 Wearing type audio acquisition assembly
CN113421570A (en) * 2021-06-21 2021-09-21 紫优科技(深圳)有限公司 Intelligent earphone identity authentication method and device
CN113380251A (en) * 2021-06-22 2021-09-10 紫优科技(深圳)有限公司 Mobile voice interaction method and device based on intelligent earphone
CN113411709A (en) * 2021-06-28 2021-09-17 紫优科技(深圳)有限公司 Design method and system of cloud intelligent earphone system

Also Published As

Publication number Publication date
WO2020114181A1 (en) 2020-06-11
CN111276135B (en) 2023-06-20

Similar Documents

Publication Publication Date Title
CN111276135B (en) Network voice recognition method, network service interaction method and intelligent earphone
CN208507180U (en) A kind of portable intelligent interactive voice control equipment
CN110769394B (en) Video call method and electronic equipment
CN208689384U (en) A kind of band has the smartwatch of voice interactive function earphone
WO2019218368A1 (en) Tws headphones, wrist-worn ai voice interaction device and system
WO2020133112A1 (en) Method for automatically switching bluetooth audio encoding method and electronic apparatus
CN110278205B (en) Bluetooth sound box base and control method and system thereof
WO2020132818A1 (en) Wireless short-distance audio sharing method and electronic device
CN109890018A (en) Blue-tooth transmission method, bluetooth transceiver and the computer readable storage medium of audio
CN201629037U (en) Remote controller
JP7442647B2 (en) bluetooth communication method and device
CN110177363B (en) Bluetooth data transmission method and device, Bluetooth chip and storage medium
CN110189753B (en) Bluetooth sound box and control method, system and storage medium thereof
CN109286725A (en) Interpretation method and terminal
CN109712631A (en) Audio data transfer control method, device, system and readable storage medium storing program for executing
EP3552508B1 (en) Smart helmet having remote control, and remote control method thereof
CN103794215A (en) Speech control-based handheld terminal, system and speech control-based control method
CN109102816A (en) Coding control method, device and electronic equipment
CN115150501A (en) Voice interaction method and electronic equipment
CN109194998A (en) Data transmission method, device, electronic equipment and computer-readable medium
CN104754261A (en) Projection equipment and projection method
CN204291130U (en) Mobile network is utilized to realize equipment that is single and multi-person speech message
CN108399918B (en) Intelligent device connection method, intelligent device and terminal
CN213906675U (en) Portable wireless bluetooth recording equipment
CN111556406B (en) Audio processing method, audio processing device and earphone

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant