CN111276135A - Network voice recognition method, network service interaction method and intelligent earphone - Google Patents
Network voice recognition method, network service interaction method and intelligent earphone Download PDFInfo
- Publication number
- CN111276135A CN111276135A CN201811465464.4A CN201811465464A CN111276135A CN 111276135 A CN111276135 A CN 111276135A CN 201811465464 A CN201811465464 A CN 201811465464A CN 111276135 A CN111276135 A CN 111276135A
- Authority
- CN
- China
- Prior art keywords
- charging box
- voice command
- earphone
- headset
- cloud server
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 73
- 230000003993 interaction Effects 0.000 title claims abstract description 33
- 238000004891 communication Methods 0.000 claims abstract description 38
- 238000004590 computer program Methods 0.000 claims description 8
- 230000006870 function Effects 0.000 description 28
- 230000002354 daily effect Effects 0.000 description 5
- 230000008878 coupling Effects 0.000 description 3
- 238000010168 coupling process Methods 0.000 description 3
- 238000005859 coupling reaction Methods 0.000 description 3
- 238000010586 diagram Methods 0.000 description 3
- 238000004146 energy storage Methods 0.000 description 3
- 230000008569 process Effects 0.000 description 3
- 230000002618 waking effect Effects 0.000 description 2
- 230000000386 athletic effect Effects 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 238000013500 data storage Methods 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 230000003203 everyday effect Effects 0.000 description 1
- 239000000835 fiber Substances 0.000 description 1
- 230000002452 interceptive effect Effects 0.000 description 1
- 230000006855 networking Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 239000007787 solid Substances 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L65/00—Network arrangements, protocols or services for supporting real-time applications in data packet communication
- H04L65/40—Support for services or applications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/26—Speech to text systems
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/28—Constructional details of speech recognition systems
- G10L15/30—Distributed recognition, e.g. in client-server systems, for mobile phones or network applications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L19/00—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L67/00—Network arrangements or protocols for supporting network services or applications
- H04L67/01—Protocols
- H04L67/02—Protocols based on web technology, e.g. hypertext transfer protocol [HTTP]
- H04L67/025—Protocols based on web technology, e.g. hypertext transfer protocol [HTTP] for remote control or remote monitoring of applications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L67/00—Network arrangements or protocols for supporting network services or applications
- H04L67/14—Session management
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R1/00—Details of transducers, loudspeakers or microphones
- H04R1/10—Earpieces; Attachments therefor ; Earphones; Monophonic headphones
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04W—WIRELESS COMMUNICATION NETWORKS
- H04W88/00—Devices specially adapted for wireless communication networks, e.g. terminals, base stations or access point devices
- H04W88/02—Terminal devices
- H04W88/06—Terminal devices adapted for operation in multiple networks or having at least two operational modes, e.g. multi-mode terminals
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
- G10L2015/223—Execution procedure of a spoken command
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Physics & Mathematics (AREA)
- Acoustics & Sound (AREA)
- Computational Linguistics (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Human Computer Interaction (AREA)
- Computer Networks & Wireless Communication (AREA)
- Telephonic Communication Services (AREA)
- Charge And Discharge Circuits For Batteries Or The Like (AREA)
Abstract
The application discloses a network voice recognition method, a network service interaction method and an intelligent headset. The intelligent earphone is in communication connection with the cloud server through the charging box in the intelligent earphone, the charging box sends the voice command sent by the earphone to the cloud server, and the cloud server performs voice command recognition and network service interaction with the cloud server, so that the intelligent earphone can achieve network voice recognition and network service interaction.
Description
Technical Field
The application relates to the technical field of intelligent terminals, in particular to a network voice recognition method, a network service interaction method and an intelligent headset.
Background
As shown in fig. 1, the wireless smart headset 100 includes two components, namely a wireless headset 11 and a charging box 12. The wireless headset 11 includes an audio capture/play/codec unit 111 and a wireless audio unit 122. The audio collecting/playing/encoding/decoding unit 111 is configured to collect audio data and play the received audio data, and the wireless audio unit 122 is configured to implement wireless communication with the intelligent terminal, establish an audio data transmission channel, and transmit the audio data collected by the wireless headset and the audio data to be played. The charging box 12 includes a charging/key control unit 121 and an energy storage/power supply unit 122, and the energy storage/power supply unit 122 is used for realizing battery energy storage, supplying power to the earphone, and the like.
At present, network voice recognition or network service interaction is popular with consumers, and solutions for supporting network voice recognition or network service interaction by intelligent terminals are sought. At present, a smart speaker/mobile phone can already support network voice recognition and service interaction, however, the smart speaker is an external audio unit and has poor privacy. The size of the smart phone is limited by the screen, and the smart phone cannot meet the experience requirements of a scene with a carrying requirement, such as a sports scene. The intelligent headset has good privacy and is convenient to carry, however, no intelligent headset can realize network voice recognition or network service interaction at present.
Therefore, how to implement network voice recognition or network service interaction by the intelligent headset becomes a problem to be solved urgently at present.
Disclosure of Invention
The application provides a network voice recognition method, a network service interaction method and an intelligent headset, so that the intelligent headset can realize network voice recognition or network service interaction.
One aspect of the present application provides a network voice recognition method applied to an intelligent headset, where the intelligent headset includes a headset and a charging box, and the method includes: the charging box receives voice commands from the headset; the charging box sends the voice command to a cloud server; the charging box receives a voice command recognition result from the cloud server; the charging box sends the voice command recognition result to the earphone, or the charging box executes the voice command and sends a voice command execution result to the earphone; and the earphone plays the voice command recognition result or the voice command execution result.
In the aspect, the charging box sends the voice command received from the earphone to the cloud server, and then receives the voice command recognition result sent by the cloud server, so that the intelligent earphone can realize network voice recognition, and the charging box can execute the recognized voice command.
In one implementation, the method further comprises: the earphone collects the voice command; and the earphone sends the collected voice command to the charging box.
In another implementation, the method further comprises: the earphone establishes a communication connection with the charging box; and the charging box establishes a communication connection with the cloud server.
In yet another implementation, after the charging box sends the voice command recognition result to the headset, the method further includes: the headset executes the voice command.
In this implementation, the recognized voice command is executed by the headset.
Another aspect of the present application provides a network service interaction method, which is applied to an intelligent headset, where the intelligent headset includes a headset and a charging box, and the method includes: the charging box receives a voice command from the earphone, wherein the voice command is used for indicating to acquire audio data; the charging box sends the voice command to a cloud server; the charging box receives audio data from the cloud server; the charging box sends the audio data to the headset; and the earphone decodes and plays the audio data.
In the aspect, the voice instruction for acquiring the audio data sent by the earphone is received by the charging box and sent to the cloud server, the cloud server identifies the voice instruction and executes the voice instruction, the audio data is sent to the charging box, and the charging box sends the audio data to the earphone for decoding and playing, so that the intelligent earphone can realize network service interaction.
In one implementation, the method further comprises: the charging box decodes audio data received from the cloud server; and the charging box sending the audio data to the headset, including: the charging box sends the decoded audio data to the headset.
In this implementation, when the formats of the audio data sent between the charging box and the headset are inconsistent, the charging box further decodes the audio data received from the cloud server, and sends the decoded audio data to the headset for re-decoding.
In another implementation, after the charging box sends the voice command to a cloud server, the method further includes: the charging box receives a voice command recognition result from the cloud server; and the charging box sends the voice command recognition result to the earphone.
In the implementation mode, when the voice recognition and the network service interaction are not integrated in the cloud server, the charging box also receives a voice command recognition result sent by the cloud server and sends the voice command recognition result to the earphone.
In yet another implementation, the method further comprises: the earphone collects the voice command; and the earphone sends the collected voice command to the charging box.
In yet another implementation, the method further comprises: the earphone establishes a communication connection with the charging box; and the charging box establishes a communication connection with the cloud server.
Yet another aspect of the present application provides a network service interaction method applied to a smart headset, where the smart headset includes a headset and a charging box, and the method includes: the charging box receives voice commands from the headset; the charging box sends the voice command to a cloud server; the charging box receives a voice command execution result from the cloud server; the charging box sends the voice command execution result to the earphone; and the earphone plays the voice command execution result.
In the aspect, the charging box receives the voice command sent by the earphone and sends the voice command to the cloud server, and the cloud server recognizes and executes the voice command and returns a voice command execution result to the charging box, so that the intelligent earphone realizes network service interaction.
Yet another aspect of the present application provides a smart headset comprising a headset and a charging box for receiving voice commands from the headset; the charging box is further used for sending the voice command to a cloud server; the charging box is further used for receiving a voice command recognition result from the cloud server; the charging box is also used for sending the voice command recognition result to the earphone, or the charging box is also used for executing the voice command and sending a voice command execution result to the earphone; and the earphone is used for playing the voice command recognition result or the voice command execution result.
In one implementation, the headset is further configured to capture the voice command; and the earphone is also used for sending the collected voice command to the charging box.
In another implementation, the headset is further configured to establish a communication connection with the charging box; and the charging box is also used for establishing communication connection with the cloud server.
In yet another implementation, the headset is further configured to execute the voice command.
Yet another aspect of the present application provides a smart headset comprising a headset and a charging box for receiving a voice command from the headset, the voice command being indicative of acquiring audio data; the charging box is further used for sending the voice command to a cloud server; the charging box is also used for receiving audio data from the cloud server; the charging box is further used for sending the audio data to the earphone; and the earphone is used for decoding and playing the audio data.
In one implementation, the charging box is further configured to decode audio data received from the cloud server; and the charging box is further configured to send the decoded audio data to the headset.
In another implementation, the charging box is further configured to receive a voice command recognition result from the cloud server; and the charging box is also used for sending the voice command recognition result to the earphone.
In yet another implementation, the headset is further configured to capture the voice command; and the earphone is also used for sending the collected voice command to the charging box.
In yet another implementation, the headset is further configured to establish a communication connection with the charging box; and the charging box is also used for establishing communication connection with the cloud server.
Yet another aspect of the present application provides a smart headset comprising a headset and a charging box for receiving voice commands from the headset; the charging box is further used for sending the voice command to a cloud server; the charging box is also used for receiving a voice command execution result from the cloud server; the charging box is also used for sending the voice command execution result to the earphone; and the earphone is used for playing the voice command execution result.
Yet another aspect of the present application provides a computer-readable storage medium having stored therein instructions, which when executed on a computer, cause the computer to perform the method of the above-described aspects.
Yet another aspect of the present application provides a computer program product containing instructions which, when run on a computer, cause the computer to perform the method of the above-described aspects.
Drawings
Fig. 1 is a schematic structural diagram of a conventional smart headset;
fig. 2 is a general structural schematic diagram of an intelligent headset according to an embodiment of the present disclosure;
fig. 3 is a schematic flowchart of a network speech recognition method according to an embodiment of the present application;
fig. 4 is a schematic flowchart of a network service interaction provided in an embodiment of the present application;
fig. 5 is a schematic flow chart of another network service interaction provided in the embodiment of the present application.
Detailed Description
The embodiments of the present application will be described below with reference to the drawings.
The network voice recognition is a function that the terminal collects a user voice command, transmits the user voice command to the cloud server by using a network, recognizes the user voice command, returns the user voice command to the terminal, and is executed by the terminal.
The network interaction service means that a terminal sends a service requirement to a cloud server, and the cloud server responds to the terminal requirement and sends an execution result to the terminal.
In the application, the network voice recognition service and the network interaction service may be integrated in one server, or may be executed by different servers.
Referring to fig. 2, fig. 2 is a schematic diagram of a general structure of an intelligent headset according to an embodiment of the present disclosure, where the intelligent headset includes a headset 21 and a charging box 22. Generally, the earphone 21 is wirelessly connected with the charging box 22. The structure of the headphone 21 is the same as that shown in fig. 1, and includes an audio capture/play/codec unit 211 and a wireless audio unit 212. The audio capturing/playing/encoding/decoding unit 211 further specifically includes functions of voice wakeup, local audio playing, audio capturing, audio encoding and decoding, and the like. It should be noted that the audio capture/play/codec unit 211 may include different functions for different network functions. For example, when the smart headset is used to implement network voice recognition, the audio acquisition/playing/codec unit 211 may include functions such as voice wakeup, audio acquisition, and audio playing; when the smart headset is used to implement network interaction services, such as acquiring audio data, the audio collecting/playing/codec unit 211 includes functions of voice wakeup, audio playing, audio collecting, audio codec, and the like. When the smart headset is used to implement both the network voice recognition and the network interactive service, the audio collection/play/codec unit 211 may include all the above functions. The wireless audio unit 212 can perform Bluetooth (BT) communication or the like with the wireless audio unit 223 of the charging box 22.
The charging box 22 includes an internet service unit 221, an audio codec unit 222, a wireless audio unit 223, a wireless network unit 224, a charging/key control unit 225, and a power storage/supply 226, etc. Since the charging box 22 has a function of connecting to a cloud server, it may also be referred to as a networking charging box. The internet service unit 221 includes a streaming media service Software Development Kit (SDK) and a speech engine cloud SDK. The streaming media service SDK is streaming media software for connecting the cloud server, and the voice engine cloud end SDK is voice software for connecting the cloud server. According to the function implementation of the smart headset, the two SDKs can be included, or one SDK can be included. The audio codec unit 222 includes functions of local playing, streaming media playing, and audio codec. The audio codec unit 222 is optional according to the functional implementation of the smart headset. The wireless network unit 224 is configured to implement communication connection with the cloud server, and may establish communication connection with the server in a communication connection manner such as 4G/3G/2G, Wi-Fi and TCP/IP. The functions of the charging/key control unit 225 and the power storing/supplying unit 226 are the same as those of the charging/key control unit 121 and the power storing/supplying unit 122 of the embodiment shown in fig. 1, respectively.
Referring to fig. 3, fig. 3 is a schematic flowchart of a network voice recognition method according to an embodiment of the present application, and is applied to the smart headset shown in fig. 2. The method comprises the following steps:
s101, the earphone sends a voice command to the charging box.
In this step, after the user wears the earphone, the user gives a voice command by waking up the earphone through voice or manually. The headset collects the voice command of the user through the microphone, and optionally before S101, the method further includes: the earphone collects the voice command; and the earphone sends the collected voice command to the charging box.
The headset may also retrieve local voice commands.
The headset itself cannot recognize the voice command. Therefore, after the headset collects the voice command, the voice command is converted into voice command data, and the voice command data is transmitted to the wireless audio unit 223 of the charging box through the wireless audio unit 212. The voice command may be a control command or may be another command.
Optionally, before S101, the method further includes:
the earphone establishes a communication connection with the charging box;
the charging box establishes a communication connection with the cloud server.
In particular, the headset establishes a communication connection, which may be a bluetooth connection, for example, with the wireless audio unit 223 of the charging box through the wireless audio unit 212.
The charging box establishes a communication connection with the cloud server through the wireless network unit 224. For example, communication connection modes such as 4G/3G/2G, Wi-Fi and TCP/IP can be adopted.
And S102, after receiving the voice command, the charging box sends the voice command to a cloud server.
After the wireless audio unit 223 of the charging box receives the voice command transmitted from the wireless audio unit 212 of the headset, it cannot recognize the voice command by itself, and transmits the voice command received from the headset to the cloud server through the wireless network unit 224.
In another embodiment, the charging box may also have a local language command recognition capability, and after receiving the voice command sent by the headset, the charging box performs local recognition, if the voice command cannot be locally recognized, and forwards the voice command to the cloud server for network recognition.
And S103, after receiving the voice command, the cloud server identifies the voice command.
The cloud server has a voice recognition function, and after receiving the voice command sent by the charging box, recognizes the voice command to obtain a voice command recognition result. The speech recognition process can refer to the existing speech recognition technology, and is not described in detail herein.
And S104, the cloud server sends a voice command recognition result to the charging box.
And the cloud server sends a voice command recognition result to the charging box through a wireless network.
As one implementation, the voice command may be executed by the charging box when the voice command is a control command. And after S104, proceeding to S105, after the charging box receives the voice command recognition result, executing the voice command.
For example, the voice command is "play song" love Nostol ", and the charging box calls the song" love Nostol "from the memory or acquires the song" love Nostol "from the server after receiving the voice command recognition result sent by the cloud server.
And S106, the charging box sends a voice command execution result to the earphone.
As shown in the above example, the result of the voice command execution may be audio data for the song.
And S107, after receiving the voice command execution result, the earphone plays the voice command execution result.
As shown in the above example, the headset plays the song after receiving the audio data for the song.
Optionally, after S107, the method further includes: the headset executes the voice command.
As another implementation, the voice command may also be executed by the headset when the voice command is a control command. After S104, proceeding to S106, after the charging box receives the voice command recognition result, the charging box sends the voice command recognition result to the headset.
For example, the voice command is "play song" love noun ", and the charging box sends the voice command recognition result to the headset" — play song "love noun".
And S107, after receiving the voice command recognition result, the earphone plays the voice command recognition result.
As shown in the above example, after receiving the voice command recognition result, "play song" love noun ", the headset acquires the audio data of the song from the charging box through the wireless audio unit 212, or acquires the audio data of the song from the local memory, and plays the song.
Due to the limitation of the size and the power supply of the earphone, the network voice recognition function of the earphone and the cloud server cannot be realized, and in the embodiment, the network voice recognition function of the earphone and the cloud server can be realized by integrating the charging box with the unit. The intelligent earphone with the integrated network voice recognition function can be realized by combining the original earphone.
According to the network voice recognition method provided by the embodiment of the application, the charging box sends the voice command received from the earphone to the cloud server, and then the charging box receives the voice command recognition result sent by the cloud server, so that the intelligent earphone can realize network voice recognition, and the charging box can execute the recognized voice command.
Referring to fig. 4, fig. 4 is a schematic flowchart of a network service interaction method according to an embodiment of the present application, and is applied to the smart headset shown in fig. 2. The method comprises the following steps:
s201, the earphone sends a voice command to the charging box, and the voice command is used for indicating to acquire audio data.
In this step, after the user wears the earphone, the user gives a voice command by waking up the earphone through voice or manually. The headset collects the voice command of the user through the microphone, and optionally before S201, the method further includes: the earphone collects the voice command; and the earphone sends the collected voice command to the charging box.
The headset may also retrieve local voice commands.
The headset itself cannot recognize the voice command. Therefore, after the headset collects the voice command, the voice command is converted into voice command data, and the voice command data is transmitted to the wireless audio unit 223 of the charging box through the wireless audio unit 212. The voice command is used to indicate data audio data. For example, audio data of song "noun of love" is acquired.
Optionally, before S201, the method further includes:
the earphone establishes a communication connection with the charging box;
the charging box establishes a communication connection with the cloud server.
In particular, the headset establishes a communication connection, which may be a bluetooth connection, for example, with the wireless audio unit 223 of the charging box through the wireless audio unit 212.
The charging box establishes a communication connection with the cloud server through the wireless network unit 224. For example, communication connection modes such as 4G/3G/2G, Wi-Fi and TCP/IP can be adopted.
And S202, after receiving the voice command, the charging box sends the voice command to a cloud server.
After the wireless audio unit 223 of the charging box receives the voice command transmitted from the wireless audio unit 212 of the headset, it cannot recognize the voice command by itself, and transmits the voice command received from the headset to the cloud server through the wireless network unit 224.
In another embodiment, the charging box has a function of storing a local media resource file, and when a command for playing the local media resource is identified (local identification or network identification), the local media resource file is read, decoded and sent to the earphone for playing.
S203, the cloud server identifies the voice command and acquires the audio data indicated by the voice command.
The cloud server has a voice recognition function and a network audio service function. Alternatively, the voice recognition function and the network audio service function may be integrated in one server, or may be implemented by different servers.
When the voice recognition function and the network audio service function are integrated in one server, the cloud server recognizes the voice command after receiving the voice command sent by the charging box, and acquires audio data indicated by the voice command.
When the voice recognition function and the network audio service function are realized by different servers, after recognizing the voice command, the cloud server sends a request for acquiring audio data to another server with the network audio service function, and receives audio data sent by another server. Optionally, after S203, the method further includes: the cloud server sends a voice command recognition result to the charging box; and after receiving the voice command recognition result from the cloud server, the charging box sends the voice command recognition result to the earphone.
S204, the cloud server sends the audio data to the charging box.
S205, after receiving the audio data sent by the cloud server, the charging box sends the audio data to the earphone.
Specifically, the wireless network unit 224 of the charging box receives audio data transmitted from the cloud server and transmits the audio data to the wireless audio unit 212 of the headset through the wireless audio unit 223.
S206, the earphone decodes and plays the audio data.
When the format of the audio data sent by the charging box is consistent with the format of the audio data played by the earphone, the charging box can directly send the audio data to the earphone after receiving the audio data sent by the cloud server, and the audio data is decoded and played by the earphone.
Optionally, when the format of the audio data sent by the charging box is not consistent with the format of the audio data played by the earphone, after S204, the method further includes: the charging box decodes audio data received from the cloud server.
S205 specifically comprises: the charging box sends the decoded audio data to the headset.
Specifically, since the format of the audio data sent by the charging box is inconsistent with the format of the audio data played by the headset, the charging box needs to decode the audio data received from the cloud server and send the decoded audio data to the headset. And after receiving the audio data decoded by the charging box, the earphone decodes the audio data according to the playing format of the earphone and plays the audio data decoded by the earphone.
According to the network service interaction method provided by the embodiment of the application, the charging box receives the voice instruction for acquiring the audio data sent by the earphone and sends the voice instruction to the cloud server, the cloud server recognizes the voice instruction and executes the voice instruction, the audio data are sent to the charging box, and the charging box sends the audio data to the earphone for decoding and playing, so that the intelligent earphone can realize network service interaction.
Referring to fig. 5, fig. 5 is a schematic flowchart of a network service interaction method according to an embodiment of the present application, and is applied to the smart headset shown in fig. 2. The method comprises the following steps:
s301, the earphone sends a voice command to the charging box.
Unlike the embodiment shown in fig. 4, the voice command may be a network traffic service command instructing the cloud server to proceed. For example, the user's daily number of athletic steps is uploaded to the cloud server. When the user walks quickly at the current moment, the user sends a voice instruction to the earphone, and the cloud server is instructed to prompt when the exercise step number of the user exceeds the average value of the exercise step numbers of the user every day. For example, the user sends a voice command "do i exceed average daily steps today? "
S302, after the charging box receives the voice command from the earphone, the charging box sends the voice command to a cloud server.
After the wireless audio unit 223 of the charging box receives the voice command transmitted from the wireless audio unit 212 of the headset, it cannot recognize the voice command by itself, and transmits the voice command received from the headset to the cloud server through the wireless network unit 224.
And S303, after receiving the voice command, the cloud server identifies the voice command and executes the voice command.
The cloud server has a voice recognition function, and recognizes a voice command of a user, "do i exceed the average number of steps per day today? The cloud server executes the voice command, compares the current step number of the user uploaded by the earphone or the charging box with the daily average step number of the user, and obtains a voice command execution result that the current step number of the user exceeds the daily average step number of the user, namely that the current step number of the user exceeds the daily average step number of the user! "
S304, the cloud server sends a voice command execution result to the charging box.
S305, after the charging box receives a voice command execution result from the cloud server, the charging box sends the voice command execution result to the earphone.
Specifically, the wireless network unit 224 of the charging box receives the voice command execution result transmitted from the cloud server and transmits the voice command execution result to the wireless audio unit 212 of the headset through the wireless audio unit 223.
S306, after the earphone receives the voice command execution result, playing the voice command execution result.
The wireless audio unit 212 of the headset plays the voice command execution result after receiving the voice command execution result.
Of course, after the cloud server executes the voice command, the command execution result may be in another prompting manner, for example, the charging box performs vibration prompting after receiving the command execution result.
According to the network service interaction method provided by the embodiment of the application, the charging box receives the voice command sent by the earphone and sends the voice command to the cloud server, and the cloud server identifies and executes the voice command and returns the voice command execution result to the charging box, so that the intelligent earphone realizes network service interaction.
The embodiment of the application also provides an intelligent headset, and the structure of the intelligent headset is shown in fig. 2. Specifically, the method comprises the following steps:
the charging box is used for receiving voice commands from the earphone;
the charging box is further used for sending the voice command to a cloud server;
the charging box is further used for receiving a voice command recognition result from the cloud server;
the charging box is also used for sending the voice command recognition result to the earphone, or the charging box is also used for executing the voice command and sending a voice command execution result to the earphone;
the earphone is used for playing the voice command recognition result or the voice command execution result.
Optionally, the headset is further configured to capture the voice command; and the earphone is also used for sending the collected voice command to the charging box.
Optionally, the headset is further configured to establish a communication connection with the charging box; and the charging box is also used for establishing communication connection with the cloud server.
Optionally, the headset is further configured to execute the voice command.
Specific functional implementation can refer to the embodiment shown in fig. 3.
According to the intelligent earphone provided by the embodiment of the application, the charging box sends the voice command received from the earphone to the cloud server, and then the charging box receives the voice command recognition result sent by the cloud server, so that the intelligent earphone can realize network voice recognition, and the charging box can execute the recognized voice command.
The embodiment of the application also provides another intelligent headset, and the structure of the intelligent headset is shown in fig. 2. Specifically, the method comprises the following steps:
the charging box is used for receiving a voice command from the earphone, and the voice command is used for indicating to acquire audio data;
the charging box is further used for sending the voice command to a cloud server;
the charging box is also used for receiving audio data from the cloud server;
the charging box is further used for sending the audio data to the earphone;
the earphone is used for decoding and playing the audio data.
Optionally, the charging box is further configured to decode audio data received from the cloud server; and the charging box is further configured to send the decoded audio data to the headset.
Optionally, the charging box is further configured to receive a voice command recognition result from the cloud server; and the charging box is also used for sending the voice command recognition result to the earphone.
Optionally, the headset is further configured to capture the voice command; and the earphone is also used for sending the collected voice command to the charging box.
Optionally, the headset is further configured to establish a communication connection with the charging box; and the charging box is also used for establishing communication connection with the cloud server.
Specific functional implementation can refer to the embodiment shown in fig. 4.
According to the intelligent earphone provided by the embodiment of the application, the voice instruction for acquiring the audio data sent by the earphone is received by the charging box and sent to the cloud server, the cloud server identifies the voice instruction and executes the voice instruction, the audio data is sent to the charging box, and the charging box sends the audio data to the earphone for decoding and playing, so that the intelligent earphone can realize network service interaction.
The embodiment of the application also provides another intelligent headset, and the structure of the intelligent headset is shown in fig. 2. Specifically, the method comprises the following steps:
the charging box is used for receiving voice commands from the earphone;
the charging box is further used for sending the voice command to a cloud server;
the charging box is also used for receiving a voice command execution result from the cloud server;
the charging box is also used for sending the voice command execution result to the earphone;
the earphone is used for playing the voice command execution result.
Specific functional implementation can refer to the embodiment shown in fig. 5.
According to the intelligent earphone provided by the embodiment of the application, the charging box receives the voice command sent by the earphone and sends the voice command to the cloud server, the cloud server identifies and executes the voice command, and the voice command execution result is returned to the charging box, so that the intelligent earphone realizes network service interaction.
Embodiments of the present application further provide a computer-readable storage medium, on which a computer program is stored, and when the computer program is executed by a processor, the computer program implements the method performed by the smart headset in the embodiments shown in fig. 3 to 5.
It is clear to those skilled in the art that, for convenience and brevity of description, the specific working processes of the above-described systems, apparatuses and units may refer to the corresponding processes in the foregoing method embodiments, and are not described herein again.
In the several embodiments provided in the present application, it should be understood that the disclosed system, apparatus and method may be implemented in other ways. For example, the division of the unit is only one logical function division, and other division may be implemented in practice, for example, a plurality of units or components may be combined or integrated into another system, or some features may be omitted, or not executed. The shown or discussed mutual coupling, direct coupling or communication connection may be an indirect coupling or communication connection of devices or units through some interfaces, and may be in an electrical, mechanical or other form.
Units described as separate parts may or may not be physically separate, and parts displayed as units may or may not be physical units, may be located in one place, or may be distributed on a plurality of network units. Some or all of the units can be selected according to actual needs to achieve the purpose of the solution of the embodiment.
In the above embodiments, the implementation may be wholly or partially realized by software, hardware, firmware, or any combination thereof. When implemented in software, may be implemented in whole or in part in the form of a computer program product. The computer program product includes one or more computer instructions. The procedures or functions according to the embodiments of the present application are wholly or partially generated when the computer program instructions are loaded and executed on a computer. The computer may be a general purpose computer, a special purpose computer, a network of computers, or other programmable device. The computer instructions may be stored on or transmitted over a computer-readable storage medium. The computer instructions may be transmitted from one website, computer, server, or data center to another website, computer, server, or data center by wire (e.g., coaxial cable, fiber optic, Digital Subscriber Line (DSL)), or wirelessly (e.g., infrared, wireless, microwave, etc.). The computer-readable storage medium can be any available medium that can be accessed by a computer or a data storage device, such as a server, a data center, etc., that includes one or more of the available media. The usable medium may be a read-only memory (ROM), or a Random Access Memory (RAM), or a magnetic medium, such as a floppy disk, a hard disk, a magnetic tape, a magnetic disk, or an optical medium, such as a Digital Versatile Disk (DVD), or a semiconductor medium, such as a Solid State Disk (SSD).
Claims (21)
1. A network voice recognition method is applied to an intelligent headset, the intelligent headset comprises a headset body and a charging box, and the method comprises the following steps:
the charging box receives voice commands from the headset;
the charging box sends the voice command to a cloud server;
the charging box receives a voice command recognition result from the cloud server;
the charging box sends the voice command recognition result to the earphone, or the charging box executes the voice command and sends a voice command execution result to the earphone;
and the earphone plays the voice command recognition result or the voice command execution result.
2. The method of claim 1, wherein the method further comprises:
the earphone collects the voice command;
and the earphone sends the collected voice command to the charging box.
3. The method of claim 1 or 2, wherein the method further comprises:
the earphone establishes a communication connection with the charging box;
the charging box establishes a communication connection with the cloud server.
4. The method of claim 1, wherein after the charging box sends the voice command recognition result to the headset, the method further comprises:
the headset executes the voice command.
5. A network service interaction method is applied to an intelligent headset, wherein the intelligent headset comprises a headset body and a charging box, and the method comprises the following steps:
the charging box receives a voice command from the earphone, wherein the voice command is used for indicating to acquire audio data;
the charging box sends the voice command to a cloud server;
the charging box receives audio data from the cloud server;
the charging box sends the audio data to the headset;
the headphones decode and play the audio data.
6. The method of claim 5, wherein the method further comprises:
the charging box decodes audio data received from the cloud server;
the charging box sending the audio data to the headset, comprising:
the charging box sends the decoded audio data to the headset.
7. The method of claim 5, wherein after the charging box sends the voice command to a cloud server, the method further comprises:
the charging box receives a voice command recognition result from the cloud server;
the charging box sends the voice command recognition result to the headset.
8. The method of any one of claims 5 to 7, further comprising:
the earphone collects the voice command;
and the earphone sends the collected voice command to the charging box.
9. The method of any one of claims 5 to 8, further comprising:
the earphone establishes a communication connection with the charging box;
the charging box establishes a communication connection with the cloud server.
10. A network service interaction method is applied to an intelligent headset, wherein the intelligent headset comprises a headset body and a charging box, and the method comprises the following steps:
the charging box receives voice commands from the headset;
the charging box sends the voice command to a cloud server;
the charging box receives a voice command execution result from the cloud server;
the charging box sends the voice command execution result to the earphone;
and the earphone plays the voice command execution result.
11. The utility model provides an intelligent earphone, intelligent earphone includes the earphone and charges the box, its characterized in that:
the charging box is used for receiving voice commands from the earphone;
the charging box is further used for sending the voice command to a cloud server;
the charging box is further used for receiving a voice command recognition result from the cloud server;
the charging box is also used for sending the voice command recognition result to the earphone, or the charging box is also used for executing the voice command and sending a voice command execution result to the earphone;
the earphone is used for playing the voice command recognition result or the voice command execution result.
12. The smart headset as recited in claim 11, wherein:
the earphone is also used for collecting the voice command;
the earphone is also used for sending the collected voice command to the charging box.
13. The smart headset as recited in claim 11 or 12, wherein:
the earphone is also used for establishing communication connection with the charging box;
the charging box is also used for establishing communication connection with the cloud server.
14. The smart headset as recited in claim 11, wherein:
the headset is also configured to execute the voice command.
15. The utility model provides an intelligent earphone, intelligent earphone includes the earphone and charges the box, its characterized in that:
the charging box is used for receiving a voice command from the earphone, and the voice command is used for indicating to acquire audio data;
the charging box is further used for sending the voice command to a cloud server;
the charging box is also used for receiving audio data from the cloud server;
the charging box is further used for sending the audio data to the earphone;
the earphone is used for decoding and playing the audio data.
16. The smart headset as recited in claim 15, wherein:
the charging box is further used for decoding audio data received from the cloud server;
the charging box is further configured to send the decoded audio data to the headset.
17. The smart headset as recited in claim 15, wherein:
the charging box is further used for receiving a voice command recognition result from the cloud server;
the charging box is further configured to send the voice command recognition result to the headset.
18. The smart headset as recited in any one of claims 15-17, wherein:
the earphone is also used for collecting the voice command;
the earphone is also used for sending the collected voice command to the charging box.
19. The smart headset as recited in any one of claims 15-18, wherein:
the earphone is also used for establishing communication connection with the charging box;
the charging box is also used for establishing communication connection with the cloud server.
20. The utility model provides an intelligent earphone, intelligent earphone includes the earphone and charges the box, its characterized in that:
the charging box is used for receiving voice commands from the earphone;
the charging box is further used for sending the voice command to a cloud server;
the charging box is also used for receiving a voice command execution result from the cloud server;
the charging box is also used for sending the voice command execution result to the earphone;
the earphone is used for playing the voice command execution result.
21. A computer-readable storage medium, on which a computer program is stored, which program, when being executed by a processor, carries out the method of any one of claims 1 to 4, or the method of any one of claims 5 to 9, or the method of claim 10.
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201811465464.4A CN111276135B (en) | 2018-12-03 | 2018-12-03 | Network voice recognition method, network service interaction method and intelligent earphone |
PCT/CN2019/115873 WO2020114181A1 (en) | 2018-12-03 | 2019-11-06 | Network voice recognition method, network service interaction method and intelligent earphone |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201811465464.4A CN111276135B (en) | 2018-12-03 | 2018-12-03 | Network voice recognition method, network service interaction method and intelligent earphone |
Publications (2)
Publication Number | Publication Date |
---|---|
CN111276135A true CN111276135A (en) | 2020-06-12 |
CN111276135B CN111276135B (en) | 2023-06-20 |
Family
ID=70974054
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201811465464.4A Active CN111276135B (en) | 2018-12-03 | 2018-12-03 | Network voice recognition method, network service interaction method and intelligent earphone |
Country Status (2)
Country | Link |
---|---|
CN (1) | CN111276135B (en) |
WO (1) | WO2020114181A1 (en) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113286212A (en) * | 2021-05-20 | 2021-08-20 | 北京明略软件***有限公司 | Wearing type audio acquisition assembly |
CN113380251A (en) * | 2021-06-22 | 2021-09-10 | 紫优科技(深圳)有限公司 | Mobile voice interaction method and device based on intelligent earphone |
CN113411709A (en) * | 2021-06-28 | 2021-09-17 | 紫优科技(深圳)有限公司 | Design method and system of cloud intelligent earphone system |
CN113421570A (en) * | 2021-06-21 | 2021-09-21 | 紫优科技(深圳)有限公司 | Intelligent earphone identity authentication method and device |
Citations (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP3119248U (en) * | 2005-12-06 | 2006-02-16 | ▲シウ▼瑩企業有限公司 | Wireless earphone device and charging base assembly |
CN102594988A (en) * | 2012-02-10 | 2012-07-18 | 深圳市中兴移动通信有限公司 | Method and system capable of achieving automatic pairing connection of Bluetooth earphones by speech recognition |
US20140119554A1 (en) * | 2012-10-25 | 2014-05-01 | Elwha Llc | Methods and systems for non-volatile memory in wireless headsets |
WO2014197635A2 (en) * | 2013-06-07 | 2014-12-11 | Apple Inc. | Intelligent automated assistant |
US20170110124A1 (en) * | 2015-10-20 | 2017-04-20 | Bragi GmbH | Wearable Earpiece Voice Command Control System and Method |
CN106850847A (en) * | 2017-03-10 | 2017-06-13 | 上海斐讯数据通信技术有限公司 | Voice messaging sharing method and its intelligent earphone based on cloud platform |
CN107333201A (en) * | 2017-07-24 | 2017-11-07 | 歌尔科技有限公司 | One kind translation earphone storage box, wireless translation earphone and wireless translation system |
CN206977651U (en) * | 2017-06-05 | 2018-02-06 | 广东朝阳电子科技股份有限公司 | The circuit structure of TWS Bluetooth earphone devices with WiFi function |
CN207518810U (en) * | 2017-11-20 | 2018-06-19 | 深圳市胜兴旺精密科技有限公司 | Charging box |
CN108509428A (en) * | 2018-02-26 | 2018-09-07 | 深圳市百泰实业股份有限公司 | Earphone interpretation method and system |
CN108550367A (en) * | 2018-05-18 | 2018-09-18 | 深圳傲智天下信息科技有限公司 | A kind of portable intelligent interactive voice control device, method and system |
CN108549206A (en) * | 2018-07-12 | 2018-09-18 | 深圳傲智天下信息科技有限公司 | A kind of band has the smartwatch of voice interactive function earphone |
CN108564949A (en) * | 2018-05-18 | 2018-09-21 | 深圳傲智天下信息科技有限公司 | A kind of TWS earphones, Wrist belt-type AI voice interaction devices and system |
CN108900945A (en) * | 2018-09-29 | 2018-11-27 | 上海与德科技有限公司 | Bluetooth headset box and audio recognition method, server and storage medium |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9402120B2 (en) * | 2014-09-05 | 2016-07-26 | Epickal AB | Wireless earbuds |
-
2018
- 2018-12-03 CN CN201811465464.4A patent/CN111276135B/en active Active
-
2019
- 2019-11-06 WO PCT/CN2019/115873 patent/WO2020114181A1/en active Application Filing
Patent Citations (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP3119248U (en) * | 2005-12-06 | 2006-02-16 | ▲シウ▼瑩企業有限公司 | Wireless earphone device and charging base assembly |
CN102594988A (en) * | 2012-02-10 | 2012-07-18 | 深圳市中兴移动通信有限公司 | Method and system capable of achieving automatic pairing connection of Bluetooth earphones by speech recognition |
US20140119554A1 (en) * | 2012-10-25 | 2014-05-01 | Elwha Llc | Methods and systems for non-volatile memory in wireless headsets |
WO2014197635A2 (en) * | 2013-06-07 | 2014-12-11 | Apple Inc. | Intelligent automated assistant |
US20170110124A1 (en) * | 2015-10-20 | 2017-04-20 | Bragi GmbH | Wearable Earpiece Voice Command Control System and Method |
CN106850847A (en) * | 2017-03-10 | 2017-06-13 | 上海斐讯数据通信技术有限公司 | Voice messaging sharing method and its intelligent earphone based on cloud platform |
CN206977651U (en) * | 2017-06-05 | 2018-02-06 | 广东朝阳电子科技股份有限公司 | The circuit structure of TWS Bluetooth earphone devices with WiFi function |
CN107333201A (en) * | 2017-07-24 | 2017-11-07 | 歌尔科技有限公司 | One kind translation earphone storage box, wireless translation earphone and wireless translation system |
CN207518810U (en) * | 2017-11-20 | 2018-06-19 | 深圳市胜兴旺精密科技有限公司 | Charging box |
CN108509428A (en) * | 2018-02-26 | 2018-09-07 | 深圳市百泰实业股份有限公司 | Earphone interpretation method and system |
CN108550367A (en) * | 2018-05-18 | 2018-09-18 | 深圳傲智天下信息科技有限公司 | A kind of portable intelligent interactive voice control device, method and system |
CN108564949A (en) * | 2018-05-18 | 2018-09-21 | 深圳傲智天下信息科技有限公司 | A kind of TWS earphones, Wrist belt-type AI voice interaction devices and system |
CN108549206A (en) * | 2018-07-12 | 2018-09-18 | 深圳傲智天下信息科技有限公司 | A kind of band has the smartwatch of voice interactive function earphone |
CN108900945A (en) * | 2018-09-29 | 2018-11-27 | 上海与德科技有限公司 | Bluetooth headset box and audio recognition method, server and storage medium |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113286212A (en) * | 2021-05-20 | 2021-08-20 | 北京明略软件***有限公司 | Wearing type audio acquisition assembly |
CN113286212B (en) * | 2021-05-20 | 2022-07-12 | 北京明略软件***有限公司 | Wearing type audio acquisition assembly |
CN113421570A (en) * | 2021-06-21 | 2021-09-21 | 紫优科技(深圳)有限公司 | Intelligent earphone identity authentication method and device |
CN113380251A (en) * | 2021-06-22 | 2021-09-10 | 紫优科技(深圳)有限公司 | Mobile voice interaction method and device based on intelligent earphone |
CN113411709A (en) * | 2021-06-28 | 2021-09-17 | 紫优科技(深圳)有限公司 | Design method and system of cloud intelligent earphone system |
Also Published As
Publication number | Publication date |
---|---|
WO2020114181A1 (en) | 2020-06-11 |
CN111276135B (en) | 2023-06-20 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN111276135B (en) | Network voice recognition method, network service interaction method and intelligent earphone | |
CN208507180U (en) | A kind of portable intelligent interactive voice control equipment | |
CN110769394B (en) | Video call method and electronic equipment | |
CN208689384U (en) | A kind of band has the smartwatch of voice interactive function earphone | |
WO2019218368A1 (en) | Tws headphones, wrist-worn ai voice interaction device and system | |
WO2020133112A1 (en) | Method for automatically switching bluetooth audio encoding method and electronic apparatus | |
CN110278205B (en) | Bluetooth sound box base and control method and system thereof | |
WO2020132818A1 (en) | Wireless short-distance audio sharing method and electronic device | |
CN109890018A (en) | Blue-tooth transmission method, bluetooth transceiver and the computer readable storage medium of audio | |
CN201629037U (en) | Remote controller | |
JP7442647B2 (en) | bluetooth communication method and device | |
CN110177363B (en) | Bluetooth data transmission method and device, Bluetooth chip and storage medium | |
CN110189753B (en) | Bluetooth sound box and control method, system and storage medium thereof | |
CN109286725A (en) | Interpretation method and terminal | |
CN109712631A (en) | Audio data transfer control method, device, system and readable storage medium storing program for executing | |
EP3552508B1 (en) | Smart helmet having remote control, and remote control method thereof | |
CN103794215A (en) | Speech control-based handheld terminal, system and speech control-based control method | |
CN109102816A (en) | Coding control method, device and electronic equipment | |
CN115150501A (en) | Voice interaction method and electronic equipment | |
CN109194998A (en) | Data transmission method, device, electronic equipment and computer-readable medium | |
CN104754261A (en) | Projection equipment and projection method | |
CN204291130U (en) | Mobile network is utilized to realize equipment that is single and multi-person speech message | |
CN108399918B (en) | Intelligent device connection method, intelligent device and terminal | |
CN213906675U (en) | Portable wireless bluetooth recording equipment | |
CN111556406B (en) | Audio processing method, audio processing device and earphone |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |