CN107077845A - A kind of speech output method and device - Google Patents
A kind of speech output method and device Download PDFInfo
- Publication number
- CN107077845A CN107077845A CN201680002958.1A CN201680002958A CN107077845A CN 107077845 A CN107077845 A CN 107077845A CN 201680002958 A CN201680002958 A CN 201680002958A CN 107077845 A CN107077845 A CN 107077845A
- Authority
- CN
- China
- Prior art keywords
- voice input
- user
- input content
- degree
- content
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 61
- 230000019771 cognition Effects 0.000 claims abstract description 294
- 238000004590 computer program Methods 0.000 claims description 12
- 239000000284 extract Substances 0.000 claims description 6
- 230000006870 function Effects 0.000 abstract description 11
- 238000010586 diagram Methods 0.000 description 16
- 238000007493 shaping process Methods 0.000 description 14
- 238000004891 communication Methods 0.000 description 9
- 230000008859 change Effects 0.000 description 7
- 238000005516 engineering process Methods 0.000 description 7
- 238000012545 processing Methods 0.000 description 6
- 238000012986 modification Methods 0.000 description 3
- 230000004048 modification Effects 0.000 description 3
- 230000009286 beneficial effect Effects 0.000 description 2
- 230000003287 optical effect Effects 0.000 description 2
- KLDZYURQCUYZBL-UHFFFAOYSA-N 2-[3-[(2-hydroxyphenyl)methylideneamino]propyliminomethyl]phenol Chemical compound OC1=CC=CC=C1C=NCCCN=CC1=CC=CC=C1O KLDZYURQCUYZBL-UHFFFAOYSA-N 0.000 description 1
- BSYNRYMUTXBXSQ-UHFFFAOYSA-N Aspirin Chemical compound CC(=O)OC1=CC=CC=C1C(O)=O BSYNRYMUTXBXSQ-UHFFFAOYSA-N 0.000 description 1
- 229960001138 acetylsalicylic acid Drugs 0.000 description 1
- 238000004378 air conditioning Methods 0.000 description 1
- 238000004458 analytical method Methods 0.000 description 1
- 238000013500 data storage Methods 0.000 description 1
- 201000001098 delayed sleep phase syndrome Diseases 0.000 description 1
- 208000033921 delayed sleep phase type circadian rhythm sleep disease Diseases 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 239000003814 drug Substances 0.000 description 1
- 230000010365 information processing Effects 0.000 description 1
- 238000007726 management method Methods 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 230000008569 process Effects 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
Landscapes
- Engineering & Computer Science (AREA)
- Computational Linguistics (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- User Interface Of Digital Computer (AREA)
- Telephone Function (AREA)
Abstract
A kind of speech output method and device.Methods described includes:Receive the phonetic entry content (S11) of user's input;According to the phonetic entry content, cognition degree of the user to the phonetic entry content generic is determined, the cognition degree is the professional knowledge degree of awareness (S12) of the user to the classification;From at least one voice output content corresponding with the phonetic entry content, the voice output content (S13) matched with the cognition degree is obtained and exported.The technical scheme can be according to cognition degree of the user to the phonetic entry content generic of input, the voice output content matched with its cognition degree is selected to be exported for user, so that voice output content more conforms to the demand of user, provide the user the voice output function of personalization, improve the accuracy of voice output, allow users to get the information content of maximum from voice output content, improve the Experience Degree of user.
Description
The application is based on that the applying date is September in 2015 8, application No. is the application for a patent for invention propositions of CN201510568430.8, entitled " a kind of speech output method and device ", and the priority of the application for a patent for invention is required, the application is incorporated herein as reference in the full content of the application for a patent for invention.
The present invention relates to technical field of information processing more particularly to a kind of speech output methods and device.
Currently, voice input is increasingly praised highly by people with the development of electronics technology, voice input is that the Content Transformation that people is spoken by speech recognition is a kind of input mode of text.With intelligent terminal popularizing in people's lives, more and more intelligent terminals gradually have the function of voice service, such as, user can input proposition problem by voice, the voice that voice software on intelligent terminal passes through analysis user, the problem of user is equally answered in a manner of voice services to provide help for user.However, although this method is that user brings great convenience, answer is obtained by cumbersome online inquiry without user, but only has a kind of answering model in current voice service software, that is, different users puts question to identical problem (main contents of problem are identical), then exports identical help information.And the technical level or patent ability of different user be all it is different, for a user, it may be necessary to different help informations, or different answer-mode, therefore, the above method cannot be distinguished from different technical needs to provide voice help for user, not have specific aim.
Summary of the invention
The embodiment of the present invention provides a kind of speech output method and device.The technical solution is as follows:
In a first aspect, providing a kind of speech output method, comprising the following steps:
Receive the voice input content of user's input;
According to the voice input content, the user is determined to the cognition degree of the voice input content generic, the cognition degree is the professional knowledge degree of awareness of the user to the classification;
From at least one voice output content corresponding with the voice input content, obtains and export the voice output content to match with the cognition degree.
Some beneficial effects of the embodiment of the present invention may include:
Above-mentioned technical proposal, it can be according to user to the cognition degree of the voice input content generic of input, the voice output content to match for user's selection with its cognition degree exports, so that voice output content is more in line with the demand of user, to provide more personalized voice output function for user, the accuracy for improving voice output simultaneously, allows users to get maximum information content from voice output content, improves the Experience Degree of user.
In one embodiment, described according to the voice input content, determine the user to the cognition degree of the voice input content generic, comprising:
Identify the voiceprint of the user;
According to the voiceprint, the voice input content for receiving the user for the first time is judged whether it is;
When the voice input content to receive the user for the first time, determine that the user is to preset minimum cognition degree to the cognition degree of the voice input content generic.
In the embodiment, according to whether being that the voice output content that user's selection matches exports to receive the voice input content of user for the first time, so that voice output content is more in line with the demand of user, to provide more personalized voice output function for user, the accuracy of voice output is improved simultaneously, it allows users to get maximum information content from voice output content, improves the Experience Degree of user.
In one embodiment, the method also includes:
Record the input time of the voice input content and using duration, when use a length of duration received between the voice input content and the output voice output content.
In the embodiment, by recording the input time of voice input content and using duration, so that when the subsequent output voice output content for user, determine that the foundation of user cognition degree is more abundant, to more accurately determine the cognition degree of user, and then the voice output content more accurate, personalized for user's output.
In one embodiment, described according to the voice input content, determine the user to the cognition degree of the voice input content generic, comprising:
Identify the voiceprint of the user;
According to the voiceprint of the user, judge whether the adjacent voice input content received twice is inputted by same user;
When the adjacent voice input content received twice is inputted by same user, according to the input time of the adjacent voice input content received twice and duration is used, calculates the time interval between the adjacent voice input content received twice;
According to the time interval, determine the user to the cognition degree of the voice input content generic;Wherein, the time interval is longer, and the cognition degree is lower.
In the embodiment, the time interval between voice input content by calculating the adjacent same user received twice, so that determining that the foundation of user cognition degree is more abundant, to more accurately determine the cognition degree of user, and then the voice output content more accurate, personalized for user's output.
In one embodiment, described according to the voice input content, determine the user to the cognition degree of the voice input content generic, comprising:
Identify the voiceprint of the user;
According to the voiceprint of the user, history input record information corresponding with the user is obtained, the history input record information includes that history is accumulative using at least one of time, the accumulative input number of history and history input frequency information;
According to the history input record information, determine the user to the cognition degree of the voice input content generic;Wherein, the history is accumulative longer using the time, and the cognition degree is higher;The accumulative input number of the history is more, and the cognition degree is higher;The history input frequency is higher, and the cognition degree is higher.
In the embodiment, the cognition degree of user is determined according to the corresponding history input record information of user, enables the terminal to the cognition degree for more accurately determining user, and then the voice output content more accurate, personalized for user's output.
In one embodiment, described according to the voice input content, determine the user to the cognition degree of the voice input content generic, comprising:
Extract the keyword in the voice input content;
Determine the matching degree of the keyword and predetermined keyword in the voice input content;
According to the matching degree of keyword and predetermined keyword in the voice input content, determine the user to the cognition degree of the voice input content generic;Wherein, the matching degree of the professional keyword in the keyword and predetermined keyword in the voice input content is higher, and the cognition degree is higher;The matching degree of the amateur keyword in keyword and predetermined keyword in the voice input content is higher, and the cognition degree is lower.
In the embodiment, the cognition degree of user is determined according to the matching degree of keyword and predetermined keyword in voice input content, so that the determination of user cognition degree is more accurate, personalized, thus the voice output content more accurate, personalized for user's output.
In one embodiment, described according to the voice input content, determine the user to the cognition degree of the voice input content generic, comprising:
Determine that the sentence structure type of the voice input content, the sentence structure type include professional sentence structure type or amateur sentence structure type;
According to the sentence structure type of the voice input content, determine the user to the cognition degree of the voice input content generic;Wherein, the user is higher than the cognition degree to the voice input content generic of the amateur sentence structure type to the cognition degree of the voice input content generic of the professional sentence structure type.
In the embodiment, the cognition degree of user is determined according to the sentence structure type of voice input content, so that the determination of user cognition degree is more accurate, personalized, thus the voice output content more accurate, personalized for user's output.
In one embodiment, described according to the voice input content, determine the user to the cognition degree of the voice input content generic, comprising:
When determining that the adjacent voice input content received twice is inputted by same user, according to the keyword in the adjacent voice input content received twice, the degree of association between the adjacent voice input content received twice is determined;
According to the degree of association between the adjacent voice input content received twice, determine the user to the cognition degree of the voice input content generic;Wherein, the degree of association is higher, and the cognition degree is lower.
In the embodiment, the cognition degree of user is determined according to the degree of association between the voice input content of the adjacent same user received twice, so that the determination of user cognition degree is more accurate, personalized, thus the voice output content more accurate, personalized for user's output.
In one embodiment, described according to the voice input content, determine the user to the cognition degree of the voice input content generic, comprising:
According to the voice input content, determine at least two voices input parameters of the voice input content, the voice input parameter include: the voiceprint of the user, same user the adjacent voice input content inputted twice between time interval, history input record information corresponding with the user, the matching degree of the keyword in the voice input content and predetermined keyword, the pass between the sentence structure type of the voice input content and the adjacent voice input content inputted twice of same user
Connection degree;
The weight that parameter is inputted according to preset each single item voice, calculates the user to the cognition degree of the voice input content generic.
In the embodiment, the different weights of parameter are inputted according to the voice of multinomial voice input content, user is calculated to the cognition degree of voice input content generic, so that the determination of user cognition degree is more accurate, personalized, thus the voice output content more accurate, personalized for user's output.
In one embodiment, described according to the voice input content, determine the user to the cognition degree of the voice input content generic, comprising:
When the voice that can not determine the voice input content inputs parameter, determine that the user is to preset minimum cognition degree to the cognition degree of the voice input content generic.
In the embodiment, for that can not determine the voice input content of voice input parameter, the voice output content that output matches with the voice input content, to provide more accurate and personalized voice output function for user, it allows users to get more useful information from voice output content, improves the Experience Degree of user.
In one embodiment, described from least one voice output content corresponding with the voice input content, it obtains and exports the voice output content to match with the cognition degree, comprising:
According to the corresponding relationship between cognition degree and cognition grade, the corresponding cognition grade of the cognition degree is determined;
According to the corresponding relationship between cognition grade and voice output content, voice output content corresponding with the cognition grade is obtained;
Export the voice output content.
In the embodiment, matched voice output content is selected to export for user according to the corresponding relationship between cognition grade and voice output content, it is exported to select the voice output content to match with user cognition degree for user, so that voice output content is more in line with the demand of user, improve the accuracy of voice output, it allows users to get maximum information content from voice output content, improves the Experience Degree of user.
In one embodiment, the method also includes:
According to the input time of the voice input content and duration is used, updates the history input record information.
In the embodiment, by the update to history input record information, when so that exporting voice output content again for user, the cognition degree of user can be determined according to accurate history input record, to export more accurate voice output content for user.
In one embodiment, the method also includes:
The user is stored to the cognition degree of the voice input content generic;
It is described according to the voice input content, determine the user to the cognition degree of the voice input content generic, comprising:
Identify the voiceprint of the user;
The user is inquired to the cognition degree of the voice input content generic according to the voiceprint of the user.
In the embodiment, by inquiring the cognition degree of user, user more convenient can be quickly determined out to voice input content
The cognition degree of generic, so that the voice output content more quickly and accurately to match for user's selection exports.
Second aspect provides a kind of instantaneous speech power, comprising:
Receiving module, for receiving the voice input content of user's input;
Determining module, for determining the user to the cognition degree of the voice input content generic, the cognition degree is the professional knowledge degree of awareness of the user to the classification according to the voice input content;
Output module, for obtaining and exporting the voice output content to match with the cognition degree from least one voice output content corresponding with the voice input content.
In one embodiment, the determining module includes:
First identifies submodule, for identification the voiceprint of the user;
First judging submodule, for judging whether it is the voice input content for receiving the user for the first time according to the voiceprint;
Second determines submodule, is to preset minimum cognition degree to the cognition degree of the voice input content generic for when the voice input content to receive the user for the first time, determining the user.
In one embodiment, described device further include:
Logging modle, for recording the input time of the voice input content and using duration, when use a length of duration received between the voice input content and the output voice output content.
In one embodiment, the determining module includes:
Second identifies submodule, for identification the voiceprint of the user;
Second judgment submodule judges whether the adjacent voice input content received twice is inputted by same user for the voiceprint according to the user;
First computational submodule, for when the adjacent voice input content received twice is inputted by same user, according to the input time of the adjacent voice input content received twice and duration is used, calculates the time interval between the adjacent voice input content received twice;
Third determines submodule, for determining the user to the cognition degree of the voice input content generic according to the time interval;Wherein, the time interval is longer, and the cognition degree is lower.
In one embodiment, the determining module includes:
Third identifies submodule, for identification the voiceprint of the user;
First acquisition submodule, for the voiceprint according to the user, history input record information corresponding with the user is obtained, the history input record information includes that history is accumulative using at least one of time, the accumulative input number of history and history input frequency information;
4th determines submodule, for determining the user to the cognition degree of the voice input content generic according to the history input record information;Wherein, the history is accumulative longer using the time, and the cognition degree is higher;The accumulative input number of the history is more, and the cognition degree is higher;The history input frequency is higher, and the cognition degree is higher.
In one embodiment, the determining module includes:
Extracting sub-module, for extracting the keyword in the voice input content;
5th determines submodule, for determining the matching degree of keyword and predetermined keyword in the voice input content;
6th determines that submodule determines the user to the cognition degree of the voice input content generic for the matching degree according to keyword and predetermined keyword in the voice input content;Wherein, the matching degree of the professional keyword in the keyword and predetermined keyword in the voice input content is higher, and the cognition degree is higher;The matching degree of the amateur keyword in keyword and predetermined keyword in the voice input content is higher, and the cognition degree is lower.
In one embodiment, the determining module includes:
7th determines submodule, for determining that the sentence structure type of the voice input content, the sentence structure type include professional sentence structure type or amateur sentence structure type;
8th determines that submodule determines the user to the cognition degree of the voice input content generic for the sentence structure type according to the voice input content;Wherein, the user is higher than the cognition degree to the voice input content generic of the amateur sentence structure type to the cognition degree of the voice input content generic of the professional sentence structure type.
In one embodiment, the determining module includes:
9th determines submodule, for when determining that the adjacent voice input content received twice is inputted by same user, according to the keyword in the adjacent voice input content received twice, the degree of association between the adjacent voice input content received twice is determined;
Tenth determines submodule, for determining the user to the cognition degree of the voice input content generic according to the degree of association between the adjacent voice input content received twice;Wherein, the degree of association is higher, and the cognition degree is lower.
In one embodiment, the determining module includes:
11st determines submodule, for according to the voice input content, determine at least two voices input parameters of the voice input content, the voice input parameter include: the voiceprint of the user, same user the adjacent voice input content inputted twice between time interval, history input record information corresponding with the user, the matching degree of the keyword in the voice input content and predetermined keyword, the degree of association between the sentence structure type of the voice input content and the adjacent voice input content inputted twice of same user;
Computational submodule calculates the user to the cognition degree of the voice input content generic for inputting the weight of parameter according to preset each single item voice.
In one embodiment, the determining module includes:
12nd determines submodule, for when the voice that can not determine the voice input content inputs parameter, determining that the user is to preset minimum cognition degree to the cognition degree of the voice input content generic.
In one embodiment, the output module includes:
13rd determines submodule, for determining the corresponding cognition grade of the cognition degree according to the corresponding relationship between cognition degree and cognition grade;
Second acquisition submodule, for obtaining voice output content corresponding with the cognition grade according to the corresponding relationship between cognition grade and voice output content;
Output sub-module, for exporting the voice output content.
In one embodiment, described device further include:
Update module updates the history input record information for the input time according to the voice input content and using duration.
In one embodiment, described device further include:
Memory module, for storing the user to the cognition degree of the voice input content generic;
The determining module includes:
4th identifies submodule, for identification the voiceprint of the user;
Submodule is inquired, for inquiring the user to the cognition degree of the voice input content generic according to the voiceprint of the user.
Some beneficial effects of the embodiment of the present invention may include:
Above-mentioned apparatus, it can be according to user to the cognition degree of the voice input content generic of input, the voice output content to match for user's selection with its cognition degree exports, so that voice output content is more in line with the demand of user, to provide more personalized voice output function for user, the accuracy for improving voice output simultaneously, allows users to get maximum information content from voice output content, improves the Experience Degree of user.
The third aspect provides a kind of instantaneous speech power, which is characterized in that described device includes:
Processor;
For storing the memory of the processor-executable instruction;
Wherein, the processor is configured to:
Receive the voice input content of user's input;
According to the voice input content, the user is determined to the cognition degree of the voice input content generic, the cognition degree is the professional knowledge degree of awareness of the user to the classification;
From at least one voice output content corresponding with the voice input content, obtains and export the voice output content to match with the cognition degree.
Above-mentioned processor is also configured to
Identify the voiceprint of the user;
According to the voiceprint, the voice input content for receiving the user for the first time is judged whether it is;
When the voice input content to receive the user for the first time, determine that the user is to preset minimum cognition degree to the cognition degree of the voice input content generic.
Above-mentioned processor is also configured to
Record the input time of the voice input content and using duration, when use a length of duration received between the voice input content and the output voice output content.
Above-mentioned processor is also configured to
Identify the voiceprint of the user;
According to the voiceprint of the user, judge whether the adjacent voice input content received twice is inputted by same user;
When the adjacent voice input content received twice is inputted by same user, according to the adjacent language received twice
The input time of sound input content and duration is used, calculates the time interval between the adjacent voice input content received twice;
According to the time interval, determine the user to the cognition degree of the voice input content generic;Wherein, the time interval is longer, and the cognition degree is lower.
Above-mentioned processor is also configured to
Identify the voiceprint of the user;
According to the voiceprint of the user, history input record information corresponding with the user is obtained, the history input record information includes that history is accumulative using at least one of time, the accumulative input number of history and history input frequency information;
According to the history input record information, determine the user to the cognition degree of the voice input content generic;Wherein, the history is accumulative longer using the time, and the cognition degree is higher;The accumulative input number of the history is more, and the cognition degree is higher;The history input frequency is higher, and the cognition degree is higher.
Above-mentioned processor is also configured to
Extract the keyword in the voice input content;
Determine the matching degree of the keyword and predetermined keyword in the voice input content;
According to the matching degree of keyword and predetermined keyword in the voice input content, determine the user to the cognition degree of the voice input content generic;Wherein, the matching degree of the professional keyword in the keyword and predetermined keyword in the voice input content is higher, and the cognition degree is higher;The matching degree of the amateur keyword in keyword and predetermined keyword in the voice input content is higher, and the cognition degree is lower.
Above-mentioned processor is also configured to
Determine that the sentence structure type of the voice input content, the sentence structure type include professional sentence structure type or amateur sentence structure type;
According to the sentence structure type of the voice input content, determine the user to the cognition degree of the voice input content generic;Wherein, the user is higher than the cognition degree to the voice input content generic of the amateur sentence structure type to the cognition degree of the voice input content generic of the professional sentence structure type.
Above-mentioned processor is also configured to
When determining that the adjacent voice input content received twice is inputted by same user, according to the keyword in the adjacent voice input content received twice, the degree of association between the adjacent voice input content received twice is determined;
According to the degree of association between the adjacent voice input content received twice, determine the user to the cognition degree of the voice input content generic;Wherein, the degree of association is higher, and the cognition degree is lower.
Above-mentioned processor is also configured to
According to the voice input content, determine at least two voices input parameters of the voice input content, the voice input parameter include: the voiceprint of the user, same user the adjacent voice input content inputted twice between time interval, history input record information corresponding with the user, the matching degree of the keyword in the voice input content and predetermined keyword, the degree of association between the sentence structure type of the voice input content and the adjacent voice input content inputted twice of same user;
The weight that parameter is inputted according to preset each single item voice, calculates the user to the voice input content generic
Cognition degree.
Above-mentioned processor is also configured to
When the voice that can not determine the voice input content inputs parameter, determine that the user is to preset minimum cognition degree to the cognition degree of the voice input content generic.
Above-mentioned processor is also configured to
According to the corresponding relationship between cognition degree and cognition grade, the corresponding cognition grade of the cognition degree is determined;
According to the corresponding relationship between cognition grade and voice output content, voice output content corresponding with the cognition grade is obtained;
Export the voice output content.
Above-mentioned processor is also configured to
According to the input time of the voice input content and duration is used, updates the history input record information.
Above-mentioned processor is also configured to
The user is stored to the cognition degree of the voice input content generic;
It is described according to the voice input content, determine the user to the cognition degree of the voice input content generic, comprising:
Identify the voiceprint of the user;
The user is inquired to the cognition degree of the voice input content generic according to the voiceprint of the user.
Fourth aspect provides a kind of non-transitory computer readable recording medium, and record has computer program on the medium, and described program includes the instruction for executing the method as described in the first aspect of the embodiment of the present invention.
5th aspect, provides a kind of computer program, and described program includes: the instruction for executing the method as described in the first aspect of the embodiment of the present invention when described program is computer-executed.
Other features and advantages of the present invention will be illustrated in the following description, also, partly as will become apparent from the description, or understand through the implementation of the invention.The objectives and other advantages of the invention can be achieved and obtained by structure specifically indicated in the written description, claims, and drawings.
Below by drawings and examples, technical scheme of the present invention will be described in further detail.
Attached drawing is used to provide further understanding of the present invention, and constitutes part of specification, is used to explain the present invention together with embodiments of the present invention, is not construed as limiting the invention.In the accompanying drawings:
Fig. 1 is a kind of flow chart of speech output method in the embodiment of the present invention;
The flow chart that Fig. 2 is step S12 in a kind of speech output method in the embodiment of the present invention;
The flow chart that Fig. 3 is step S12 in a kind of speech output method in the embodiment of the present invention;
The flow chart that Fig. 4 is step S12 in a kind of speech output method in the embodiment of the present invention;
The flow chart that Fig. 5 is step S12 in a kind of speech output method in the embodiment of the present invention;
The flow chart that Fig. 6 is step S13 in a kind of speech output method in the embodiment of the present invention;
Fig. 7 is a kind of block diagram of instantaneous speech power in the embodiment of the present invention;
Fig. 8 is a kind of block diagram of determining module in instantaneous speech power in the embodiment of the present invention;
Fig. 9 is a kind of block diagram of determining module in instantaneous speech power in the embodiment of the present invention;
Figure 10 is a kind of block diagram of determining module in instantaneous speech power in the embodiment of the present invention;
Figure 11 is a kind of block diagram of determining module in instantaneous speech power in the embodiment of the present invention;
Figure 12 is a kind of block diagram of output module in instantaneous speech power in the embodiment of the present invention;
Figure 13 is a kind of block diagram of instantaneous speech power in the embodiment of the present invention;
Figure 14 is a kind of block diagram of the device of executable speech output method in the embodiment of the present invention.
Hereinafter, preferred embodiments of the present invention will be described with reference to the accompanying drawings, it should be understood that preferred embodiments described herein are only used to illustrate and explain the present invention, is not intended to limit the present invention.
Fig. 1 is a kind of flow chart of speech output method in the embodiment of the present invention.As shown in Figure 1, this method is in terminal, terminal can be mobile phone, computer, digital broadcasting terminal, messaging device, game console, tablet device, Medical Devices, body-building equipment, and personal digital assistant etc. includes the following steps S11-S13:
Step S11 receives the voice input content of user's input.
In the step, user can input voice input content by way of typing sound.
Step S12 determines user to the cognition degree of voice input content generic according to voice input content;The cognition degree is the professional knowledge degree of awareness of the user to voice input content generic.
For example, user inputs voice input content " how air-conditioner temperature being arranged ", then user is the professional knowledge degree of awareness of the user to air-conditioning class to the cognition degree of voice input content generic;User inputs voice input content " what medicine aspirin is ", then user is the professional knowledge degree of awareness of the user to pharmaceutical to the cognition degree of voice input content generic.Terminal can be by extracting the keyword in voice input content, to determine classification belonging to voice input content.
Step S13 is obtained from least one voice output content corresponding with voice input content and is exported the voice output content to match with cognition degree.
Using technical solution provided in an embodiment of the present invention, it can be according to user to the cognition degree of the voice input content generic of input, the voice output content to match for user's selection with its cognition degree exports, so that voice output content is more in line with the demand of user, to provide more personalized voice output function for user, the accuracy for improving voice output simultaneously, allows users to get maximum information content from voice output content, improves the Experience Degree of user.
In step s 12, user can determine the cognition degree of voice input content generic in several ways.It can determine the voice input parameter of voice input content first according to voice input content, determine user to the cognition degree of voice input content generic further according to voice input parameter.Wherein, the difference of parameter is inputted according to voice, the method of determination of cognition degree is also different, voice input parameter may include the voiceprint of user, same user the adjacent voice input content inputted twice between time interval, history input record information corresponding to the user, the association between the keyword voice input content inputted twice adjacent with the matching degree of predetermined keyword, the sentence structure type of voice input content and same user in voice input content
Degree, etc..Illustrate the embodiment of step S12 below by way of different embodiments.
In one embodiment, as shown in Fig. 2, step S12 may be embodied as following steps S21-S23:
Step S21 identifies the voiceprint of user.
Step S22 judges whether it is the voice input content for receiving user for the first time according to voiceprint.
Step S23 determines that user is to preset minimum cognition degree to the cognition degree of voice input content generic when the voice input content to receive user for the first time.
In the present embodiment, the corresponding voiceprint of different user is stored in terminal, when user inputs voice input content, if terminal can inquire the voiceprint of the user in pre-stored voiceprint, illustrate not to be the voice input content for receiving the user for the first time, and if terminal fails to inquire the voiceprint of the user in pre-stored voiceprint, illustrate that terminal is the voice input content for receiving the user for the first time.When the voice input content not to receive user for the first time, then terminal continues to determine that other voices input parameter according to voice input content, and inputs parameter according to other voices and execute step S12.In the terminal, the corresponding relationship being previously stored between cognition degree and voice input content, wherein including voice input content corresponding with minimum cognition degree is preset.
In one embodiment, the above method is further comprising the steps of: record voice input content input time and use duration, this using when it is a length of receive voice input content and export voice output content between duration.Therefore, as shown in figure 3, step S12 may be embodied as following steps S31-S34:
Step S31 identifies the voiceprint of user.
Step S32 judges whether the adjacent voice input content received twice is inputted by same user according to the voiceprint of user.
Step S33 according to the input time of the adjacent voice input content received twice and uses duration, calculates the time interval between the adjacent voice input content received twice when the adjacent voice input content received twice is inputted by same user.
Step S34 determines user to the cognition degree of voice input content generic according to time interval;Wherein, time interval is longer, and cognition degree is lower.
In the present embodiment, when the adjacent voice input content received twice is inputted by same user, time interval between the so adjacent voice input content received twice can reflect out the reaction duration for the upper voice output content that user exports terminal, in addition, user can also receive the time interval voice input content to this by last time output voice output content to the reaction duration for the upper voice output content that terminal exports to characterize.For example, the voice input content that the terminal last time receives is " how air-conditioner temperature being arranged ", and for the voice input content, it is " being introduced into temperature shaping modes, then change temperature " that terminal, which exports corresponding voice output content,;This voice input content received of terminal is " how entering temperature shaping modes ", when terminal determines that the adjacent voice input content received twice is inputted by same user, then can be used receive voice input content " how air-conditioner temperature is set " and receive the time interval between voice input content " how to enter temperature shaping modes " come characterize user to a upper voice output content " be introduced into temperature shaping modes; then change temperature " reaction duration, and then determine user to the cognition degree of voice input content generic.Alternatively, can also be used output voice output content " being introduced into temperature shaping modes, then change temperature " and this receive voice input content " such as
What enters temperature shaping modes " between time interval characterize user to the reaction duration of a upper voice output content " being introduced into temperature shaping modes, then change temperature ", and then determine user to the cognition degree of voice input content generic.Time interval is longer, as soon as illustrate that user is longer to the reaction duration of upper voice output content, then cognition degree is lower.
Furthermore, a prefixed time interval can also be preset, when the adjacent voice input content received twice is inputted by same user and time interval between the adjacent voice input content received twice is more than prefixed time interval, it is to preset minimum cognition degree, and obtain and preset the voice output content that minimum cognition degree matches and exported to the cognition degree of voice input content generic that terminal, which can directly determine user,.
In one embodiment, as shown in figure 4, step S12 may be embodied as following steps S41-S43:
Step S41 identifies the voiceprint of user.
Step S42 obtains history input record information corresponding to the user according to the voiceprint of user;History input record information includes that history is accumulative using at least one of time, the accumulative input number of history and history input frequency information.
Step S43 determines user to the cognition degree of voice input content generic according to history input record information;Wherein, history is accumulative longer using the time, and cognition degree is higher;The accumulative input number of history is more, and cognition degree is higher;History input frequency is higher, and cognition degree is higher.
In the present embodiment, terminal receives the voice input content of user's input every time, just will record the input time of voice input content and using duration, this using when a length of receive voice input content and export the duration between voice output content.Terminal can count history input record information corresponding to the user according to the input time of record and using duration, wherein the accumulative summation using duration as recorded each time using the time of history.In addition, the above method is further comprising the steps of: according to the input time of voice input content and using duration, more new historical input record information.In this way, when terminal determines cognition degree of the user to voice input content generic according to history input record information corresponding to the user, based on history input record information it is more abundant accurate, so as to be exported for more accurate and personalization the voice output content of user's selection.
In one embodiment, as shown in figure 5, step S12 may be embodied as following steps S51-S53:
Step S51 extracts the keyword in voice input content.
Step S52 determines the matching degree of the keyword and predetermined keyword in voice input content.
Step S53 determines user to the cognition degree of voice input content generic according to the matching degree of keyword and predetermined keyword in voice input content;Wherein, the matching degree of the professional keyword in the keyword and predetermined keyword in voice input content is higher, and cognition degree is higher;The matching degree of amateur keyword in keyword and predetermined keyword in voice input content is higher, and cognition degree is lower.
In the present embodiment, the predetermined keyword prestored in terminal includes two types of professional keyword and amateur keyword, when executing step S52, need to determine the matching degree between the keyword and professional keyword in voice input content, and the matching degree between amateur keyword respectively.Such as, professional keyword includes " setting path ", amateur keyword includes " how using ", if the voice input content that terminal receives is " ... setting path ", it so can determine that the matching degree between the keyword and professional keyword in voice input content is higher, therefore user is also higher to the cognition degree of voice input content generic;If the voice input content that terminal receives is " ... how to use ", it can determine that the matching degree between the keyword and amateur keyword in voice input content is higher, therefore user also gets over the cognition degree of voice input content generic
It is low.
In one embodiment, step S12 may be embodied as following steps A1-A2:
Step A1 determines that the sentence structure type of voice input content, sentence structure type include professional sentence structure type or amateur sentence structure type.
Step A2 determines user to the cognition degree of voice input content generic according to the sentence structure type of voice input content;Wherein, user is higher than the cognition degree to the voice input content generic of amateur sentence structure type to the cognition degree of the voice input content generic of professional sentence structure type.
Sentence structure type is prestored in the present embodiment, in terminal, sentence structure type can be embodied by regular expression.Wherein, the regular expression of professional sentence structure type is such as: adjective+noun+verb;The regular expression of amateur sentence structure type is such as: pronoun+verb.It should be pointed out that the manifestation mode of sentence structure type is not limited to regular expression, can also be embodied in such a way that other can embody sentence structure.It is exemplified below, the voice input content that terminal receives is " what the step of booting is ", terminal is by analyzing the voice input content, the sentence structure type for determining the voice input content is " adjective+noun+verb+pronoun ", it can determine the sentence structure type of the voice input content so as professional sentence structure type, user is higher to the cognition degree of the voice input content generic.For another example, the voice input content that terminal receives is " how this thing is used ", terminal is by analyzing the voice input content, the sentence structure type for determining the voice input content is " pronoun+verb ", the sentence structure type that so can determine the voice input content is amateur sentence structure type, and user is lower to the cognition degree of the voice input content generic.
In one embodiment, step S12 may be embodied as following steps B1-B2:
Step B1, according to the keyword in the adjacent voice input content received twice, determines the degree of association between the adjacent voice input content received twice when determining that the adjacent voice input content received twice is inputted by same user.
Step B2 determines user to the cognition degree of voice input content generic according to the degree of association between the adjacent voice input content received twice;Wherein, the degree of association is higher, and cognition degree is lower.
In the present embodiment, when the adjacent voice input content received twice is inputted by same user, the degree of association between the so adjacent voice input content received twice can reflect out user to the degree of understanding of a upper voice output content, therefore the degree of association between the adjacent voice input content received twice is higher, as soon as illustrating that user is lower to the degree of understanding of upper voice output content, user is also lower to the cognition degree of voice input content generic;The degree of association between the adjacent voice input content received twice is lower, as soon as illustrating that user is higher to the degree of understanding of upper voice output content, user is also higher to the cognition degree of voice input content generic.Such as, the voice input content that the terminal last time receives is " how air-conditioner temperature being arranged ", this voice input content received of terminal is " how entering temperature shaping modes " simultaneously, when terminal determines that the adjacent voice input content received twice is inputted by same user, it can extract the keyword in the adjacent voice input content received twice, such as keyword " air-conditioner temperature ", " temperature shaping modes ", the degree of association between the adjacent voice input content received twice is determined by the degree of association between keyword " air-conditioner temperature " and keyword " temperature shaping modes ", since " air-conditioner temperature " and " temperature shaping modes " is all keyword related with temperature, therefore the degree of association between the two is higher.For another example, the voice input content that the terminal last time receives is " how air-conditioner temperature being arranged ", while terminal this voice input content for receiving is " what the step of booting is ", when terminal determines that the adjacent voice input content received twice is same
When user is inputted, extracting the keyword in two neighboring voice input content respectively is " air-conditioner temperature " and " booting ", due to the keyword that the two keywords are two unrelated types, therefore the degree of association between the two is almost nil, that is to say, the degree of association between the bright adjacent voice input content received twice is very low, user is higher to the degree of understanding of a upper voice output content, and it is higher to the cognition degree of voice input content generic to further relate to user.
In one embodiment, for the various ways of execution step S12 in above-described embodiment, also multinomial voice can be inputted parameter and combined, and user is calculated to the cognition degree of voice input content generic according to pre-set weight.Therefore, above-mentioned steps S12 can also be embodied as following steps: according to voice input content, determine at least two voice input parameters of voice input content, wherein, voice input parameter include: the voiceprint of user, same user the adjacent voice input content inputted twice between time interval, history input record information corresponding to the user, the degree of association between the keyword voice input content inputted twice adjacent with the matching degree of predetermined keyword, the sentence structure type of voice input content and same user in voice input content;The weight of parameter is inputted according to preset each single item voice, calculates user to the cognition degree of voice input content generic.
In one embodiment, the above method is further comprising the steps of: when the voice that can not determine voice input content inputs parameter, determining that user is to preset minimum cognition degree to the cognition degree of voice input content generic.In the present embodiment, when receiving the voice input content of user's input, for that can not determine the voice input content of voice input content, it is to preset minimum cognition degree to the cognition degree of the voice input content generic that terminal, which can directly determine user, therefore, even the voice input content of voice input parameter can not be determined, user can also get matched voice output content, to improve the Experience Degree of user.
In one embodiment, the above method is further comprising the steps of: cognition degree of the storage user to voice input content generic.At this point, step S12 may be embodied as following steps: identifying the voiceprint of user;User is inquired to the cognition degree of voice input content generic according to the voiceprint of user.In the present embodiment, by inquiring the cognition degree of user, user more convenient can be quickly determined out to the cognition degree of voice input content generic, so that the voice output content more quickly and accurately to match for user's selection exports.
In one embodiment, as shown in fig. 6, step S13 is implementable for following steps S61-S63:
Step S61 determines the corresponding cognition grade of cognition degree according to the corresponding relationship between cognition degree and cognition grade.
Step S62 obtains voice output content corresponding with cognition grade according to the corresponding relationship between cognition grade and voice output content.
Step S63 exports voice output content.
In the present embodiment, terminal prestores cognition degree and recognizes the corresponding relationship between grade, and the corresponding relationship between cognition grade and voice output content, such as, cognition grade can be divided into low cognition grade, middle cognition grade, high cognition grade three grades as needed, the low cognition grade of correspondence of the cognition degree between " 0%~30% ", cognition degree recognize grade, corresponding high cognition grade of the cognition degree between " 71%~100% " in the correspondence between " 31%~70% ".Voice output content corresponding with low cognition grade is detailed version voice output content, voice output content corresponding with middle cognition grade is standard edition voice output content, voice output content corresponding with height cognition grade is succinct version voice output content, for each voice input content, terminal can all store three kinds of corresponding detailed version, succinct version, standard edition voice output contents.For example, for voice
How input content " is arranged air-conditioner temperature ", voice output content corresponding thereto includes: detailed version " clicking the mode button in first row middle position; click and enter temperature shaping modes twice; click ' +/- ' the change temperature of second row left button; click primary, 1 degree of temperature ' +/- ' ";Standard edition " clicks mode button and enters temperature shaping modes, click ' +/- ' the change temperature of button ";Succinct version " being introduced into temperature shaping modes, then change temperature ".Furthermore, cognition grade corresponding with minimum cognition degree is preset can be low cognition grade, therefore, for that can not determine the voice input content of voice input parameter or for the first time receive the voice input content of user, terminal can directly export detailed version voice output content.It can be seen that, using the technical solution of the present embodiment, make terminal when exporting voice output content for user, it can be by determining that user analyzes the current demand of user to the cognition degree of voice input content generic, and matched voice output content is exported according to the current demand of user, it allows users to get more more accurate information from voice output content.
Corresponding to a kind of above-mentioned speech output method, the embodiment of the present invention also provides a kind of instantaneous speech power, and the device is to execute the above method.
Fig. 7 is a kind of block diagram of instantaneous speech power in the embodiment of the present invention.As shown in fig. 7, the device includes:
Receiving module 71, for receiving the voice input content of user's input;
Determining module 72, for determining user to the cognition degree of voice input content generic, cognition degree is the professional knowledge degree of awareness of the user to classification according to voice input content;
Output module 73, for obtaining and exporting the voice output content to match with cognition degree from least one voice output content corresponding with voice input content.
In one embodiment, as shown in figure 8, determining module 72 includes:
First identifies submodule 721, for identification the voiceprint of user;
First judging submodule 722, for according to voiceprint, judging whether it is the voice input content for receiving user for the first time;
Second determines submodule 723, for when the voice input content to receive user for the first time, determining that user is to preset minimum cognition degree to the cognition degree of voice input content generic.
In one embodiment, above-mentioned apparatus further include:
Logging modle, for recording the input time of voice input content and using duration, when use is a length of to be received voice input content and exports the duration between voice output content.
In one embodiment, as shown in figure 9, determining module 72 includes:
Second identifies submodule 724, for identification the voiceprint of user;
Second judgment submodule 725 judges whether the adjacent voice input content received twice is inputted by same user for the voiceprint according to user;
First computational submodule 726, for when the adjacent voice input content received twice is inputted by same user, according to the input time of the adjacent voice input content received twice and duration is used, calculates the time interval between the adjacent voice input content received twice;
Third determines submodule 727, for determining user to the cognition degree of voice input content generic according to time interval;Wherein, time interval is longer, and cognition degree is lower.
In one embodiment, as shown in Figure 10, determining module 72 includes:
Third identifies submodule 728, for identification the voiceprint of user;
First acquisition submodule 729 obtains history input record information corresponding to the user for the voiceprint according to user, and history input record information includes that history is accumulative using at least one of time, the accumulative input number of history and history input frequency information;
4th determines submodule 7210, for determining user to the cognition degree of voice input content generic according to history input record information;Wherein, history is accumulative longer using the time, and cognition degree is higher;The accumulative input number of history is more, and cognition degree is higher;History input frequency is higher, and cognition degree is higher.
In one embodiment, as shown in figure 11, determining module 72 includes:
Extracting sub-module 7211, for extracting the keyword in voice input content;
5th determines submodule 7212, for determining the matching degree of keyword and predetermined keyword in voice input content;
6th determines that submodule 7213 determines user to the cognition degree of voice input content generic for the matching degree according to keyword and predetermined keyword in voice input content;Wherein, the matching degree of the professional keyword in the keyword and predetermined keyword in voice input content is higher, and cognition degree is higher;The matching degree of amateur keyword in keyword and predetermined keyword in voice input content is higher, and cognition degree is lower.
In one embodiment, determining module 72 includes:
7th determines submodule, for determining that the sentence structure type of voice input content, sentence structure type include professional sentence structure type or amateur sentence structure type;
8th determines that submodule determines user to the cognition degree of voice input content generic for the sentence structure type according to voice input content;Wherein, user is higher than the cognition degree to the voice input content generic of amateur sentence structure type to the cognition degree of the voice input content generic of professional sentence structure type.
In one embodiment, determining module 72 includes:
9th determines submodule, for according to the keyword in the adjacent voice input content received twice, determining the degree of association between the adjacent voice input content received twice when determining that the adjacent voice input content received twice is inputted by same user;
Tenth determines submodule, for determining user to the cognition degree of voice input content generic according to the degree of association between the adjacent voice input content received twice;Wherein, the degree of association is higher, and cognition degree is lower.
In one embodiment, determining module 72 includes:
11st determines submodule, for according to voice input content, determine at least two voices input parameters of voice input content, voice input parameter include: the voiceprint of user, same user the adjacent voice input content inputted twice between time interval, history input record information corresponding to the user, the degree of association between the keyword voice input content inputted twice adjacent with the matching degree of predetermined keyword, the sentence structure type of voice input content and same user in voice input content;
Computational submodule calculates user to the cognition degree of voice input content generic for inputting the weight of parameter according to preset each single item voice.
In one embodiment, determining module 72 includes:
12nd determines submodule, for determining user to voice when the voice that can not determine voice input content inputs parameter
The cognition degree of input content generic is to preset minimum cognition degree.
In one embodiment, as shown in figure 12, output module 73 includes:
13rd determines submodule 731, for determining the corresponding cognition grade of cognition degree according to the corresponding relationship between cognition degree and cognition grade;
Second acquisition submodule 732, for obtaining voice output content corresponding with cognition grade according to the corresponding relationship between cognition grade and voice output content;
Output sub-module 733, for exporting voice output content.
In one embodiment, as shown in figure 13, above-mentioned apparatus further include:
Update module 74, for according to input time of voice input content and using duration, more new historical input record information.
Memory module 75, for storing user to the cognition degree of voice input content generic.
In one embodiment, determining module 72 includes:
4th identifies submodule, for identification the voiceprint of user;
Submodule is inquired, for inquiring user to the cognition degree of voice input content generic according to the voiceprint of user.
Using device provided in an embodiment of the present invention, it can be according to user to the cognition degree of the voice input content generic of input, the voice output content to match for user's selection with its cognition degree exports, so that voice output content is more in line with the demand of user, to provide more personalized voice output function for user, the accuracy for improving voice output simultaneously, allows users to get maximum information content from voice output content, improves the Experience Degree of user.
Figure 14 is a kind of block diagram of the device of executable speech output method shown according to an exemplary embodiment.For example, device 1600 can be mobile phone, computer, digital broadcasting terminal, messaging device, game console, tablet device, Medical Devices, body-building equipment, personal digital assistant etc..
Referring to Fig.1 4, device 1600 may include following one or more components: processor 1601, memory 1602 and communication component 1603.
The integrated operation of the usual control device 1600 of processor 1601 such as operates associated operation with display, telephone call, data communication, camera operation and record.Processor 1601 can execute instruction, to perform all or part of the steps of the methods described above.
Memory 1602 is configured as storing various types of data to support the operation in device 1600.The example of these data includes the instruction of any application or method for operating on device 1600, contact data, telephone book data, message, picture, video etc..Memory 1602 can be realized by any kind of volatibility or non-volatile memory device or their combination, such as static random access memory (SRAM), electrically erasable programmable read-only memory (EEPROM), Erasable Programmable Read Only Memory EPROM (EPROM), programmable read only memory (PROM), read-only memory (ROM), magnetic memory, flash memory, disk or CD.
Communication component 1603 is configured to facilitate the communication of wired or wireless way between device 1600 and other equipment.Device 1600 can access the wireless network based on communication standard, such as Wi-Fi, 2G or 3G or their combination.In one exemplary embodiment, communication component 1603 receives broadcast singal or broadcast related information from external broadcasting management system via broadcast channel.In one exemplary embodiment, communication component 1603 further includes near-field communication (NFC) module, to promote short distance
Communication.For example, can be realized based on radio frequency identification (RFID) technology, Infrared Data Association (IrDA) technology, ultra wide band (UWB) technology, bluetooth (BT) technology and other technologies in NFC module.
In the exemplary embodiment, device 1600 can be realized by one or more application specific integrated circuit (ASIC), digital signal processor (DSP), digital signal processing appts (DSPD), programmable logic device (PLD), field programmable gate array (FPGA), controller, microcontroller, microprocessor or other electronic components, for executing above-mentioned speech output method.
In the exemplary embodiment, a kind of non-transitorycomputer readable storage medium including instruction, the memory 1602 for example including instruction are additionally provided, above-metioned instruction can be executed by the processor 1601 of device 1600 to complete above-mentioned speech output method.For example, non-transitorycomputer readable storage medium can be ROM, random access memory (RAM), CD-ROM, tape, floppy disk and optical data storage devices etc..
The present invention also provides a kind of non-transitory computer readable recording medium, record has computer program on the medium, and described program includes the instruction for executing speech output method as described in the above embodiment the present invention.
The present invention also provides a kind of computer program, described program includes: the instruction for executing speech output method as described in the above embodiment the present invention when described program is computer-executed.
It should be understood by those skilled in the art that, the embodiment of the present invention can provide as method, system or computer program product.Therefore, the form of complete hardware embodiment, complete software embodiment or embodiment combining software and hardware aspects can be used in the present invention.Moreover, the form for the computer program product implemented in the computer-usable storage medium (including but not limited to magnetic disk storage and optical memory etc.) that one or more wherein includes computer usable program code can be used in the present invention.
The present invention be referring to according to the method for the embodiment of the present invention, the flowchart and/or the block diagram of equipment (system) and computer program product describes.It should be understood that the combination of process and/or box in each flow and/or block and flowchart and/or the block diagram that can be realized by computer program instructions in flowchart and/or the block diagram.These computer program instructions be can provide to the processor of general purpose computer, special purpose computer, Embedded Processor or other programmable data processing devices to generate a machine, so that generating by the instruction that computer or the processor of other programmable data processing devices execute for realizing the device for the function of specifying in one or more flows of the flowchart and/or one or more blocks of the block diagram.
These computer program instructions, which may also be stored in, to be able to guide in computer or other programmable data processing devices computer-readable memory operate in a specific manner, so that instruction stored in the computer readable memory generates the manufacture including command device, which realizes the function of specifying in one or more flows of the flowchart and/or one or more blocks of the block diagram.
These computer program instructions also can be loaded onto a computer or other programmable data processing device, so that series of operation steps are executed on a computer or other programmable device to generate computer implemented processing, thus the step of instruction executed on a computer or other programmable device is provided for realizing the function of specifying in one or more flows of the flowchart and/or one or more blocks of the block diagram.
Obviously, those skilled in the art various changes and modifications can be made to the invention without departing from the spirit and scope of the present invention.If then the present invention is also intended to include these modifications and variations in this way, these modifications and changes of the present invention is within the scope of the claims of the present invention and its equivalent technology.
Claims (20)
- A kind of speech output method characterized by comprisingReceive the voice input content of user's input;According to the voice input content, the user is determined to the cognition degree of the voice input content generic, the cognition degree is the professional knowledge degree of awareness of the user to the classification;From at least one voice output content corresponding with the voice input content, obtains and export the voice output content to match with the cognition degree.
- The method according to claim 1, wherein it is described according to the voice input content, determine the user to the cognition degree of the voice input content generic, comprising:Identify the voiceprint of the user;According to the voiceprint, the voice input content for receiving the user for the first time is judged whether it is;When the voice input content to receive the user for the first time, determine that the user is to preset minimum cognition degree to the cognition degree of the voice input content generic.
- The method according to claim 1, wherein the method also includes:Record the input time of the voice input content and using duration, when use a length of duration received between the voice input content and the output voice output content.
- According to the method described in claim 3, determining the user to the cognition degree of the voice input content generic it is characterized in that, described according to the voice input content, comprising:Identify the voiceprint of the user;According to the voiceprint of the user, judge whether the adjacent voice input content received twice is inputted by same user;When the adjacent voice input content received twice is inputted by same user, according to the input time of the adjacent voice input content received twice and duration is used, calculates the time interval between the adjacent voice input content received twice;According to the time interval, determine the user to the cognition degree of the voice input content generic;Wherein, the time interval is longer, and the cognition degree is lower.
- According to the method described in claim 3, determining the user to the cognition degree of the voice input content generic it is characterized in that, described according to the voice input content, comprising:Identify the voiceprint of the user;According to the voiceprint of the user, history input record information corresponding with the user is obtained, the history input record information includes that history is accumulative using at least one of time, the accumulative input number of history and history input frequency information;According to the history input record information, determine the user to the cognition degree of the voice input content generic;Wherein, the history is accumulative longer using the time, and the cognition degree is higher;The accumulative input number of the history is more, and the cognition degree is higher;The history input frequency is higher, and the cognition degree is higher.
- The method according to claim 1, wherein it is described according to the voice input content, determine the user to the cognition degree of the voice input content generic, comprising:Extract the keyword in the voice input content;Determine the matching degree of the keyword and predetermined keyword in the voice input content;According to the matching degree of keyword and predetermined keyword in the voice input content, determine the user to the cognition degree of the voice input content generic;Wherein, the matching degree of the professional keyword in the keyword and predetermined keyword in the voice input content is higher, and the cognition degree is higher;The matching degree of the amateur keyword in keyword and predetermined keyword in the voice input content is higher, and the cognition degree is lower.
- The method according to claim 1, wherein it is described according to the voice input content, determine the user to the cognition degree of the voice input content generic, comprising:Determine that the sentence structure type of the voice input content, the sentence structure type include professional sentence structure type or amateur sentence structure type;According to the sentence structure type of the voice input content, determine the user to the cognition degree of the voice input content generic;Wherein, the user is higher than the cognition degree to the voice input content generic of the amateur sentence structure type to the cognition degree of the voice input content generic of the professional sentence structure type.
- The method according to claim 1, wherein it is described according to the voice input content, determine the user to the cognition degree of the voice input content generic, comprising:When determining that the adjacent voice input content received twice is inputted by same user, according to the keyword in the adjacent voice input content received twice, the degree of association between the adjacent voice input content received twice is determined;According to the degree of association between the adjacent voice input content received twice, determine the user to the cognition degree of the voice input content generic;Wherein, the degree of association is higher, and the cognition degree is lower.
- The method according to claim 1, wherein it is described according to the voice input content, determine the user to the cognition degree of the voice input content generic, comprising:According to the voice input content, determine at least two voices input parameters of the voice input content, the voice input parameter include: the voiceprint of the user, same user the adjacent voice input content inputted twice between time interval, history input record information corresponding with the user, the matching degree of the keyword in the voice input content and predetermined keyword, the degree of association between the sentence structure type of the voice input content and the adjacent voice input content inputted twice of same user;The weight that parameter is inputted according to preset each single item voice, calculates the user to the cognition degree of the voice input content generic.
- According to the method described in claim 9, determining the user to the cognition degree of the voice input content generic it is characterized in that, described according to the voice input content, comprising:When the voice that can not determine the voice input content inputs parameter, determine that the user is to preset minimum cognition degree to the cognition degree of the voice input content generic.
- The method according to claim 1, wherein described from corresponding with the voice input content In at least one voice output content, obtains and exports the voice output content to match with the cognition degree, comprising:According to the corresponding relationship between cognition degree and cognition grade, the corresponding cognition grade of the cognition degree is determined;According to the corresponding relationship between cognition grade and voice output content, voice output content corresponding with the cognition grade is obtained;Export the voice output content.
- According to the method described in claim 5, it is characterized in that, the method also includes:According to the input time of the voice input content and duration is used, updates the history input record information.
- The method according to claim 1, wherein the method also includes:The user is stored to the cognition degree of the voice input content generic;It is described according to the voice input content, determine the user to the cognition degree of the voice input content generic, comprising:Identify the voiceprint of the user;The user is inquired to the cognition degree of the voice input content generic according to the voiceprint of the user.
- A kind of instantaneous speech power, which is characterized in that described device includes:Processor;For storing the memory of the processor-executable instruction;Wherein, the processor is configured to executing a kind of speech output method, it is described the described method includes:Receive the voice input content of user's input;According to the voice input content, the user is determined to the cognition degree of the voice input content generic, the cognition degree is the professional knowledge degree of awareness of the user to the classification;From at least one voice output content corresponding with the voice input content, obtains and export the voice output content to match with the cognition degree.
- Device according to claim 14, wherein the processor is also configured toIdentify the voiceprint of the user;According to the voiceprint, the voice input content for receiving the user for the first time is judged whether it is;When the voice input content to receive the user for the first time, determine that the user is to preset minimum cognition degree to the cognition degree of the voice input content generic.
- Device according to claim 14, wherein the processor is also configured toRecord the input time of the voice input content and using duration, when use a length of duration received between the voice input content and the output voice output content.
- Device according to claim 16, wherein the processor is also configured toIdentify the voiceprint of the user;According to the voiceprint of the user, judge whether the adjacent voice input content received twice is inputted by same user;When the adjacent voice input content received twice is inputted by same user, adjacent received twice according to described The input time of voice input content and duration is used, calculates the time interval between the adjacent voice input content received twice;According to the time interval, determine the user to the cognition degree of the voice input content generic;Wherein, the time interval is longer, and the cognition degree is lower.
- Device according to claim 16, wherein the processor is also configured toIdentify the voiceprint of the user;According to the voiceprint of the user, history input record information corresponding with the user is obtained, the history input record information includes that history is accumulative using at least one of time, the accumulative input number of history and history input frequency information;According to the history input record information, determine the user to the cognition degree of the voice input content generic;Wherein, the history is accumulative longer using the time, and the cognition degree is higher;The accumulative input number of the history is more, and the cognition degree is higher;The history input frequency is higher, and the cognition degree is higher.
- Device according to claim 14, wherein the processor is also configured toExtract the keyword in the voice input content;Determine the matching degree of the keyword and predetermined keyword in the voice input content;According to the matching degree of keyword and predetermined keyword in the voice input content, determine the user to the cognition degree of the voice input content generic;Wherein, the matching degree of the professional keyword in the keyword and predetermined keyword in the voice input content is higher, and the cognition degree is higher;The matching degree of the amateur keyword in keyword and predetermined keyword in the voice input content is higher, and the cognition degree is lower.
- A kind of non-transitory computer readable recording medium, record has computer program on the medium, and described program includes the instruction for executing a kind of speech output method, which comprises receives the voice input content of user's input;According to the voice input content, the user is determined to the cognition degree of the voice input content generic, the cognition degree is the professional knowledge degree of awareness of the user to the classification;From at least one voice output content corresponding with the voice input content, obtains and export the voice output content to match with the cognition degree.
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN2015105684308 | 2015-09-08 | ||
CN201510568430.8A CN105304082B (en) | 2015-09-08 | 2015-09-08 | A kind of speech output method and device |
PCT/CN2016/082427 WO2017041510A1 (en) | 2015-09-08 | 2016-05-18 | Voice output method and device |
Publications (2)
Publication Number | Publication Date |
---|---|
CN107077845A true CN107077845A (en) | 2017-08-18 |
CN107077845B CN107077845B (en) | 2020-07-17 |
Family
ID=55201255
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201510568430.8A Active CN105304082B (en) | 2015-09-08 | 2015-09-08 | A kind of speech output method and device |
CN201680002958.1A Active CN107077845B (en) | 2015-09-08 | 2016-05-18 | Voice output method and device |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201510568430.8A Active CN105304082B (en) | 2015-09-08 | 2015-09-08 | A kind of speech output method and device |
Country Status (2)
Country | Link |
---|---|
CN (2) | CN105304082B (en) |
WO (1) | WO2017041510A1 (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109035896A (en) * | 2018-08-13 | 2018-12-18 | 广东小天才科技有限公司 | Oral training method and learning equipment |
CN110018843A (en) * | 2018-01-09 | 2019-07-16 | 北京小度互娱科技有限公司 | The test method and device of application program operation reserve |
Families Citing this family (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105304082B (en) * | 2015-09-08 | 2018-12-28 | 北京云知声信息技术有限公司 | A kind of speech output method and device |
CN106251862A (en) * | 2016-07-19 | 2016-12-21 | 东莞市优陌儿智护电子科技有限公司 | The implementation method of complete semantic intelligence intercommunication and system thereof |
CN106649698B (en) * | 2016-12-19 | 2020-12-22 | 宇龙计算机通信科技(深圳)有限公司 | Information processing method and information processing device |
CN107767869B (en) * | 2017-09-26 | 2021-03-12 | 百度在线网络技术(北京)有限公司 | Method and apparatus for providing voice service |
CN107863108B (en) * | 2017-11-16 | 2021-03-23 | 百度在线网络技术(北京)有限公司 | Information output method and device |
CN110619870B (en) * | 2018-06-04 | 2022-05-06 | 佛山市顺德区美的电热电器制造有限公司 | Man-machine conversation method and device, household appliance and computer storage medium |
CN109036386B (en) * | 2018-09-14 | 2021-03-16 | 北京网众共创科技有限公司 | Voice processing method and device |
CN109766411A (en) * | 2019-01-14 | 2019-05-17 | 广东小天才科技有限公司 | Method and system for analyzing search problems |
CN111782782B (en) * | 2020-06-09 | 2023-04-18 | 苏宁金融科技(南京)有限公司 | Consultation reply method and device for intelligent customer service, computer equipment and storage medium |
CN114398514B (en) * | 2021-12-24 | 2022-11-22 | 北京达佳互联信息技术有限公司 | Video display method and device and electronic equipment |
Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN1553381A (en) * | 2003-05-26 | 2004-12-08 | 杨宏惠 | Multi-language correspondent list style language database and synchronous computer inter-transtation and communication |
CN1838159A (en) * | 2006-02-14 | 2006-09-27 | 北京未名博思生物智能科技开发有限公司 | Cognition logic machine and its information processing method |
CN101616221A (en) * | 2008-06-25 | 2009-12-30 | 富士通株式会社 | Guidance information display unit and guidance information display packing |
US7725308B2 (en) * | 2006-06-07 | 2010-05-25 | Motorola, Inc. | Interactive tool for semi-automatic generation of a natural language grammar from a device descriptor |
WO2012155079A2 (en) * | 2011-05-12 | 2012-11-15 | Johnson Controls Technology Company | Adaptive voice recognition systems and methods |
CN104408099A (en) * | 2014-11-18 | 2015-03-11 | 百度在线网络技术(北京)有限公司 | Searching method and searching device |
CN104509084A (en) * | 2012-07-18 | 2015-04-08 | Tw移动电子株式会社 | System for providing contact number information having added search function, and method for same |
CN104574251A (en) * | 2015-01-06 | 2015-04-29 | 熊国顺 | Intelligent public safety information system and application method |
CN104637007A (en) * | 2013-11-07 | 2015-05-20 | 大连东方之星信息技术有限公司 | Statistical analysis system employing degree-of-cognition system |
Family Cites Families (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7542971B2 (en) * | 2004-02-02 | 2009-06-02 | Fuji Xerox Co., Ltd. | Systems and methods for collaborative note-taking |
JP2006208483A (en) * | 2005-01-25 | 2006-08-10 | Sony Corp | Device, method, and program for assisting survey of interesting matter of listener, and recording medium |
CN101304457A (en) * | 2007-05-10 | 2008-11-12 | 许罗迈 | Method and apparatus for implementing automatic spoken language training based on voice telephone |
US20090216757A1 (en) * | 2008-02-27 | 2009-08-27 | Robi Sen | System and Method for Performing Frictionless Collaboration for Criteria Search |
CN103578469A (en) * | 2012-08-08 | 2014-02-12 | 百度在线网络技术(北京)有限公司 | Method and device for showing voice recognition result |
CN103680222B (en) * | 2012-09-19 | 2017-10-24 | 镇江诺尼基智能技术有限公司 | Children stories question and answer exchange method |
CN103000173B (en) * | 2012-12-11 | 2015-06-17 | 优视科技有限公司 | Voice interaction method and device |
US9269354B2 (en) * | 2013-03-11 | 2016-02-23 | Nuance Communications, Inc. | Semantic re-ranking of NLU results in conversational dialogue applications |
CN103594086B (en) * | 2013-10-25 | 2016-08-17 | 海菲曼(天津)科技有限公司 | Speech processing system, device and method |
CN105304082B (en) * | 2015-09-08 | 2018-12-28 | 北京云知声信息技术有限公司 | A kind of speech output method and device |
-
2015
- 2015-09-08 CN CN201510568430.8A patent/CN105304082B/en active Active
-
2016
- 2016-05-18 CN CN201680002958.1A patent/CN107077845B/en active Active
- 2016-05-18 WO PCT/CN2016/082427 patent/WO2017041510A1/en active Application Filing
Patent Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN1553381A (en) * | 2003-05-26 | 2004-12-08 | 杨宏惠 | Multi-language correspondent list style language database and synchronous computer inter-transtation and communication |
CN1838159A (en) * | 2006-02-14 | 2006-09-27 | 北京未名博思生物智能科技开发有限公司 | Cognition logic machine and its information processing method |
US7725308B2 (en) * | 2006-06-07 | 2010-05-25 | Motorola, Inc. | Interactive tool for semi-automatic generation of a natural language grammar from a device descriptor |
CN101616221A (en) * | 2008-06-25 | 2009-12-30 | 富士通株式会社 | Guidance information display unit and guidance information display packing |
WO2012155079A2 (en) * | 2011-05-12 | 2012-11-15 | Johnson Controls Technology Company | Adaptive voice recognition systems and methods |
CN104509084A (en) * | 2012-07-18 | 2015-04-08 | Tw移动电子株式会社 | System for providing contact number information having added search function, and method for same |
CN104637007A (en) * | 2013-11-07 | 2015-05-20 | 大连东方之星信息技术有限公司 | Statistical analysis system employing degree-of-cognition system |
CN104408099A (en) * | 2014-11-18 | 2015-03-11 | 百度在线网络技术(北京)有限公司 | Searching method and searching device |
CN104574251A (en) * | 2015-01-06 | 2015-04-29 | 熊国顺 | Intelligent public safety information system and application method |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110018843A (en) * | 2018-01-09 | 2019-07-16 | 北京小度互娱科技有限公司 | The test method and device of application program operation reserve |
CN110018843B (en) * | 2018-01-09 | 2022-08-30 | 北京小度互娱科技有限公司 | Method and device for testing application program operation strategy |
CN109035896A (en) * | 2018-08-13 | 2018-12-18 | 广东小天才科技有限公司 | Oral training method and learning equipment |
Also Published As
Publication number | Publication date |
---|---|
CN105304082B (en) | 2018-12-28 |
WO2017041510A1 (en) | 2017-03-16 |
CN107077845B (en) | 2020-07-17 |
CN105304082A (en) | 2016-02-03 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN107077845A (en) | A kind of speech output method and device | |
CN107644642B (en) | Semantic recognition method and device, storage medium and electronic equipment | |
JP6482911B2 (en) | Device control method and electric device | |
KR20200017249A (en) | Apparatus and method for providing feedback for confirming intent of a user in an electronic device | |
KR102489914B1 (en) | Electronic Device and method for controlling the electronic device | |
US20190221208A1 (en) | Method, user interface, and device for audio-based emoji input | |
CN106773742A (en) | Sound control method and speech control system | |
CN109215638B (en) | Voice learning method and device, voice equipment and storage medium | |
CN107077510A (en) | A kind of information output method and device | |
CN108038102A (en) | Recommendation method, apparatus, terminal and the storage medium of facial expression image | |
CN111210844B (en) | Method, device and equipment for determining speech emotion recognition model and storage medium | |
CN108509412A (en) | A kind of data processing method, device, electronic equipment and storage medium | |
CN109002184A (en) | A kind of association method and device of input method candidate word | |
CN108320208A (en) | Vehicle recommends method and device | |
KR20210032875A (en) | Voice information processing method, apparatus, program and storage medium | |
CN111984180A (en) | Terminal screen reading method, device, equipment and computer readable storage medium | |
US11315573B2 (en) | Speaker recognizing method, speaker recognizing apparatus, recording medium recording speaker recognizing program, database making method, database making apparatus, and recording medium recording database making program | |
CN111199730B (en) | Voice recognition method, device, terminal and storage medium | |
KR20200027753A (en) | Electronic Device and the Method for Operating Task corresponding to Shortened Command | |
US11948567B2 (en) | Electronic device and control method therefor | |
CN109684443B (en) | Intelligent interaction method and device | |
KR102656521B1 (en) | Display device and controlling method of display device | |
CN105302335B (en) | Vocabulary recommends method and apparatus and computer readable storage medium | |
CN107133280A (en) | The response method and device of feedback | |
KR20200016774A (en) | System for processing user utterance and operating method thereof |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant | ||
CP03 | Change of name, title or address |
Address after: Room 101, 1st floor, building 1, Xisanqi building materials City, Haidian District, Beijing 100096 Patentee after: Yunzhisheng Intelligent Technology Co.,Ltd. Address before: 100028 12 / F, Guanjie building, building 1, yard 16, Taiyanggong Middle Road, Chaoyang District, Beijing Patentee before: BEIJING UNISOUND INFORMATION TECHNOLOGY Co.,Ltd. |
|
CP03 | Change of name, title or address |