CN107077845A - A kind of speech output method and device - Google Patents

A kind of speech output method and device Download PDF

Info

Publication number
CN107077845A
CN107077845A CN201680002958.1A CN201680002958A CN107077845A CN 107077845 A CN107077845 A CN 107077845A CN 201680002958 A CN201680002958 A CN 201680002958A CN 107077845 A CN107077845 A CN 107077845A
Authority
CN
China
Prior art keywords
voice input
user
input content
degree
content
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201680002958.1A
Other languages
Chinese (zh)
Other versions
CN107077845B (en
Inventor
王天
王天一
刘升平
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Unisound Intelligent Technology Co Ltd
Original Assignee
Beijing Yunzhisheng Information Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Yunzhisheng Information Technology Co Ltd filed Critical Beijing Yunzhisheng Information Technology Co Ltd
Publication of CN107077845A publication Critical patent/CN107077845A/en
Application granted granted Critical
Publication of CN107077845B publication Critical patent/CN107077845B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue

Landscapes

  • Engineering & Computer Science (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • User Interface Of Digital Computer (AREA)
  • Telephone Function (AREA)

Abstract

A kind of speech output method and device.Methods described includes:Receive the phonetic entry content (S11) of user's input;According to the phonetic entry content, cognition degree of the user to the phonetic entry content generic is determined, the cognition degree is the professional knowledge degree of awareness (S12) of the user to the classification;From at least one voice output content corresponding with the phonetic entry content, the voice output content (S13) matched with the cognition degree is obtained and exported.The technical scheme can be according to cognition degree of the user to the phonetic entry content generic of input, the voice output content matched with its cognition degree is selected to be exported for user, so that voice output content more conforms to the demand of user, provide the user the voice output function of personalization, improve the accuracy of voice output, allow users to get the information content of maximum from voice output content, improve the Experience Degree of user.

Description

A kind of speech output method and device
The application is based on that the applying date is September in 2015 8, application No. is the application for a patent for invention propositions of CN201510568430.8, entitled " a kind of speech output method and device ", and the priority of the application for a patent for invention is required, the application is incorporated herein as reference in the full content of the application for a patent for invention.
Technical field
The present invention relates to technical field of information processing more particularly to a kind of speech output methods and device.
Background technique
Currently, voice input is increasingly praised highly by people with the development of electronics technology, voice input is that the Content Transformation that people is spoken by speech recognition is a kind of input mode of text.With intelligent terminal popularizing in people's lives, more and more intelligent terminals gradually have the function of voice service, such as, user can input proposition problem by voice, the voice that voice software on intelligent terminal passes through analysis user, the problem of user is equally answered in a manner of voice services to provide help for user.However, although this method is that user brings great convenience, answer is obtained by cumbersome online inquiry without user, but only has a kind of answering model in current voice service software, that is, different users puts question to identical problem (main contents of problem are identical), then exports identical help information.And the technical level or patent ability of different user be all it is different, for a user, it may be necessary to different help informations, or different answer-mode, therefore, the above method cannot be distinguished from different technical needs to provide voice help for user, not have specific aim.
Summary of the invention
The embodiment of the present invention provides a kind of speech output method and device.The technical solution is as follows:
In a first aspect, providing a kind of speech output method, comprising the following steps:
Receive the voice input content of user's input;
According to the voice input content, the user is determined to the cognition degree of the voice input content generic, the cognition degree is the professional knowledge degree of awareness of the user to the classification;
From at least one voice output content corresponding with the voice input content, obtains and export the voice output content to match with the cognition degree.
Some beneficial effects of the embodiment of the present invention may include:
Above-mentioned technical proposal, it can be according to user to the cognition degree of the voice input content generic of input, the voice output content to match for user's selection with its cognition degree exports, so that voice output content is more in line with the demand of user, to provide more personalized voice output function for user, the accuracy for improving voice output simultaneously, allows users to get maximum information content from voice output content, improves the Experience Degree of user.
In one embodiment, described according to the voice input content, determine the user to the cognition degree of the voice input content generic, comprising:
Identify the voiceprint of the user;
According to the voiceprint, the voice input content for receiving the user for the first time is judged whether it is;
When the voice input content to receive the user for the first time, determine that the user is to preset minimum cognition degree to the cognition degree of the voice input content generic.
In the embodiment, according to whether being that the voice output content that user's selection matches exports to receive the voice input content of user for the first time, so that voice output content is more in line with the demand of user, to provide more personalized voice output function for user, the accuracy of voice output is improved simultaneously, it allows users to get maximum information content from voice output content, improves the Experience Degree of user.
In one embodiment, the method also includes:
Record the input time of the voice input content and using duration, when use a length of duration received between the voice input content and the output voice output content.
In the embodiment, by recording the input time of voice input content and using duration, so that when the subsequent output voice output content for user, determine that the foundation of user cognition degree is more abundant, to more accurately determine the cognition degree of user, and then the voice output content more accurate, personalized for user's output.
In one embodiment, described according to the voice input content, determine the user to the cognition degree of the voice input content generic, comprising:
Identify the voiceprint of the user;
According to the voiceprint of the user, judge whether the adjacent voice input content received twice is inputted by same user;
When the adjacent voice input content received twice is inputted by same user, according to the input time of the adjacent voice input content received twice and duration is used, calculates the time interval between the adjacent voice input content received twice;
According to the time interval, determine the user to the cognition degree of the voice input content generic;Wherein, the time interval is longer, and the cognition degree is lower.
In the embodiment, the time interval between voice input content by calculating the adjacent same user received twice, so that determining that the foundation of user cognition degree is more abundant, to more accurately determine the cognition degree of user, and then the voice output content more accurate, personalized for user's output.
In one embodiment, described according to the voice input content, determine the user to the cognition degree of the voice input content generic, comprising:
Identify the voiceprint of the user;
According to the voiceprint of the user, history input record information corresponding with the user is obtained, the history input record information includes that history is accumulative using at least one of time, the accumulative input number of history and history input frequency information;
According to the history input record information, determine the user to the cognition degree of the voice input content generic;Wherein, the history is accumulative longer using the time, and the cognition degree is higher;The accumulative input number of the history is more, and the cognition degree is higher;The history input frequency is higher, and the cognition degree is higher.
In the embodiment, the cognition degree of user is determined according to the corresponding history input record information of user, enables the terminal to the cognition degree for more accurately determining user, and then the voice output content more accurate, personalized for user's output.
In one embodiment, described according to the voice input content, determine the user to the cognition degree of the voice input content generic, comprising:
Extract the keyword in the voice input content;
Determine the matching degree of the keyword and predetermined keyword in the voice input content;
According to the matching degree of keyword and predetermined keyword in the voice input content, determine the user to the cognition degree of the voice input content generic;Wherein, the matching degree of the professional keyword in the keyword and predetermined keyword in the voice input content is higher, and the cognition degree is higher;The matching degree of the amateur keyword in keyword and predetermined keyword in the voice input content is higher, and the cognition degree is lower.
In the embodiment, the cognition degree of user is determined according to the matching degree of keyword and predetermined keyword in voice input content, so that the determination of user cognition degree is more accurate, personalized, thus the voice output content more accurate, personalized for user's output.
In one embodiment, described according to the voice input content, determine the user to the cognition degree of the voice input content generic, comprising:
Determine that the sentence structure type of the voice input content, the sentence structure type include professional sentence structure type or amateur sentence structure type;
According to the sentence structure type of the voice input content, determine the user to the cognition degree of the voice input content generic;Wherein, the user is higher than the cognition degree to the voice input content generic of the amateur sentence structure type to the cognition degree of the voice input content generic of the professional sentence structure type.
In the embodiment, the cognition degree of user is determined according to the sentence structure type of voice input content, so that the determination of user cognition degree is more accurate, personalized, thus the voice output content more accurate, personalized for user's output.
In one embodiment, described according to the voice input content, determine the user to the cognition degree of the voice input content generic, comprising:
When determining that the adjacent voice input content received twice is inputted by same user, according to the keyword in the adjacent voice input content received twice, the degree of association between the adjacent voice input content received twice is determined;
According to the degree of association between the adjacent voice input content received twice, determine the user to the cognition degree of the voice input content generic;Wherein, the degree of association is higher, and the cognition degree is lower.
In the embodiment, the cognition degree of user is determined according to the degree of association between the voice input content of the adjacent same user received twice, so that the determination of user cognition degree is more accurate, personalized, thus the voice output content more accurate, personalized for user's output.
In one embodiment, described according to the voice input content, determine the user to the cognition degree of the voice input content generic, comprising:
According to the voice input content, determine at least two voices input parameters of the voice input content, the voice input parameter include: the voiceprint of the user, same user the adjacent voice input content inputted twice between time interval, history input record information corresponding with the user, the matching degree of the keyword in the voice input content and predetermined keyword, the pass between the sentence structure type of the voice input content and the adjacent voice input content inputted twice of same user Connection degree;
The weight that parameter is inputted according to preset each single item voice, calculates the user to the cognition degree of the voice input content generic.
In the embodiment, the different weights of parameter are inputted according to the voice of multinomial voice input content, user is calculated to the cognition degree of voice input content generic, so that the determination of user cognition degree is more accurate, personalized, thus the voice output content more accurate, personalized for user's output.
In one embodiment, described according to the voice input content, determine the user to the cognition degree of the voice input content generic, comprising:
When the voice that can not determine the voice input content inputs parameter, determine that the user is to preset minimum cognition degree to the cognition degree of the voice input content generic.
In the embodiment, for that can not determine the voice input content of voice input parameter, the voice output content that output matches with the voice input content, to provide more accurate and personalized voice output function for user, it allows users to get more useful information from voice output content, improves the Experience Degree of user.
In one embodiment, described from least one voice output content corresponding with the voice input content, it obtains and exports the voice output content to match with the cognition degree, comprising:
According to the corresponding relationship between cognition degree and cognition grade, the corresponding cognition grade of the cognition degree is determined;
According to the corresponding relationship between cognition grade and voice output content, voice output content corresponding with the cognition grade is obtained;
Export the voice output content.
In the embodiment, matched voice output content is selected to export for user according to the corresponding relationship between cognition grade and voice output content, it is exported to select the voice output content to match with user cognition degree for user, so that voice output content is more in line with the demand of user, improve the accuracy of voice output, it allows users to get maximum information content from voice output content, improves the Experience Degree of user.
In one embodiment, the method also includes:
According to the input time of the voice input content and duration is used, updates the history input record information.
In the embodiment, by the update to history input record information, when so that exporting voice output content again for user, the cognition degree of user can be determined according to accurate history input record, to export more accurate voice output content for user.
In one embodiment, the method also includes:
The user is stored to the cognition degree of the voice input content generic;
It is described according to the voice input content, determine the user to the cognition degree of the voice input content generic, comprising:
Identify the voiceprint of the user;
The user is inquired to the cognition degree of the voice input content generic according to the voiceprint of the user.
In the embodiment, by inquiring the cognition degree of user, user more convenient can be quickly determined out to voice input content The cognition degree of generic, so that the voice output content more quickly and accurately to match for user's selection exports.
Second aspect provides a kind of instantaneous speech power, comprising:
Receiving module, for receiving the voice input content of user's input;
Determining module, for determining the user to the cognition degree of the voice input content generic, the cognition degree is the professional knowledge degree of awareness of the user to the classification according to the voice input content;
Output module, for obtaining and exporting the voice output content to match with the cognition degree from least one voice output content corresponding with the voice input content.
In one embodiment, the determining module includes:
First identifies submodule, for identification the voiceprint of the user;
First judging submodule, for judging whether it is the voice input content for receiving the user for the first time according to the voiceprint;
Second determines submodule, is to preset minimum cognition degree to the cognition degree of the voice input content generic for when the voice input content to receive the user for the first time, determining the user.
In one embodiment, described device further include:
Logging modle, for recording the input time of the voice input content and using duration, when use a length of duration received between the voice input content and the output voice output content.
In one embodiment, the determining module includes:
Second identifies submodule, for identification the voiceprint of the user;
Second judgment submodule judges whether the adjacent voice input content received twice is inputted by same user for the voiceprint according to the user;
First computational submodule, for when the adjacent voice input content received twice is inputted by same user, according to the input time of the adjacent voice input content received twice and duration is used, calculates the time interval between the adjacent voice input content received twice;
Third determines submodule, for determining the user to the cognition degree of the voice input content generic according to the time interval;Wherein, the time interval is longer, and the cognition degree is lower.
In one embodiment, the determining module includes:
Third identifies submodule, for identification the voiceprint of the user;
First acquisition submodule, for the voiceprint according to the user, history input record information corresponding with the user is obtained, the history input record information includes that history is accumulative using at least one of time, the accumulative input number of history and history input frequency information;
4th determines submodule, for determining the user to the cognition degree of the voice input content generic according to the history input record information;Wherein, the history is accumulative longer using the time, and the cognition degree is higher;The accumulative input number of the history is more, and the cognition degree is higher;The history input frequency is higher, and the cognition degree is higher.
In one embodiment, the determining module includes:
Extracting sub-module, for extracting the keyword in the voice input content;
5th determines submodule, for determining the matching degree of keyword and predetermined keyword in the voice input content;
6th determines that submodule determines the user to the cognition degree of the voice input content generic for the matching degree according to keyword and predetermined keyword in the voice input content;Wherein, the matching degree of the professional keyword in the keyword and predetermined keyword in the voice input content is higher, and the cognition degree is higher;The matching degree of the amateur keyword in keyword and predetermined keyword in the voice input content is higher, and the cognition degree is lower.
In one embodiment, the determining module includes:
7th determines submodule, for determining that the sentence structure type of the voice input content, the sentence structure type include professional sentence structure type or amateur sentence structure type;
8th determines that submodule determines the user to the cognition degree of the voice input content generic for the sentence structure type according to the voice input content;Wherein, the user is higher than the cognition degree to the voice input content generic of the amateur sentence structure type to the cognition degree of the voice input content generic of the professional sentence structure type.
In one embodiment, the determining module includes:
9th determines submodule, for when determining that the adjacent voice input content received twice is inputted by same user, according to the keyword in the adjacent voice input content received twice, the degree of association between the adjacent voice input content received twice is determined;
Tenth determines submodule, for determining the user to the cognition degree of the voice input content generic according to the degree of association between the adjacent voice input content received twice;Wherein, the degree of association is higher, and the cognition degree is lower.
In one embodiment, the determining module includes:
11st determines submodule, for according to the voice input content, determine at least two voices input parameters of the voice input content, the voice input parameter include: the voiceprint of the user, same user the adjacent voice input content inputted twice between time interval, history input record information corresponding with the user, the matching degree of the keyword in the voice input content and predetermined keyword, the degree of association between the sentence structure type of the voice input content and the adjacent voice input content inputted twice of same user;
Computational submodule calculates the user to the cognition degree of the voice input content generic for inputting the weight of parameter according to preset each single item voice.
In one embodiment, the determining module includes:
12nd determines submodule, for when the voice that can not determine the voice input content inputs parameter, determining that the user is to preset minimum cognition degree to the cognition degree of the voice input content generic.
In one embodiment, the output module includes:
13rd determines submodule, for determining the corresponding cognition grade of the cognition degree according to the corresponding relationship between cognition degree and cognition grade;
Second acquisition submodule, for obtaining voice output content corresponding with the cognition grade according to the corresponding relationship between cognition grade and voice output content;
Output sub-module, for exporting the voice output content.
In one embodiment, described device further include:
Update module updates the history input record information for the input time according to the voice input content and using duration.
In one embodiment, described device further include:
Memory module, for storing the user to the cognition degree of the voice input content generic;
The determining module includes:
4th identifies submodule, for identification the voiceprint of the user;
Submodule is inquired, for inquiring the user to the cognition degree of the voice input content generic according to the voiceprint of the user.
Some beneficial effects of the embodiment of the present invention may include:
Above-mentioned apparatus, it can be according to user to the cognition degree of the voice input content generic of input, the voice output content to match for user's selection with its cognition degree exports, so that voice output content is more in line with the demand of user, to provide more personalized voice output function for user, the accuracy for improving voice output simultaneously, allows users to get maximum information content from voice output content, improves the Experience Degree of user.
The third aspect provides a kind of instantaneous speech power, which is characterized in that described device includes:
Processor;
For storing the memory of the processor-executable instruction;
Wherein, the processor is configured to:
Receive the voice input content of user's input;
According to the voice input content, the user is determined to the cognition degree of the voice input content generic, the cognition degree is the professional knowledge degree of awareness of the user to the classification;
From at least one voice output content corresponding with the voice input content, obtains and export the voice output content to match with the cognition degree.
Above-mentioned processor is also configured to
Identify the voiceprint of the user;
According to the voiceprint, the voice input content for receiving the user for the first time is judged whether it is;
When the voice input content to receive the user for the first time, determine that the user is to preset minimum cognition degree to the cognition degree of the voice input content generic.
Above-mentioned processor is also configured to
Record the input time of the voice input content and using duration, when use a length of duration received between the voice input content and the output voice output content.
Above-mentioned processor is also configured to
Identify the voiceprint of the user;
According to the voiceprint of the user, judge whether the adjacent voice input content received twice is inputted by same user;
When the adjacent voice input content received twice is inputted by same user, according to the adjacent language received twice The input time of sound input content and duration is used, calculates the time interval between the adjacent voice input content received twice;
According to the time interval, determine the user to the cognition degree of the voice input content generic;Wherein, the time interval is longer, and the cognition degree is lower.
Above-mentioned processor is also configured to
Identify the voiceprint of the user;
According to the voiceprint of the user, history input record information corresponding with the user is obtained, the history input record information includes that history is accumulative using at least one of time, the accumulative input number of history and history input frequency information;
According to the history input record information, determine the user to the cognition degree of the voice input content generic;Wherein, the history is accumulative longer using the time, and the cognition degree is higher;The accumulative input number of the history is more, and the cognition degree is higher;The history input frequency is higher, and the cognition degree is higher.
Above-mentioned processor is also configured to
Extract the keyword in the voice input content;
Determine the matching degree of the keyword and predetermined keyword in the voice input content;
According to the matching degree of keyword and predetermined keyword in the voice input content, determine the user to the cognition degree of the voice input content generic;Wherein, the matching degree of the professional keyword in the keyword and predetermined keyword in the voice input content is higher, and the cognition degree is higher;The matching degree of the amateur keyword in keyword and predetermined keyword in the voice input content is higher, and the cognition degree is lower.
Above-mentioned processor is also configured to
Determine that the sentence structure type of the voice input content, the sentence structure type include professional sentence structure type or amateur sentence structure type;
According to the sentence structure type of the voice input content, determine the user to the cognition degree of the voice input content generic;Wherein, the user is higher than the cognition degree to the voice input content generic of the amateur sentence structure type to the cognition degree of the voice input content generic of the professional sentence structure type.
Above-mentioned processor is also configured to
When determining that the adjacent voice input content received twice is inputted by same user, according to the keyword in the adjacent voice input content received twice, the degree of association between the adjacent voice input content received twice is determined;
According to the degree of association between the adjacent voice input content received twice, determine the user to the cognition degree of the voice input content generic;Wherein, the degree of association is higher, and the cognition degree is lower.
Above-mentioned processor is also configured to
According to the voice input content, determine at least two voices input parameters of the voice input content, the voice input parameter include: the voiceprint of the user, same user the adjacent voice input content inputted twice between time interval, history input record information corresponding with the user, the matching degree of the keyword in the voice input content and predetermined keyword, the degree of association between the sentence structure type of the voice input content and the adjacent voice input content inputted twice of same user;
The weight that parameter is inputted according to preset each single item voice, calculates the user to the voice input content generic Cognition degree.
Above-mentioned processor is also configured to
When the voice that can not determine the voice input content inputs parameter, determine that the user is to preset minimum cognition degree to the cognition degree of the voice input content generic.
Above-mentioned processor is also configured to
According to the corresponding relationship between cognition degree and cognition grade, the corresponding cognition grade of the cognition degree is determined;
According to the corresponding relationship between cognition grade and voice output content, voice output content corresponding with the cognition grade is obtained;
Export the voice output content.
Above-mentioned processor is also configured to
According to the input time of the voice input content and duration is used, updates the history input record information.
Above-mentioned processor is also configured to
The user is stored to the cognition degree of the voice input content generic;
It is described according to the voice input content, determine the user to the cognition degree of the voice input content generic, comprising:
Identify the voiceprint of the user;
The user is inquired to the cognition degree of the voice input content generic according to the voiceprint of the user.
Fourth aspect provides a kind of non-transitory computer readable recording medium, and record has computer program on the medium, and described program includes the instruction for executing the method as described in the first aspect of the embodiment of the present invention.
5th aspect, provides a kind of computer program, and described program includes: the instruction for executing the method as described in the first aspect of the embodiment of the present invention when described program is computer-executed.
Other features and advantages of the present invention will be illustrated in the following description, also, partly as will become apparent from the description, or understand through the implementation of the invention.The objectives and other advantages of the invention can be achieved and obtained by structure specifically indicated in the written description, claims, and drawings.
Below by drawings and examples, technical scheme of the present invention will be described in further detail.
Detailed description of the invention
Attached drawing is used to provide further understanding of the present invention, and constitutes part of specification, is used to explain the present invention together with embodiments of the present invention, is not construed as limiting the invention.In the accompanying drawings:
Fig. 1 is a kind of flow chart of speech output method in the embodiment of the present invention;
The flow chart that Fig. 2 is step S12 in a kind of speech output method in the embodiment of the present invention;
The flow chart that Fig. 3 is step S12 in a kind of speech output method in the embodiment of the present invention;
The flow chart that Fig. 4 is step S12 in a kind of speech output method in the embodiment of the present invention;
The flow chart that Fig. 5 is step S12 in a kind of speech output method in the embodiment of the present invention;
The flow chart that Fig. 6 is step S13 in a kind of speech output method in the embodiment of the present invention;
Fig. 7 is a kind of block diagram of instantaneous speech power in the embodiment of the present invention;
Fig. 8 is a kind of block diagram of determining module in instantaneous speech power in the embodiment of the present invention;
Fig. 9 is a kind of block diagram of determining module in instantaneous speech power in the embodiment of the present invention;
Figure 10 is a kind of block diagram of determining module in instantaneous speech power in the embodiment of the present invention;
Figure 11 is a kind of block diagram of determining module in instantaneous speech power in the embodiment of the present invention;
Figure 12 is a kind of block diagram of output module in instantaneous speech power in the embodiment of the present invention;
Figure 13 is a kind of block diagram of instantaneous speech power in the embodiment of the present invention;
Figure 14 is a kind of block diagram of the device of executable speech output method in the embodiment of the present invention.
Specific embodiment
Hereinafter, preferred embodiments of the present invention will be described with reference to the accompanying drawings, it should be understood that preferred embodiments described herein are only used to illustrate and explain the present invention, is not intended to limit the present invention.
Fig. 1 is a kind of flow chart of speech output method in the embodiment of the present invention.As shown in Figure 1, this method is in terminal, terminal can be mobile phone, computer, digital broadcasting terminal, messaging device, game console, tablet device, Medical Devices, body-building equipment, and personal digital assistant etc. includes the following steps S11-S13:
Step S11 receives the voice input content of user's input.
In the step, user can input voice input content by way of typing sound.
Step S12 determines user to the cognition degree of voice input content generic according to voice input content;The cognition degree is the professional knowledge degree of awareness of the user to voice input content generic.
For example, user inputs voice input content " how air-conditioner temperature being arranged ", then user is the professional knowledge degree of awareness of the user to air-conditioning class to the cognition degree of voice input content generic;User inputs voice input content " what medicine aspirin is ", then user is the professional knowledge degree of awareness of the user to pharmaceutical to the cognition degree of voice input content generic.Terminal can be by extracting the keyword in voice input content, to determine classification belonging to voice input content.
Step S13 is obtained from least one voice output content corresponding with voice input content and is exported the voice output content to match with cognition degree.
Using technical solution provided in an embodiment of the present invention, it can be according to user to the cognition degree of the voice input content generic of input, the voice output content to match for user's selection with its cognition degree exports, so that voice output content is more in line with the demand of user, to provide more personalized voice output function for user, the accuracy for improving voice output simultaneously, allows users to get maximum information content from voice output content, improves the Experience Degree of user.
In step s 12, user can determine the cognition degree of voice input content generic in several ways.It can determine the voice input parameter of voice input content first according to voice input content, determine user to the cognition degree of voice input content generic further according to voice input parameter.Wherein, the difference of parameter is inputted according to voice, the method of determination of cognition degree is also different, voice input parameter may include the voiceprint of user, same user the adjacent voice input content inputted twice between time interval, history input record information corresponding to the user, the association between the keyword voice input content inputted twice adjacent with the matching degree of predetermined keyword, the sentence structure type of voice input content and same user in voice input content Degree, etc..Illustrate the embodiment of step S12 below by way of different embodiments.
In one embodiment, as shown in Fig. 2, step S12 may be embodied as following steps S21-S23:
Step S21 identifies the voiceprint of user.
Step S22 judges whether it is the voice input content for receiving user for the first time according to voiceprint.
Step S23 determines that user is to preset minimum cognition degree to the cognition degree of voice input content generic when the voice input content to receive user for the first time.
In the present embodiment, the corresponding voiceprint of different user is stored in terminal, when user inputs voice input content, if terminal can inquire the voiceprint of the user in pre-stored voiceprint, illustrate not to be the voice input content for receiving the user for the first time, and if terminal fails to inquire the voiceprint of the user in pre-stored voiceprint, illustrate that terminal is the voice input content for receiving the user for the first time.When the voice input content not to receive user for the first time, then terminal continues to determine that other voices input parameter according to voice input content, and inputs parameter according to other voices and execute step S12.In the terminal, the corresponding relationship being previously stored between cognition degree and voice input content, wherein including voice input content corresponding with minimum cognition degree is preset.
In one embodiment, the above method is further comprising the steps of: record voice input content input time and use duration, this using when it is a length of receive voice input content and export voice output content between duration.Therefore, as shown in figure 3, step S12 may be embodied as following steps S31-S34:
Step S31 identifies the voiceprint of user.
Step S32 judges whether the adjacent voice input content received twice is inputted by same user according to the voiceprint of user.
Step S33 according to the input time of the adjacent voice input content received twice and uses duration, calculates the time interval between the adjacent voice input content received twice when the adjacent voice input content received twice is inputted by same user.
Step S34 determines user to the cognition degree of voice input content generic according to time interval;Wherein, time interval is longer, and cognition degree is lower.
In the present embodiment, when the adjacent voice input content received twice is inputted by same user, time interval between the so adjacent voice input content received twice can reflect out the reaction duration for the upper voice output content that user exports terminal, in addition, user can also receive the time interval voice input content to this by last time output voice output content to the reaction duration for the upper voice output content that terminal exports to characterize.For example, the voice input content that the terminal last time receives is " how air-conditioner temperature being arranged ", and for the voice input content, it is " being introduced into temperature shaping modes, then change temperature " that terminal, which exports corresponding voice output content,;This voice input content received of terminal is " how entering temperature shaping modes ", when terminal determines that the adjacent voice input content received twice is inputted by same user, then can be used receive voice input content " how air-conditioner temperature is set " and receive the time interval between voice input content " how to enter temperature shaping modes " come characterize user to a upper voice output content " be introduced into temperature shaping modes; then change temperature " reaction duration, and then determine user to the cognition degree of voice input content generic.Alternatively, can also be used output voice output content " being introduced into temperature shaping modes, then change temperature " and this receive voice input content " such as What enters temperature shaping modes " between time interval characterize user to the reaction duration of a upper voice output content " being introduced into temperature shaping modes, then change temperature ", and then determine user to the cognition degree of voice input content generic.Time interval is longer, as soon as illustrate that user is longer to the reaction duration of upper voice output content, then cognition degree is lower.
Furthermore, a prefixed time interval can also be preset, when the adjacent voice input content received twice is inputted by same user and time interval between the adjacent voice input content received twice is more than prefixed time interval, it is to preset minimum cognition degree, and obtain and preset the voice output content that minimum cognition degree matches and exported to the cognition degree of voice input content generic that terminal, which can directly determine user,.
In one embodiment, as shown in figure 4, step S12 may be embodied as following steps S41-S43:
Step S41 identifies the voiceprint of user.
Step S42 obtains history input record information corresponding to the user according to the voiceprint of user;History input record information includes that history is accumulative using at least one of time, the accumulative input number of history and history input frequency information.
Step S43 determines user to the cognition degree of voice input content generic according to history input record information;Wherein, history is accumulative longer using the time, and cognition degree is higher;The accumulative input number of history is more, and cognition degree is higher;History input frequency is higher, and cognition degree is higher.
In the present embodiment, terminal receives the voice input content of user's input every time, just will record the input time of voice input content and using duration, this using when a length of receive voice input content and export the duration between voice output content.Terminal can count history input record information corresponding to the user according to the input time of record and using duration, wherein the accumulative summation using duration as recorded each time using the time of history.In addition, the above method is further comprising the steps of: according to the input time of voice input content and using duration, more new historical input record information.In this way, when terminal determines cognition degree of the user to voice input content generic according to history input record information corresponding to the user, based on history input record information it is more abundant accurate, so as to be exported for more accurate and personalization the voice output content of user's selection.
In one embodiment, as shown in figure 5, step S12 may be embodied as following steps S51-S53:
Step S51 extracts the keyword in voice input content.
Step S52 determines the matching degree of the keyword and predetermined keyword in voice input content.
Step S53 determines user to the cognition degree of voice input content generic according to the matching degree of keyword and predetermined keyword in voice input content;Wherein, the matching degree of the professional keyword in the keyword and predetermined keyword in voice input content is higher, and cognition degree is higher;The matching degree of amateur keyword in keyword and predetermined keyword in voice input content is higher, and cognition degree is lower.
In the present embodiment, the predetermined keyword prestored in terminal includes two types of professional keyword and amateur keyword, when executing step S52, need to determine the matching degree between the keyword and professional keyword in voice input content, and the matching degree between amateur keyword respectively.Such as, professional keyword includes " setting path ", amateur keyword includes " how using ", if the voice input content that terminal receives is " ... setting path ", it so can determine that the matching degree between the keyword and professional keyword in voice input content is higher, therefore user is also higher to the cognition degree of voice input content generic;If the voice input content that terminal receives is " ... how to use ", it can determine that the matching degree between the keyword and amateur keyword in voice input content is higher, therefore user also gets over the cognition degree of voice input content generic It is low.
In one embodiment, step S12 may be embodied as following steps A1-A2:
Step A1 determines that the sentence structure type of voice input content, sentence structure type include professional sentence structure type or amateur sentence structure type.
Step A2 determines user to the cognition degree of voice input content generic according to the sentence structure type of voice input content;Wherein, user is higher than the cognition degree to the voice input content generic of amateur sentence structure type to the cognition degree of the voice input content generic of professional sentence structure type.
Sentence structure type is prestored in the present embodiment, in terminal, sentence structure type can be embodied by regular expression.Wherein, the regular expression of professional sentence structure type is such as: adjective+noun+verb;The regular expression of amateur sentence structure type is such as: pronoun+verb.It should be pointed out that the manifestation mode of sentence structure type is not limited to regular expression, can also be embodied in such a way that other can embody sentence structure.It is exemplified below, the voice input content that terminal receives is " what the step of booting is ", terminal is by analyzing the voice input content, the sentence structure type for determining the voice input content is " adjective+noun+verb+pronoun ", it can determine the sentence structure type of the voice input content so as professional sentence structure type, user is higher to the cognition degree of the voice input content generic.For another example, the voice input content that terminal receives is " how this thing is used ", terminal is by analyzing the voice input content, the sentence structure type for determining the voice input content is " pronoun+verb ", the sentence structure type that so can determine the voice input content is amateur sentence structure type, and user is lower to the cognition degree of the voice input content generic.
In one embodiment, step S12 may be embodied as following steps B1-B2:
Step B1, according to the keyword in the adjacent voice input content received twice, determines the degree of association between the adjacent voice input content received twice when determining that the adjacent voice input content received twice is inputted by same user.
Step B2 determines user to the cognition degree of voice input content generic according to the degree of association between the adjacent voice input content received twice;Wherein, the degree of association is higher, and cognition degree is lower.
In the present embodiment, when the adjacent voice input content received twice is inputted by same user, the degree of association between the so adjacent voice input content received twice can reflect out user to the degree of understanding of a upper voice output content, therefore the degree of association between the adjacent voice input content received twice is higher, as soon as illustrating that user is lower to the degree of understanding of upper voice output content, user is also lower to the cognition degree of voice input content generic;The degree of association between the adjacent voice input content received twice is lower, as soon as illustrating that user is higher to the degree of understanding of upper voice output content, user is also higher to the cognition degree of voice input content generic.Such as, the voice input content that the terminal last time receives is " how air-conditioner temperature being arranged ", this voice input content received of terminal is " how entering temperature shaping modes " simultaneously, when terminal determines that the adjacent voice input content received twice is inputted by same user, it can extract the keyword in the adjacent voice input content received twice, such as keyword " air-conditioner temperature ", " temperature shaping modes ", the degree of association between the adjacent voice input content received twice is determined by the degree of association between keyword " air-conditioner temperature " and keyword " temperature shaping modes ", since " air-conditioner temperature " and " temperature shaping modes " is all keyword related with temperature, therefore the degree of association between the two is higher.For another example, the voice input content that the terminal last time receives is " how air-conditioner temperature being arranged ", while terminal this voice input content for receiving is " what the step of booting is ", when terminal determines that the adjacent voice input content received twice is same When user is inputted, extracting the keyword in two neighboring voice input content respectively is " air-conditioner temperature " and " booting ", due to the keyword that the two keywords are two unrelated types, therefore the degree of association between the two is almost nil, that is to say, the degree of association between the bright adjacent voice input content received twice is very low, user is higher to the degree of understanding of a upper voice output content, and it is higher to the cognition degree of voice input content generic to further relate to user.
In one embodiment, for the various ways of execution step S12 in above-described embodiment, also multinomial voice can be inputted parameter and combined, and user is calculated to the cognition degree of voice input content generic according to pre-set weight.Therefore, above-mentioned steps S12 can also be embodied as following steps: according to voice input content, determine at least two voice input parameters of voice input content, wherein, voice input parameter include: the voiceprint of user, same user the adjacent voice input content inputted twice between time interval, history input record information corresponding to the user, the degree of association between the keyword voice input content inputted twice adjacent with the matching degree of predetermined keyword, the sentence structure type of voice input content and same user in voice input content;The weight of parameter is inputted according to preset each single item voice, calculates user to the cognition degree of voice input content generic.
In one embodiment, the above method is further comprising the steps of: when the voice that can not determine voice input content inputs parameter, determining that user is to preset minimum cognition degree to the cognition degree of voice input content generic.In the present embodiment, when receiving the voice input content of user's input, for that can not determine the voice input content of voice input content, it is to preset minimum cognition degree to the cognition degree of the voice input content generic that terminal, which can directly determine user, therefore, even the voice input content of voice input parameter can not be determined, user can also get matched voice output content, to improve the Experience Degree of user.
In one embodiment, the above method is further comprising the steps of: cognition degree of the storage user to voice input content generic.At this point, step S12 may be embodied as following steps: identifying the voiceprint of user;User is inquired to the cognition degree of voice input content generic according to the voiceprint of user.In the present embodiment, by inquiring the cognition degree of user, user more convenient can be quickly determined out to the cognition degree of voice input content generic, so that the voice output content more quickly and accurately to match for user's selection exports.
In one embodiment, as shown in fig. 6, step S13 is implementable for following steps S61-S63:
Step S61 determines the corresponding cognition grade of cognition degree according to the corresponding relationship between cognition degree and cognition grade.
Step S62 obtains voice output content corresponding with cognition grade according to the corresponding relationship between cognition grade and voice output content.
Step S63 exports voice output content.
In the present embodiment, terminal prestores cognition degree and recognizes the corresponding relationship between grade, and the corresponding relationship between cognition grade and voice output content, such as, cognition grade can be divided into low cognition grade, middle cognition grade, high cognition grade three grades as needed, the low cognition grade of correspondence of the cognition degree between " 0%~30% ", cognition degree recognize grade, corresponding high cognition grade of the cognition degree between " 71%~100% " in the correspondence between " 31%~70% ".Voice output content corresponding with low cognition grade is detailed version voice output content, voice output content corresponding with middle cognition grade is standard edition voice output content, voice output content corresponding with height cognition grade is succinct version voice output content, for each voice input content, terminal can all store three kinds of corresponding detailed version, succinct version, standard edition voice output contents.For example, for voice How input content " is arranged air-conditioner temperature ", voice output content corresponding thereto includes: detailed version " clicking the mode button in first row middle position; click and enter temperature shaping modes twice; click ' +/- ' the change temperature of second row left button; click primary, 1 degree of temperature ' +/- ' ";Standard edition " clicks mode button and enters temperature shaping modes, click ' +/- ' the change temperature of button ";Succinct version " being introduced into temperature shaping modes, then change temperature ".Furthermore, cognition grade corresponding with minimum cognition degree is preset can be low cognition grade, therefore, for that can not determine the voice input content of voice input parameter or for the first time receive the voice input content of user, terminal can directly export detailed version voice output content.It can be seen that, using the technical solution of the present embodiment, make terminal when exporting voice output content for user, it can be by determining that user analyzes the current demand of user to the cognition degree of voice input content generic, and matched voice output content is exported according to the current demand of user, it allows users to get more more accurate information from voice output content.
Corresponding to a kind of above-mentioned speech output method, the embodiment of the present invention also provides a kind of instantaneous speech power, and the device is to execute the above method.
Fig. 7 is a kind of block diagram of instantaneous speech power in the embodiment of the present invention.As shown in fig. 7, the device includes:
Receiving module 71, for receiving the voice input content of user's input;
Determining module 72, for determining user to the cognition degree of voice input content generic, cognition degree is the professional knowledge degree of awareness of the user to classification according to voice input content;
Output module 73, for obtaining and exporting the voice output content to match with cognition degree from least one voice output content corresponding with voice input content.
In one embodiment, as shown in figure 8, determining module 72 includes:
First identifies submodule 721, for identification the voiceprint of user;
First judging submodule 722, for according to voiceprint, judging whether it is the voice input content for receiving user for the first time;
Second determines submodule 723, for when the voice input content to receive user for the first time, determining that user is to preset minimum cognition degree to the cognition degree of voice input content generic.
In one embodiment, above-mentioned apparatus further include:
Logging modle, for recording the input time of voice input content and using duration, when use is a length of to be received voice input content and exports the duration between voice output content.
In one embodiment, as shown in figure 9, determining module 72 includes:
Second identifies submodule 724, for identification the voiceprint of user;
Second judgment submodule 725 judges whether the adjacent voice input content received twice is inputted by same user for the voiceprint according to user;
First computational submodule 726, for when the adjacent voice input content received twice is inputted by same user, according to the input time of the adjacent voice input content received twice and duration is used, calculates the time interval between the adjacent voice input content received twice;
Third determines submodule 727, for determining user to the cognition degree of voice input content generic according to time interval;Wherein, time interval is longer, and cognition degree is lower.
In one embodiment, as shown in Figure 10, determining module 72 includes:
Third identifies submodule 728, for identification the voiceprint of user;
First acquisition submodule 729 obtains history input record information corresponding to the user for the voiceprint according to user, and history input record information includes that history is accumulative using at least one of time, the accumulative input number of history and history input frequency information;
4th determines submodule 7210, for determining user to the cognition degree of voice input content generic according to history input record information;Wherein, history is accumulative longer using the time, and cognition degree is higher;The accumulative input number of history is more, and cognition degree is higher;History input frequency is higher, and cognition degree is higher.
In one embodiment, as shown in figure 11, determining module 72 includes:
Extracting sub-module 7211, for extracting the keyword in voice input content;
5th determines submodule 7212, for determining the matching degree of keyword and predetermined keyword in voice input content;
6th determines that submodule 7213 determines user to the cognition degree of voice input content generic for the matching degree according to keyword and predetermined keyword in voice input content;Wherein, the matching degree of the professional keyword in the keyword and predetermined keyword in voice input content is higher, and cognition degree is higher;The matching degree of amateur keyword in keyword and predetermined keyword in voice input content is higher, and cognition degree is lower.
In one embodiment, determining module 72 includes:
7th determines submodule, for determining that the sentence structure type of voice input content, sentence structure type include professional sentence structure type or amateur sentence structure type;
8th determines that submodule determines user to the cognition degree of voice input content generic for the sentence structure type according to voice input content;Wherein, user is higher than the cognition degree to the voice input content generic of amateur sentence structure type to the cognition degree of the voice input content generic of professional sentence structure type.
In one embodiment, determining module 72 includes:
9th determines submodule, for according to the keyword in the adjacent voice input content received twice, determining the degree of association between the adjacent voice input content received twice when determining that the adjacent voice input content received twice is inputted by same user;
Tenth determines submodule, for determining user to the cognition degree of voice input content generic according to the degree of association between the adjacent voice input content received twice;Wherein, the degree of association is higher, and cognition degree is lower.
In one embodiment, determining module 72 includes:
11st determines submodule, for according to voice input content, determine at least two voices input parameters of voice input content, voice input parameter include: the voiceprint of user, same user the adjacent voice input content inputted twice between time interval, history input record information corresponding to the user, the degree of association between the keyword voice input content inputted twice adjacent with the matching degree of predetermined keyword, the sentence structure type of voice input content and same user in voice input content;
Computational submodule calculates user to the cognition degree of voice input content generic for inputting the weight of parameter according to preset each single item voice.
In one embodiment, determining module 72 includes:
12nd determines submodule, for determining user to voice when the voice that can not determine voice input content inputs parameter The cognition degree of input content generic is to preset minimum cognition degree.
In one embodiment, as shown in figure 12, output module 73 includes:
13rd determines submodule 731, for determining the corresponding cognition grade of cognition degree according to the corresponding relationship between cognition degree and cognition grade;
Second acquisition submodule 732, for obtaining voice output content corresponding with cognition grade according to the corresponding relationship between cognition grade and voice output content;
Output sub-module 733, for exporting voice output content.
In one embodiment, as shown in figure 13, above-mentioned apparatus further include:
Update module 74, for according to input time of voice input content and using duration, more new historical input record information.
Memory module 75, for storing user to the cognition degree of voice input content generic.
In one embodiment, determining module 72 includes:
4th identifies submodule, for identification the voiceprint of user;
Submodule is inquired, for inquiring user to the cognition degree of voice input content generic according to the voiceprint of user.
Using device provided in an embodiment of the present invention, it can be according to user to the cognition degree of the voice input content generic of input, the voice output content to match for user's selection with its cognition degree exports, so that voice output content is more in line with the demand of user, to provide more personalized voice output function for user, the accuracy for improving voice output simultaneously, allows users to get maximum information content from voice output content, improves the Experience Degree of user.
Figure 14 is a kind of block diagram of the device of executable speech output method shown according to an exemplary embodiment.For example, device 1600 can be mobile phone, computer, digital broadcasting terminal, messaging device, game console, tablet device, Medical Devices, body-building equipment, personal digital assistant etc..
Referring to Fig.1 4, device 1600 may include following one or more components: processor 1601, memory 1602 and communication component 1603.
The integrated operation of the usual control device 1600 of processor 1601 such as operates associated operation with display, telephone call, data communication, camera operation and record.Processor 1601 can execute instruction, to perform all or part of the steps of the methods described above.
Memory 1602 is configured as storing various types of data to support the operation in device 1600.The example of these data includes the instruction of any application or method for operating on device 1600, contact data, telephone book data, message, picture, video etc..Memory 1602 can be realized by any kind of volatibility or non-volatile memory device or their combination, such as static random access memory (SRAM), electrically erasable programmable read-only memory (EEPROM), Erasable Programmable Read Only Memory EPROM (EPROM), programmable read only memory (PROM), read-only memory (ROM), magnetic memory, flash memory, disk or CD.
Communication component 1603 is configured to facilitate the communication of wired or wireless way between device 1600 and other equipment.Device 1600 can access the wireless network based on communication standard, such as Wi-Fi, 2G or 3G or their combination.In one exemplary embodiment, communication component 1603 receives broadcast singal or broadcast related information from external broadcasting management system via broadcast channel.In one exemplary embodiment, communication component 1603 further includes near-field communication (NFC) module, to promote short distance Communication.For example, can be realized based on radio frequency identification (RFID) technology, Infrared Data Association (IrDA) technology, ultra wide band (UWB) technology, bluetooth (BT) technology and other technologies in NFC module.
In the exemplary embodiment, device 1600 can be realized by one or more application specific integrated circuit (ASIC), digital signal processor (DSP), digital signal processing appts (DSPD), programmable logic device (PLD), field programmable gate array (FPGA), controller, microcontroller, microprocessor or other electronic components, for executing above-mentioned speech output method.
In the exemplary embodiment, a kind of non-transitorycomputer readable storage medium including instruction, the memory 1602 for example including instruction are additionally provided, above-metioned instruction can be executed by the processor 1601 of device 1600 to complete above-mentioned speech output method.For example, non-transitorycomputer readable storage medium can be ROM, random access memory (RAM), CD-ROM, tape, floppy disk and optical data storage devices etc..
The present invention also provides a kind of non-transitory computer readable recording medium, record has computer program on the medium, and described program includes the instruction for executing speech output method as described in the above embodiment the present invention.
The present invention also provides a kind of computer program, described program includes: the instruction for executing speech output method as described in the above embodiment the present invention when described program is computer-executed.
It should be understood by those skilled in the art that, the embodiment of the present invention can provide as method, system or computer program product.Therefore, the form of complete hardware embodiment, complete software embodiment or embodiment combining software and hardware aspects can be used in the present invention.Moreover, the form for the computer program product implemented in the computer-usable storage medium (including but not limited to magnetic disk storage and optical memory etc.) that one or more wherein includes computer usable program code can be used in the present invention.
The present invention be referring to according to the method for the embodiment of the present invention, the flowchart and/or the block diagram of equipment (system) and computer program product describes.It should be understood that the combination of process and/or box in each flow and/or block and flowchart and/or the block diagram that can be realized by computer program instructions in flowchart and/or the block diagram.These computer program instructions be can provide to the processor of general purpose computer, special purpose computer, Embedded Processor or other programmable data processing devices to generate a machine, so that generating by the instruction that computer or the processor of other programmable data processing devices execute for realizing the device for the function of specifying in one or more flows of the flowchart and/or one or more blocks of the block diagram.
These computer program instructions, which may also be stored in, to be able to guide in computer or other programmable data processing devices computer-readable memory operate in a specific manner, so that instruction stored in the computer readable memory generates the manufacture including command device, which realizes the function of specifying in one or more flows of the flowchart and/or one or more blocks of the block diagram.
These computer program instructions also can be loaded onto a computer or other programmable data processing device, so that series of operation steps are executed on a computer or other programmable device to generate computer implemented processing, thus the step of instruction executed on a computer or other programmable device is provided for realizing the function of specifying in one or more flows of the flowchart and/or one or more blocks of the block diagram.
Obviously, those skilled in the art various changes and modifications can be made to the invention without departing from the spirit and scope of the present invention.If then the present invention is also intended to include these modifications and variations in this way, these modifications and changes of the present invention is within the scope of the claims of the present invention and its equivalent technology.

Claims (20)

  1. A kind of speech output method characterized by comprising
    Receive the voice input content of user's input;
    According to the voice input content, the user is determined to the cognition degree of the voice input content generic, the cognition degree is the professional knowledge degree of awareness of the user to the classification;
    From at least one voice output content corresponding with the voice input content, obtains and export the voice output content to match with the cognition degree.
  2. The method according to claim 1, wherein it is described according to the voice input content, determine the user to the cognition degree of the voice input content generic, comprising:
    Identify the voiceprint of the user;
    According to the voiceprint, the voice input content for receiving the user for the first time is judged whether it is;
    When the voice input content to receive the user for the first time, determine that the user is to preset minimum cognition degree to the cognition degree of the voice input content generic.
  3. The method according to claim 1, wherein the method also includes:
    Record the input time of the voice input content and using duration, when use a length of duration received between the voice input content and the output voice output content.
  4. According to the method described in claim 3, determining the user to the cognition degree of the voice input content generic it is characterized in that, described according to the voice input content, comprising:
    Identify the voiceprint of the user;
    According to the voiceprint of the user, judge whether the adjacent voice input content received twice is inputted by same user;
    When the adjacent voice input content received twice is inputted by same user, according to the input time of the adjacent voice input content received twice and duration is used, calculates the time interval between the adjacent voice input content received twice;
    According to the time interval, determine the user to the cognition degree of the voice input content generic;Wherein, the time interval is longer, and the cognition degree is lower.
  5. According to the method described in claim 3, determining the user to the cognition degree of the voice input content generic it is characterized in that, described according to the voice input content, comprising:
    Identify the voiceprint of the user;
    According to the voiceprint of the user, history input record information corresponding with the user is obtained, the history input record information includes that history is accumulative using at least one of time, the accumulative input number of history and history input frequency information;
    According to the history input record information, determine the user to the cognition degree of the voice input content generic;Wherein, the history is accumulative longer using the time, and the cognition degree is higher;The accumulative input number of the history is more, and the cognition degree is higher;The history input frequency is higher, and the cognition degree is higher.
  6. The method according to claim 1, wherein it is described according to the voice input content, determine the user to the cognition degree of the voice input content generic, comprising:
    Extract the keyword in the voice input content;
    Determine the matching degree of the keyword and predetermined keyword in the voice input content;
    According to the matching degree of keyword and predetermined keyword in the voice input content, determine the user to the cognition degree of the voice input content generic;Wherein, the matching degree of the professional keyword in the keyword and predetermined keyword in the voice input content is higher, and the cognition degree is higher;The matching degree of the amateur keyword in keyword and predetermined keyword in the voice input content is higher, and the cognition degree is lower.
  7. The method according to claim 1, wherein it is described according to the voice input content, determine the user to the cognition degree of the voice input content generic, comprising:
    Determine that the sentence structure type of the voice input content, the sentence structure type include professional sentence structure type or amateur sentence structure type;
    According to the sentence structure type of the voice input content, determine the user to the cognition degree of the voice input content generic;Wherein, the user is higher than the cognition degree to the voice input content generic of the amateur sentence structure type to the cognition degree of the voice input content generic of the professional sentence structure type.
  8. The method according to claim 1, wherein it is described according to the voice input content, determine the user to the cognition degree of the voice input content generic, comprising:
    When determining that the adjacent voice input content received twice is inputted by same user, according to the keyword in the adjacent voice input content received twice, the degree of association between the adjacent voice input content received twice is determined;
    According to the degree of association between the adjacent voice input content received twice, determine the user to the cognition degree of the voice input content generic;Wherein, the degree of association is higher, and the cognition degree is lower.
  9. The method according to claim 1, wherein it is described according to the voice input content, determine the user to the cognition degree of the voice input content generic, comprising:
    According to the voice input content, determine at least two voices input parameters of the voice input content, the voice input parameter include: the voiceprint of the user, same user the adjacent voice input content inputted twice between time interval, history input record information corresponding with the user, the matching degree of the keyword in the voice input content and predetermined keyword, the degree of association between the sentence structure type of the voice input content and the adjacent voice input content inputted twice of same user;
    The weight that parameter is inputted according to preset each single item voice, calculates the user to the cognition degree of the voice input content generic.
  10. According to the method described in claim 9, determining the user to the cognition degree of the voice input content generic it is characterized in that, described according to the voice input content, comprising:
    When the voice that can not determine the voice input content inputs parameter, determine that the user is to preset minimum cognition degree to the cognition degree of the voice input content generic.
  11. The method according to claim 1, wherein described from corresponding with the voice input content In at least one voice output content, obtains and exports the voice output content to match with the cognition degree, comprising:
    According to the corresponding relationship between cognition degree and cognition grade, the corresponding cognition grade of the cognition degree is determined;
    According to the corresponding relationship between cognition grade and voice output content, voice output content corresponding with the cognition grade is obtained;
    Export the voice output content.
  12. According to the method described in claim 5, it is characterized in that, the method also includes:
    According to the input time of the voice input content and duration is used, updates the history input record information.
  13. The method according to claim 1, wherein the method also includes:
    The user is stored to the cognition degree of the voice input content generic;
    It is described according to the voice input content, determine the user to the cognition degree of the voice input content generic, comprising:
    Identify the voiceprint of the user;
    The user is inquired to the cognition degree of the voice input content generic according to the voiceprint of the user.
  14. A kind of instantaneous speech power, which is characterized in that described device includes:
    Processor;
    For storing the memory of the processor-executable instruction;
    Wherein, the processor is configured to executing a kind of speech output method, it is described the described method includes:
    Receive the voice input content of user's input;
    According to the voice input content, the user is determined to the cognition degree of the voice input content generic, the cognition degree is the professional knowledge degree of awareness of the user to the classification;
    From at least one voice output content corresponding with the voice input content, obtains and export the voice output content to match with the cognition degree.
  15. Device according to claim 14, wherein the processor is also configured to
    Identify the voiceprint of the user;
    According to the voiceprint, the voice input content for receiving the user for the first time is judged whether it is;
    When the voice input content to receive the user for the first time, determine that the user is to preset minimum cognition degree to the cognition degree of the voice input content generic.
  16. Device according to claim 14, wherein the processor is also configured to
    Record the input time of the voice input content and using duration, when use a length of duration received between the voice input content and the output voice output content.
  17. Device according to claim 16, wherein the processor is also configured to
    Identify the voiceprint of the user;
    According to the voiceprint of the user, judge whether the adjacent voice input content received twice is inputted by same user;
    When the adjacent voice input content received twice is inputted by same user, adjacent received twice according to described The input time of voice input content and duration is used, calculates the time interval between the adjacent voice input content received twice;
    According to the time interval, determine the user to the cognition degree of the voice input content generic;Wherein, the time interval is longer, and the cognition degree is lower.
  18. Device according to claim 16, wherein the processor is also configured to
    Identify the voiceprint of the user;
    According to the voiceprint of the user, history input record information corresponding with the user is obtained, the history input record information includes that history is accumulative using at least one of time, the accumulative input number of history and history input frequency information;
    According to the history input record information, determine the user to the cognition degree of the voice input content generic;Wherein, the history is accumulative longer using the time, and the cognition degree is higher;The accumulative input number of the history is more, and the cognition degree is higher;The history input frequency is higher, and the cognition degree is higher.
  19. Device according to claim 14, wherein the processor is also configured to
    Extract the keyword in the voice input content;
    Determine the matching degree of the keyword and predetermined keyword in the voice input content;
    According to the matching degree of keyword and predetermined keyword in the voice input content, determine the user to the cognition degree of the voice input content generic;Wherein, the matching degree of the professional keyword in the keyword and predetermined keyword in the voice input content is higher, and the cognition degree is higher;The matching degree of the amateur keyword in keyword and predetermined keyword in the voice input content is higher, and the cognition degree is lower.
  20. A kind of non-transitory computer readable recording medium, record has computer program on the medium, and described program includes the instruction for executing a kind of speech output method, which comprises receives the voice input content of user's input;
    According to the voice input content, the user is determined to the cognition degree of the voice input content generic, the cognition degree is the professional knowledge degree of awareness of the user to the classification;
    From at least one voice output content corresponding with the voice input content, obtains and export the voice output content to match with the cognition degree.
CN201680002958.1A 2015-09-08 2016-05-18 Voice output method and device Active CN107077845B (en)

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
CN2015105684308 2015-09-08
CN201510568430.8A CN105304082B (en) 2015-09-08 2015-09-08 A kind of speech output method and device
PCT/CN2016/082427 WO2017041510A1 (en) 2015-09-08 2016-05-18 Voice output method and device

Publications (2)

Publication Number Publication Date
CN107077845A true CN107077845A (en) 2017-08-18
CN107077845B CN107077845B (en) 2020-07-17

Family

ID=55201255

Family Applications (2)

Application Number Title Priority Date Filing Date
CN201510568430.8A Active CN105304082B (en) 2015-09-08 2015-09-08 A kind of speech output method and device
CN201680002958.1A Active CN107077845B (en) 2015-09-08 2016-05-18 Voice output method and device

Family Applications Before (1)

Application Number Title Priority Date Filing Date
CN201510568430.8A Active CN105304082B (en) 2015-09-08 2015-09-08 A kind of speech output method and device

Country Status (2)

Country Link
CN (2) CN105304082B (en)
WO (1) WO2017041510A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109035896A (en) * 2018-08-13 2018-12-18 广东小天才科技有限公司 Oral training method and learning equipment
CN110018843A (en) * 2018-01-09 2019-07-16 北京小度互娱科技有限公司 The test method and device of application program operation reserve

Families Citing this family (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105304082B (en) * 2015-09-08 2018-12-28 北京云知声信息技术有限公司 A kind of speech output method and device
CN106251862A (en) * 2016-07-19 2016-12-21 东莞市优陌儿智护电子科技有限公司 The implementation method of complete semantic intelligence intercommunication and system thereof
CN106649698B (en) * 2016-12-19 2020-12-22 宇龙计算机通信科技(深圳)有限公司 Information processing method and information processing device
CN107767869B (en) * 2017-09-26 2021-03-12 百度在线网络技术(北京)有限公司 Method and apparatus for providing voice service
CN107863108B (en) * 2017-11-16 2021-03-23 百度在线网络技术(北京)有限公司 Information output method and device
CN110619870B (en) * 2018-06-04 2022-05-06 佛山市顺德区美的电热电器制造有限公司 Man-machine conversation method and device, household appliance and computer storage medium
CN109036386B (en) * 2018-09-14 2021-03-16 北京网众共创科技有限公司 Voice processing method and device
CN109766411A (en) * 2019-01-14 2019-05-17 广东小天才科技有限公司 Method and system for analyzing search problems
CN111782782B (en) * 2020-06-09 2023-04-18 苏宁金融科技(南京)有限公司 Consultation reply method and device for intelligent customer service, computer equipment and storage medium
CN114398514B (en) * 2021-12-24 2022-11-22 北京达佳互联信息技术有限公司 Video display method and device and electronic equipment

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1553381A (en) * 2003-05-26 2004-12-08 杨宏惠 Multi-language correspondent list style language database and synchronous computer inter-transtation and communication
CN1838159A (en) * 2006-02-14 2006-09-27 北京未名博思生物智能科技开发有限公司 Cognition logic machine and its information processing method
CN101616221A (en) * 2008-06-25 2009-12-30 富士通株式会社 Guidance information display unit and guidance information display packing
US7725308B2 (en) * 2006-06-07 2010-05-25 Motorola, Inc. Interactive tool for semi-automatic generation of a natural language grammar from a device descriptor
WO2012155079A2 (en) * 2011-05-12 2012-11-15 Johnson Controls Technology Company Adaptive voice recognition systems and methods
CN104408099A (en) * 2014-11-18 2015-03-11 百度在线网络技术(北京)有限公司 Searching method and searching device
CN104509084A (en) * 2012-07-18 2015-04-08 Tw移动电子株式会社 System for providing contact number information having added search function, and method for same
CN104574251A (en) * 2015-01-06 2015-04-29 熊国顺 Intelligent public safety information system and application method
CN104637007A (en) * 2013-11-07 2015-05-20 大连东方之星信息技术有限公司 Statistical analysis system employing degree-of-cognition system

Family Cites Families (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7542971B2 (en) * 2004-02-02 2009-06-02 Fuji Xerox Co., Ltd. Systems and methods for collaborative note-taking
JP2006208483A (en) * 2005-01-25 2006-08-10 Sony Corp Device, method, and program for assisting survey of interesting matter of listener, and recording medium
CN101304457A (en) * 2007-05-10 2008-11-12 许罗迈 Method and apparatus for implementing automatic spoken language training based on voice telephone
US20090216757A1 (en) * 2008-02-27 2009-08-27 Robi Sen System and Method for Performing Frictionless Collaboration for Criteria Search
CN103578469A (en) * 2012-08-08 2014-02-12 百度在线网络技术(北京)有限公司 Method and device for showing voice recognition result
CN103680222B (en) * 2012-09-19 2017-10-24 镇江诺尼基智能技术有限公司 Children stories question and answer exchange method
CN103000173B (en) * 2012-12-11 2015-06-17 优视科技有限公司 Voice interaction method and device
US9269354B2 (en) * 2013-03-11 2016-02-23 Nuance Communications, Inc. Semantic re-ranking of NLU results in conversational dialogue applications
CN103594086B (en) * 2013-10-25 2016-08-17 海菲曼(天津)科技有限公司 Speech processing system, device and method
CN105304082B (en) * 2015-09-08 2018-12-28 北京云知声信息技术有限公司 A kind of speech output method and device

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1553381A (en) * 2003-05-26 2004-12-08 杨宏惠 Multi-language correspondent list style language database and synchronous computer inter-transtation and communication
CN1838159A (en) * 2006-02-14 2006-09-27 北京未名博思生物智能科技开发有限公司 Cognition logic machine and its information processing method
US7725308B2 (en) * 2006-06-07 2010-05-25 Motorola, Inc. Interactive tool for semi-automatic generation of a natural language grammar from a device descriptor
CN101616221A (en) * 2008-06-25 2009-12-30 富士通株式会社 Guidance information display unit and guidance information display packing
WO2012155079A2 (en) * 2011-05-12 2012-11-15 Johnson Controls Technology Company Adaptive voice recognition systems and methods
CN104509084A (en) * 2012-07-18 2015-04-08 Tw移动电子株式会社 System for providing contact number information having added search function, and method for same
CN104637007A (en) * 2013-11-07 2015-05-20 大连东方之星信息技术有限公司 Statistical analysis system employing degree-of-cognition system
CN104408099A (en) * 2014-11-18 2015-03-11 百度在线网络技术(北京)有限公司 Searching method and searching device
CN104574251A (en) * 2015-01-06 2015-04-29 熊国顺 Intelligent public safety information system and application method

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110018843A (en) * 2018-01-09 2019-07-16 北京小度互娱科技有限公司 The test method and device of application program operation reserve
CN110018843B (en) * 2018-01-09 2022-08-30 北京小度互娱科技有限公司 Method and device for testing application program operation strategy
CN109035896A (en) * 2018-08-13 2018-12-18 广东小天才科技有限公司 Oral training method and learning equipment

Also Published As

Publication number Publication date
CN105304082B (en) 2018-12-28
WO2017041510A1 (en) 2017-03-16
CN107077845B (en) 2020-07-17
CN105304082A (en) 2016-02-03

Similar Documents

Publication Publication Date Title
CN107077845A (en) A kind of speech output method and device
CN107644642B (en) Semantic recognition method and device, storage medium and electronic equipment
JP6482911B2 (en) Device control method and electric device
KR20200017249A (en) Apparatus and method for providing feedback for confirming intent of a user in an electronic device
KR102489914B1 (en) Electronic Device and method for controlling the electronic device
US20190221208A1 (en) Method, user interface, and device for audio-based emoji input
CN106773742A (en) Sound control method and speech control system
CN109215638B (en) Voice learning method and device, voice equipment and storage medium
CN107077510A (en) A kind of information output method and device
CN108038102A (en) Recommendation method, apparatus, terminal and the storage medium of facial expression image
CN111210844B (en) Method, device and equipment for determining speech emotion recognition model and storage medium
CN108509412A (en) A kind of data processing method, device, electronic equipment and storage medium
CN109002184A (en) A kind of association method and device of input method candidate word
CN108320208A (en) Vehicle recommends method and device
KR20210032875A (en) Voice information processing method, apparatus, program and storage medium
CN111984180A (en) Terminal screen reading method, device, equipment and computer readable storage medium
US11315573B2 (en) Speaker recognizing method, speaker recognizing apparatus, recording medium recording speaker recognizing program, database making method, database making apparatus, and recording medium recording database making program
CN111199730B (en) Voice recognition method, device, terminal and storage medium
KR20200027753A (en) Electronic Device and the Method for Operating Task corresponding to Shortened Command
US11948567B2 (en) Electronic device and control method therefor
CN109684443B (en) Intelligent interaction method and device
KR102656521B1 (en) Display device and controlling method of display device
CN105302335B (en) Vocabulary recommends method and apparatus and computer readable storage medium
CN107133280A (en) The response method and device of feedback
KR20200016774A (en) System for processing user utterance and operating method thereof

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
CP03 Change of name, title or address

Address after: Room 101, 1st floor, building 1, Xisanqi building materials City, Haidian District, Beijing 100096

Patentee after: Yunzhisheng Intelligent Technology Co.,Ltd.

Address before: 100028 12 / F, Guanjie building, building 1, yard 16, Taiyanggong Middle Road, Chaoyang District, Beijing

Patentee before: BEIJING UNISOUND INFORMATION TECHNOLOGY Co.,Ltd.

CP03 Change of name, title or address