CN113205729A - Foreign student-oriented speech evaluation method, device and system - Google Patents
Foreign student-oriented speech evaluation method, device and system Download PDFInfo
- Publication number
- CN113205729A CN113205729A CN202110389484.3A CN202110389484A CN113205729A CN 113205729 A CN113205729 A CN 113205729A CN 202110389484 A CN202110389484 A CN 202110389484A CN 113205729 A CN113205729 A CN 113205729A
- Authority
- CN
- China
- Prior art keywords
- speech
- characteristic
- evaluation
- data
- text
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000011156 evaluation Methods 0.000 title claims abstract description 106
- 230000014759 maintenance of location Effects 0.000 claims abstract description 13
- 238000000034 method Methods 0.000 claims description 55
- 239000013598 vector Substances 0.000 claims description 38
- 238000004458 analytical method Methods 0.000 claims description 23
- 230000008451 emotion Effects 0.000 claims description 20
- 230000036541 health Effects 0.000 claims description 8
- 238000004364 calculation method Methods 0.000 claims description 7
- 238000000605 extraction Methods 0.000 claims description 7
- 238000012549 training Methods 0.000 claims description 7
- 238000001514 detection method Methods 0.000 claims description 4
- 230000011218 segmentation Effects 0.000 claims description 4
- 230000002996 emotional effect Effects 0.000 claims 1
- 238000011161 development Methods 0.000 abstract description 4
- 230000006872 improvement Effects 0.000 abstract description 3
- 230000009897 systematic effect Effects 0.000 abstract description 3
- 230000000694 effects Effects 0.000 abstract description 2
- 238000006243 chemical reaction Methods 0.000 description 9
- 238000013461 design Methods 0.000 description 7
- 230000018109 developmental process Effects 0.000 description 3
- 238000010586 diagram Methods 0.000 description 3
- 230000008569 process Effects 0.000 description 3
- 238000012550 audit Methods 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 241000393496 Electra Species 0.000 description 1
- 238000013528 artificial neural network Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000001427 coherent effect Effects 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 238000012937 correction Methods 0.000 description 1
- 238000013136 deep learning model Methods 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 210000001061 forehead Anatomy 0.000 description 1
- 230000002262 irrigation Effects 0.000 description 1
- 238000003973 irrigation Methods 0.000 description 1
- 239000000463 material Substances 0.000 description 1
- 230000008520 organization Effects 0.000 description 1
- 238000012545 processing Methods 0.000 description 1
- 230000001737 promoting effect Effects 0.000 description 1
- 230000000306 recurrent effect Effects 0.000 description 1
- 230000007480 spreading Effects 0.000 description 1
- 239000013589 supplement Substances 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09B—EDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
- G09B19/00—Teaching not covered by other main groups of this subclass
- G09B19/06—Foreign languages
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/20—Natural language analysis
- G06F40/205—Parsing
- G06F40/211—Syntactic parsing, e.g. based on context-free grammar [CFG] or unification grammars
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/20—Natural language analysis
- G06F40/205—Parsing
- G06F40/216—Parsing using statistical methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/20—Natural language analysis
- G06F40/279—Recognition of textual entities
- G06F40/284—Lexical analysis, e.g. tokenisation or collocates
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q10/00—Administration; Management
- G06Q10/06—Resources, workflows, human or project management; Enterprise or organisation planning; Enterprise or organisation modelling
- G06Q10/063—Operations research, analysis or management
- G06Q10/0639—Performance analysis of employees; Performance analysis of enterprise or organisation operations
- G06Q10/06393—Score-carding, benchmarking or key performance indicator [KPI] analysis
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
- G10L25/48—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
- G10L25/51—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
- G10L25/78—Detection of presence or absence of voice signals
- G10L25/87—Detection of discrete points within a voice signal
Landscapes
- Engineering & Computer Science (AREA)
- Business, Economics & Management (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Computational Linguistics (AREA)
- Human Resources & Organizations (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Health & Medical Sciences (AREA)
- Artificial Intelligence (AREA)
- Entrepreneurship & Innovation (AREA)
- Educational Administration (AREA)
- Acoustics & Sound (AREA)
- Economics (AREA)
- Strategic Management (AREA)
- Signal Processing (AREA)
- Human Computer Interaction (AREA)
- Multimedia (AREA)
- Development Economics (AREA)
- General Business, Economics & Management (AREA)
- Game Theory and Decision Science (AREA)
- Tourism & Hospitality (AREA)
- Probability & Statistics with Applications (AREA)
- Quality & Reliability (AREA)
- Operations Research (AREA)
- Marketing (AREA)
- Educational Technology (AREA)
- Electrically Operated Instructional Devices (AREA)
Abstract
The invention discloses a lecture evaluation method, a lecture evaluation device and a lecture evaluation system for foreign retention students, wherein the lecture evaluation method, the lecture evaluation device and the lecture evaluation system comprise the following steps: acquiring speech data of a lecturer; carrying out voice recognition on the speech voice data to obtain recognition text data; respectively extracting features of the speech data and the recognition text data to obtain quantifiable evaluation features, wherein the quantifiable evaluation features comprise speech scoring features and speech piece scoring features; and evaluating the speech data of the speech according to the quantifiable evaluation characteristics to obtain an evaluation result of the speech. The invention provides objective, professional, universal and systematic spoken language ability evaluation standards for Chinese learners, and has promotion effect on the improvement of the abilities of the Chinese learners and the development of the Chinese education industry.
Description
Technical Field
The invention relates to the field of Chinese education industry and the technical field of language signal processing, in particular to a speech evaluation method, device and system for foreign students.
Background
In the spreading period of Chinese development, most Chinese learners study systematically in schools, both textbook textbooks and teacher vocabularies basically have more written languages, and the teaching and the use of spoken languages are relatively few. Therefore, the following two problems are common to Chinese learners when learning Chinese: first, there is a lack of opportunity for spoken language practice; and secondly, some Chinese learners want to do oral exercises, but cannot insist on the exercises because of no evaluation feedback. The traditional Chinese spoken language practice method has few platforms and resources in the market, because the spoken language is a very targeted course and does not have a uniform reference standard. Each individual will have a different expression and thus the evaluation feedback will be different accordingly.
Disclosure of Invention
The invention mainly aims to provide a speech evaluation method, a speech evaluation device and a speech evaluation system for foreign retention students, which provide objective, professional, universal and systematic spoken language ability evaluation standards for Chinese learners and have a promoting effect on the improvement of the self ability of the Chinese learners and the development of the Chinese education industry.
The invention adopts the following technical scheme:
in a first aspect, a speech evaluation method for foreign retention students includes:
acquiring speech data of a lecturer;
carrying out voice recognition on the speech voice data to obtain recognition text data;
respectively extracting features of the speech data and the recognition text data to obtain quantifiable evaluation features, wherein the quantifiable evaluation features comprise speech scoring features and speech piece scoring features; the voice scoring characteristics comprise fluency characteristics, effectiveness characteristics, speech speed characteristics and voice basic scoring characteristics; the sentence scoring characteristics comprise a principal and predicate analysis characteristic, a mixed sentence pattern characteristic, a proper vocabulary use characteristic, a language framework scoring characteristic, a healthy emotion characteristic, a consistent theme characteristic, a spoken language expression thinking characteristic and a sentence basic scoring characteristic;
and evaluating the speech data of the speech according to the quantifiable evaluation characteristics to obtain an evaluation result of the speech.
Preferably, the method for extracting fluency features comprises the following steps:
counting the time TD of pause occurrence in the speech data; the pause refers to the fact that the voice data of the speech are subjected to endpoint detection by adopting a VAD algorithm, the position of an endpoint of the voice is obtained, the duration between two connected endpoints is calculated, and the duration exceeds a set threshold value htIndicating that a pause occurred;
designing and outputting speech fluency S based on the number TD of pause occurrences; the value of the voice fluency S is in inverse proportion to the pause times, namely the smaller the pause times, the larger the value of the voice fluency S is;
preferably, the method for extracting the effectiveness characteristics comprises the following steps:
counting the number of pause words in the recognized text data and the occurrence times of invalid repeated voice texts; specifically, counting the number I of words paused in the text based on a pausing word list, and counting the occurrence times J of invalid repeated voice texts in the text based on a rule method;
the number of stop words and the number of occurrences of invalid repeated speech text are taken as validity characteristics.
Preferably, the method for extracting the speech rate features includes:
acquiring the text length L of the recognition text data and the audio length T of the speech voice data, and calculating the speech speed L/T;
and taking the speech rate L/T as the speech rate characteristic.
Preferably, the method for extracting the speech basic feature comprises the following steps:
obtaining the audio length T of the speech data, obtaining the audio length T and a preset threshold hvtThe ratio of (A) to (B); the preset threshold hvtThe audio length of the speech of the voice basic score can be obtained for the speaker;
the audio length T is compared with a preset threshold value hvtThe ratio of (a) is used as a speech basis feature.
Preferably, the method for extracting the syntactic analysis characteristics of the principal and predicate object comprises the following steps:
counting the number G of sentences which accord with the grammar structure information standard in the N sentences of the recognized text data to obtain the text sentence structure standard rate G/N;
and taking the standard rate G/N of the structure of the text sentence as a syntactic analysis characteristic of the principal and predicate object.
Preferably, the method for extracting the mixed sentence pattern features comprises the following steps:
counting the number M of sentences which accord with a standard sentence pattern in the N-sentence texts of the identification text data to obtain the standard rate M/N of the text sentence pattern;
and taking the standard rate M/N of the text sentence pattern as a syntactic analysis characteristic of the principal and predicate object.
Preferably, the method for extracting proper or improper vocabulary use features comprises the following steps:
counting the error quantity W of the text in the N sentences of text of the identification text data;
the text error number W is used as a word whether proper features are used.
Preferably, the method for extracting the language framework score features comprises the following steps:
splitting the recognized text data into an array by taking sentences as units, dividing words by taking words as units, matching the array after word division with a language frame dictionary, and calculating the number F of sentences conforming to the language frame;
and taking the number F of sentences which accord with the language frame as the language frame score characteristic.
Preferably, the method for extracting the health characteristics of the emotion comprises the following steps:
firstly, a large-scale corpus is used for training word vectors through word2vec, then word segmentation operation is carried out on the recognition text data, and word vectors word corresponding to each word are found outiThe difference between the directions of the two vectors of the word vector in the recognized text data and the word vector in the violation dictionary is measured through cosine distance, and the calculation method is thatWhere D is the word vector dimension, word1kBeing a vector of text words, word2kThe word vectors are word vectors in the violation dictionary library;
and taking the cosine distance as the health characteristic of the emotion.
Preferably, the method for extracting whether the theme is consistent with the features comprises the following steps:
according to the number P of the subject words matched with the identification text data and the number P of all the subject wordsallCounting the subject coincidence rate P/P of the identification text data and the subjectall;
Matching the topic compliance P/PallAs to whether the subject conforms to the characteristics.
Preferably, the method for extracting the features of the thought expressed by the spoken language comprises the following steps:
analyzing and matching the text data with a written language dictionary in the unit of sentences, and counting the number Wr of the sentences of the written language;
the sentence number Wr of the written language is taken as a spoken language expression thinking characteristic.
Preferably, the method for extracting the basic feature of the speech piece comprises the following steps:
taking the length of the identification text data as a basic score Q of a basic part of the user language, and regarding the identification text data, if the standard text length isLstandardThe standard basis of speech is scoretext_basicIdentifying the length of the text data as L, and calculating the basic score of the user language piece
And taking the user speech piece basic part Q as the speech piece basic part characteristic.
Preferably, according to the quantifiable evaluation feature, evaluating the speech data to obtain an evaluation result of the speech, including:
and respectively obtaining corresponding scores based on the extracted fluency feature, effectiveness feature, speech speed feature, speech basic feature, chief and predicate analysis feature, mixed sentence pattern feature, proper vocabulary use feature, language framework score feature, healthy emotion feature, consistent theme feature, spoken language expression thinking feature and speech basic feature, and calculating a total score through summation to realize evaluation of the speech data to obtain an evaluation result of the speech.
On the other hand, a speech evaluation device for foreign retention students comprises:
the voice data acquisition module is used for acquiring the speech data of the speaker;
the speech data recognition module is used for carrying out speech recognition on the speech data of the speech to obtain recognition text data;
the feature extraction module is used for respectively extracting features of the speech data and the recognition text data to obtain quantifiable evaluation features, and the quantifiable evaluation features comprise speech scoring features and speech piece scoring features; the voice scoring characteristics comprise fluency characteristics, effectiveness characteristics, speech speed characteristics and voice basic scoring characteristics; the sentence scoring characteristics comprise a principal and predicate analysis characteristic, a mixed sentence pattern characteristic, a proper vocabulary use characteristic, a language framework scoring characteristic, a healthy emotion characteristic, a consistent theme characteristic, a spoken language expression thinking characteristic and a sentence basic scoring characteristic;
and the evaluation module is used for evaluating the speech data according to the quantifiable evaluation characteristics to obtain the evaluation result of the speech.
In another aspect, a speech evaluation system for foreign retention students includes:
the client is used for acquiring the speech data of the speaker;
the server is used for the speech data of the speech sent by the client; carrying out voice recognition on the speech voice data to obtain recognition text data; respectively extracting features of the speech data and the recognition text data to obtain quantifiable evaluation features, wherein the quantifiable evaluation features comprise speech scoring features and speech piece scoring features; the voice scoring characteristics comprise fluency characteristics, effectiveness characteristics, speech speed characteristics and voice basic scoring characteristics; the sentence scoring characteristics comprise a principal and predicate analysis characteristic, a mixed sentence pattern characteristic, a proper vocabulary use characteristic, a language framework scoring characteristic, a healthy emotion characteristic, a consistent theme characteristic, a spoken language expression thinking characteristic and a sentence basic scoring characteristic; evaluating the speech data according to the quantifiable evaluation characteristics to obtain an evaluation result of the speech; and sending the evaluation result to a client for display.
Compared with the prior art, the invention has the following beneficial effects:
(1) the spoken language ability method for the Chinese learner is objective in standard, professional, universal and systematic, and a method based on a corpus is adopted to design a calculation method of each evaluation characteristic; the invention can promote the self-ability improvement of Chinese learners and the development of Chinese education industry;
(2) after learning the oral practice requirements of the current Chinese learner, the device and the system of the invention are designed according to the learning concept of 'the language is not learned but is practiced'; the invention is a spoken language practice system with convenient operation, flexible application and integrated design, and the practice can be carried out on the WeChat small program at any time and any place as long as one smart phone is provided; the invention can meet the individual requirements of the current Chinese learner and guides the spoken language of the user in a targeted manner; the device is different from common spoken language dialogue practice which has the characteristic of random use and aims to realize simple daily communication; the invention can not only train the spoken language ability of Chinese learners, but also improve the level of other skills such as Chinese thinking ability of the learners; the invention aims to provide students with the opportunity to speak Chinese, avoid the phenomenon of 'dumb Chinese', train the spoken language expression ability of the students and finally achieve the effect of fully developing four language skills of listening, speaking, reading and writing of Chinese learners.
The above description is only an overview of the technical solutions of the present invention, and the embodiments of the present invention are described below in order to make the technical means of the present invention more clearly understood and to make the description of the technical means more comprehensible.
The above and other objects, advantages and features of the present invention will become more apparent to those skilled in the art from the following detailed description of specific embodiments thereof, taken in conjunction with the accompanying drawings.
Drawings
Fig. 1 is an interface schematic diagram of a lecture evaluation method for foreign retention students according to an embodiment of the present invention;
FIG. 2 is a flowchart of a lecture evaluation method for foreign reservation students according to an embodiment of the present invention;
fig. 3 is a block diagram of the speech evaluation device for foreign retention students according to the embodiment of the present invention;
fig. 4 is a block diagram of a speech evaluation system for foreign reservation students according to an embodiment of the present invention.
Detailed Description
The invention will be further illustrated with reference to the following specific examples. It should be understood that these examples are for illustrative purposes only and are not intended to limit the scope of the present invention. Further, it should be understood that various changes or modifications of the present invention may be made by those skilled in the art after reading the teaching of the present invention, and such equivalents may fall within the scope of the present invention as defined in the appended claims.
Referring to fig. 1 and 2, the speech evaluation method for foreign retention students of the present invention includes the following steps: (1) acquiring text data of user voice; (2) and calculating the speech capability evaluation index.
The method comprises the following steps: text data of a user's voice is acquired.
The method specifically comprises the following steps:
s201, obtaining speech data of a speaker;
s202, carrying out voice recognition on the speech voice data to obtain recognition text data.
In the invention, based on a lecture evaluation method for foreign students, scene corpora corresponding to 3 evaluation dimensions are designed according to different levels of topics, related prompts of the topics, and model texts and written expression corpora of the topics.
1. Theme design of different levels. 46 corpora are designed (it should be noted that the corpus can be expanded according to needs, and the following is only an example), and the types such as "my weekend", "my vacation arrangement" and the like are simple in one degree and difficult in another degree. The theme design at different levels is shown in table 1 below, where difficiltly columns represent different levels and the me column represents a theme.
TABLE 1
2. Corpus design of relevant tips of subjects. 5 sets of related prompt corpuses of different subjects are designed (it should be noted that the related prompt corpuses of the subjects can be expanded according to needs, and the following is only an example). The tips 1-5 columns represent the prompt corpuses associated with the corresponding topics, respectively, as shown in Table 2 below.
TABLE 2
3. The model and written expression language material design of the subject. A set of corpora is designed 46, and the corpus schematic of the design is shown in the following table (it should be noted that the corpus schematic can be expanded as desired, and the following is only an example). Where Model is a Model and expression is a written expression for similarity comparison between the user's speech and text, as shown in table 3 below.
TABLE 3
And respectively displaying the designed small speech contents to the user, thereby obtaining 3 different conversation data sets such as corresponding speech audio, text evaluation, speech score and the like. Taking the partial information of the user session data set as an example, the content topic is "my weekend", as shown in table 4 below:
TABLE 4
As shown in the above table, the results of the actual performance of the 3 users in the speech evaluation system are included, different users can select different topics, and the language expression capabilities and levels of different users are different, so that the total word number and the speech speed of speech are different, the contents are also significantly different, and the final scores and the total score are also different.
The second step is that: and calculating the speech capability evaluation index.
The method specifically comprises the following steps:
s203, respectively carrying out feature extraction on the speech data and the recognition text data to obtain quantifiable evaluation features, wherein the quantifiable evaluation features comprise speech scoring features and speech piece scoring features; the voice scoring characteristics comprise fluency characteristics, effectiveness characteristics, speech speed characteristics and voice basic scoring characteristics; the sentence scoring characteristics comprise a principal and predicate analysis characteristic, a mixed sentence pattern characteristic, a proper vocabulary use characteristic, a language framework scoring characteristic, a healthy emotion characteristic, a consistent theme characteristic, a spoken language expression thinking characteristic and a sentence basic scoring characteristic;
and S204, evaluating the speech data according to the quantifiable evaluation characteristics to obtain the evaluation result of the speech.
In this embodiment, the method for extracting fluency features includes:
counting the times of pause occurrence in the speech data; specifically, the speech data is subjected to end point detection by VAD algorithm, and a threshold value h is settTime between endpoints over htA pause is indicated;
outputting voice fluency S as fluency characteristics based on the number of times of pause occurrence; the value range of the voice fluency S is [0,1], and is inversely proportional to the pause times, namely, the smaller the pause times, the larger the value of the voice fluency S.
Specifically, fluency characteristics (indexes) measure the thinking power and language organization ability of Chinese learners, and are used for carrying out fluency scoring on spoken texts spoken by users and input voices of the users, and scoring dimensionality is mainly carried out according to pause, speech speed and intonation used for the voices. Performing endpoint detection on voice information by adopting VAD algorithm to obtain the endpoint position of voice, and solving the time length between two connected endpoints, wherein the time length exceeds a set threshold value htIndicating that a pause has occurred. Big data training and calculation are carried out on the existing data set through a recurrent neural network to obtain a fluency judging model, the user input voice is given, the voice fluency S is output, and the value range of the voice fluency is [0,1]]The larger the value, the better.
In this embodiment, the method for extracting the validity characteristic includes:
counting the number of pause words in the recognized text data and the occurrence times of invalid repeated voice texts; specifically, counting the number I of words paused in the text based on a pausing word list, and counting the occurrence times J of invalid repeated voice texts in the text based on a rule method;
the number of stop words and the number of occurrences of invalid repeated speech text are taken as validity characteristics.
Specifically, the validity characteristics (indexes) determine the occurrence times of the stop words and the invalid repeated voice texts of the Chinese learner in the speaking process. For Chinese speech, when the Chinese learner does not express smoothly, the Chinese speech utters words such as 'forehead', 'click' and 'o', which are called pause words, and speech texts which appear continuously for multiple times are called invalid repeated speech texts. For the text converted from the user voice, counting the number I of pause words in the text based on the pause word list, and counting the occurrence times J of invalid repeated voice text in the text based on a rule method. The numeric area of the number of the stop words and the occurrence frequency of the invalid repeated voice text is [0, ∞ ], and the smaller the value is, the better the value is.
In this embodiment, the method for extracting speech rate features includes:
acquiring the text length L of the recognition text data and the audio length T of the speech voice data, and calculating the speech speed L/T;
and taking the speech rate L/T as the speech rate characteristic.
Specifically, the speech speed characteristics (indexes) are used for judging the Chinese oral proficiency of the Chinese learner. Obtaining an evaluation text of a user, obtaining the text length L input by the user and the audio length T input by the user, and calculating the speech speed L/T, wherein the value range of the speech speed is [0, ∞]Having a value of [ v ]low,vhigh]Is preferred wherein vlowIs the required minimum speech rate threshold, vhighIs the required maximum speech rate threshold.
In this embodiment, the method for extracting the speech basic feature includes:
obtaining the audio length T of the speech data, obtaining the audio length T and a preset threshold hvtThe ratio of (A) to (B); the preset threshold hvtThe audio length of the speech of the voice basic score can be obtained for the speaker;
the audio length T is compared with a preset threshold value hvtThe ratio of (a) is used as a speech basis feature.
Specifically, the speech basic sub-characteristics (indexes) measure the basic ability of spoken speech of the Chinese learner. Setting a threshold h for the length of audio input by a uservtWhen the user audio length exceeds a threshold hvtGiving the user full score of the voice base, the voice length value range is [0, ∞]The larger the value, the better.
In this embodiment, the method for extracting the syntactic analysis feature of the principal and predicate object includes:
counting the number G of sentences which accord with the grammar structure information standard in the N sentences of the recognized text data to obtain the text sentence structure standard rate G/N;
and taking the standard rate G/N of the structure of the text sentence as a syntactic analysis characteristic of the principal and predicate object.
Specifically, the syntactic analysis characteristics (indexes) of the principal and predicate guest measure the Chinese grammar ability of the user, the evaluation text input by the user is divided into an array by taking a sentence as a unit, and meanwhile, the word is divided by taking a word as a unit. The sentence is then parsed through a dependency syntax tree to determine the syntactic structure of the sentence or the dependency between the words in the sentence. And further extracting the main and subordinate object shape complementing relationship of the sentence according to the dependency relationship. And a correct syntactic structure dictionary is counted by combining a specialist method with a technical counting mode, the main meaning object shape complementing structure information analyzed by taking a sentence as a unit is sent to the dictionary for matching of the inclusion relationship, if the main meaning object shape complementing structure information analyzed by the sentence has matching information in the dictionary, the main meaning object shape complementing structure information is regarded as the syntactic structure information standard of the sentence, and if the main meaning object shape complementing structure information does not have matching information, the system judges that the syntactic structure information has a problem. And for N sentences of texts input by the user, calculating the sentence quantity G meeting the grammar structure information standard to obtain the text sentence structure standard rate G/N, wherein the value range of the text sentence structure standard rate is [0,1], and the larger the value is, the better the value is.
In this embodiment, the method for extracting the features of the mixed sentence pattern includes:
counting the number M of sentences which accord with a standard sentence pattern in the N-sentence texts of the identification text data to obtain the standard rate M/N of the text sentence pattern;
and taking the standard rate M/N of the text sentence pattern as a syntactic analysis characteristic of the principal and predicate object.
Specifically, the mixed sentence pattern characteristics (indexes) judge the Chinese grammar application capability of the user. The method comprises the steps of counting a dictionary of a standard sentence pattern according to the standard sentence pattern given in International Chinese teaching general course outline, analyzing a text input by a user in a sentence unit, dividing a sentence into words, analyzing the words from three angles of main meaning object fixed form supplement, word part of speech and key words, sending information of three dimensions into the dictionary of the standard sentence pattern for accurate matching, and if the standard sentence pattern is matched, regarding the sentence as the standard sentence pattern, otherwise, not the standard sentence pattern. For N sentence texts input by a user, calculating the number M of sentences conforming to the standard sentence pattern to obtain the standard rate M/N of the text sentence pattern, wherein the value range of the standard rate of the text sentence pattern is [0,1], and the larger the value is, the better the value is.
In this embodiment, the method for extracting proper vocabulary use features includes:
counting the error quantity W of the text in the N sentences of text of the identification text data;
the text error number W is used as a word whether proper features are used.
Specifically, whether proper characteristics (indexes) are used for measuring the phrase correlation ability of a user, the text input by the user is corrected through an expert dictionary method and an open source framework pycorector, the expert dictionary method is used for counting error-prone word collocation, such as wrong collocation of 'wearing clothes' and 'wearing hat', and the like, by experts in the Chinese education field, further, when a program is processed, whether error collocation exists in the text is traversed, and if the error collocation exists, the text is regarded as a text error. The open source framework pycorrctor detects the position of the wrongly written character according to the language model, and corrects the wrongly written character through the phonetic sound similar characteristic, the stroke five-stroke editing distance characteristic and the language model confusion characteristic. The technology uses a statistical language model kenlm tool, and simultaneously uses rnn _ attention, rnn _ crf, seq2seq _ attention, transformer, conv _ seq2seq and electra pre-training models to perform joint training, and a deep learning model for text error correction is trained to be used for detecting text error information input by a user. And for N sentences of texts input by the user, calculating the text error number W, wherein the text error number has a value range of [0, ∞ ], and the smaller the value is, the better the value is.
In this embodiment, the method for extracting the language framework score feature includes:
splitting the recognized text data into an array by taking sentences as units, dividing words by taking words as units, matching the array after word division with a language frame dictionary, and calculating the number F of sentences conforming to the language frame;
and taking the number F of sentences which accord with the language frame as the language frame score characteristic.
Specifically, the language framework score features (indexes) measure the language logic ability of the user, and a language framework dictionary is arranged by a specialist method, such as "though. And splitting the evaluation text input by the user into an array by taking the sentence as a unit, dividing the word by taking the word as a unit, matching the array after word division with a language frame dictionary, and calculating the number F of the sentences conforming to the language frame. The number of the language frame sentences is in a value range of [0, N ], and the larger the value is, the better the value is.
In this embodiment, the method for extracting whether emotion is healthy includes:
firstly, a large-scale corpus is used for training word vectors through word2vec, then word segmentation operation is carried out on the recognition text data, and word vectors word corresponding to each word are found outiThe difference between the directions of the two vectors of the word vector in the recognized text data and the word vector in the violation dictionary is measured through cosine distance, and the calculation method is thatWhere D is the word vector dimension, word1kBeing a vector of text words, word2kThe word vectors are word vectors in the violation dictionary library;
and taking the cosine distance as the health characteristic of the emotion.
Specifically, judging whether the expressed sentences of the user accord with the core value view of social connotation by emotion health characteristics (indexes), performing text health audit aiming at texts input by the user, covering rich audit dimensions such as political involvement, yellow involvement, terrorism involvement, malicious promotion, low-quality irrigation, official illegal content library and the like, and sorting out an illegal dictionary library by a specialist law; firstly, a large-scale corpus is used for training word vectors through word2vec, then word segmentation operation is carried out on input texts, and word vectors word corresponding to each word are found outiThe difference between the direction of the word vector in the input text and the direction of the two vectors of the word vector in the violation dictionary is measured by cosine distance, and the calculation method is thatWhere D is the word vector dimension, word1kBeing a vector of text words, word2kAnd the word vectors are word vectors in the violation dictionary library. The cosine value range of the specified included angle is [0,1]]The larger the cosine value is, the smaller the included angle between the two vectors is, namely the two vectors are more similar, and when the similarity is larger than the threshold hdIt indicates that the word violates the core value view of social meaning, with smaller values being better.
In this embodiment, the method for extracting whether the theme conforms to the feature includes:
according toThe number P of the subject words matched with the identification text data and the number P of all the subject wordsallCounting the subject coincidence rate P/P of the identification text data and the subjectall;
Matching the topic compliance P/PallAs to whether the subject conforms to the characteristics.
Specifically, whether the theme conforms to the characteristics (indexes) or not is evaluated, whether the statement stated by a user is the same as the theme of the theme given by the system or not is judged, the subject words of the theme are counted according to the corresponding theme through a professional method, the subject words are stored by dictionary information, if the evaluation text input by the user is accurately matched with the subject word information listed by an expert, the number P of the subject words matched with the evaluation text of the user and the number P of all the subject words are evaluated according to the userallFurther, the topic coincidence rate P/P of the text and the topic is countedallThe subject match rate value range is [0,1]]The larger the value, the better.
In this embodiment, the method for extracting the thought characteristics of spoken language expression includes:
analyzing and matching the text data with a written language dictionary in the unit of sentences, and counting the number Wr of the sentences of the written language;
the sentence number Wr of the written language is taken as a spoken language expression thinking characteristic.
Specifically, the spoken language expression thinking characteristics (indexes) judge the thinking strength of the user, the spoken language expression thinking of the user is analyzed according to the question examination time of the user, the spoken language proportion of the user and the use proportion dimensionality of written languages, the question reading time of the user is captured according to the front end of the small program, and the added branch written languages for summarizing the problem are arranged into a written language dictionary in a specialist method. Analyzing and matching the written language dictionary for the input text of the user in the unit of sentences, and counting the sentences of the written language to obtain the written language use ratio Wr, wherein Wr is the number of the sentences of the written language. The written language is used in a ratio of [0, ∞ ], with larger values being better.
In this embodiment, the method for extracting a speech piece basic feature includes:
using the length of the identification text data as a user languageA basic score Q of the basic score, for the recognized text data, if the standard text length is LstandardThe standard basis of speech is scoretext_basicIdentifying the length of the text data as L, and calculating the basic score of the user language piece
And taking the user speech piece basic part Q as the speech piece basic part characteristic.
Specifically, the basic feature (index) of the speech piece measures the basic capability of the spoken language piece of the Chinese learner, the text length is evaluated as the basic score Q of the basic part of the speech piece of the user according to the input of the user, and for the evaluation text input by the user, if the standard text length is LstandardThe standard basis of speech is scoretext_basicInputting text with length L, calculating user speech piece basic scoreThe text length is in the range of [0, ∞]The larger the value, the better. The basic score of the language ensures that the user has a certain basic score.
For the user 162514089, based on the above feature extraction method, the related task completion capability index calculation values are respectively as follows:
(1) fluency: the index value range is [0,1], and the larger the value, the better. For the spoken voice audio input by the user 162514089, the fluency value is 0.83, which indicates that the user is fluent during expression and has good spoken language expression capability.
(2) Effectiveness: the index is calculated by the occurrence frequency of stop words and invalid repeated voice texts of Chinese learners in the speaking process, the value range of the occurrence frequency of the stop words and the invalid repeated voice texts is [0, ∞ ], and the smaller the value is, the better the value is. Aiming at the spoken voice audio frequency input by the user 162514089, the occurrence times of stop words and invalid repeated voice texts are all 0, which indicates that the user has coherent thinking during expression, smooth expression and better spoken language expression capability.
(3) The speed of speech: the index value range is [0, ∞]Having a value of [ v ]low,vhigh]Is preferred wherein vlowIs the required minimum speech rate threshold, vhighIs the required maximum speech rate threshold. For the spoken voice audio input by the user 162514089, the value of the speed of speech is 4.30 words/second, which is close to the native Chinese language user, and the spoken language expression capability of the user is better.
(4) The voice basic score is as follows: the voice basic score is measured through the voice length, the index value range is [0, ∞ ], and the larger the value is, the better the value is. For the spoken voice audio input by the user 162514089, the user voice length is 19.77 seconds, which indicates that the user expression content is not sufficient and fails to reach the suggested voice length.
(5) And (3) carrying out syntactic analysis on the major and predicate guest: the index value field is [0,1], the parsing value of the predicate object is 1.00 aiming at the text of voice audio conversion input by the user 162514089, and the result shows that no grammar error occurs to the user, the grammar knowledge of the user is rich, and the expression accords with the Chinese expression habit.
(6) Mixing sentence patterns: the index value field is [0,1], and the value of the mixed sentence pattern is 0.75 aiming at the text of the voice audio conversion input by the user 162514089, which indicates that the user uses some standard sentence patterns in the expression process and the expression mode is reasonable.
(7) Whether the vocabulary is used properly: the index value range is [0, ∞ ], and whether the vocabulary is properly used is 1.00 for the text of the voice audio conversion input by the user 162514089, which indicates that the user has no wrong word collocation and has a proper expression mode.
(8) Language framework: the index value field is [0, N ], where N is the number of sentences in the text, and the value of the language frame is 0 for the text of the voice audio conversion input by the user 162514089, which indicates that the sentence expression such as "albeit.
(9) Whether the emotion is healthy or not: the index value field is [0,1], and the value of whether emotion is healthy or not is 0.99 for the text of voice audio conversion input by the user 162514089, which indicates that the user expresses idea health and does not have content such as political, yellow, terrorism, malicious popularization and the like.
(10) Whether the topics are consistent: the index value field is [0,1], and whether the theme is consistent or not is 1.00 for the text of the voice audio conversion input by the user 162514089, so that the user can be explained on the basis of understanding the theme by closely fastening the theme during expression.
(11) Spoken language expresses thinking: the index value range is [0, ∞ ], the value of the spoken language expression thinking power for the text of the voice audio conversion input by the user 162514089 is 2, the user is explained to think during expression, some common words are replaced by higher-level words, and the richness of the words mastered by the user is reflected.
(12) The basic part of the language: the basic score of the speech segment is measured by the text length of the speech audio conversion, the index value field is [0, ∞ ], and the text length value of the text of the speech audio conversion input by the user 162514089 is 85, which indicates that the user expression content is not sufficient and cannot reach the suggested text length.
In this embodiment, evaluating the speech data according to the quantifiable evaluation feature to obtain an evaluation result of the speech, including:
and respectively obtaining corresponding scores based on the extracted fluency feature, effectiveness feature, speech speed feature, speech basic feature, chief and predicate analysis feature, mixed sentence pattern feature, proper vocabulary use feature, language framework score feature, healthy emotion feature, consistent theme feature, spoken language expression thinking feature and speech basic feature, and calculating a total score through summation to realize evaluation of the speech data to obtain an evaluation result of the speech.
Specifically, the following table 5 shows the total score evaluation method designed by the present invention.
TABLE 5
Referring to fig. 3, a speech evaluation apparatus for foreign retention students includes:
a voice data obtaining module 301, configured to obtain speech data of a speaker;
a voice data recognition module 302, configured to perform voice recognition on the speech data to obtain recognition text data;
the feature extraction module 303 is configured to perform feature extraction on the speech data and the recognition text data to obtain quantifiable evaluation features, where the quantifiable evaluation features include speech scoring features and speech piece scoring features; the voice scoring characteristics comprise fluency characteristics, effectiveness characteristics, speech speed characteristics and voice basic scoring characteristics; the sentence scoring characteristics comprise a principal and predicate analysis characteristic, a mixed sentence pattern characteristic, a proper vocabulary use characteristic, a language framework scoring characteristic, a healthy emotion characteristic, a consistent theme characteristic, a spoken language expression thinking characteristic and a sentence basic scoring characteristic;
and the evaluating module 304 is used for evaluating the speech data according to the quantifiable evaluating characteristics to obtain a speech evaluating result.
Referring to fig. 4, a speech evaluation system for foreign retention students includes:
a client 401, configured to obtain speech data of a speaker;
a server 402, configured to send the speech data of the speech by the client; carrying out voice recognition on the speech voice data to obtain recognition text data; respectively extracting features of the speech data and the recognition text data to obtain quantifiable evaluation features, wherein the quantifiable evaluation features comprise speech scoring features and speech piece scoring features; the voice scoring characteristics comprise fluency characteristics, effectiveness characteristics, speech speed characteristics and voice basic scoring characteristics; the sentence scoring characteristics comprise a principal and predicate analysis characteristic, a mixed sentence pattern characteristic, a proper vocabulary use characteristic, a language framework scoring characteristic, a healthy emotion characteristic, a consistent theme characteristic, a spoken language expression thinking characteristic and a sentence basic scoring characteristic; evaluating the speech data according to the quantifiable evaluation characteristics to obtain an evaluation result of the speech; and sending the evaluation result to a client for display.
The foregoing description of specific embodiments of the present invention has been presented. It is to be understood that the present invention is not limited to the specific embodiments described above, and that various changes and modifications may be made by one skilled in the art within the scope of the appended claims without departing from the spirit of the invention.
Claims (10)
1. A speech evaluation method for foreign retention students is characterized by comprising the following steps:
acquiring speech data of a lecturer;
carrying out voice recognition on the speech voice data to obtain recognition text data;
respectively extracting features of the speech data and the recognition text data to obtain quantifiable evaluation features, wherein the quantifiable evaluation features comprise speech scoring features and speech piece scoring features; the voice scoring characteristics comprise fluency characteristics, effectiveness characteristics, speech speed characteristics and voice basic scoring characteristics; the sentence scoring characteristics comprise a principal and predicate analysis characteristic, a mixed sentence pattern characteristic, a proper vocabulary use characteristic, a language framework scoring characteristic, a healthy emotion characteristic, a consistent theme characteristic, a spoken language expression thinking characteristic and a sentence basic scoring characteristic;
and evaluating the speech data of the speech according to the quantifiable evaluation characteristics to obtain an evaluation result of the speech.
2. The foreign reservation student oriented speech evaluation method according to claim 1, wherein the fluency feature extraction method comprises:
counting the time TD of pause occurrence in the speech data; pause refers to a pairPerforming endpoint detection on the speech data by adopting a VAD algorithm to acquire the endpoint position of the speech, and solving the time length between two connected endpoints which exceeds a set threshold value htIndicating that a pause occurred;
designing and outputting speech fluency S based on the number TD of pause occurrences; the value of the voice fluency S is in inverse proportion to the pause times, namely the smaller the pause times, the larger the value of the voice fluency S is;
a method of extracting significance signatures, comprising:
counting the number of pause words in the recognized text data and the occurrence times of invalid repeated voice texts; specifically, counting the number I of words paused in the text based on a pausing word list, and counting the occurrence times J of invalid repeated voice texts in the text based on a rule method;
the number of stop words and the number of occurrences of invalid repeated speech text are taken as validity characteristics.
3. The foreign reservation student oriented speech evaluation method according to claim 1, wherein the method for extracting speech rate features comprises:
acquiring the text length L of the recognition text data and the audio length T of the speech voice data, and calculating the speech speed L/T;
taking the speech rate L/T as a speech rate characteristic;
the method for extracting the voice basic feature comprises the following steps:
obtaining the audio length T of the speech data, obtaining the audio length T and a preset threshold hvtThe ratio of (A) to (B); the preset threshold hvtThe audio length of the speech of the voice basic score can be obtained for the speaker;
the audio length T is compared with a preset threshold value hvtThe ratio of (a) is used as a speech basis feature.
4. The foreign reservation student oriented speech evaluation method according to claim 1, wherein the method for extracting the syntactic analysis characteristics of the principal and predicate guest comprises the following steps:
counting the number G of sentences which accord with the grammar structure information standard in the N sentences of the recognized text data to obtain the text sentence structure standard rate G/N;
taking the standard rate G/N of the structure of the text sentence as the syntactic analysis characteristic of the principal and predicate object;
the method for extracting the mixed sentence pattern features comprises the following steps:
counting the number M of sentences which accord with a standard sentence pattern in the N-sentence texts of the identification text data to obtain the standard rate M/N of the text sentence pattern;
and taking the standard rate M/N of the text sentence pattern as a syntactic analysis characteristic of the principal and predicate object.
5. The speech evaluation method for the foreign reservation student according to claim 1, wherein the method for extracting the proper feature of vocabulary use comprises:
counting the error quantity W of the text in the N sentences of text of the identification text data;
using the text error number W as a word to determine whether the word is proper or not;
the method for extracting the language framework score features comprises the following steps:
splitting the recognized text data into an array by taking sentences as units, dividing words by taking words as units, matching the array after word division with a language frame dictionary, and calculating the number F of sentences conforming to the language frame;
and taking the number F of sentences which accord with the language frame as the language frame score characteristic.
6. The foreign reservation student oriented speech evaluation method according to claim 1, wherein the method for extracting the emotional health characteristic comprises the following steps:
firstly, a large-scale corpus is used for training word vectors through word2vec, then word segmentation operation is carried out on the recognition text data, and word vectors word corresponding to each word are found outiThe difference between the directions of the two vectors of the word vector in the recognized text data and the word vector in the violation dictionary is measured through cosine distance, and the calculation method is thatWhere D is the word vector dimension, word1kBeing a vector of text words, word2kThe word vectors are word vectors in the violation dictionary library;
taking the cosine distance as the health characteristic of the emotion;
the method for extracting whether the theme conforms to the characteristics comprises the following steps:
according to the number P of the subject words matched with the identification text data and the number P of all the subject wordsallCounting the subject coincidence rate P/P of the identification text data and the subjectall;
Matching the topic compliance P/PallAs to whether the subject conforms to the characteristics.
7. The foreign reservation student oriented speech evaluation method according to claim 1, wherein the method of extracting the spoken language expression thinking power feature comprises:
analyzing and matching the text data with a written language dictionary in the unit of sentences, and counting the number Wr of the sentences of the written language;
taking the sentence number Wr of the written language as the thinking characteristic of the spoken language expression;
the method for extracting the basic feature of the speech piece comprises the following steps:
taking the length of the identification text data as a basic score Q of a basic part of the user language, and regarding the identification text data, if the standard text length is LstandardThe standard basis of speech is scoretext_basicIdentifying the length of the text data as L, and calculating the basic score of the user language piece
And taking the user speech piece basic part Q as the speech piece basic part characteristic.
8. The lecture evaluation method for foreign retention students according to claim 1, wherein the evaluation of the lecture voice data according to the quantifiable evaluation features to obtain the evaluation result of the lecture comprises:
and respectively obtaining corresponding scores based on the extracted fluency feature, effectiveness feature, speech speed feature, speech basic feature, chief and predicate analysis feature, mixed sentence pattern feature, proper vocabulary use feature, language framework score feature, healthy emotion feature, consistent theme feature, spoken language expression thinking feature and speech basic feature, and calculating a total score through summation to realize evaluation of the speech data to obtain an evaluation result of the speech.
9. The utility model provides a speech evaluation device towards foreign student, its characterized in that includes:
the voice data acquisition module is used for acquiring the speech data of the speaker;
the speech data recognition module is used for carrying out speech recognition on the speech data of the speech to obtain recognition text data;
the feature extraction module is used for respectively extracting features of the speech data and the recognition text data to obtain quantifiable evaluation features, and the quantifiable evaluation features comprise speech scoring features and speech piece scoring features; the voice scoring characteristics comprise fluency characteristics, effectiveness characteristics, speech speed characteristics and voice basic scoring characteristics; the sentence scoring characteristics comprise a principal and predicate analysis characteristic, a mixed sentence pattern characteristic, a proper vocabulary use characteristic, a language framework scoring characteristic, a healthy emotion characteristic, a consistent theme characteristic, a spoken language expression thinking characteristic and a sentence basic scoring characteristic;
and the evaluation module is used for evaluating the speech data according to the quantifiable evaluation characteristics to obtain the evaluation result of the speech.
10. A speech evaluation system for foreign retention students is characterized by comprising:
the client is used for acquiring the speech data of the speaker;
the server is used for the speech data of the speech sent by the client; carrying out voice recognition on the speech voice data to obtain recognition text data; respectively extracting features of the speech data and the recognition text data to obtain quantifiable evaluation features, wherein the quantifiable evaluation features comprise speech scoring features and speech piece scoring features; the voice scoring characteristics comprise fluency characteristics, effectiveness characteristics, speech speed characteristics and voice basic scoring characteristics; the sentence scoring characteristics comprise a principal and predicate analysis characteristic, a mixed sentence pattern characteristic, a proper vocabulary use characteristic, a language framework scoring characteristic, a healthy emotion characteristic, a consistent theme characteristic, a spoken language expression thinking characteristic and a sentence basic scoring characteristic; evaluating the speech data according to the quantifiable evaluation characteristics to obtain an evaluation result of the speech; and sending the evaluation result to a client for display.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110389484.3A CN113205729A (en) | 2021-04-12 | 2021-04-12 | Foreign student-oriented speech evaluation method, device and system |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110389484.3A CN113205729A (en) | 2021-04-12 | 2021-04-12 | Foreign student-oriented speech evaluation method, device and system |
Publications (1)
Publication Number | Publication Date |
---|---|
CN113205729A true CN113205729A (en) | 2021-08-03 |
Family
ID=77026561
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202110389484.3A Pending CN113205729A (en) | 2021-04-12 | 2021-04-12 | Foreign student-oriented speech evaluation method, device and system |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN113205729A (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114187544A (en) * | 2021-11-30 | 2022-03-15 | 厦门大学 | College English speaking multi-mode automatic scoring method |
CN117787921A (en) * | 2024-02-27 | 2024-03-29 | 北京烽火万家科技有限公司 | Intelligent education training management method and identity anti-counterfeiting method for intelligent education training |
Citations (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20090258333A1 (en) * | 2008-03-17 | 2009-10-15 | Kai Yu | Spoken language learning systems |
CN101740024A (en) * | 2008-11-19 | 2010-06-16 | 中国科学院自动化研究所 | Method for automatic evaluation based on generalized fluent spoken language fluency |
CN101739868A (en) * | 2008-11-19 | 2010-06-16 | 中国科学院自动化研究所 | Automatic evaluation and diagnosis method of text reading level for oral test |
CN101826263A (en) * | 2009-03-04 | 2010-09-08 | 中国科学院自动化研究所 | Objective standard based automatic oral evaluation system |
CN106847263A (en) * | 2017-01-13 | 2017-06-13 | 科大讯飞股份有限公司 | Speech level evaluation method and apparatus and system |
JP2018045062A (en) * | 2016-09-14 | 2018-03-22 | Kddi株式会社 | Program, device and method automatically grading from dictation voice of learner |
CN108122561A (en) * | 2017-12-19 | 2018-06-05 | 广东小天才科技有限公司 | Spoken language voice evaluation method based on electronic equipment and electronic equipment |
CN109147765A (en) * | 2018-11-16 | 2019-01-04 | 安徽听见科技有限公司 | Audio quality comprehensive evaluating method and system |
CN109785698A (en) * | 2017-11-13 | 2019-05-21 | 上海流利说信息技术有限公司 | Method, apparatus, electronic equipment and medium for spoken language proficiency evaluation and test |
CN110069784A (en) * | 2019-05-05 | 2019-07-30 | 广东电网有限责任公司 | A kind of voice quality inspection methods of marking, device, terminal and can storage medium |
CN110136721A (en) * | 2019-04-09 | 2019-08-16 | 北京大米科技有限公司 | A kind of scoring generation method, device, storage medium and electronic equipment |
CN110675292A (en) * | 2019-09-23 | 2020-01-10 | 浙江优学智能科技有限公司 | Child language ability evaluation method based on artificial intelligence |
WO2020149621A1 (en) * | 2019-01-14 | 2020-07-23 | 김주혁 | English speaking evaluation system and method |
CN111833853A (en) * | 2020-07-01 | 2020-10-27 | 腾讯科技(深圳)有限公司 | Voice processing method and device, electronic equipment and computer readable storage medium |
US20210050001A1 (en) * | 2019-08-16 | 2021-02-18 | Ponddy Education Inc. | Systems and Methods for Comprehensive Chinese Speech Scoring and Diagnosis |
-
2021
- 2021-04-12 CN CN202110389484.3A patent/CN113205729A/en active Pending
Patent Citations (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20090258333A1 (en) * | 2008-03-17 | 2009-10-15 | Kai Yu | Spoken language learning systems |
CN101740024A (en) * | 2008-11-19 | 2010-06-16 | 中国科学院自动化研究所 | Method for automatic evaluation based on generalized fluent spoken language fluency |
CN101739868A (en) * | 2008-11-19 | 2010-06-16 | 中国科学院自动化研究所 | Automatic evaluation and diagnosis method of text reading level for oral test |
CN101826263A (en) * | 2009-03-04 | 2010-09-08 | 中国科学院自动化研究所 | Objective standard based automatic oral evaluation system |
JP2018045062A (en) * | 2016-09-14 | 2018-03-22 | Kddi株式会社 | Program, device and method automatically grading from dictation voice of learner |
CN106847263A (en) * | 2017-01-13 | 2017-06-13 | 科大讯飞股份有限公司 | Speech level evaluation method and apparatus and system |
CN109785698A (en) * | 2017-11-13 | 2019-05-21 | 上海流利说信息技术有限公司 | Method, apparatus, electronic equipment and medium for spoken language proficiency evaluation and test |
CN108122561A (en) * | 2017-12-19 | 2018-06-05 | 广东小天才科技有限公司 | Spoken language voice evaluation method based on electronic equipment and electronic equipment |
CN109147765A (en) * | 2018-11-16 | 2019-01-04 | 安徽听见科技有限公司 | Audio quality comprehensive evaluating method and system |
WO2020149621A1 (en) * | 2019-01-14 | 2020-07-23 | 김주혁 | English speaking evaluation system and method |
CN110136721A (en) * | 2019-04-09 | 2019-08-16 | 北京大米科技有限公司 | A kind of scoring generation method, device, storage medium and electronic equipment |
CN110069784A (en) * | 2019-05-05 | 2019-07-30 | 广东电网有限责任公司 | A kind of voice quality inspection methods of marking, device, terminal and can storage medium |
US20210050001A1 (en) * | 2019-08-16 | 2021-02-18 | Ponddy Education Inc. | Systems and Methods for Comprehensive Chinese Speech Scoring and Diagnosis |
CN110675292A (en) * | 2019-09-23 | 2020-01-10 | 浙江优学智能科技有限公司 | Child language ability evaluation method based on artificial intelligence |
CN111833853A (en) * | 2020-07-01 | 2020-10-27 | 腾讯科技(深圳)有限公司 | Voice processing method and device, electronic equipment and computer readable storage medium |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114187544A (en) * | 2021-11-30 | 2022-03-15 | 厦门大学 | College English speaking multi-mode automatic scoring method |
CN117787921A (en) * | 2024-02-27 | 2024-03-29 | 北京烽火万家科技有限公司 | Intelligent education training management method and identity anti-counterfeiting method for intelligent education training |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN105845134B (en) | Spoken language evaluation method and system for freely reading question types | |
CN103594087B (en) | Improve the method and system of oral evaluation performance | |
CN110797010A (en) | Question-answer scoring method, device, equipment and storage medium based on artificial intelligence | |
CN106558252B (en) | Spoken language practice method and device realized by computer | |
CN103151042A (en) | Full-automatic oral language evaluating management and scoring system and scoring method thereof | |
CN101551947A (en) | Computer system for assisting spoken language learning | |
US20140141392A1 (en) | Systems and Methods for Evaluating Difficulty of Spoken Text | |
Shen et al. | CECOS: A Chinese-English code-switching speech database | |
KR20160008949A (en) | Apparatus and method for foreign language learning based on spoken dialogue | |
CN108280065B (en) | Foreign text evaluation method and device | |
Inoue et al. | A Study of Objective Measurement of Comprehensibility through Native Speakers' Shadowing of Learners' Utterances. | |
CN113205729A (en) | Foreign student-oriented speech evaluation method, device and system | |
KR100995847B1 (en) | Language training method and system based sound analysis on internet | |
CN110148413B (en) | Voice evaluation method and related device | |
Loukina et al. | Automated scoring across different modalities | |
Liao et al. | A prototype of an adaptive Chinese pronunciation training system | |
CN110675292A (en) | Child language ability evaluation method based on artificial intelligence | |
Yoon et al. | Word-embedding based content features for automated oral proficiency scoring | |
CN112668883A (en) | Small speech practice system for integrating Chinese speech and speech piece evaluation | |
Yoon et al. | A comparison of grammatical proficiency measures in the automated assessment of spontaneous speech | |
WO2019075827A1 (en) | Voice evaluation method and device | |
CN112309429A (en) | Method, device and equipment for explosion loss detection and computer readable storage medium | |
Shufang | Design of an automatic english pronunciation error correction system based on radio magnetic pronunciation recording devices | |
Zechner et al. | Automatic scoring of children’s read-aloud text passages and word lists | |
CN114241835A (en) | Student spoken language quality evaluation method and device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20210803 |