CN113205729A - Foreign student-oriented speech evaluation method, device and system - Google Patents

Foreign student-oriented speech evaluation method, device and system Download PDF

Info

Publication number
CN113205729A
CN113205729A CN202110389484.3A CN202110389484A CN113205729A CN 113205729 A CN113205729 A CN 113205729A CN 202110389484 A CN202110389484 A CN 202110389484A CN 113205729 A CN113205729 A CN 113205729A
Authority
CN
China
Prior art keywords
speech
characteristic
evaluation
data
text
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202110389484.3A
Other languages
Chinese (zh)
Inventor
李会法
沈莺英
谈遥新
张恒彰
王华珍
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Huaqiao University
Original Assignee
Huaqiao University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Huaqiao University filed Critical Huaqiao University
Priority to CN202110389484.3A priority Critical patent/CN113205729A/en
Publication of CN113205729A publication Critical patent/CN113205729A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09BEDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
    • G09B19/00Teaching not covered by other main groups of this subclass
    • G09B19/06Foreign languages
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/205Parsing
    • G06F40/211Syntactic parsing, e.g. based on context-free grammar [CFG] or unification grammars
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/205Parsing
    • G06F40/216Parsing using statistical methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/279Recognition of textual entities
    • G06F40/284Lexical analysis, e.g. tokenisation or collocates
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/06Resources, workflows, human or project management; Enterprise or organisation planning; Enterprise or organisation modelling
    • G06Q10/063Operations research, analysis or management
    • G06Q10/0639Performance analysis of employees; Performance analysis of enterprise or organisation operations
    • G06Q10/06393Score-carding, benchmarking or key performance indicator [KPI] analysis
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/48Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
    • G10L25/51Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/78Detection of presence or absence of voice signals
    • G10L25/87Detection of discrete points within a voice signal

Landscapes

  • Engineering & Computer Science (AREA)
  • Business, Economics & Management (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Computational Linguistics (AREA)
  • Human Resources & Organizations (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Health & Medical Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Entrepreneurship & Innovation (AREA)
  • Educational Administration (AREA)
  • Acoustics & Sound (AREA)
  • Economics (AREA)
  • Strategic Management (AREA)
  • Signal Processing (AREA)
  • Human Computer Interaction (AREA)
  • Multimedia (AREA)
  • Development Economics (AREA)
  • General Business, Economics & Management (AREA)
  • Game Theory and Decision Science (AREA)
  • Tourism & Hospitality (AREA)
  • Probability & Statistics with Applications (AREA)
  • Quality & Reliability (AREA)
  • Operations Research (AREA)
  • Marketing (AREA)
  • Educational Technology (AREA)
  • Electrically Operated Instructional Devices (AREA)

Abstract

The invention discloses a lecture evaluation method, a lecture evaluation device and a lecture evaluation system for foreign retention students, wherein the lecture evaluation method, the lecture evaluation device and the lecture evaluation system comprise the following steps: acquiring speech data of a lecturer; carrying out voice recognition on the speech voice data to obtain recognition text data; respectively extracting features of the speech data and the recognition text data to obtain quantifiable evaluation features, wherein the quantifiable evaluation features comprise speech scoring features and speech piece scoring features; and evaluating the speech data of the speech according to the quantifiable evaluation characteristics to obtain an evaluation result of the speech. The invention provides objective, professional, universal and systematic spoken language ability evaluation standards for Chinese learners, and has promotion effect on the improvement of the abilities of the Chinese learners and the development of the Chinese education industry.

Description

Foreign student-oriented speech evaluation method, device and system
Technical Field
The invention relates to the field of Chinese education industry and the technical field of language signal processing, in particular to a speech evaluation method, device and system for foreign students.
Background
In the spreading period of Chinese development, most Chinese learners study systematically in schools, both textbook textbooks and teacher vocabularies basically have more written languages, and the teaching and the use of spoken languages are relatively few. Therefore, the following two problems are common to Chinese learners when learning Chinese: first, there is a lack of opportunity for spoken language practice; and secondly, some Chinese learners want to do oral exercises, but cannot insist on the exercises because of no evaluation feedback. The traditional Chinese spoken language practice method has few platforms and resources in the market, because the spoken language is a very targeted course and does not have a uniform reference standard. Each individual will have a different expression and thus the evaluation feedback will be different accordingly.
Disclosure of Invention
The invention mainly aims to provide a speech evaluation method, a speech evaluation device and a speech evaluation system for foreign retention students, which provide objective, professional, universal and systematic spoken language ability evaluation standards for Chinese learners and have a promoting effect on the improvement of the self ability of the Chinese learners and the development of the Chinese education industry.
The invention adopts the following technical scheme:
in a first aspect, a speech evaluation method for foreign retention students includes:
acquiring speech data of a lecturer;
carrying out voice recognition on the speech voice data to obtain recognition text data;
respectively extracting features of the speech data and the recognition text data to obtain quantifiable evaluation features, wherein the quantifiable evaluation features comprise speech scoring features and speech piece scoring features; the voice scoring characteristics comprise fluency characteristics, effectiveness characteristics, speech speed characteristics and voice basic scoring characteristics; the sentence scoring characteristics comprise a principal and predicate analysis characteristic, a mixed sentence pattern characteristic, a proper vocabulary use characteristic, a language framework scoring characteristic, a healthy emotion characteristic, a consistent theme characteristic, a spoken language expression thinking characteristic and a sentence basic scoring characteristic;
and evaluating the speech data of the speech according to the quantifiable evaluation characteristics to obtain an evaluation result of the speech.
Preferably, the method for extracting fluency features comprises the following steps:
counting the time TD of pause occurrence in the speech data; the pause refers to the fact that the voice data of the speech are subjected to endpoint detection by adopting a VAD algorithm, the position of an endpoint of the voice is obtained, the duration between two connected endpoints is calculated, and the duration exceeds a set threshold value htIndicating that a pause occurred;
designing and outputting speech fluency S based on the number TD of pause occurrences; the value of the voice fluency S is in inverse proportion to the pause times, namely the smaller the pause times, the larger the value of the voice fluency S is;
preferably, the method for extracting the effectiveness characteristics comprises the following steps:
counting the number of pause words in the recognized text data and the occurrence times of invalid repeated voice texts; specifically, counting the number I of words paused in the text based on a pausing word list, and counting the occurrence times J of invalid repeated voice texts in the text based on a rule method;
the number of stop words and the number of occurrences of invalid repeated speech text are taken as validity characteristics.
Preferably, the method for extracting the speech rate features includes:
acquiring the text length L of the recognition text data and the audio length T of the speech voice data, and calculating the speech speed L/T;
and taking the speech rate L/T as the speech rate characteristic.
Preferably, the method for extracting the speech basic feature comprises the following steps:
obtaining the audio length T of the speech data, obtaining the audio length T and a preset threshold hvtThe ratio of (A) to (B); the preset threshold hvtThe audio length of the speech of the voice basic score can be obtained for the speaker;
the audio length T is compared with a preset threshold value hvtThe ratio of (a) is used as a speech basis feature.
Preferably, the method for extracting the syntactic analysis characteristics of the principal and predicate object comprises the following steps:
counting the number G of sentences which accord with the grammar structure information standard in the N sentences of the recognized text data to obtain the text sentence structure standard rate G/N;
and taking the standard rate G/N of the structure of the text sentence as a syntactic analysis characteristic of the principal and predicate object.
Preferably, the method for extracting the mixed sentence pattern features comprises the following steps:
counting the number M of sentences which accord with a standard sentence pattern in the N-sentence texts of the identification text data to obtain the standard rate M/N of the text sentence pattern;
and taking the standard rate M/N of the text sentence pattern as a syntactic analysis characteristic of the principal and predicate object.
Preferably, the method for extracting proper or improper vocabulary use features comprises the following steps:
counting the error quantity W of the text in the N sentences of text of the identification text data;
the text error number W is used as a word whether proper features are used.
Preferably, the method for extracting the language framework score features comprises the following steps:
splitting the recognized text data into an array by taking sentences as units, dividing words by taking words as units, matching the array after word division with a language frame dictionary, and calculating the number F of sentences conforming to the language frame;
and taking the number F of sentences which accord with the language frame as the language frame score characteristic.
Preferably, the method for extracting the health characteristics of the emotion comprises the following steps:
firstly, a large-scale corpus is used for training word vectors through word2vec, then word segmentation operation is carried out on the recognition text data, and word vectors word corresponding to each word are found outiThe difference between the directions of the two vectors of the word vector in the recognized text data and the word vector in the violation dictionary is measured through cosine distance, and the calculation method is that
Figure BDA0003015953280000031
Where D is the word vector dimension, word1kBeing a vector of text words, word2kThe word vectors are word vectors in the violation dictionary library;
and taking the cosine distance as the health characteristic of the emotion.
Preferably, the method for extracting whether the theme is consistent with the features comprises the following steps:
according to the number P of the subject words matched with the identification text data and the number P of all the subject wordsallCounting the subject coincidence rate P/P of the identification text data and the subjectall
Matching the topic compliance P/PallAs to whether the subject conforms to the characteristics.
Preferably, the method for extracting the features of the thought expressed by the spoken language comprises the following steps:
analyzing and matching the text data with a written language dictionary in the unit of sentences, and counting the number Wr of the sentences of the written language;
the sentence number Wr of the written language is taken as a spoken language expression thinking characteristic.
Preferably, the method for extracting the basic feature of the speech piece comprises the following steps:
taking the length of the identification text data as a basic score Q of a basic part of the user language, and regarding the identification text data, if the standard text length isLstandardThe standard basis of speech is scoretext_basicIdentifying the length of the text data as L, and calculating the basic score of the user language piece
Figure BDA0003015953280000032
And taking the user speech piece basic part Q as the speech piece basic part characteristic.
Preferably, according to the quantifiable evaluation feature, evaluating the speech data to obtain an evaluation result of the speech, including:
and respectively obtaining corresponding scores based on the extracted fluency feature, effectiveness feature, speech speed feature, speech basic feature, chief and predicate analysis feature, mixed sentence pattern feature, proper vocabulary use feature, language framework score feature, healthy emotion feature, consistent theme feature, spoken language expression thinking feature and speech basic feature, and calculating a total score through summation to realize evaluation of the speech data to obtain an evaluation result of the speech.
On the other hand, a speech evaluation device for foreign retention students comprises:
the voice data acquisition module is used for acquiring the speech data of the speaker;
the speech data recognition module is used for carrying out speech recognition on the speech data of the speech to obtain recognition text data;
the feature extraction module is used for respectively extracting features of the speech data and the recognition text data to obtain quantifiable evaluation features, and the quantifiable evaluation features comprise speech scoring features and speech piece scoring features; the voice scoring characteristics comprise fluency characteristics, effectiveness characteristics, speech speed characteristics and voice basic scoring characteristics; the sentence scoring characteristics comprise a principal and predicate analysis characteristic, a mixed sentence pattern characteristic, a proper vocabulary use characteristic, a language framework scoring characteristic, a healthy emotion characteristic, a consistent theme characteristic, a spoken language expression thinking characteristic and a sentence basic scoring characteristic;
and the evaluation module is used for evaluating the speech data according to the quantifiable evaluation characteristics to obtain the evaluation result of the speech.
In another aspect, a speech evaluation system for foreign retention students includes:
the client is used for acquiring the speech data of the speaker;
the server is used for the speech data of the speech sent by the client; carrying out voice recognition on the speech voice data to obtain recognition text data; respectively extracting features of the speech data and the recognition text data to obtain quantifiable evaluation features, wherein the quantifiable evaluation features comprise speech scoring features and speech piece scoring features; the voice scoring characteristics comprise fluency characteristics, effectiveness characteristics, speech speed characteristics and voice basic scoring characteristics; the sentence scoring characteristics comprise a principal and predicate analysis characteristic, a mixed sentence pattern characteristic, a proper vocabulary use characteristic, a language framework scoring characteristic, a healthy emotion characteristic, a consistent theme characteristic, a spoken language expression thinking characteristic and a sentence basic scoring characteristic; evaluating the speech data according to the quantifiable evaluation characteristics to obtain an evaluation result of the speech; and sending the evaluation result to a client for display.
Compared with the prior art, the invention has the following beneficial effects:
(1) the spoken language ability method for the Chinese learner is objective in standard, professional, universal and systematic, and a method based on a corpus is adopted to design a calculation method of each evaluation characteristic; the invention can promote the self-ability improvement of Chinese learners and the development of Chinese education industry;
(2) after learning the oral practice requirements of the current Chinese learner, the device and the system of the invention are designed according to the learning concept of 'the language is not learned but is practiced'; the invention is a spoken language practice system with convenient operation, flexible application and integrated design, and the practice can be carried out on the WeChat small program at any time and any place as long as one smart phone is provided; the invention can meet the individual requirements of the current Chinese learner and guides the spoken language of the user in a targeted manner; the device is different from common spoken language dialogue practice which has the characteristic of random use and aims to realize simple daily communication; the invention can not only train the spoken language ability of Chinese learners, but also improve the level of other skills such as Chinese thinking ability of the learners; the invention aims to provide students with the opportunity to speak Chinese, avoid the phenomenon of 'dumb Chinese', train the spoken language expression ability of the students and finally achieve the effect of fully developing four language skills of listening, speaking, reading and writing of Chinese learners.
The above description is only an overview of the technical solutions of the present invention, and the embodiments of the present invention are described below in order to make the technical means of the present invention more clearly understood and to make the description of the technical means more comprehensible.
The above and other objects, advantages and features of the present invention will become more apparent to those skilled in the art from the following detailed description of specific embodiments thereof, taken in conjunction with the accompanying drawings.
Drawings
Fig. 1 is an interface schematic diagram of a lecture evaluation method for foreign retention students according to an embodiment of the present invention;
FIG. 2 is a flowchart of a lecture evaluation method for foreign reservation students according to an embodiment of the present invention;
fig. 3 is a block diagram of the speech evaluation device for foreign retention students according to the embodiment of the present invention;
fig. 4 is a block diagram of a speech evaluation system for foreign reservation students according to an embodiment of the present invention.
Detailed Description
The invention will be further illustrated with reference to the following specific examples. It should be understood that these examples are for illustrative purposes only and are not intended to limit the scope of the present invention. Further, it should be understood that various changes or modifications of the present invention may be made by those skilled in the art after reading the teaching of the present invention, and such equivalents may fall within the scope of the present invention as defined in the appended claims.
Referring to fig. 1 and 2, the speech evaluation method for foreign retention students of the present invention includes the following steps: (1) acquiring text data of user voice; (2) and calculating the speech capability evaluation index.
The method comprises the following steps: text data of a user's voice is acquired.
The method specifically comprises the following steps:
s201, obtaining speech data of a speaker;
s202, carrying out voice recognition on the speech voice data to obtain recognition text data.
In the invention, based on a lecture evaluation method for foreign students, scene corpora corresponding to 3 evaluation dimensions are designed according to different levels of topics, related prompts of the topics, and model texts and written expression corpora of the topics.
1. Theme design of different levels. 46 corpora are designed (it should be noted that the corpus can be expanded according to needs, and the following is only an example), and the types such as "my weekend", "my vacation arrangement" and the like are simple in one degree and difficult in another degree. The theme design at different levels is shown in table 1 below, where difficiltly columns represent different levels and the me column represents a theme.
TABLE 1
Figure BDA0003015953280000061
Figure BDA0003015953280000071
2. Corpus design of relevant tips of subjects. 5 sets of related prompt corpuses of different subjects are designed (it should be noted that the related prompt corpuses of the subjects can be expanded according to needs, and the following is only an example). The tips 1-5 columns represent the prompt corpuses associated with the corresponding topics, respectively, as shown in Table 2 below.
TABLE 2
Figure BDA0003015953280000072
Figure BDA0003015953280000081
Figure BDA0003015953280000091
Figure BDA0003015953280000101
3. The model and written expression language material design of the subject. A set of corpora is designed 46, and the corpus schematic of the design is shown in the following table (it should be noted that the corpus schematic can be expanded as desired, and the following is only an example). Where Model is a Model and expression is a written expression for similarity comparison between the user's speech and text, as shown in table 3 below.
TABLE 3
Figure BDA0003015953280000102
Figure BDA0003015953280000111
Figure BDA0003015953280000121
And respectively displaying the designed small speech contents to the user, thereby obtaining 3 different conversation data sets such as corresponding speech audio, text evaluation, speech score and the like. Taking the partial information of the user session data set as an example, the content topic is "my weekend", as shown in table 4 below:
TABLE 4
Figure BDA0003015953280000131
As shown in the above table, the results of the actual performance of the 3 users in the speech evaluation system are included, different users can select different topics, and the language expression capabilities and levels of different users are different, so that the total word number and the speech speed of speech are different, the contents are also significantly different, and the final scores and the total score are also different.
The second step is that: and calculating the speech capability evaluation index.
The method specifically comprises the following steps:
s203, respectively carrying out feature extraction on the speech data and the recognition text data to obtain quantifiable evaluation features, wherein the quantifiable evaluation features comprise speech scoring features and speech piece scoring features; the voice scoring characteristics comprise fluency characteristics, effectiveness characteristics, speech speed characteristics and voice basic scoring characteristics; the sentence scoring characteristics comprise a principal and predicate analysis characteristic, a mixed sentence pattern characteristic, a proper vocabulary use characteristic, a language framework scoring characteristic, a healthy emotion characteristic, a consistent theme characteristic, a spoken language expression thinking characteristic and a sentence basic scoring characteristic;
and S204, evaluating the speech data according to the quantifiable evaluation characteristics to obtain the evaluation result of the speech.
In this embodiment, the method for extracting fluency features includes:
counting the times of pause occurrence in the speech data; specifically, the speech data is subjected to end point detection by VAD algorithm, and a threshold value h is settTime between endpoints over htA pause is indicated;
outputting voice fluency S as fluency characteristics based on the number of times of pause occurrence; the value range of the voice fluency S is [0,1], and is inversely proportional to the pause times, namely, the smaller the pause times, the larger the value of the voice fluency S.
Specifically, fluency characteristics (indexes) measure the thinking power and language organization ability of Chinese learners, and are used for carrying out fluency scoring on spoken texts spoken by users and input voices of the users, and scoring dimensionality is mainly carried out according to pause, speech speed and intonation used for the voices. Performing endpoint detection on voice information by adopting VAD algorithm to obtain the endpoint position of voice, and solving the time length between two connected endpoints, wherein the time length exceeds a set threshold value htIndicating that a pause has occurred. Big data training and calculation are carried out on the existing data set through a recurrent neural network to obtain a fluency judging model, the user input voice is given, the voice fluency S is output, and the value range of the voice fluency is [0,1]]The larger the value, the better.
In this embodiment, the method for extracting the validity characteristic includes:
counting the number of pause words in the recognized text data and the occurrence times of invalid repeated voice texts; specifically, counting the number I of words paused in the text based on a pausing word list, and counting the occurrence times J of invalid repeated voice texts in the text based on a rule method;
the number of stop words and the number of occurrences of invalid repeated speech text are taken as validity characteristics.
Specifically, the validity characteristics (indexes) determine the occurrence times of the stop words and the invalid repeated voice texts of the Chinese learner in the speaking process. For Chinese speech, when the Chinese learner does not express smoothly, the Chinese speech utters words such as 'forehead', 'click' and 'o', which are called pause words, and speech texts which appear continuously for multiple times are called invalid repeated speech texts. For the text converted from the user voice, counting the number I of pause words in the text based on the pause word list, and counting the occurrence times J of invalid repeated voice text in the text based on a rule method. The numeric area of the number of the stop words and the occurrence frequency of the invalid repeated voice text is [0, ∞ ], and the smaller the value is, the better the value is.
In this embodiment, the method for extracting speech rate features includes:
acquiring the text length L of the recognition text data and the audio length T of the speech voice data, and calculating the speech speed L/T;
and taking the speech rate L/T as the speech rate characteristic.
Specifically, the speech speed characteristics (indexes) are used for judging the Chinese oral proficiency of the Chinese learner. Obtaining an evaluation text of a user, obtaining the text length L input by the user and the audio length T input by the user, and calculating the speech speed L/T, wherein the value range of the speech speed is [0, ∞]Having a value of [ v ]low,vhigh]Is preferred wherein vlowIs the required minimum speech rate threshold, vhighIs the required maximum speech rate threshold.
In this embodiment, the method for extracting the speech basic feature includes:
obtaining the audio length T of the speech data, obtaining the audio length T and a preset threshold hvtThe ratio of (A) to (B); the preset threshold hvtThe audio length of the speech of the voice basic score can be obtained for the speaker;
the audio length T is compared with a preset threshold value hvtThe ratio of (a) is used as a speech basis feature.
Specifically, the speech basic sub-characteristics (indexes) measure the basic ability of spoken speech of the Chinese learner. Setting a threshold h for the length of audio input by a uservtWhen the user audio length exceeds a threshold hvtGiving the user full score of the voice base, the voice length value range is [0, ∞]The larger the value, the better.
In this embodiment, the method for extracting the syntactic analysis feature of the principal and predicate object includes:
counting the number G of sentences which accord with the grammar structure information standard in the N sentences of the recognized text data to obtain the text sentence structure standard rate G/N;
and taking the standard rate G/N of the structure of the text sentence as a syntactic analysis characteristic of the principal and predicate object.
Specifically, the syntactic analysis characteristics (indexes) of the principal and predicate guest measure the Chinese grammar ability of the user, the evaluation text input by the user is divided into an array by taking a sentence as a unit, and meanwhile, the word is divided by taking a word as a unit. The sentence is then parsed through a dependency syntax tree to determine the syntactic structure of the sentence or the dependency between the words in the sentence. And further extracting the main and subordinate object shape complementing relationship of the sentence according to the dependency relationship. And a correct syntactic structure dictionary is counted by combining a specialist method with a technical counting mode, the main meaning object shape complementing structure information analyzed by taking a sentence as a unit is sent to the dictionary for matching of the inclusion relationship, if the main meaning object shape complementing structure information analyzed by the sentence has matching information in the dictionary, the main meaning object shape complementing structure information is regarded as the syntactic structure information standard of the sentence, and if the main meaning object shape complementing structure information does not have matching information, the system judges that the syntactic structure information has a problem. And for N sentences of texts input by the user, calculating the sentence quantity G meeting the grammar structure information standard to obtain the text sentence structure standard rate G/N, wherein the value range of the text sentence structure standard rate is [0,1], and the larger the value is, the better the value is.
In this embodiment, the method for extracting the features of the mixed sentence pattern includes:
counting the number M of sentences which accord with a standard sentence pattern in the N-sentence texts of the identification text data to obtain the standard rate M/N of the text sentence pattern;
and taking the standard rate M/N of the text sentence pattern as a syntactic analysis characteristic of the principal and predicate object.
Specifically, the mixed sentence pattern characteristics (indexes) judge the Chinese grammar application capability of the user. The method comprises the steps of counting a dictionary of a standard sentence pattern according to the standard sentence pattern given in International Chinese teaching general course outline, analyzing a text input by a user in a sentence unit, dividing a sentence into words, analyzing the words from three angles of main meaning object fixed form supplement, word part of speech and key words, sending information of three dimensions into the dictionary of the standard sentence pattern for accurate matching, and if the standard sentence pattern is matched, regarding the sentence as the standard sentence pattern, otherwise, not the standard sentence pattern. For N sentence texts input by a user, calculating the number M of sentences conforming to the standard sentence pattern to obtain the standard rate M/N of the text sentence pattern, wherein the value range of the standard rate of the text sentence pattern is [0,1], and the larger the value is, the better the value is.
In this embodiment, the method for extracting proper vocabulary use features includes:
counting the error quantity W of the text in the N sentences of text of the identification text data;
the text error number W is used as a word whether proper features are used.
Specifically, whether proper characteristics (indexes) are used for measuring the phrase correlation ability of a user, the text input by the user is corrected through an expert dictionary method and an open source framework pycorector, the expert dictionary method is used for counting error-prone word collocation, such as wrong collocation of 'wearing clothes' and 'wearing hat', and the like, by experts in the Chinese education field, further, when a program is processed, whether error collocation exists in the text is traversed, and if the error collocation exists, the text is regarded as a text error. The open source framework pycorrctor detects the position of the wrongly written character according to the language model, and corrects the wrongly written character through the phonetic sound similar characteristic, the stroke five-stroke editing distance characteristic and the language model confusion characteristic. The technology uses a statistical language model kenlm tool, and simultaneously uses rnn _ attention, rnn _ crf, seq2seq _ attention, transformer, conv _ seq2seq and electra pre-training models to perform joint training, and a deep learning model for text error correction is trained to be used for detecting text error information input by a user. And for N sentences of texts input by the user, calculating the text error number W, wherein the text error number has a value range of [0, ∞ ], and the smaller the value is, the better the value is.
In this embodiment, the method for extracting the language framework score feature includes:
splitting the recognized text data into an array by taking sentences as units, dividing words by taking words as units, matching the array after word division with a language frame dictionary, and calculating the number F of sentences conforming to the language frame;
and taking the number F of sentences which accord with the language frame as the language frame score characteristic.
Specifically, the language framework score features (indexes) measure the language logic ability of the user, and a language framework dictionary is arranged by a specialist method, such as "though. And splitting the evaluation text input by the user into an array by taking the sentence as a unit, dividing the word by taking the word as a unit, matching the array after word division with a language frame dictionary, and calculating the number F of the sentences conforming to the language frame. The number of the language frame sentences is in a value range of [0, N ], and the larger the value is, the better the value is.
In this embodiment, the method for extracting whether emotion is healthy includes:
firstly, a large-scale corpus is used for training word vectors through word2vec, then word segmentation operation is carried out on the recognition text data, and word vectors word corresponding to each word are found outiThe difference between the directions of the two vectors of the word vector in the recognized text data and the word vector in the violation dictionary is measured through cosine distance, and the calculation method is that
Figure BDA0003015953280000161
Where D is the word vector dimension, word1kBeing a vector of text words, word2kThe word vectors are word vectors in the violation dictionary library;
and taking the cosine distance as the health characteristic of the emotion.
Specifically, judging whether the expressed sentences of the user accord with the core value view of social connotation by emotion health characteristics (indexes), performing text health audit aiming at texts input by the user, covering rich audit dimensions such as political involvement, yellow involvement, terrorism involvement, malicious promotion, low-quality irrigation, official illegal content library and the like, and sorting out an illegal dictionary library by a specialist law; firstly, a large-scale corpus is used for training word vectors through word2vec, then word segmentation operation is carried out on input texts, and word vectors word corresponding to each word are found outiThe difference between the direction of the word vector in the input text and the direction of the two vectors of the word vector in the violation dictionary is measured by cosine distance, and the calculation method is that
Figure BDA0003015953280000171
Where D is the word vector dimension, word1kBeing a vector of text words, word2kAnd the word vectors are word vectors in the violation dictionary library. The cosine value range of the specified included angle is [0,1]]The larger the cosine value is, the smaller the included angle between the two vectors is, namely the two vectors are more similar, and when the similarity is larger than the threshold hdIt indicates that the word violates the core value view of social meaning, with smaller values being better.
In this embodiment, the method for extracting whether the theme conforms to the feature includes:
according toThe number P of the subject words matched with the identification text data and the number P of all the subject wordsallCounting the subject coincidence rate P/P of the identification text data and the subjectall
Matching the topic compliance P/PallAs to whether the subject conforms to the characteristics.
Specifically, whether the theme conforms to the characteristics (indexes) or not is evaluated, whether the statement stated by a user is the same as the theme of the theme given by the system or not is judged, the subject words of the theme are counted according to the corresponding theme through a professional method, the subject words are stored by dictionary information, if the evaluation text input by the user is accurately matched with the subject word information listed by an expert, the number P of the subject words matched with the evaluation text of the user and the number P of all the subject words are evaluated according to the userallFurther, the topic coincidence rate P/P of the text and the topic is countedallThe subject match rate value range is [0,1]]The larger the value, the better.
In this embodiment, the method for extracting the thought characteristics of spoken language expression includes:
analyzing and matching the text data with a written language dictionary in the unit of sentences, and counting the number Wr of the sentences of the written language;
the sentence number Wr of the written language is taken as a spoken language expression thinking characteristic.
Specifically, the spoken language expression thinking characteristics (indexes) judge the thinking strength of the user, the spoken language expression thinking of the user is analyzed according to the question examination time of the user, the spoken language proportion of the user and the use proportion dimensionality of written languages, the question reading time of the user is captured according to the front end of the small program, and the added branch written languages for summarizing the problem are arranged into a written language dictionary in a specialist method. Analyzing and matching the written language dictionary for the input text of the user in the unit of sentences, and counting the sentences of the written language to obtain the written language use ratio Wr, wherein Wr is the number of the sentences of the written language. The written language is used in a ratio of [0, ∞ ], with larger values being better.
In this embodiment, the method for extracting a speech piece basic feature includes:
using the length of the identification text data as a user languageA basic score Q of the basic score, for the recognized text data, if the standard text length is LstandardThe standard basis of speech is scoretext_basicIdentifying the length of the text data as L, and calculating the basic score of the user language piece
Figure BDA0003015953280000181
And taking the user speech piece basic part Q as the speech piece basic part characteristic.
Specifically, the basic feature (index) of the speech piece measures the basic capability of the spoken language piece of the Chinese learner, the text length is evaluated as the basic score Q of the basic part of the speech piece of the user according to the input of the user, and for the evaluation text input by the user, if the standard text length is LstandardThe standard basis of speech is scoretext_basicInputting text with length L, calculating user speech piece basic score
Figure BDA0003015953280000182
The text length is in the range of [0, ∞]The larger the value, the better. The basic score of the language ensures that the user has a certain basic score.
For the user 162514089, based on the above feature extraction method, the related task completion capability index calculation values are respectively as follows:
(1) fluency: the index value range is [0,1], and the larger the value, the better. For the spoken voice audio input by the user 162514089, the fluency value is 0.83, which indicates that the user is fluent during expression and has good spoken language expression capability.
(2) Effectiveness: the index is calculated by the occurrence frequency of stop words and invalid repeated voice texts of Chinese learners in the speaking process, the value range of the occurrence frequency of the stop words and the invalid repeated voice texts is [0, ∞ ], and the smaller the value is, the better the value is. Aiming at the spoken voice audio frequency input by the user 162514089, the occurrence times of stop words and invalid repeated voice texts are all 0, which indicates that the user has coherent thinking during expression, smooth expression and better spoken language expression capability.
(3) The speed of speech: the index value range is [0, ∞]Having a value of [ v ]low,vhigh]Is preferred wherein vlowIs the required minimum speech rate threshold, vhighIs the required maximum speech rate threshold. For the spoken voice audio input by the user 162514089, the value of the speed of speech is 4.30 words/second, which is close to the native Chinese language user, and the spoken language expression capability of the user is better.
(4) The voice basic score is as follows: the voice basic score is measured through the voice length, the index value range is [0, ∞ ], and the larger the value is, the better the value is. For the spoken voice audio input by the user 162514089, the user voice length is 19.77 seconds, which indicates that the user expression content is not sufficient and fails to reach the suggested voice length.
(5) And (3) carrying out syntactic analysis on the major and predicate guest: the index value field is [0,1], the parsing value of the predicate object is 1.00 aiming at the text of voice audio conversion input by the user 162514089, and the result shows that no grammar error occurs to the user, the grammar knowledge of the user is rich, and the expression accords with the Chinese expression habit.
(6) Mixing sentence patterns: the index value field is [0,1], and the value of the mixed sentence pattern is 0.75 aiming at the text of the voice audio conversion input by the user 162514089, which indicates that the user uses some standard sentence patterns in the expression process and the expression mode is reasonable.
(7) Whether the vocabulary is used properly: the index value range is [0, ∞ ], and whether the vocabulary is properly used is 1.00 for the text of the voice audio conversion input by the user 162514089, which indicates that the user has no wrong word collocation and has a proper expression mode.
(8) Language framework: the index value field is [0, N ], where N is the number of sentences in the text, and the value of the language frame is 0 for the text of the voice audio conversion input by the user 162514089, which indicates that the sentence expression such as "albeit.
(9) Whether the emotion is healthy or not: the index value field is [0,1], and the value of whether emotion is healthy or not is 0.99 for the text of voice audio conversion input by the user 162514089, which indicates that the user expresses idea health and does not have content such as political, yellow, terrorism, malicious popularization and the like.
(10) Whether the topics are consistent: the index value field is [0,1], and whether the theme is consistent or not is 1.00 for the text of the voice audio conversion input by the user 162514089, so that the user can be explained on the basis of understanding the theme by closely fastening the theme during expression.
(11) Spoken language expresses thinking: the index value range is [0, ∞ ], the value of the spoken language expression thinking power for the text of the voice audio conversion input by the user 162514089 is 2, the user is explained to think during expression, some common words are replaced by higher-level words, and the richness of the words mastered by the user is reflected.
(12) The basic part of the language: the basic score of the speech segment is measured by the text length of the speech audio conversion, the index value field is [0, ∞ ], and the text length value of the text of the speech audio conversion input by the user 162514089 is 85, which indicates that the user expression content is not sufficient and cannot reach the suggested text length.
In this embodiment, evaluating the speech data according to the quantifiable evaluation feature to obtain an evaluation result of the speech, including:
and respectively obtaining corresponding scores based on the extracted fluency feature, effectiveness feature, speech speed feature, speech basic feature, chief and predicate analysis feature, mixed sentence pattern feature, proper vocabulary use feature, language framework score feature, healthy emotion feature, consistent theme feature, spoken language expression thinking feature and speech basic feature, and calculating a total score through summation to realize evaluation of the speech data to obtain an evaluation result of the speech.
Specifically, the following table 5 shows the total score evaluation method designed by the present invention.
TABLE 5
Figure BDA0003015953280000191
Figure BDA0003015953280000201
Figure BDA0003015953280000211
Referring to fig. 3, a speech evaluation apparatus for foreign retention students includes:
a voice data obtaining module 301, configured to obtain speech data of a speaker;
a voice data recognition module 302, configured to perform voice recognition on the speech data to obtain recognition text data;
the feature extraction module 303 is configured to perform feature extraction on the speech data and the recognition text data to obtain quantifiable evaluation features, where the quantifiable evaluation features include speech scoring features and speech piece scoring features; the voice scoring characteristics comprise fluency characteristics, effectiveness characteristics, speech speed characteristics and voice basic scoring characteristics; the sentence scoring characteristics comprise a principal and predicate analysis characteristic, a mixed sentence pattern characteristic, a proper vocabulary use characteristic, a language framework scoring characteristic, a healthy emotion characteristic, a consistent theme characteristic, a spoken language expression thinking characteristic and a sentence basic scoring characteristic;
and the evaluating module 304 is used for evaluating the speech data according to the quantifiable evaluating characteristics to obtain a speech evaluating result.
Referring to fig. 4, a speech evaluation system for foreign retention students includes:
a client 401, configured to obtain speech data of a speaker;
a server 402, configured to send the speech data of the speech by the client; carrying out voice recognition on the speech voice data to obtain recognition text data; respectively extracting features of the speech data and the recognition text data to obtain quantifiable evaluation features, wherein the quantifiable evaluation features comprise speech scoring features and speech piece scoring features; the voice scoring characteristics comprise fluency characteristics, effectiveness characteristics, speech speed characteristics and voice basic scoring characteristics; the sentence scoring characteristics comprise a principal and predicate analysis characteristic, a mixed sentence pattern characteristic, a proper vocabulary use characteristic, a language framework scoring characteristic, a healthy emotion characteristic, a consistent theme characteristic, a spoken language expression thinking characteristic and a sentence basic scoring characteristic; evaluating the speech data according to the quantifiable evaluation characteristics to obtain an evaluation result of the speech; and sending the evaluation result to a client for display.
The foregoing description of specific embodiments of the present invention has been presented. It is to be understood that the present invention is not limited to the specific embodiments described above, and that various changes and modifications may be made by one skilled in the art within the scope of the appended claims without departing from the spirit of the invention.

Claims (10)

1. A speech evaluation method for foreign retention students is characterized by comprising the following steps:
acquiring speech data of a lecturer;
carrying out voice recognition on the speech voice data to obtain recognition text data;
respectively extracting features of the speech data and the recognition text data to obtain quantifiable evaluation features, wherein the quantifiable evaluation features comprise speech scoring features and speech piece scoring features; the voice scoring characteristics comprise fluency characteristics, effectiveness characteristics, speech speed characteristics and voice basic scoring characteristics; the sentence scoring characteristics comprise a principal and predicate analysis characteristic, a mixed sentence pattern characteristic, a proper vocabulary use characteristic, a language framework scoring characteristic, a healthy emotion characteristic, a consistent theme characteristic, a spoken language expression thinking characteristic and a sentence basic scoring characteristic;
and evaluating the speech data of the speech according to the quantifiable evaluation characteristics to obtain an evaluation result of the speech.
2. The foreign reservation student oriented speech evaluation method according to claim 1, wherein the fluency feature extraction method comprises:
counting the time TD of pause occurrence in the speech data; pause refers to a pairPerforming endpoint detection on the speech data by adopting a VAD algorithm to acquire the endpoint position of the speech, and solving the time length between two connected endpoints which exceeds a set threshold value htIndicating that a pause occurred;
designing and outputting speech fluency S based on the number TD of pause occurrences; the value of the voice fluency S is in inverse proportion to the pause times, namely the smaller the pause times, the larger the value of the voice fluency S is;
a method of extracting significance signatures, comprising:
counting the number of pause words in the recognized text data and the occurrence times of invalid repeated voice texts; specifically, counting the number I of words paused in the text based on a pausing word list, and counting the occurrence times J of invalid repeated voice texts in the text based on a rule method;
the number of stop words and the number of occurrences of invalid repeated speech text are taken as validity characteristics.
3. The foreign reservation student oriented speech evaluation method according to claim 1, wherein the method for extracting speech rate features comprises:
acquiring the text length L of the recognition text data and the audio length T of the speech voice data, and calculating the speech speed L/T;
taking the speech rate L/T as a speech rate characteristic;
the method for extracting the voice basic feature comprises the following steps:
obtaining the audio length T of the speech data, obtaining the audio length T and a preset threshold hvtThe ratio of (A) to (B); the preset threshold hvtThe audio length of the speech of the voice basic score can be obtained for the speaker;
the audio length T is compared with a preset threshold value hvtThe ratio of (a) is used as a speech basis feature.
4. The foreign reservation student oriented speech evaluation method according to claim 1, wherein the method for extracting the syntactic analysis characteristics of the principal and predicate guest comprises the following steps:
counting the number G of sentences which accord with the grammar structure information standard in the N sentences of the recognized text data to obtain the text sentence structure standard rate G/N;
taking the standard rate G/N of the structure of the text sentence as the syntactic analysis characteristic of the principal and predicate object;
the method for extracting the mixed sentence pattern features comprises the following steps:
counting the number M of sentences which accord with a standard sentence pattern in the N-sentence texts of the identification text data to obtain the standard rate M/N of the text sentence pattern;
and taking the standard rate M/N of the text sentence pattern as a syntactic analysis characteristic of the principal and predicate object.
5. The speech evaluation method for the foreign reservation student according to claim 1, wherein the method for extracting the proper feature of vocabulary use comprises:
counting the error quantity W of the text in the N sentences of text of the identification text data;
using the text error number W as a word to determine whether the word is proper or not;
the method for extracting the language framework score features comprises the following steps:
splitting the recognized text data into an array by taking sentences as units, dividing words by taking words as units, matching the array after word division with a language frame dictionary, and calculating the number F of sentences conforming to the language frame;
and taking the number F of sentences which accord with the language frame as the language frame score characteristic.
6. The foreign reservation student oriented speech evaluation method according to claim 1, wherein the method for extracting the emotional health characteristic comprises the following steps:
firstly, a large-scale corpus is used for training word vectors through word2vec, then word segmentation operation is carried out on the recognition text data, and word vectors word corresponding to each word are found outiThe difference between the directions of the two vectors of the word vector in the recognized text data and the word vector in the violation dictionary is measured through cosine distance, and the calculation method is that
Figure FDA0003015953270000021
Where D is the word vector dimension, word1kBeing a vector of text words, word2kThe word vectors are word vectors in the violation dictionary library;
taking the cosine distance as the health characteristic of the emotion;
the method for extracting whether the theme conforms to the characteristics comprises the following steps:
according to the number P of the subject words matched with the identification text data and the number P of all the subject wordsallCounting the subject coincidence rate P/P of the identification text data and the subjectall
Matching the topic compliance P/PallAs to whether the subject conforms to the characteristics.
7. The foreign reservation student oriented speech evaluation method according to claim 1, wherein the method of extracting the spoken language expression thinking power feature comprises:
analyzing and matching the text data with a written language dictionary in the unit of sentences, and counting the number Wr of the sentences of the written language;
taking the sentence number Wr of the written language as the thinking characteristic of the spoken language expression;
the method for extracting the basic feature of the speech piece comprises the following steps:
taking the length of the identification text data as a basic score Q of a basic part of the user language, and regarding the identification text data, if the standard text length is LstandardThe standard basis of speech is scoretext_basicIdentifying the length of the text data as L, and calculating the basic score of the user language piece
Figure FDA0003015953270000031
And taking the user speech piece basic part Q as the speech piece basic part characteristic.
8. The lecture evaluation method for foreign retention students according to claim 1, wherein the evaluation of the lecture voice data according to the quantifiable evaluation features to obtain the evaluation result of the lecture comprises:
and respectively obtaining corresponding scores based on the extracted fluency feature, effectiveness feature, speech speed feature, speech basic feature, chief and predicate analysis feature, mixed sentence pattern feature, proper vocabulary use feature, language framework score feature, healthy emotion feature, consistent theme feature, spoken language expression thinking feature and speech basic feature, and calculating a total score through summation to realize evaluation of the speech data to obtain an evaluation result of the speech.
9. The utility model provides a speech evaluation device towards foreign student, its characterized in that includes:
the voice data acquisition module is used for acquiring the speech data of the speaker;
the speech data recognition module is used for carrying out speech recognition on the speech data of the speech to obtain recognition text data;
the feature extraction module is used for respectively extracting features of the speech data and the recognition text data to obtain quantifiable evaluation features, and the quantifiable evaluation features comprise speech scoring features and speech piece scoring features; the voice scoring characteristics comprise fluency characteristics, effectiveness characteristics, speech speed characteristics and voice basic scoring characteristics; the sentence scoring characteristics comprise a principal and predicate analysis characteristic, a mixed sentence pattern characteristic, a proper vocabulary use characteristic, a language framework scoring characteristic, a healthy emotion characteristic, a consistent theme characteristic, a spoken language expression thinking characteristic and a sentence basic scoring characteristic;
and the evaluation module is used for evaluating the speech data according to the quantifiable evaluation characteristics to obtain the evaluation result of the speech.
10. A speech evaluation system for foreign retention students is characterized by comprising:
the client is used for acquiring the speech data of the speaker;
the server is used for the speech data of the speech sent by the client; carrying out voice recognition on the speech voice data to obtain recognition text data; respectively extracting features of the speech data and the recognition text data to obtain quantifiable evaluation features, wherein the quantifiable evaluation features comprise speech scoring features and speech piece scoring features; the voice scoring characteristics comprise fluency characteristics, effectiveness characteristics, speech speed characteristics and voice basic scoring characteristics; the sentence scoring characteristics comprise a principal and predicate analysis characteristic, a mixed sentence pattern characteristic, a proper vocabulary use characteristic, a language framework scoring characteristic, a healthy emotion characteristic, a consistent theme characteristic, a spoken language expression thinking characteristic and a sentence basic scoring characteristic; evaluating the speech data according to the quantifiable evaluation characteristics to obtain an evaluation result of the speech; and sending the evaluation result to a client for display.
CN202110389484.3A 2021-04-12 2021-04-12 Foreign student-oriented speech evaluation method, device and system Pending CN113205729A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110389484.3A CN113205729A (en) 2021-04-12 2021-04-12 Foreign student-oriented speech evaluation method, device and system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110389484.3A CN113205729A (en) 2021-04-12 2021-04-12 Foreign student-oriented speech evaluation method, device and system

Publications (1)

Publication Number Publication Date
CN113205729A true CN113205729A (en) 2021-08-03

Family

ID=77026561

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110389484.3A Pending CN113205729A (en) 2021-04-12 2021-04-12 Foreign student-oriented speech evaluation method, device and system

Country Status (1)

Country Link
CN (1) CN113205729A (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114187544A (en) * 2021-11-30 2022-03-15 厦门大学 College English speaking multi-mode automatic scoring method
CN117787921A (en) * 2024-02-27 2024-03-29 北京烽火万家科技有限公司 Intelligent education training management method and identity anti-counterfeiting method for intelligent education training

Citations (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090258333A1 (en) * 2008-03-17 2009-10-15 Kai Yu Spoken language learning systems
CN101740024A (en) * 2008-11-19 2010-06-16 中国科学院自动化研究所 Method for automatic evaluation based on generalized fluent spoken language fluency
CN101739868A (en) * 2008-11-19 2010-06-16 中国科学院自动化研究所 Automatic evaluation and diagnosis method of text reading level for oral test
CN101826263A (en) * 2009-03-04 2010-09-08 中国科学院自动化研究所 Objective standard based automatic oral evaluation system
CN106847263A (en) * 2017-01-13 2017-06-13 科大讯飞股份有限公司 Speech level evaluation method and apparatus and system
JP2018045062A (en) * 2016-09-14 2018-03-22 Kddi株式会社 Program, device and method automatically grading from dictation voice of learner
CN108122561A (en) * 2017-12-19 2018-06-05 广东小天才科技有限公司 Spoken language voice evaluation method based on electronic equipment and electronic equipment
CN109147765A (en) * 2018-11-16 2019-01-04 安徽听见科技有限公司 Audio quality comprehensive evaluating method and system
CN109785698A (en) * 2017-11-13 2019-05-21 上海流利说信息技术有限公司 Method, apparatus, electronic equipment and medium for spoken language proficiency evaluation and test
CN110069784A (en) * 2019-05-05 2019-07-30 广东电网有限责任公司 A kind of voice quality inspection methods of marking, device, terminal and can storage medium
CN110136721A (en) * 2019-04-09 2019-08-16 北京大米科技有限公司 A kind of scoring generation method, device, storage medium and electronic equipment
CN110675292A (en) * 2019-09-23 2020-01-10 浙江优学智能科技有限公司 Child language ability evaluation method based on artificial intelligence
WO2020149621A1 (en) * 2019-01-14 2020-07-23 김주혁 English speaking evaluation system and method
CN111833853A (en) * 2020-07-01 2020-10-27 腾讯科技(深圳)有限公司 Voice processing method and device, electronic equipment and computer readable storage medium
US20210050001A1 (en) * 2019-08-16 2021-02-18 Ponddy Education Inc. Systems and Methods for Comprehensive Chinese Speech Scoring and Diagnosis

Patent Citations (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090258333A1 (en) * 2008-03-17 2009-10-15 Kai Yu Spoken language learning systems
CN101740024A (en) * 2008-11-19 2010-06-16 中国科学院自动化研究所 Method for automatic evaluation based on generalized fluent spoken language fluency
CN101739868A (en) * 2008-11-19 2010-06-16 中国科学院自动化研究所 Automatic evaluation and diagnosis method of text reading level for oral test
CN101826263A (en) * 2009-03-04 2010-09-08 中国科学院自动化研究所 Objective standard based automatic oral evaluation system
JP2018045062A (en) * 2016-09-14 2018-03-22 Kddi株式会社 Program, device and method automatically grading from dictation voice of learner
CN106847263A (en) * 2017-01-13 2017-06-13 科大讯飞股份有限公司 Speech level evaluation method and apparatus and system
CN109785698A (en) * 2017-11-13 2019-05-21 上海流利说信息技术有限公司 Method, apparatus, electronic equipment and medium for spoken language proficiency evaluation and test
CN108122561A (en) * 2017-12-19 2018-06-05 广东小天才科技有限公司 Spoken language voice evaluation method based on electronic equipment and electronic equipment
CN109147765A (en) * 2018-11-16 2019-01-04 安徽听见科技有限公司 Audio quality comprehensive evaluating method and system
WO2020149621A1 (en) * 2019-01-14 2020-07-23 김주혁 English speaking evaluation system and method
CN110136721A (en) * 2019-04-09 2019-08-16 北京大米科技有限公司 A kind of scoring generation method, device, storage medium and electronic equipment
CN110069784A (en) * 2019-05-05 2019-07-30 广东电网有限责任公司 A kind of voice quality inspection methods of marking, device, terminal and can storage medium
US20210050001A1 (en) * 2019-08-16 2021-02-18 Ponddy Education Inc. Systems and Methods for Comprehensive Chinese Speech Scoring and Diagnosis
CN110675292A (en) * 2019-09-23 2020-01-10 浙江优学智能科技有限公司 Child language ability evaluation method based on artificial intelligence
CN111833853A (en) * 2020-07-01 2020-10-27 腾讯科技(深圳)有限公司 Voice processing method and device, electronic equipment and computer readable storage medium

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114187544A (en) * 2021-11-30 2022-03-15 厦门大学 College English speaking multi-mode automatic scoring method
CN117787921A (en) * 2024-02-27 2024-03-29 北京烽火万家科技有限公司 Intelligent education training management method and identity anti-counterfeiting method for intelligent education training

Similar Documents

Publication Publication Date Title
CN105845134B (en) Spoken language evaluation method and system for freely reading question types
CN103594087B (en) Improve the method and system of oral evaluation performance
CN110797010A (en) Question-answer scoring method, device, equipment and storage medium based on artificial intelligence
CN106558252B (en) Spoken language practice method and device realized by computer
CN103151042A (en) Full-automatic oral language evaluating management and scoring system and scoring method thereof
CN101551947A (en) Computer system for assisting spoken language learning
US20140141392A1 (en) Systems and Methods for Evaluating Difficulty of Spoken Text
Shen et al. CECOS: A Chinese-English code-switching speech database
KR20160008949A (en) Apparatus and method for foreign language learning based on spoken dialogue
CN108280065B (en) Foreign text evaluation method and device
Inoue et al. A Study of Objective Measurement of Comprehensibility through Native Speakers' Shadowing of Learners' Utterances.
CN113205729A (en) Foreign student-oriented speech evaluation method, device and system
KR100995847B1 (en) Language training method and system based sound analysis on internet
CN110148413B (en) Voice evaluation method and related device
Loukina et al. Automated scoring across different modalities
Liao et al. A prototype of an adaptive Chinese pronunciation training system
CN110675292A (en) Child language ability evaluation method based on artificial intelligence
Yoon et al. Word-embedding based content features for automated oral proficiency scoring
CN112668883A (en) Small speech practice system for integrating Chinese speech and speech piece evaluation
Yoon et al. A comparison of grammatical proficiency measures in the automated assessment of spontaneous speech
WO2019075827A1 (en) Voice evaluation method and device
CN112309429A (en) Method, device and equipment for explosion loss detection and computer readable storage medium
Shufang Design of an automatic english pronunciation error correction system based on radio magnetic pronunciation recording devices
Zechner et al. Automatic scoring of children’s read-aloud text passages and word lists
CN114241835A (en) Student spoken language quality evaluation method and device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20210803